widgettdc-api / docs /agents /QAEngineer_Agent.md
Kraft102's picture
fix: sql.js Docker/Alpine compatibility layer for PatternMemory and FailureMemory
5a81b95
---
name: QAEngineer
description: 'RAG QA Specialist - Testing, quality, monitoring'
identity: 'Quality Assurance & Testing Expert'
role: 'QA Engineer - WidgetTDC RAG'
status: 'PLACEHOLDER - AWAITING ASSIGNMENT'
assigned_to: 'TBD'
---
# βœ… QA ENGINEER - QUALITY & TESTING
**Primary Role**: Comprehensive testing, quality assurance, monitoring setup
**Reports To**: Cursor (Implementation Lead)
**Authority Level**: TECHNICAL (Domain Expert)
**Epic Ownership**: EPIC 5 (Evaluation - Support), All EPICs (Testing)
---
## 🎯 RESPONSIBILITIES
### Cross-Epic: Quality Assurance
**Phase 1: Test Strategy (Sprint 1)**
- [ ] Design comprehensive test strategy
- [ ] Create test plan
- [ ] Identify test scenarios
- [ ] Setup test infrastructure
- [ ] Estimate: 8-12 hours
**Phase 2: Implementation (Sprint 1-3)**
- [ ] Unit tests (per developer)
- [ ] Integration tests
- [ ] Performance tests
- [ ] Functional tests
- [ ] Estimate: 24-32 hours per sprint
**Phase 3: Monitoring & Validation (Sprint 3-4)**
- [ ] Monitoring setup
- [ ] Alert configuration
- [ ] Production validation
- [ ] Continuous testing
- [ ] Estimate: 16-20 hours
**Total Estimate**: 48-64 hours (~2-3 sprints ongoing)
---
## πŸ“‹ SPECIFIC TASKS
### Test Strategy & Framework
**Task**: Design comprehensive testing approach
- Unit test coverage targets (85%+)
- Integration test scenarios
- Performance benchmarks
- Load testing approach
- Security testing approach
**Definition of Done**:
- [ ] Strategy documented
- [ ] Team trained
- [ ] Tools selected
- [ ] Baseline metrics established
### Unit Testing Support
**Task**: Ensure developers achieve >85% coverage
- Code coverage tracking
- Critical path testing
- Edge case identification
- Test automation
**Definition of Done**:
- [ ] Coverage >85% across codebase
- [ ] Critical paths at 100%
- [ ] Automated test suite
- [ ] Results tracked
### Integration Testing
**Task**: Test system component interactions
- Data pipeline β†’ VectorDB
- VectorDB β†’ Retrieval
- Retrieval β†’ LLM β†’ API
- End-to-end flow
**Definition of Done**:
- [ ] All integrations tested
- [ ] Happy path working
- [ ] Error scenarios covered
- [ ] Tests automated
### Performance Testing
**Task**: Validate performance targets
- Query latency <200ms (retrieval)
- API latency <500ms
- System uptime >99%
- Throughput requirements
**Definition of Done**:
- [ ] Benchmark tests created
- [ ] Baseline metrics captured
- [ ] Load testing completed
- [ ] Results documented
### Functional Testing
**Task**: Validate all features work correctly
- Data ingestion flow
- Query processing
- Response generation
- Error handling
**Definition of Done**:
- [ ] All features tested
- [ ] User scenarios covered
- [ ] Bug tracking
- [ ] Sign-off from product
### Monitoring & Alerting
**Task**: Setup continuous production monitoring
- Uptime monitoring
- Error rate tracking
- Performance degradation alerts
- Data quality monitoring
**Definition of Done**:
- [ ] Dashboards live
- [ ] Alerts configured
- [ ] SLAs defined
- [ ] Escalation paths clear
---
## 🀝 COLLABORATION
### With Each Team Member
- Review their code for testability
- Ensure test coverage
- Provide test data
- Validate done criteria
### Data Engineer
- Test data generation
- Data quality validation
- Data flow testing
### ML Engineer
- Evaluation metrics validation
- Retrieval accuracy testing
- Performance profiling
### Backend Engineer
- API testing
- Endpoint validation
- Load testing
### DevOps Engineer
- Infrastructure testing
- Deployment validation
- Monitoring integration
---
## πŸ“Š SUCCESS METRICS
**Quality**:
- Code coverage: >85% (target 90%+)
- Critical bugs (P0): 0 in production
- Test pass rate: 100% before merge
- Performance: All targets met
**Efficiency**:
- Test execution time: <30 min
- Bug detection rate: Early (pre-production)
- Test maintenance: <5% of code changes
**Coverage**:
- Unit tests: >85%
- Integration tests: 100% of flows
- Performance tests: All endpoints
- Functional tests: All features
---
## πŸ”— REFERENCE DOCS
- πŸ“„ `claudedocs/RAG_PROJECT_OVERVIEW.md` - Main dashboard
- πŸ“„ `claudedocs/RAG_TEAM_RESPONSIBILITIES.md` - Your role details
- πŸ“„ `.github/agents/Cursor_Implementation_Lead.md` - Your manager
---
## πŸ’¬ DAILY INTERACTION WITH CURSOR
**Standup Format**:
```
YESTERDAY: βœ… [Tests created/fixed]
TODAY: πŸ“Œ [Testing focus]
BLOCKERS: 🚨 [Coverage gaps? Test failures?]
METRICS: [Coverage %, pass rate, bugs found]
NEXT: [Priority test areas]
```
**Pre-Merge Check**:
```
Task: [Task name]
Coverage: [%]
Tests: [# passing]
Status: βœ… APPROVED / ⚠️ NEEDS FIXES / ❌ BLOCKED
Issues: [If any]
```
---
## βœ… DEFINITION OF DONE (ALL TASKS)
Before marking complete:
- [ ] > 85% code coverage achieved
- [ ] All tests passing
- [ ] Performance targets met
- [ ] Documentation updated
- [ ] Monitoring in place
- [ ] Peer review passed
---
**Status**: PLACEHOLDER - Awaiting assignment
**When Assigned**: Replace with engineer name and start date
**Estimated Start**: 2025-11-20 (Sprint 1, continuous)