
phase-06-quality-security
by Harery
Enterprise 8-Phase Software Development Lifecycle Framework with Quality Gates, Multi-Agent Orchestration, and AI-Assisted Development using Claude Code. Perfect for regulated industries (HIPAA, SOC 2, PCI DSS, SOX, GDPR).
SKILL.md
name: "phase_06_quality_security" description: "Validate quality and security through comprehensive testing. Functional testing, integration/system testing, performance testing, security testing, compliance validation, user acceptance testing (UAT), and defect management." phase: 6 phase_name: "Quality & Security Validation" owner: "QA Lead" secondary_owner: "Security Lead" participants: ["QA Engineers", "Performance Engineer", "Security Lead", "Compliance Officer", "Product Owner", "Users", "Developers"] entry_criteria: [ "All planned features implemented", "Unit tests passing", "Code reviews completed", "Build artifacts ready" ] exit_criteria: [ "Functional tests passing", "Integration tests passing", "Performance tests meeting SLAs", "Security tests passing with no critical vulnerabilities", "Compliance validation complete", "UAT signed off", "Known defects documented and prioritized", "Test report completed", "Go/no-go decision to proceed to Phase 7" ] estimated_duration: "4-8 weeks" dependencies: ["phase_05_development_execution"] outputs: [ "Test Plan and Test Cases", "Test Execution Results", "Performance Test Report", "Security Test Report", "Penetration Test Report", "Compliance Validation Report", "UAT Sign-off", "Defect Register", "Test Summary Report" ] next_phase: "phase_07_deployment_release"
Phase 6: Quality & Security Validation
Your Safety Net Before Production
Welcome to Phase 6 - the deep breath before the big dive. You've built something great in Phase 5, and now it's time to make absolutely sure it's ready for the real world.
Think of this phase as your pre-flight checklist. It's not about finding fault - it's about catching issues now, when they're cheap to fix, rather than in production when they're expensive and stressful.
The mindset here: Test thoroughly, fix confidently, deploy safely.
What You'll Achieve
By the end of this phase, you'll have:
- Confidence that your software works as intended
- Proof that it can handle the load your users will throw at it
- Assurance that security vulnerabilities have been caught and addressed
- Validation that you're meeting all compliance requirements
- Real user feedback that you're building something people actually want
What happens next: Phase 7 will deploy everything to production, knowing you've done your due diligence.
Who's Driving This Phase?
| Role | What They're Responsible For |
|---|---|
| QA Lead | Your quality champion and testing strategist |
| Security Lead | Your security shield, catching what dev might have missed |
| Performance Engineer | Making sure speed doesn't suffer as scale grows |
| Compliance Officer | Ensuring you're playing by the rules |
| Product Owner | Connecting quality to what users actually need |
| Users | The ultimate judges - does this solve their problem? |
The Testing Journey
Testing in Phase 6 isn't a checklist - it's a discovery process. You're not just confirming what you know works; you're finding out what doesn't.
Functional Tests → Integration Tests → Performance Tests → Security Tests → Compliance Checks → UAT
Each layer catches different issues. Skip a layer, and you risk expensive surprises later.
Step 1: Functional Testing (Does It Work?)
Time needed: 2-3 weeks Led by: QA Lead + QA Engineers
What You're Testing
Functional testing answers the question: "Does the software do what it's supposed to do?" You'll test every feature, every workflow, and every edge case you can think of.
The Testing Approach
- Test the happy path - What happens when everything goes right?
- Test the unhappy paths - What happens when things go wrong?
- Test edge cases - What happens at boundaries (empty input, max values, etc.)?
- Test workflows - Do real user journeys work end-to-end?
- Test integrations - Does your software play nicely with other systems?
A Good Test Case Has:
- Clear preconditions (what state should the system be in?)
- Specific steps (what exactly should the tester do?)
- Expected results (what should happen?)
- Actual results (what actually happened?)
- Status (pass/fail)
Test Case Example:
Title: User login with valid credentials
Preconditions:
- User account exists with email: test@example.com
- User password is: SecurePass123
Steps:
1. Navigate to login page
2. Enter email: test@example.com
3. Enter password: SecurePass123
4. Click "Login" button
Expected Result:
- User is redirected to dashboard
- Welcome message displays: "Welcome back!"
- User session is established
Status: [ ] Pass [ ] Fail
Pro tip: Write test cases during Phase 4 (Development Planning), not Phase 6. You'll thank yourself later.
What You'll Have
Expected Output:
- Functional test results (pass/fail for each test case)
- Defect reports for any issues found
- Test coverage report (what percentage of features were tested)
- Feature validation matrix (linking requirements to tests)
Who Approves:
- QA Lead confirms functional testing is complete
- Product Owner confirms acceptance criteria are met
Step 2: Integration & System Testing (Do the Pieces Work Together?)
Time needed: 2 weeks Led by: QA Lead
What You're Testing
Unit tests (from Phase 5) check that individual pieces work. Integration testing checks that those pieces work together. System testing checks that the whole system works as a complete product.
What to Test
- API integrations - Does your backend talk correctly to external services?
- Database integrations - Are queries returning the right data? Are transactions atomic?
- Third-party services - Do payment gateways, email services, auth providers work?
- Data flow - Does data move correctly through the system?
- Error recovery - If one service fails, does the system handle it gracefully?
Data Format Mismatches:
- Date formats (YYYY-MM-DD vs MM/DD/YYYY)
- Number formats (1,000.5 vs 1.000,5)
- Character encodings (UTF-8 vs ASCII)
Timing Issues:
- Race conditions (two processes updating the same data)
- Timeout settings (one system gives up before another responds)
- Async processing (assuming something is done when it's still in queue)
API Changes:
- Version mismatches (you're calling v1, they're on v2)
- Deprecated fields (something you rely on is going away)
- Breaking changes (the API contract changed)
Pro tip: Contract testing (using tools like Pact) can catch many of these issues before they reach production.
What You'll Have
Expected Output:
- Integration test results
- System test results
- API test results
- Integration defect reports (with priorities)
Who Approves:
- QA Lead validates integration testing
- Tech Lead validates technical integration points
Step 3: Performance Testing (Will It Handle the Load?)
Time needed: 2 weeks Led by: Performance Engineer
Why Performance Testing Matters
Your software might work perfectly with one user. But what happens when 100 users try it at once? Or 1,000? Or 100,000? Performance testing finds the breaking point before your users do.
Types of Performance Testing
- Load Testing - Normal expected load (e.g., 1,000 concurrent users)
- Stress Testing - Beyond normal load (e.g., 5,000 concurrent users)
- Endurance Testing - Sustained load over time (e.g., 24 hours at peak load)
- Spike Testing - Sudden load increases (e.g., flash sale traffic)
| Metric | What It Means | Target |
|---|---|---|
| Response Time (p95) | 95% of requests complete in this time | <200ms |
| Response Time (p99) | 99% of requests complete in this time | <500ms |
| Throughput | Requests per second the system can handle | 1000+ req/sec |
| Concurrent Users | Users using the system simultaneously | 10,000+ |
| Error Rate | Percentage of requests that fail | <0.1% |
What to Do If You Don't Meet Targets:
- Profile the code (find bottlenecks)
- Optimize database queries (add indexes, rewrite queries)
- Add caching (Redis, Memcached)
- Scale horizontally (add more servers)
- Scale vertically (bigger servers)
Remember: Performance is a feature. Slow software feels broken.
What You'll Have
Expected Output:
- Performance test plan
- Performance test results (with metrics)
- Performance analysis report (bottlenecks identified)
- Optimization recommendations
Who Approves:
- Performance Engineer validates testing methodology
- SRE confirms the system is operable under load
- Tech Lead validates optimization recommendations
Step 4: Security Testing (Is It Safe?)
Time needed: 2-3 weeks Led by: Security Lead
Why Security Testing Is Non-Negotiable
In 2026, security breaches aren't just embarrassing - they're existential threats. One vulnerability can cost you customers, revenue, and reputation. Security testing is your insurance policy.
What You're Testing For
- OWASP Top 10 - The most critical security risks (injection, broken auth, XSS, etc.)
- Authentication & Authorization - Can only the right people access the right things?
- Input Validation - Are user inputs sanitized (no SQL injection, no XSS)?
- Session Management - Are sessions secure (timeout, secure cookies)?
- Encryption - Is data encrypted at rest and in transit?
- Secrets Management - Are API keys, passwords properly secured?
SAST (Static Application Security Testing):
- Scans source code for security issues
- Tools: SonarQube, Semgrep, Checkmarx
- Run on every pull request
DAST (Dynamic Application Security Testing):
- Tests running application for vulnerabilities
- Tools: OWASP ZAP, Burp Suite
- Finds issues SAST might miss
SCA (Software Composition Analysis):
- Scans dependencies for known vulnerabilities
- Tools: Snyk, Dependabot, OWASP Dependency-Check
- Critical in 2026's dependency-heavy world
Penetration Testing:
- Human testers attempt to breach your system
- Finds logic issues automated tools miss
- Consider hiring a professional firm
Pro tip: Security is a process, not a phase. But Phase 6 is your last chance to catch issues before they become incidents.
What You'll Have
Expected Output:
- Security test plan
- SAST scan results (with vulnerability counts)
- DAST scan results (with exploitable issues)
- SCA scan results (with vulnerable dependencies)
- Penetration test report (if conducted)
- Remediation plan (how you'll fix what you found)
Who Approves:
- Security Lead validates the testing approach
- CISO approves the security posture before production
- Compliance Officer validates regulatory security requirements
Step 5: Compliance Validation (Are You Following the Rules?)
Time needed: 2 weeks Led by: Compliance Officer
Why Compliance Matters
Depending on your industry and location, you might be subject to regulations like HIPAA (healthcare), PCI DSS (payments), GDPR (EU data), SOC 2 (security), or others. Non-compliance isn't just a fine - it can shut down your business.
What You're Validating
- Data protection - Is sensitive data properly secured?
- Access controls - Can only authorized people access sensitive data?
- Audit trails - Are all actions logged and traceable?
- Retention policies - Is data kept for the required duration (and no longer)?
- User rights - Can users access, correct, or delete their data (GDPR)?
- Breach notification - Do you have processes to report breaches within required timeframes?
| Framework | Industry | Key Focus |
|---|---|---|
| HIPAA | Healthcare | PHI protection, breach notification |
| PCI DSS | Payments | Card data security, vulnerability scanning |
| SOC 2 | SaaS | Security, availability, privacy controls |
| GDPR | EU data | Data subject rights, breach notification |
| SOX | Public companies | Financial reporting controls, audit trail |
| CMMC | Defense | Controlled technical data, incident response |
Compliance Testing Activities:
- Review audit logs for completeness
- Verify access control implementations
- Validate encryption standards
- Test data retention and deletion processes
- Conduct mock audits to prepare for real ones
Pro tip: Compliance isn't a checkbox - it's a culture of doing things the right way, every time.
What You'll Have
Expected Output:
- Compliance validation report
- Gap analysis (what's missing, if anything)
- Remediation plan (how to fix gaps)
- Compliance sign-off
Who Approves:
- Compliance Officer owns the validation
- Legal reviews the compliance position
- CISO approves security compliance aspects
Step 6: User Acceptance Testing (Do Users Love It?)
Time needed: 2-3 weeks Led by: Product Owner
Why UAT Matters
All the testing so far has been technical. UAT is different - it's about whether real users can actually use your software to solve their problems. It's the reality check before production.
How UAT Works
- Select UAT participants - Real users (or people who think like real users)
- Prepare UAT environment - Production-like setup with test data
- Train participants - Show them how to use the software
- Define UAT scenarios - Real-world tasks users need to accomplish
- Execute testing - Users work through scenarios
- Collect feedback - What works? What's confusing? What's missing?
- Address issues - Fix critical problems (nice-to-haves can wait)
Do This:
- Keep UAT focused on critical user journeys
- Provide clear, simple test instructions
- Make it easy to give feedback (forms, surveys, interviews)
- Thank participants for their time
- Act on the feedback (even if it's explaining why you can't)
Avoid This:
- Making UAT too technical (it's not about the code, it's about the experience)
- Ignoring feedback because "users don't understand"
- Waiting until the last minute to plan UAT
- Overloading participants with too many scenarios
Red Flag: If users can't complete basic tasks without asking questions, your UX needs work.
What You'll Have
Expected Output:
- UAT plan (scenarios, participants, schedule)
- UAT test results (what passed, what failed)
- User feedback (comments, suggestions, complaints)
- UAT sign-off (users agree the software is ready)
Who Approves:
- Product Owner coordinates UAT
- Users provide the actual sign-off
- Business stakeholders give final approval
Step 7: Defect Management (Fix What Matters)
Time needed: Continuous throughout Phase 6 Led by: QA Lead
How to Prioritize Defects
Not all bugs are created equal. Some are showstoppers; others are nice to fix eventually. You need a system to prioritize.
Defect Priority Levels
| Priority | Description | Examples | Fix Target |
|---|---|---|---|
| P1 - Critical | System down, data loss, security breach | Login doesn't work, database crashes, data corruption | 4 hours |
| P2 - High | Major feature broken, significant impact | Can't save work, key workflow fails | 24 hours |
| P3 - Medium | Workaround available, partial impact | UI issue, non-critical bug | 1 week |
| P4 - Low | Cosmetic, enhancement | Spelling error, nice-to-have improvement | Next release |
When Logging a Defect:
- Describe what you did (steps to reproduce)
- Describe what you expected to happen
- Describe what actually happened
- Include screenshots, logs, or recordings if possible
- Suggest priority (but let QA lead make final call)
When Fixing a Defect:
- Write a test that reproduces the issue (TDD for bugs)
- Fix the issue
- Verify the fix works
- Check for similar issues (if this bug exists here, might it exist elsewhere?)
- Document the fix (why did it break? how did you fix it?)
When Closing a Defect:
- Confirm it's actually fixed
- Get confirmation from the person who found it
- Update documentation if needed
- Close the ticket with a clear resolution
Pro tip: Every defect is a learning opportunity. Ask "why did this happen?" and "how can we prevent it in the future?"
What You'll Have
Expected Output:
- Defect register (all logged defects with priorities)
- Defect metrics (how many found, how many fixed, how fast)
- Root cause analysis for critical defects
- Known issues document (defects that won't be fixed before release)
Who Approves:
- QA Lead manages the defect register
- Tech Lead assigns technical defects
- Product Owner decides which defects block release
Step 8: Go/No-Go Decision (Are You Ready?)
Time needed: 1 week Led by: Project Manager + Product Owner
Making the Decision
At the end of Phase 6, you face a critical question: Are we ready to deploy to production? This isn't a formality - it's a deliberate decision based on evidence.
The Go/No-Go Assessment
Go Criteria (must all be true):
- All critical (P1) and high (P2) defects are fixed
- Functional tests are passing (or failures are understood and accepted)
- Performance meets SLAs (or degradation is understood and accepted)
- Security tests show no critical vulnerabilities
- Compliance requirements are met (or exceptions are documented)
- UAT participants have signed off
- Known issues are documented and prioritized
No-Go Triggers (any one means stop):
- Critical security vulnerabilities unaddressed
- Major compliance gaps
- Core functionality not working
- Performance below minimum thresholds
- Users unable to complete critical tasks
- Inadequate monitoring or rollback plans
A No-Go isn't failure - it's wisdom.
If you're not ready, you're not ready. Deploying when you're not ready is how production incidents happen.
When You Get a No-Go:
- Don't panic (this is why you have Phase 6)
- Identify what's blocking (specific, actionable items)
- Estimate what it takes to unblock (time, resources)
- Make a plan (what needs to happen, who does it, when)
- Communicate (stakeholders need to know)
- Execute the plan
- Reassess (run go/no-go again when ready)
Pro tip: It's better to delay a release by two weeks than to cause a production incident that takes two months to recover from.
What You'll Have
Expected Output:
- Test summary report (all test results in one place)
- Quality gates assessment (are criteria met?)
- Security assessment (current posture)
- Compliance assessment (any gaps?)
- Risk assessment (what could still go wrong?)
- Go/no-go recommendation
- Go/no-go approval (or decision to defer)
Who Approves:
- Project Manager compiles the recommendation
- All functional leads (Tech, QA, Security) validate readiness
- Executive Sponsor makes the final call
Quality Gates: Are You Ready for Phase 7?
Before you can move to Phase 7 (Deployment & Release), confirm:
- Functional tests are passing - Core functionality works
- Integration tests are passing - Systems work together
- Performance tests meet SLAs - System can handle the load
- Security tests pass - No critical vulnerabilities
- Compliance validation is complete - Regulatory requirements met
- UAT is signed off - Users approve
- Known defects are documented - Transparency about what isn't perfect
- Go/no-go decision is made - Clear path forward
If anything is missing: Address it before proceeding. Quality gates are your friends - they prevent bad things from happening in production.
Testing Tools for 2026
| Type | Recommended Tools | Why These Tools? |
|---|---|---|
| E2E Testing | Playwright, Cypress | Fast, reliable, modern browser support |
| API Testing | Postman, REST Assured | Comprehensive API validation |
| Performance | k6, JMeter | Scriptable, scalable, cloud-native |
| Security (SAST) | SonarQube, Semgrep | Fast, accurate, CI/CD integrated |
| Security (DAST) | OWASP ZAP, Burp Suite | Industry standard, comprehensive |
| Security (SCA) | Snyk, Dependabot | Essential for dependency-heavy apps |
| Test Management | Jira, TestRail | Centralized test tracking and reporting |
Words of Encouragement
Phase 6 can feel stressful. You're looking for problems, which means you're finding problems. That's not failure - that's success.
Remember:
- Every bug found in Phase 6 is a bug that won't cause a production incident
- Security vulnerabilities found now are breaches prevented
- Performance issues found now are outages avoided
- User feedback now is angry users later avoided
- A No-Go decision now is a disaster prevented later
You're doing important work. The thoroughness of Phase 6 is what separates professional software development from amateur hour.
What's Next
Once you have go/no-go approval, you'll move to Phase 7: Deployment & Release. There, you'll carefully and confidently deploy everything to production, knowing you've done your due diligence.
Think of Phase 6 as the safety inspection before a rocket launch. It's tedious, it's rigorous, but it's what makes the launch successful.
Next up: Deploying with confidence, not with fingers crossed.
Previous Phase: Phase 5: Development Execution Next Phase: Phase 7: Deployment & Release
Version 1.0.0 | OCTALUME Enterprise Lifecycle Framework
Score
Total Score
Based on repository quality metrics
SKILL.mdファイルが含まれている
ライセンスが設定されている
100文字以上の説明がある
GitHub Stars 100以上
1ヶ月以内に更新
10回以上フォークされている
オープンIssueが50未満
プログラミング言語が設定されている
1つ以上のタグが設定されている
Reviews
Reviews coming soon
