Files
antigravity-skills-reference/skills/loki-mode/references/sdlc-phases.md

9.3 KiB

SDLC Phases Reference

All phases with detailed workflows and testing procedures.


Phase Overview

Bootstrap -> Discovery -> Architecture -> Infrastructure
     |           |            |              |
  (Setup)   (Analyze PRD)  (Design)    (Cloud/DB Setup)
                                             |
Development <- QA <- Deployment <- Business Ops <- Growth Loop
     |         |         |            |            |
 (Build)    (Test)   (Release)    (Monitor)    (Iterate)

Phase 0: Bootstrap

Purpose: Initialize Loki Mode environment

Actions:

  1. Create .loki/ directory structure
  2. Initialize orchestrator state in .loki/state/orchestrator.json
  3. Validate PRD exists and is readable
  4. Spawn initial agent pool (3-5 agents)
  5. Create CONTINUITY.md

Directory Structure Created:

.loki/
+-- CONTINUITY.md
+-- state/
|   +-- orchestrator.json
|   +-- agents/
|   +-- circuit-breakers/
+-- queue/
|   +-- pending.json
|   +-- in-progress.json
|   +-- completed.json
|   +-- dead-letter.json
+-- specs/
+-- memory/
+-- artifacts/

Phase 1: Discovery

Purpose: Understand requirements and market context

Actions:

  1. Parse PRD, extract requirements
  2. Spawn biz-analytics agent for competitive research
  3. Web search competitors, extract features, reviews
  4. Identify market gaps and opportunities
  5. Generate task backlog with priorities and dependencies

Output:

  • Requirements document
  • Competitive analysis
  • Initial task backlog in .loki/queue/pending.json

Phase 2: Architecture

Purpose: Design system architecture and generate specs

SPEC-FIRST WORKFLOW

Step 1: Extract API Requirements from PRD

  • Parse PRD for user stories and functionality
  • Map to REST/GraphQL operations
  • Document data models and relationships

Step 2: Generate OpenAPI 3.1 Specification

openapi: 3.1.0
info:
  title: Product API
  version: 1.0.0
paths:
  /auth/login:
    post:
      summary: Authenticate user and return JWT
      requestBody:
        required: true
        content:
          application/json:
            schema:
              type: object
              required: [email, password]
              properties:
                email: { type: string, format: email }
                password: { type: string, minLength: 8 }
      responses:
        200:
          description: Success
          content:
            application/json:
              schema:
                type: object
                properties:
                  token: { type: string }
                  expiresAt: { type: string, format: date-time }
        401:
          description: Invalid credentials

Step 3: Validate Spec

npm install -g @stoplight/spectral-cli
spectral lint .loki/specs/openapi.yaml
swagger-cli validate .loki/specs/openapi.yaml

Step 4: Generate Artifacts from Spec

# TypeScript types
npx openapi-typescript .loki/specs/openapi.yaml --output src/types/api.ts

# Client SDK
npx openapi-generator-cli generate \
  -i .loki/specs/openapi.yaml \
  -g typescript-axios \
  -o src/clients/api

# Server stubs
npx openapi-generator-cli generate \
  -i .loki/specs/openapi.yaml \
  -g nodejs-express-server \
  -o backend/generated

# Documentation
npx redoc-cli bundle .loki/specs/openapi.yaml -o docs/api.html

Step 5: Select Tech Stack

  • Spawn eng-backend + eng-frontend architects
  • Both agents review spec and propose stack
  • Consensus required (both must agree)
  • Self-reflection checkpoint with evidence

Step 6: Create Project Scaffolding

  • Initialize project with tech stack
  • Install dependencies
  • Configure linters
  • Setup contract testing framework

Phase 3: Infrastructure

Purpose: Provision cloud resources and CI/CD

Actions:

  1. Spawn ops-devops agent
  2. Provision cloud resources (see references/deployment.md)
  3. Set up CI/CD pipelines
  4. Configure monitoring and alerting
  5. Create staging and production environments

CI/CD Pipeline:

name: CI/CD Pipeline
on: [push, pull_request]
jobs:
  test:
    - Lint
    - Type check
    - Unit tests
    - Contract tests
    - Security scan
  deploy-staging:
    needs: test
    - Deploy to staging
    - Smoke tests
  deploy-production:
    needs: deploy-staging
    - Blue-green deploy
    - Health checks
    - Auto-rollback on errors

Phase 4: Development

Purpose: Implement features with quality gates

Workflow Per Task:

1. Dispatch implementation subagent (Task tool, model: sonnet)
2. Subagent implements with TDD, commits, reports back
3. Dispatch 3 reviewers IN PARALLEL (single message, 3 Task calls):
   - code-reviewer (opus)
   - business-logic-reviewer (opus)
   - security-reviewer (opus)
4. Aggregate findings by severity
5. IF Critical/High/Medium found:
   - Dispatch fix subagent
   - Re-run ALL 3 reviewers
   - Loop until all PASS
6. Add TODO comments for Low issues
7. Add FIXME comments for Cosmetic issues
8. Mark task complete with git checkpoint

Implementation Rules:

  • Agents implement ONLY what's in the spec
  • Must validate against openapi.yaml schema
  • Must return responses matching spec
  • Performance targets from spec x-performance extension

Phase 5: Quality Assurance

Purpose: Comprehensive testing and security audit

Testing Phases:

UNIT Phase:

npm run test:unit
# or
pytest tests/unit/
  • Coverage: >80% required
  • All tests must pass

INTEGRATION Phase:

npm run test:integration
  • Test API endpoints against actual database
  • Test external service integrations
  • Verify data flows end-to-end

E2E Phase:

npx playwright test
# or
npx cypress run
  • Test complete user flows
  • Cross-browser testing
  • Mobile responsive testing

CONTRACT Phase:

npm run test:contract
  • Validate implementation matches OpenAPI spec
  • Test request/response schemas
  • Breaking change detection

SECURITY Phase:

npm audit
npx snyk test
semgrep --config=auto .
  • OWASP Top 10 checks
  • Dependency vulnerabilities
  • Static analysis

PERFORMANCE Phase:

npx k6 run tests/load.js
npx lighthouse http://localhost:3000
  • Load testing: 100 concurrent users for 1 minute
  • Stress testing: 500 concurrent users for 30 seconds
  • P95 response time < 500ms required

ACCESSIBILITY Phase:

npx axe http://localhost:3000
  • WCAG 2.1 AA compliance
  • Alt text, ARIA labels, color contrast
  • Keyboard navigation, focus indicators

REGRESSION Phase:

  • Compare behavior against previous version
  • Verify no features broken by recent changes
  • Test backward compatibility of APIs

UAT Phase:

  • Create acceptance tests from PRD
  • Walk through complete user journeys
  • Verify business logic matches PRD
  • Document any UX friction points

Phase 6: Deployment

Purpose: Release to production

Actions:

  1. Spawn ops-release agent
  2. Generate semantic version, changelog
  3. Create release branch, tag
  4. Deploy to staging, run smoke tests
  5. Blue-green deploy to production
  6. Monitor for 30min, auto-rollback if errors spike

Deployment Strategies:

Blue-Green:

1. Deploy new version to "green" environment
2. Run smoke tests
3. Switch traffic from "blue" to "green"
4. Keep "blue" as rollback target

Canary:

1. Deploy to 5% of traffic
2. Monitor error rates
3. Gradually increase to 25%, 50%, 100%
4. Rollback if errors exceed threshold

Phase 7: Business Operations

Purpose: Non-technical business setup

Actions:

  1. biz-marketing: Create landing page, SEO, content
  2. biz-sales: Set up CRM, outreach templates
  3. biz-finance: Configure billing (Stripe), invoicing
  4. biz-support: Create help docs, chatbot
  5. biz-legal: Generate ToS, privacy policy

Phase 8: Growth Loop

Purpose: Continuous improvement

Cycle:

MONITOR -> ANALYZE -> OPTIMIZE -> DEPLOY -> MONITOR
    |
Customer feedback -> Feature requests -> Backlog
    |
A/B tests -> Winner -> Permanent deploy
    |
Incidents -> RCA -> Prevention -> Deploy fix

Never "Done":

  • Run performance optimizations
  • Add missing test coverage
  • Improve documentation
  • Refactor code smells
  • Update dependencies
  • Enhance user experience
  • Implement A/B test learnings

Final Review (Before Any Deployment)

1. Dispatch 3 reviewers reviewing ENTIRE implementation:
   - code-reviewer: Full codebase quality
   - business-logic-reviewer: All requirements met
   - security-reviewer: Full security audit

2. Aggregate findings across all files
3. Fix Critical/High/Medium issues
4. Re-run all 3 reviewers until all PASS
5. Generate final report in .loki/artifacts/reports/final-review.md
6. Proceed to deployment only after all PASS

Quality Gates Summary

Gate Agent Pass Criteria
Unit Tests eng-qa 100% pass
Integration Tests eng-qa 100% pass
E2E Tests eng-qa 100% pass
Coverage eng-qa > 80%
Linting eng-qa 0 errors
Type Check eng-qa 0 errors
Security Scan ops-security 0 high/critical
Dependency Audit ops-security 0 vulnerabilities
Performance eng-qa p99 < 200ms
Accessibility eng-frontend WCAG 2.1 AA
Load Test ops-devops Handles 10x expected traffic
Chaos Test ops-devops Recovers from failures
Cost Estimate ops-cost Within budget
Legal Review biz-legal Compliant