01: Task 1.1 Analyze Requirements and Design GenAI Solutions
AIP-C01 Mapping
Content Domain 1: Foundation Model Integration, Data Management, and Compliance Task 1.1: Analyze requirements and design GenAI solutions.
Task Goal
Turn ambiguous business goals into an architecture that is feasible, measurable, secure, and maintainable. In practice, this means deciding when the solution should be prompt-only, RAG-based, agentic, or customized, then validating those decisions before large-scale rollout.
Task User Story
As a GenAI solutions architect for MangaAssist, I want to translate product goals, operational constraints, and compliance requirements into a production-ready AWS architecture, So that the business gets a credible path from idea to launch instead of a demo that collapses under real traffic, governance, or cost pressure.
Task Architecture View
graph TD
A[Business Problem] --> B[Requirement Discovery]
B --> C[Capability Map]
C --> D{Pattern Choice}
D -->|Prompt-only| E[Simple FM Integration]
D -->|RAG| F[Retriever + Vector Store]
D -->|Agentic| G[Tools + Orchestrator]
D -->|Customization| H[Fine-Tuned or Adapted Model]
E --> I[POC Hypotheses]
F --> I
G --> I
H --> I
I --> J[Measure Quality Cost Latency Risk]
J --> K[Reference Architecture]
K --> L[Reusable Standards and Guardrails]
L --> M[Implementation Backlog]
Scenario Packs
Task 1.1 now uses split scenario packs instead of one oversized runbook file per skill.
| Skill | Scenario pack | Focus |
|---|---|---|
| 1.1.1 Comprehensive architectural design | Skill-1.1.1-Comprehensive-Architectural-Design/03-scenarios-and-runbooks.md | Architecture mistakes that create cost, latency, and context failures |
| 1.1.2 Technical proof-of-concept | Skill-1.1.2-Technical-Proof-of-Concept/03-scenarios-and-runbooks.md | PoC gaps that look acceptable in demos but fail in production |
| 1.1.3 Standardized technical components | Skill-1.1.3-Standardized-Technical-Components/03-scenarios-and-runbooks.md | Shared-platform failures caused by drift in clients, prompts, metrics, contracts, and IAM |
Each scenario file now includes design-review questions so you can use the pack for exam prep, architecture reviews, and interview drills.
Skill 1.1.1: Create Comprehensive Architectural Designs
User Story
As a principal solutions architect, I want to design a GenAI architecture that reflects business outcomes, user journeys, system dependencies, and technical limits, So that engineering teams build the right system the first time instead of overbuilding or under-designing the platform.
Deep Dive
The hardest part of GenAI architecture is not drawing boxes. It is making the right early decisions about where the intelligence should live.
| Design Dimension | Questions to Ask | Strong Design Choice |
|---|---|---|
| Business intent | Is the task generation, retrieval, classification, or tool orchestration? | Match the simplest capable pattern to the use case |
| Model role | Should the model answer directly, ground itself on enterprise data, or act as a planner? | Separate generation from decisioning whenever possible |
| Risk profile | What happens if the answer is wrong, late, expensive, or unsafe? | Use stricter controls for customer-facing, financial, or regulated flows |
| Operational constraints | What are the latency, throughput, availability, and region requirements? | Pick architectures that preserve headroom, not just happy-path demos |
| Compliance | Is there PII, internal policy content, or auditability needs? | Build isolation, logging, and traceability into the first design |
For MangaAssist, a strong design would avoid a single giant "chatbot" box. Instead it would split the system into:
- Intent and routing layer for deciding whether the request needs retrieval, tools, or direct generation.
- Retrieval layer for product, policy, and support knowledge.
- Generation layer using Amazon Bedrock for response synthesis.
- Operational control layer for caching, guardrails, monitoring, and evaluation.
Architecture Deliverables
- Context diagram showing user entry points, Bedrock, data sources, and safety boundaries
- Capability map separating direct FM use, RAG, tool calling, and escalation
- NFR matrix covering latency, cost, availability, privacy, and auditability
- Decision log explaining why Bedrock-native models, custom models, or hybrid patterns were chosen
- Rollout plan describing MVP vs later-stage expansion
Acceptance Signals
- Architecture clearly maps every business capability to a runtime pattern
- Constraints such as latency budget, region, cost ceiling, and compliance are explicit
- The design identifies failure boundaries and fallback paths early
- Teams can derive implementation epics directly from the architecture
Skill 1.1.2: Develop Technical Proof-of-Concept Implementations
User Story
As a GenAI program lead, I want to run focused proof-of-concept experiments on Amazon Bedrock before full-scale delivery, So that we validate feasibility, value, and operating characteristics with evidence instead of intuition.
Deep Dive
A good POC is not a miniature product. It is a disciplined experiment that removes the largest unknowns first.
| Hypothesis Type | Example for MangaAssist | POC Metric |
|---|---|---|
| Quality | Bedrock RAG can answer shipping-policy questions with grounded accuracy | Answer accuracy, hallucination rate, citation coverage |
| Latency | Bedrock + OpenSearch can stay under 2 seconds at p95 | End-to-end latency, first-token latency |
| Cost | Dynamic routing between Haiku and Sonnet reduces cost without hurting resolution | Cost per resolved session, model-routing distribution |
| Business value | GenAI-assisted support reduces human-agent load | Deflection rate, resolution rate, CSAT |
A Bedrock POC should usually include:
- A small but representative golden dataset
- One baseline design and one challenger design
- Instrumentation for latency, token usage, failure categories, and qualitative review
- Explicit exit criteria, including "do not proceed" conditions
POC Work Breakdown
- Define the top 2-3 questions the program must answer.
- Build only the minimum architecture needed to test those questions.
- Capture outputs, logs, latency, and cost at request level.
- Review results with both technical and business stakeholders.
- Convert findings into a go/no-go decision or a narrowed next phase.
Common POC Failures
- Using perfect handpicked examples instead of realistic user traffic
- Measuring only demo quality while ignoring cost and latency
- Treating a prompt success as proof that the operating model is production-ready
- Skipping negative cases such as harmful inputs, partial data, or retrieval misses
Acceptance Signals
- Each POC is tied to explicit business and technical hypotheses
- Results include success cases, failure cases, and operating constraints
- The team can say what should scale, what should change, and what should be rejected
- The POC produces reusable artifacts such as prompt templates, test sets, and telemetry hooks
Skill 1.1.3: Create Standardized Technical Components
User Story
As a platform architect, I want to define reusable GenAI building blocks and governance standards, So that multiple teams can deliver consistent, safe, and supportable solutions without reinventing the core platform each time.
Deep Dive
Standardization is how a one-team success becomes an organizational capability.
| Standard Component | Why It Matters | Typical AWS Realization |
|---|---|---|
| Model access wrapper | Prevents every team from hardcoding providers and request shapes | API Gateway + Lambda or shared SDK abstraction |
| Prompt template catalog | Keeps prompts versioned and reviewable | Bedrock Prompt Management + S3-backed repository |
| Guardrail policy | Centralizes safety and abuse controls | Bedrock Guardrails + policy documentation |
| Telemetry package | Makes cost, latency, and quality measurable everywhere | CloudWatch EMF, X-Ray, shared logging schema |
| Evaluation hooks | Enables regression testing before rollout | Golden datasets, batch evaluation runners |
| Reference deployment pattern | Creates repeatable rollout and rollback | IaC modules, CI/CD templates, AppConfig feature flags |
Use the AWS Well-Architected Framework and the Generative AI Lens as a review checklist, not a paperwork exercise. They help ask the right questions:
- Is the system resilient when a model endpoint or region is degraded?
- Is the data pipeline observable and governed?
- Are prompts, models, and retrieval indexes versioned?
- Is there a clear mechanism for rollback and audit?
Standardization Artifacts
- Reference architecture diagrams
- Reusable IaC modules for model routing, guardrails, and logging
- Shared prompt and evaluation templates
- Security and privacy control checklists
- Design-review scorecards aligned to the Well-Architected pillars
Acceptance Signals
- New GenAI projects start from a standard blueprint instead of a blank page
- Teams use common metrics, logging shapes, and deployment gates
- Design reviews can quickly spot deviations from approved patterns
- Platform teams can upgrade or harden shared components centrally
Intuition Gained After Task 1.1
The biggest lesson from Task 1.1 is that GenAI architecture is decision architecture. The real work is choosing the right operating pattern for the problem, not adding an LLM to every step.
You also learn that a POC is valuable only when it reduces uncertainty. If it does not produce a sharper decision about quality, cost, latency, safety, or rollout risk, it is only a demo.
Finally, standardization changes the scale of your impact. A strong architect does not just design one solution. They create reusable components, review habits, and operating guardrails that let many teams build well at the same time.