Capstone Decision Scenarios - MangaAssist
This capstone document shows how to choose among fine-tuning techniques for MangaAssist. It is a decision playbook, not another model recipe.
Core Question
When MangaAssist quality drops, ask:
Is the failure caused by routing, retrieval, ranking, generation, grounding,
alignment, serving constraints, or stale data?
The answer determines which fine-tuning topic is appropriate.
Scenario 1 - User Gets The Wrong Workflow
Example:
"I need a refund for a damaged volume."
Wrong behavior:
- routed to
faqorproduct_question.
Use:
- intent classifier error analysis,
- calibration and margin policy,
- business-weighted error scoring,
- multi-intent detection if both refund and damaged item apply,
- sentiment/escalation detector if frustration is present.
Do not start with LoRA. The response model cannot fix a wrong route reliably.
Scenario 2 - User Gets Bad Manga Results
Example:
"Give me quiet healing manga like Yokohama Kaidashi Kikou."
Wrong behavior:
- returns generic best sellers.
Use:
- embedding adapter,
- hard-negative mining,
- cross-encoder reranker,
- data curation for tone and theme labels,
- interpretability on embedding neighborhoods.
Scenario 3 - Answer Uses Wrong Facts
Example:
"Is this box set complete?"
Wrong behavior:
- model invents availability or volume count.
Use:
- retrieval quality check,
- RAFT if context is retrieved but ignored,
- data freshness checks,
- DPO only after factual grounding is stable.
Do not teach volatile catalog facts into model weights unless the use case truly needs it.
Scenario 4 - Answer Feels Off-Brand
Wrong behavior:
- too formal,
- too verbose,
- not spoiler-safe,
- ignores user constraints.
Use:
- prompt or prefix tuning for light behavior control,
- LoRA for deeper domain behavior,
- DPO preference alignment,
- human evaluation with a MangaAssist rubric.
Scenario 5 - Model Is Good But Too Expensive
Use:
- quantization-aware training,
- knowledge distillation,
- smaller reranker candidate count,
- MoE or adapter routing only if specialization reduces total cost.
Cost should be evaluated with quality gates, not raw dollars alone.
Technique Selection Table
| Failure | First tool | Second tool |
|---|---|---|
| wrong route | intent tuning and calibration | multi-task or multi-intent |
| unknown intent | OOD clustering | few-shot intent addition |
| weak search | embedding adapter | reranker |
| right context, wrong answer | RAFT | DPO |
| off-brand response | prompt/prefix tuning | LoRA or DPO |
| stale behavior | continual learning | data curation |
| too slow | QAT | distillation |
| many specialist workflows | adapter routing | MoE |
Final Promotion Checklist
[ ] failure source identified
[ ] dataset matches failure
[ ] baseline and candidate compared
[ ] critical slices reviewed
[ ] latency and cost measured
[ ] business-weighted risk checked
[ ] shadow deployment completed
[ ] rollback path documented
Final Decision
The best MangaAssist fine-tuning strategy is usually a sequence, not a single trick: route correctly, retrieve correctly, ground the answer, align the response, then optimize cost. This keeps the chatbot useful without turning every problem into model-weight surgery.