Correct answer: DThe few-shot vs instruction-only vs fine-tuning decision is quantitative, not stylistic. The constraints given (timeline, data volume, cost budget, distribution stability) are necessary but not sufficient. The missing information is: what accuracy gap exists between each method? What is the downstream cost if the classifier is wrong (expensive refund mistakes vs cheap labeling errors)? With 5,000 labeled examples and stable distribution, fine-tuning is technically feasible and likely optimal for sustained accuracy. Few-shot is fastest to ship but may not close an accuracy gap. Instruction-only is lean but sacrifices exemplar signal. Without knowing the actual accuracy requirements and failure costs, you cannot choose. Option D forces the learner to articulate that the decision framework requires quantitative inputs, not just constraints.
Why the other choices are wrong:- A. Few-shot is fastest, but 'fastest to implement' is not a valid engineering criterion. If you need 95% accuracy and few-shot only achieves 88%, the decision changes.
- B. Instruction-only is lean on tokens, but defining labels is not a substitute for exemplars, especially on ambivalent cases.
- C. Fine-tuning is a valid choice given the data volume and stable distribution, but calling it 'better generalization' is imprecise. The choice depends on accuracy-vs-cost tradeoffs, which must be quantified.