get-shit-done/references/thinking-models-research.md
Structured reasoning models for the researcher and synthesizer agents. Apply these at decision points during research and synthesis, not continuously. Each model counters a specific documented failure mode.
Source: Curated from thinking-partner model catalog (150+ models). Selected for direct applicability to GSD research workflow.
First Principles and Steel Man both expand scope -- run First Principles FIRST (decompose the problem), then Steel Man (strengthen alternatives). Don't run simultaneously.
Counters: Accepting surface-level explanations without decomposing into fundamental components.
Before accepting any technology recommendation or architectural pattern, decompose it to its fundamental constraints: What problem does this solve? What are the non-negotiable requirements? What are the physical/logical limits? Build your recommendation UP from these constraints rather than DOWN from conventional wisdom. If you cannot explain WHY a recommendation is correct from first principles, flag it as [LOW] regardless of source count.
Counters: Synthesizer aggregating conflicting research without checking for confounding splits.
When combining findings from multiple research documents that show contradictory results, check whether the contradiction disappears when you split by a hidden variable: framework version, deployment target, project scale, or use case category. A library that benchmarks faster overall may be slower for YOUR specific workload. Before resolving contradictions by majority vote, ask: "Is there a subgroup split that explains why both findings are correct in their own context?"
Counters: Only finding successful examples while missing failures and abandoned approaches.
After gathering evidence FOR a recommended approach, actively search for projects that ABANDONED it. Check GitHub issues for "migrated away from", "replaced X with", or "problems with X at scale". A technology with 10 success stories and 100 quiet failures looks great until you check the graveyard. Weight negative evidence (migration-away stories, deprecation notices, unresolved issues) MORE heavily than positive evidence -- failures are underreported.
Counters: Searching for evidence that confirms initial hypothesis while ignoring disconfirming evidence.
After forming your initial recommendation, spend one full research cycle searching AGAINST it. Use search terms like "{technology} problems", "{technology} alternatives", "why not {technology}", "{technology} vs {competitor}". For each piece of disconfirming evidence found, either (a) refute it with higher-confidence sources, or (b) add it as a caveat to your recommendation. If you cannot find ANY criticism of your recommendation, your search was too narrow -- widen it.
Counters: Dismissing alternative approaches without giving them their strongest possible form.
Before recommending against an alternative technology or approach, construct its STRONGEST possible case. What would a passionate advocate say? What use cases does it serve better than your recommendation? What trade-offs favor it? Present the steel-manned alternative alongside your recommendation with an honest comparison. If the steel-manned alternative is competitive, flag the decision as [NEEDS DECISION] rather than making a unilateral recommendation.
Skip structured reasoning models when the situation does not benefit from them: