Skip to content

Prompt Engineering Resources

Academic papers, platform guides, and practitioner references organized by technique. Each pattern page links to its most relevant papers; this page collects them all in one place.

These official guides are maintained by the AI platform teams and reflect current best practices.

PlatformGuideNotes
Anthropic (Claude)Prompt Engineering GuideComprehensive guide covering all major techniques
OpenAI (ChatGPT)Prompt Engineering Best PracticesStrategies and tactics for better results
OpenAI (GPT-5.4)Prompt Guidance for GPT-5.4Model-specific guidance for long-running tasks, tool use, and reliable execution
Google CloudPrompt Engineering Overview and GuideOverview of prompt engineering concepts and techniques
ResourceProviderNotes
Prompt Engineering CoursesedXUniversity-backed courses on prompt engineering
Prompt Engineering CoursesCourseraCourses from industry and academic partners
Prompting GuideDAIR.AIOpen-source guide covering techniques, applications, and research
  • Brown et al. 2020 — Language Models are Few-Shot Learnersarxiv.org/abs/2005.14165 — The GPT-3 paper that established the zero-shot paradigm, demonstrating that large language models can perform tasks from instructions alone
  • Kojima et al. 2022 — Large Language Models are Zero-Shot Reasonersarxiv.org/abs/2205.11916 — Showed that adding “Let’s think step by step” enables zero-shot chain-of-thought reasoning
  • Brown et al. 2020 — Language Models are Few-Shot Learnersarxiv.org/abs/2005.14165 — Seminal paper demonstrating few-shot learning via text demonstrations
  • Dong et al. 2022 — A Survey on In-context Learningarxiv.org/abs/2301.00234 — Comprehensive survey on why few-shot demonstrations work and how models learn in context
  • Wei et al. 2022 — Chain-of-Thought Prompting Elicits Reasoning in Large Language Modelsarxiv.org/abs/2201.11903 — The seminal chain-of-thought paper showing step-by-step reasoning improves performance on math, logic, and commonsense tasks
  • Kojima et al. 2022 — Large Language Models are Zero-Shot Reasonersarxiv.org/abs/2205.11916 — Zero-shot CoT: “Let’s think step by step”
  • Yao et al. 2023 — Tree of Thoughts: Deliberate Problem Solving with Large Language Modelsarxiv.org/abs/2305.10601 — Generalizes chain-of-thought into tree-structured reasoning with backtracking
  • Wei et al. 2021 — Finetuned Language Models Are Zero-Shot Learners (FLAN)arxiv.org/abs/2109.01652 — Foundation paper on instruction tuning
  • Ouyang et al. 2022 — Training Language Models to Follow Instructions with Human Feedbackarxiv.org/abs/2203.02155 — InstructGPT paper establishing RLHF for instruction following
  • Zhang et al. 2023 — Instruction Tuning for Large Language Models: A Surveyarxiv.org/abs/2308.10792 — Survey covering instruction tuning methods and their impact
  • Dong et al. 2022 — A Survey on In-context Learningarxiv.org/abs/2301.00234 — Defines the in-context learning paradigm
  • Lewis et al. 2020 — Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasksarxiv.org/abs/2005.11401 — RAG framework for augmenting prompts with retrieved context
  • Kong et al. 2023 — Better Zero-Shot Reasoning with Role-Play Promptingarxiv.org/abs/2308.07702 — Role-play improved ChatGPT accuracy on AQuA from 53.5% to 63.8%
  • Zheng et al. 2023 — When “A Helpful Assistant” Is Not Really Helpfularxiv.org/abs/2311.10054 — Important counterpoint: persona prompts don’t reliably improve performance across all tasks
  • Tam et al. 2024 — Let Me Speak Freely? A Study on the Impact of Format Restrictions on Performancearxiv.org/abs/2408.02442 — Shows format restrictions can degrade reasoning performance
  • Liu et al. 2024 — “We Need Structured Output”: Towards User-centered Constraints on LLM Outputarxiv.org/abs/2404.07362 — User-centered perspective on structured output constraints
  • Yi et al. 2024 — A Survey on Recent Advances in LLM-Based Multi-turn Dialogue Systemsarxiv.org/abs/2402.18013 — Survey covering dialogue management, context tracking, and coherence
  • Zheng et al. 2025 — LLMs Get Lost In Multi-Turn Conversationarxiv.org/abs/2505.06120 — Documents a 39% average performance drop in multi-turn vs. single-turn interactions
  • Wang et al. 2022 — Self-Consistency Improves Chain of Thought Reasoning in Language Modelsarxiv.org/abs/2203.11171 — Sampling diverse reasoning paths and selecting the most consistent answer
  • Shinn et al. 2023 — Reflexion: Language Agents with Verbal Reinforcement Learningarxiv.org/abs/2303.11366 — Verbal self-reflection for iterative improvement
  • Madaan et al. 2023 — Self-Refine: Iterative Refinement with Self-Feedbackarxiv.org/abs/2303.17651 — Generate-critique-refine loop without external feedback
  • Li et al. 2023 — EmotionPrompt: Leveraging Psychology for Large Language Models Enhancement via Emotional Stimulusarxiv.org/abs/2307.11760 — Showed 10%+ improvement on benchmarks using emotional stimuli, though effects vary by model and task
  • Ma et al. 2023 — Query Rewriting for Retrieval-Augmented Large Language Modelsarxiv.org/abs/2305.14283 — Rewrite-Retrieve-Read framework for query reformulation
  • Lenny Rachitsky — Five proven prompt engineering techniqueslennysnewsletter.com — Practitioner origin of the style unbundling technique
  • Liu et al. 2023 — Learning to Generate Text in Arbitrary Writing Stylesarxiv.org/abs/2312.17242 — Academic research on style decomposition and reproduction
  • Adams et al. 2023 — From Sparse to Dense: GPT-4 Summarization with Chain of Density Promptingarxiv.org/abs/2309.04269 — Iterative increasing-density summarization technique
  • Jin et al. 2024 — A Comprehensive Survey on Process-Oriented Automatic Text Summarizationarxiv.org/abs/2403.02901 — Broad survey of summarization approaches

This is primarily a practitioner pattern. The academic literature on constrained generation focuses on technical format constraints rather than business constraints in prompts. See the platform documentation above for practical guidance.

These foundational papers cover topics relevant to multiple techniques.

  • Schulhoff et al. 2024 — The Prompt Report: A Systematic Survey of Prompting Techniquesarxiv.org/abs/2406.06608 — Comprehensive taxonomy of 58 prompting techniques with a unified terminology
  • Zhou et al. 2022 — Large Language Models Are Human-Level Prompt Engineersarxiv.org/abs/2211.01910 — Automatic prompt optimization (APE)
  • White et al. 2023 — A Prompt Pattern Catalog to Enhance Prompt Engineering with ChatGPTarxiv.org/abs/2302.11382 — Pattern-based approach to prompt engineering, similar to software design patterns