LLM Prompting
LLM Prompting - In-Context-Learning🔗
[LLMLingua] Compressing Prompts for Accelerated Inference of LLMs🔗
Arxiv: https://arxiv.org/abs/2310.05736 6 Dec 2023 Microsoft
Key Components:
-
Budget Controller
- Allocates different compression ratios to prompt components
- Prioritizes instructions and questions over demonstrations
- Maintains semantic integrity under high compression
-
Token-level Iterative Compression
- Divides target prompt into segments
- Uses smaller model for perplexity distribution
- Concatenates compressed segments for accurate probability estimation
-
Instruction Tuning
- Aligns distribution between language models
- Improves compression quality
[TOT] Tree of Thoughts: Deliberate Problem Solving with LLMs🔗
Arxiv: https://arxiv.org/abs/2305.10601 17 May 2023
Key Features:
- Frames problems as search over a tree
- Each node represents a partial solution
- Four key components:
- Decomposition of intermediate process
- Generation of potential thoughts
- Heuristic evaluation of states
- Search algorithm selection
[COT] Chain-of-Thought Prompting Elicits Reasoning in LLMs🔗
Arxiv: https://arxiv.org/abs/2201.11903 10 Jan 2023
Limitations:
- No guarantee of correct reasoning paths
- Manual annotation costs for few-shot setting
- Potential for both correct and incorrect answers
[Self-Discover] LLM Self-Compose Reasoning Structures🔗
Arxiv: https://arxiv.org/abs/2402.03620 6 Feb 2024
Two-Stage Process:
-
Stage 1: Task-level reasoning structure
- Uses three actions to guide LLM
- Generates coherent reasoning structure
-
Stage 2: Instance solving
- Follows self-discovered structure
- Arrives at final answer
[Intent-based Prompt Calibration] Enhancing prompt optimization with synthetic boundary cases🔗
Arxiv: https://arxiv.org/abs/2402.03099 5 Feb 2024
Process:
- Start with initial prompt and task description
- Iteratively:
- Generate challenging boundary cases
- Evaluate current prompt
- Suggest improved prompt
- Terminate when no improvement or max iterations reached
[Text2SQL Prompting] Enhancing Few-shot Text2SQL Capabilities of LLM🔗
Arxiv: https://arxiv.org/abs/2311.16452 21 May 2023 Yale
Key Findings:
- Dual emphasis on diversity and similarity in examples
- Database knowledge augmentation benefits
- Code sequence representation for databases
- Sensitivity to number of demonstration examples
[MedPrompt] Can Generalist FM Outcompete Special-Purpose Tuning?🔗
Arxiv: https://arxiv.org/abs/2311.16452 28 Nov 2023 Microsoft
Key Features:
- Self-generated chain-of-thought
- Ensembling with self-consistency
- Choice shuffling for bias reduction
- Label verification for hallucination mitigation
[URIAL] Rethinking Alignment via In-Context Learning🔗
Arxiv: https://arxiv.org/abs/2312.01552 4 Dec 2023 Allen Institute
Key Points:
- Tuning-free alignment method
- Requires only three stylistic examples
- Supports Superficial Alignment Hypothesis
- Token distribution analysis shows minimal shifts
[CoVE] Chain-of-Verification Reduces Hallucinations in LLM Models🔗
Arxiv: https://arxiv.org/abs/2309.11495 25 Sep 2023 Meta
Process:
- Draft initial response
- Plan verification questions
- Answer questions independently
- Generate final verified response