Understanding LLM Temperature and Other Parameters
Was this section helpful?
The Curious Case of Neural Text Degeneration, Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, Yejin Choi, 2020International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1904.09751 - Introduces nucleus sampling (top-p) and analyzes its effectiveness compared to other decoding strategies for neural text generation, including temperature and top-k.
API Reference: Chat Completions, OpenAI, 2024 (OpenAI) - Official documentation detailing the generation parameters available in OpenAI's API, including temperature, top-p, max_tokens, and various penalties.
Speech and Language Processing (3rd ed. draft), Daniel Jurafsky and James H. Martin, 2025 - A standard reference for natural language processing, providing comprehensive background on language models, text generation, and the mechanics of token probability and sampling.
Generation Strategies for Language Models, Hugging Face, 2024 - Practical guide to various text generation strategies implemented in the Transformers library, offering clear explanations and examples for parameters like temperature, top-k, and top-p.