Retour aux articles
5 MIN READ

Temperature & Top-P: Controlling AI Creativity

By Learnia Team

Temperature & Top-P: Controlling AI Creativity

This article is written in English. Our training modules are available in French.

Ever noticed how ChatGPT sometimes gives creative, varied responses and other times stays strictly factual? That's not random—it's controlled by two parameters: Temperature and Top-P. Understanding them gives you precise control over AI behavior.


What Is Temperature?

Temperature controls the randomness of AI responses. It determines how likely the model is to choose unexpected words.

The Scale

0.0 ─────────────────────────────── 2.0
Deterministic                    Chaotic
Predictable                      Creative
Focused                          Random

Low Temperature (0.0 - 0.3)

The AI picks the most probable next word almost every time:

Temperature = 0
"The capital of France is ___"
→ "Paris" (99.9% of the time)

Best for: Factual answers, data extraction, code generation

Medium Temperature (0.4 - 0.7)

Balanced between predictability and variety:

Temperature = 0.5
"Write a greeting"
→ "Hello! How can I help you today?"
→ "Hi there! What brings you here?"
→ "Good day! How may I assist?"

Best for: General writing, emails, documentation

High Temperature (0.8 - 1.5)

More creative, unexpected choices:

Temperature = 1.2
"Write a creative opening"
→ "The moon whispered secrets to the tide..."
→ "Three crows sat on a digital wire..."
→ "Everything changed when the coffee machine became sentient..."

Best for: Creative writing, brainstorming, storytelling


What Is Top-P (Nucleus Sampling)?

Top-P is a different approach: instead of controlling randomness directly, it limits which words the AI can even consider.

How Top-P Works

The AI ranks all possible next words by probability:

Possible words: "Paris" (70%), "Lyon" (15%), "France" (8%), "Marseille" (5%), ...

Top-P = 0.85 → Only considers words until cumulative probability reaches 85%
→ Can choose from: "Paris", "Lyon"
→ Ignores: "France", "Marseille", and everything else

Top-P Values

0.1 → Only the single most likely word
0.5 → Top ~50% probability mass
0.9 → Most words included (default for most APIs)
1.0 → All words possible

Temperature vs Top-P: What's the Difference?

| Aspect | Temperature | Top-P | |--------|-------------|-------| | Controls | Selection randomness | Candidate pool size | | Mechanism | Scales probabilities | Filters options | | Low value | Always pick top choice | Fewer options | | High value | More random picks | More options |

A Simple Analogy

Imagine picking a restaurant:

Temperature = How adventurous your choice is

  • Low: Always pick your favorite
  • High: Might try something completely new

Top-P = Which restaurants are even on the list

  • Low: Only consider top-rated places
  • High: Consider any restaurant in town

Common Use Cases

Factual Q&A / Data Extraction

Temperature: 0.0 - 0.2
Top-P: 0.9 (or even lower)

You want consistency and accuracy:

"Extract the date from: Meeting scheduled for March 15, 2025"
→ Should always return "March 15, 2025"

Professional Writing

Temperature: 0.4 - 0.6
Top-P: 0.85 - 0.95

Balance quality with some variety:

"Draft a professional email declining a meeting request"
→ Natural variation while staying appropriate

Creative Writing

Temperature: 0.8 - 1.2
Top-P: 0.95 - 1.0

Encourage novelty and surprise:

"Write a creative story opening about time travel"
→ Unique, unexpected approaches

Code Generation

Temperature: 0.0 - 0.2
Top-P: 0.9

Code needs to be correct, not creative:

"Write a Python function to calculate factorial"
→ Standard, working implementation

Brainstorming

Temperature: 1.0 - 1.5
Top-P: 0.95

Maximize variety and unexpected ideas:

"Give me 10 creative product name ideas"
→ Wild, diverse suggestions

The Temperature/Top-P Matrix

| | Low Top-P (<0.5) | High Top-P (>0.9) | |---|---|---| | Low Temp (0-0.3) | Very focused, repetitive | Focused with slight variation | | High Temp (0.8+) | Somewhat creative | Highly creative, unpredictable |

Most APIs default to:

Temperature: 0.7, Top-P: 0.9


Practical Tips

1. Adjust One at a Time

Don't change both simultaneously—it's hard to understand the effect:

Step 1: Set Top-P to 0.9 (neutral)
Step 2: Adjust Temperature to find sweet spot

2. Match to Task Criticality

High stakes (legal, medical) → Low temperature
Low stakes (brainstorming) → Higher temperature

3. Test with the Same Prompt

Run the same prompt 5 times to see consistency:

Temperature 0.0 → Same output 5/5 times
Temperature 0.7 → Similar outputs with variation
Temperature 1.2 → Very different each time

4. Document Your Settings

When you find settings that work, save them:

{
  "use_case": "Customer support responses",
  "temperature": 0.3,
  "top_p": 0.9,
  "notes": "Professional, consistent tone"
}

Common Mistakes

1. Temperature Too High for Facts

Temperature: 1.5
"What year was the Eiffel Tower built?"
→ "1889" or "1887" or "around 1890" 😕

2. Temperature Too Low for Creativity

Temperature: 0.0
"Write a creative story"
→ Same generic story every time

3. Ignoring These Settings Entirely

Default values work often, but not always. Tune them for your use case.


Key Takeaways

  1. Temperature controls response randomness (0.0 = focused, 1.0+ = creative)
  2. Top-P filters which words are even considered
  3. Low settings for facts, code, extraction
  4. High settings for creativity, brainstorming
  5. Test and tune for your specific use case

Ready to Master LLM Parameters?

This article covered the what and why of Temperature and Top-P. But effective AI applications require understanding the full range of parameters and techniques.

In our Module 1 — Fundamentals of Prompt Engineering, you'll learn:

  • Complete parameter reference (Temperature, Top-P, Max Tokens)
  • How token prediction actually works
  • Context window management
  • Practical configuration for different use cases

Explore Module 1: Fundamentals

GO DEEPER

Module 1 — LLM Anatomy & Prompt Structure

Understand how LLMs work and construct clear, reusable prompts.