Estimating Text Temperature
By: Nikolay Mikhaylovskiy
Potential Business Impact:
Finds how "human-like" computer writing is.
Autoregressive language models typically use temperature parameter at inference to shape the probability distribution and control the randomness of the text generated. After the text was generated, this parameter can be estimated using maximum likelihood approach. Following it, we propose a procedure to estimate the temperature of any text, including ones written by humans, with respect to a given language model. We evaluate the temperature estimation capability of a wide selection of small-to-medium LLMs. We then use the best-performing Qwen3 14B to estimate temperatures of popular corpora.
Similar Papers
Exploring the Impact of Temperature on Large Language Models:Hot or Cold?
Computation and Language
Makes AI smarter by picking the best "thinking" speed.
TempTest: Local Normalization Distortion and the Detection of Machine-generated Text
Computation and Language
Finds fake writing even when it looks real.
Optimizing Temperature for Language Models with Multi-Sample Inference
Machine Learning (CS)
Finds best settings for AI to be smarter.