LLM Inference-Time Scaling Law

From GM-RKB
Jump to navigation Jump to search

A LLM Inference-Time Scaling Law is an LLM scaling law that describes how language model performance relates to inference compute, model size, and inference strategy choices during model inference.



References

2024-01-04

[1] https://proceedings.mlr.press/v202/dettmers23a.html
[2] https://arxiv.org/html/2408.00724v2
[3] https://bohrium.dp.tech/paper/arxiv/2410.16377
[4] https://openreview.net/forum?id=j7DZWSc8qu
[5] https://openreview.net/forum?id=VNckp7JEHn
[6] https://arxiv.org/abs/2410.16377
[7] https://www.interconnects.ai/p/openai-strawberry-and-inference-scaling-laws
[8] https://openai.com/index/scaling-laws-for-neural-language-models/
[9] https://akashbajwa.substack.com/p/inference-time-scaling-laws
[10] https://epoch.ai/blog/scaling-laws-literature-review

2021