LLM Inference Cost per Output Token Measure

From GM-RKB
Jump to navigation Jump to search

A LLM Inference Cost per Output Token Measure is a LLM performance measure that evaluates the computational cost associated with generating each output token during the inference process of a large language model (LLM).



References