LLM Inference Cost per Output Token Measure

From GM-RKB
Revision as of 08:24, 12 November 2024 by Gmelli (talk | contribs) (Text replacement - "]]↵----" to "]]. ----")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

A LLM Inference Cost per Output Token Measure is a LLM performance measure that evaluates the computational cost associated with generating each output token during the inference process of a large language model (LLM).



References