Photo by Diana Polekhina on Unsplash
Table with LLMs Token Output Limit
A table with information on Context Window and Maximum Output Tokens of multiple large language models (LLMs)
Summary: The table provides a comparison of Context Window and Maximum Output Tokens of various large language models (LLMs) including Claude 3, GPT-4 Turbo, Gemini, Mixtral, and Llama.
Every model has a limit to the number of tokens in the input prompt, normally called "Context Window." Additionally, each model has a limit on the number of tokens that it can generate in the output. This limit is sometimes called "maximum new tokens" or "maximum output tokens."
Let me know if you want me to add more information to the table. Other columns that might be useful, or other LLMs.