Stats about a given LLM model.
Total context window size: input (aka prompt) + output (aka completion) tokens combined.
A friendly name for the model, to display in UIs.
Maximum tokens the model can generate in a single response.
Stats about a given LLM model.