A model in Galtea represents a Large Language Model (LLM) configuration with associated cost information. This allows the platform to track, estimate, and report on the costs of using different LLMs across your products and evaluations.
Models are organization-wide and can be referenced across multiple products to ensure consistent cost tracking.
You can create, view and manage your models on the Galtea dashboard.
The cost in dollars per input token. This is the rate charged by the provider for tokens in your prompts.
Example: 0.00001 (representing $0.00001 per token)
The cost in dollars per output token. This is the rate charged by the provider for tokens in the model’s responses.
Example: 0.00003 (representing $0.00003 per token)
The cost in dollars per cached input token. Some providers offer reduced rates for cached requests.
Example: 0.000005 (representing $0.000005 per token)
The provider of the tokenizer used by the model. This is important for accurate token counting and cost estimation.
Right now, the only supported tokenizers are:
Examples:
The provider or source of the pricing model. This can be a URL to the model’s pricing page or documentation.
For instance: https://openai.com/api/pricing/