Skip to content

Conversation

@codgician
Copy link
Contributor

@codgician codgician commented Dec 12, 2025

Title

Adds github_copilot model info into model_prices_and_context_window.json, specifying max token limits and model capabilities. Values are based on the response of GET https://api.enterprise.githubcopilot.com/models.

Relevant issues

Pre-Submission checklist

Please complete all items before asking a LiteLLM maintainer to review your PR

  • [-] I have Added testing in the tests/litellm/ directory, Adding at least 1 test is a hard requirement - see details
  • [-] I have added a screenshot of my new test passing locally
  • [-] My PR passes all unit tests on make test-unit
  • My PR's scope is as isolated as possible, it only solves 1 specific problem

(No code change)

Type

🆕 New Feature

Changes

  • Adds github_copilot model info into model_prices_and_context_window.json.

Attaching the raw response of GitHub Copilot API on my side:

{
    "data": [
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-5-mini",
                "limits": {
                    "max_context_window_tokens": 264000,
                    "max_output_tokens": 64000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-5-mini",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "lightweight",
            "model_picker_enabled": true,
            "name": "GPT-5 mini",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-5 mini model from OpenAI. [Learn more about how GitHub Copilot serves GPT-5 mini](https://gh.io/copilot-openai)."
            },
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-5-mini"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "gpt-5",
                "limits": {
                    "max_context_window_tokens": 400000,
                    "max_output_tokens": 128000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-5",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "GPT-5",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-5 model from OpenAI. [Learn more about how GitHub Copilot serves GPT-5](https://gh.io/copilot-openai)."
            },
            "preview": false,
            "supported_endpoints": [
                "/chat/completions",
                "/responses"
            ],
            "vendor": "Azure OpenAI",
            "version": "gpt-5"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o-mini",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 64000
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4o-mini-2024-07-18",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o mini",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-mini-2024-07-18"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 16384,
                    "max_prompt_tokens": 64000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4o-2024-11-20",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-2024-11-20"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 16384,
                    "max_prompt_tokens": 64000
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4o-2024-08-06",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-2024-08-06"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "gpt-5.1",
                "limits": {
                    "max_context_window_tokens": 264000,
                    "max_output_tokens": 64000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-5.1",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "GPT-5.1",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-5.1 model from OpenAI. [Learn more about how GitHub Copilot serves GPT-5.1](https://gh.io/copilot-openai)."
            },
            "preview": true,
            "supported_endpoints": [
                "/chat/completions",
                "/responses"
            ],
            "vendor": "OpenAI",
            "version": "gpt-5.1"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "gpt-5.1-codex-max",
                "limits": {
                    "max_context_window_tokens": 400000,
                    "max_output_tokens": 128000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-5.1-codex-max",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "powerful",
            "model_picker_enabled": true,
            "name": "GPT-5.1-Codex-Max (Preview)",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-5.1-Codex-Max model from OpenAI. [Learn more about how GitHub Copilot serves GPT-5.1-Codex-Max](https://gh.io/copilot-openai)."
            },
            "preview": true,
            "supported_endpoints": [
                "/responses"
            ],
            "vendor": "OpenAI",
            "version": "gpt-5.1-codex-max"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "text-embedding-3-small",
                "limits": {
                    "max_inputs": 512
                },
                "object": "model_capabilities",
                "supports": {
                    "dimensions": true
                },
                "tokenizer": "cl100k_base",
                "type": "embeddings"
            },
            "id": "text-embedding-3-small",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "Embedding V3 small",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "text-embedding-3-small"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "text-embedding-3-small",
                "object": "model_capabilities",
                "supports": {
                    "dimensions": true
                },
                "tokenizer": "cl100k_base",
                "type": "embeddings"
            },
            "id": "text-embedding-3-small-inference",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "Embedding V3 small (Inference)",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "text-embedding-3-small"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "claude-sonnet-4",
                "limits": {
                    "max_context_window_tokens": 216000,
                    "max_output_tokens": 16000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 5,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32000,
                    "min_thinking_budget": 1024,
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "claude-sonnet-4",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "Claude Sonnet 4",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Claude Sonnet 4 model from Anthropic. [Learn more about how GitHub Copilot serves Claude Sonnet 4](https://docs.github.com/en/copilot/using-github-copilot/ai-models/using-claude-sonnet-in-github-copilot)."
            },
            "preview": false,
            "supported_endpoints": [
                "/chat/completions"
            ],
            "vendor": "Anthropic",
            "version": "claude-sonnet-4"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "claude-sonnet-4.5",
                "limits": {
                    "max_context_window_tokens": 144000,
                    "max_output_tokens": 16000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 5,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32000,
                    "min_thinking_budget": 1024,
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "claude-sonnet-4.5",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "Claude Sonnet 4.5",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Claude Sonnet 4.5 model from Anthropic. [Learn more about how GitHub Copilot serves Claude Sonnet 4.5](https://docs.github.com/en/copilot/using-github-copilot/ai-models/using-claude-sonnet-in-github-copilot)."
            },
            "preview": false,
            "supported_endpoints": [
                "/chat/completions"
            ],
            "vendor": "Anthropic",
            "version": "claude-sonnet-4.5"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 3,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "claude-opus-4.5",
                "limits": {
                    "max_context_window_tokens": 144000,
                    "max_output_tokens": 16000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 5,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32000,
                    "min_thinking_budget": 1024,
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "claude-opus-4.5",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "powerful",
            "model_picker_enabled": true,
            "name": "Claude Opus 4.5 (Preview)",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Claude Opus 4.5 model from Anthropic. [Learn more about how GitHub Copilot serves Claude Opus 4.5](https://gh.io/copilot-anthropic)."
            },
            "preview": true,
            "supported_endpoints": [
                "/chat/completions"
            ],
            "vendor": "Anthropic",
            "version": "claude-opus-4.5"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 10,
                "restricted_to": [
                    "pro_plus",
                    "max",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "claude-opus-4.1",
                "limits": {
                    "max_context_window_tokens": 80000,
                    "max_output_tokens": 16000,
                    "max_prompt_tokens": 80000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32000,
                    "min_thinking_budget": 1024,
                    "streaming": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "claude-opus-41",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "powerful",
            "model_picker_enabled": true,
            "name": "Claude Opus 4.1",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Claude Opus 4.1 model from Anthropic. [Learn more about how GitHub Copilot serves Claude Opus 4.1](https://docs.github.com/en/copilot/using-github-copilot/ai-models/using-claude-sonnet-in-github-copilot)."
            },
            "preview": false,
            "supported_endpoints": [
                "/chat/completions"
            ],
            "vendor": "Anthropic",
            "version": "claude-opus-41"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 0.33
            },
            "capabilities": {
                "family": "claude-haiku-4.5",
                "limits": {
                    "max_context_window_tokens": 144000,
                    "max_output_tokens": 16000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 5,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32000,
                    "min_thinking_budget": 1024,
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "claude-haiku-4.5",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "Claude Haiku 4.5",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Claude Haiku 4.5 model from Anthropic. [Learn more about how GitHub Copilot serves Claude Haiku 4.5](https://gh.io/copilot-anthropic)."
            },
            "preview": false,
            "supported_endpoints": [
                "/chat/completions"
            ],
            "vendor": "Anthropic",
            "version": "claude-haiku-4.5"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "gemini-3-pro",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 64000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/heic",
                            "image/heif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32000,
                    "min_thinking_budget": 258,
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gemini-3-pro-preview",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "powerful",
            "model_picker_enabled": true,
            "name": "Gemini 3 Pro (Preview)",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Gemini 3 Pro model from Google. [Learn more about how GitHub Copilot serves Gemini 3 Pro](https://docs.github.com/en/copilot/reference/ai-models/model-hosting#google-models)."
            },
            "preview": true,
            "vendor": "Google",
            "version": "gemini-3-pro-preview"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "gemini-2.5-pro",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 64000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/heic",
                            "image/heif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "max_thinking_budget": 32768,
                    "min_thinking_budget": 128,
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gemini-2.5-pro",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "powerful",
            "model_picker_enabled": true,
            "name": "Gemini 2.5 Pro",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest Gemini 2.5 Pro model from Google. [Learn more about how GitHub Copilot serves Gemini 2.5 Pro](https://docs.github.com/en/copilot/using-github-copilot/ai-models/choosing-the-right-ai-model-for-your-task#gemini-25-pro)."
            },
            "preview": false,
            "vendor": "Google",
            "version": "gemini-2.5-pro"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4.1",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 16384,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4.1-2025-04-14",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4.1",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-4.1 model from OpenAI. [Learn more about how GitHub Copilot serves GPT-4.1](https://docs.github.com/en/copilot/using-github-copilot/ai-models/choosing-the-right-ai-model-for-your-task#gpt-41)."
            },
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4.1-2025-04-14"
        },
        {
            "billing": {
                "is_premium": true,
                "multiplier": 1,
                "restricted_to": [
                    "pro",
                    "pro_plus",
                    "max",
                    "business",
                    "enterprise"
                ]
            },
            "capabilities": {
                "family": "gpt-5.2",
                "limits": {
                    "max_context_window_tokens": 264000,
                    "max_output_tokens": 64000,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-5.2",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "GPT-5.2 (Preview)",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-5.2 model from OpenAI. [Learn more about how GitHub Copilot serves GPT-5.2](https://gh.io/copilot-openai)."
            },
            "preview": true,
            "supported_endpoints": [
                "/chat/completions",
                "/responses"
            ],
            "vendor": "OpenAI",
            "version": "gpt-5.2"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4.1",
                "object": "model_capabilities",
                "supports": {
                    "streaming": true
                },
                "tokenizer": "o200k_base",
                "type": "completion"
            },
            "id": "gpt-41-copilot",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "GPT-4.1 Copilot",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-41-copilot"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-3.5-turbo",
                "limits": {
                    "max_context_window_tokens": 16384,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 16384
                },
                "object": "model_capabilities",
                "supports": {
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "cl100k_base",
                "type": "chat"
            },
            "id": "gpt-3.5-turbo-0613",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT 3.5 Turbo",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-3.5-turbo-0613"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4",
                "limits": {
                    "max_context_window_tokens": 32768,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 32768
                },
                "object": "model_capabilities",
                "supports": {
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "cl100k_base",
                "type": "chat"
            },
            "id": "gpt-4",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT 4",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4-0613"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4",
                "limits": {
                    "max_context_window_tokens": 32768,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 32768
                },
                "object": "model_capabilities",
                "supports": {
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "cl100k_base",
                "type": "chat"
            },
            "id": "gpt-4-0613",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT 4",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4-0613"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 64000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4o-2024-05-13",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-2024-05-13"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 64000
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4-o-preview",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-2024-05-13"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4.1",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 16384,
                    "max_prompt_tokens": 128000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "structured_outputs": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4.1",
            "is_chat_default": true,
            "is_chat_fallback": true,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "GPT-4.1",
            "object": "model",
            "policy": {
                "state": "enabled",
                "terms": "Enable access to the latest GPT-4.1 model from OpenAI. [Learn more about how GitHub Copilot serves GPT-4.1](https://docs.github.com/en/copilot/using-github-copilot/ai-models/choosing-the-right-ai-model-for-your-task#gpt-41)."
            },
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4.1-2025-04-14"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-3.5-turbo",
                "limits": {
                    "max_context_window_tokens": 16384,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 16384
                },
                "object": "model_capabilities",
                "supports": {
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "cl100k_base",
                "type": "chat"
            },
            "id": "gpt-3.5-turbo",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT 3.5 Turbo",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-3.5-turbo-0613"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o-mini",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 64000
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4o-mini",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o mini",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-mini-2024-07-18"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4",
                "limits": {
                    "max_context_window_tokens": 32768,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 32768
                },
                "object": "model_capabilities",
                "supports": {
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "cl100k_base",
                "type": "chat"
            },
            "id": "gpt-4",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT 4",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4-0613"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 64000,
                    "vision": {
                        "max_prompt_image_size": 3145728,
                        "max_prompt_images": 1,
                        "supported_media_types": [
                            "image/jpeg",
                            "image/png",
                            "image/webp",
                            "image/gif"
                        ]
                    }
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true,
                    "vision": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4o",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_category": "versatile",
            "model_picker_enabled": true,
            "name": "GPT-4o",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-2024-11-20"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "gpt-4o",
                "limits": {
                    "max_context_window_tokens": 128000,
                    "max_output_tokens": 4096,
                    "max_prompt_tokens": 64000
                },
                "object": "model_capabilities",
                "supports": {
                    "parallel_tool_calls": true,
                    "streaming": true,
                    "tool_calls": true
                },
                "tokenizer": "o200k_base",
                "type": "chat"
            },
            "id": "gpt-4-o-preview",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "GPT-4o",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "gpt-4o-2024-05-13"
        },
        {
            "billing": {
                "is_premium": false,
                "multiplier": 0
            },
            "capabilities": {
                "family": "text-embedding-ada-002",
                "limits": {
                    "max_inputs": 512
                },
                "object": "model_capabilities",
                "supports": {},
                "tokenizer": "cl100k_base",
                "type": "embeddings"
            },
            "id": "text-embedding-ada-002",
            "is_chat_default": false,
            "is_chat_fallback": false,
            "model_picker_enabled": false,
            "name": "Embedding V2 Ada",
            "object": "model",
            "preview": false,
            "vendor": "Azure OpenAI",
            "version": "text-embedding-3-small"
        }
    ],
    "object": "list"
}

@vercel
Copy link

vercel bot commented Dec 12, 2025

The latest updates on your projects. Learn more about Vercel for GitHub.

Project Deployment Preview Comments Updated (UTC)
litellm Ready Ready Preview Comment Dec 12, 2025 4:26am

@krrishdholakia
Copy link
Contributor

@codgician i don't see any cost associated with these models. Why add them to the model cost map?

@codgician
Copy link
Contributor Author

codgician commented Dec 12, 2025

@codgician i don't see any cost associated with these models. Why add them to the model cost map?

GitHub Copilot works on a subscription basis so costs can't be calculated directly. The intention of this PR is that GitHub Copilot models have slightly different max token count compared to the original models. I added these model into the file in hope that the correct token limitations and capabilities are displayed in the Admin UI.

Copy link
Contributor

@ishaan-jaff ishaan-jaff left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants