Skip to main content

Changing the AI model for Copilot Chat

Learn how to change the default LLM for Copilot Chat to a different model.

By default, Copilot Chat uses OpenAI's GPT-4o large language model. This is a highly proficient model that performs well for text generation tasks, such as summarization and knowledge-based chat. The model is also capable of reasoning, solving complex math problems and coding.

However, you are not limited to using this model. You can choose from a selection of other models, each with its own particular strengths. You may have a favorite model that you like to use, or you might prefer to use a particular model for inquiring about a specific subject.

Copilot allows you to change the model during a chat and have the alternative model used to generate responses to your prompts.

Changing the model that's used by Copilot Chat does not affect the model that's used for Copilot code completion. See Changing the AI model for Copilot code completion.

Note

Multiple model support in Copilot Chat is in public preview and is subject to change.

AI models for Copilot Chat

The following models are currently available through multi-model Copilot Chat:

  • GPT-4o: This is the default Copilot Chat model. It is a versatile, multimodal model that excels in both text and image processing and is designed to provide fast, reliable responses. It also has superior performance in non-English languages. Learn more about the model's capabilities and review the model card. GPT-4o is hosted on Azure.
  • Claude Sonnet 3.7: This model, likes its predecessor, excels across the software development lifecycle, from initial design to bug fixes, maintenance to optimizations. It also has thinking capabilities which can be enabled by selecting the thinking version of the model, which can be particularly useful in agentic scenarios. Learn more about the model's capabilities or read the model card. GitHub Copilot uses Claude Sonnet hosted on Amazon Web Services.
  • Claude Sonnet 3.5: This model excels at coding tasks across the entire software development lifecycle, from initial design to bug fixes, maintenance to optimizations. Learn more about the model's capabilities or read the model card. GitHub Copilot uses Claude Sonnet hosted on Amazon Web Services.
  • Gemini 2.0 Flash: This model has strong coding, math, and reasoning capabilities that makes it well suited to assist with software development. For information about the capabilities of Gemini 2.0 Flash, see the Google for developers blog and the Google Cloud documentation. For details of Google's data handling policy, see Generative AI and data governance on the Google website.
  • o1: This model is focused on advanced reasoning and solving complex problems, in particular in math and science. It responds more slowly than the GPT-4o model. You can make 10 requests to this model per day. Learn more about the model's capabilities and review the model card. o1 is hosted on Azure.
  • o3-mini: This model is the next generation of reasoning models, following from o1 and o1-mini. The o3-mini model outperforms o1 on coding benchmarks with response times that are comparable to o1-mini, providing improved quality at nearly the same latency. It is best suited for code generation and small context operations. You can make 50 requests to this model every 12 hours. Learn more about the model's capabilities and review the model card. o3-mini is hosted on Azure.

For more information about these models, see:

Changing your AI model

These instructions are for VS Code. For Visual Studio or for Copilot on the GitHub website, click the appropriate tab at the top of this page.

If you access Copilot Chat through a Copilot Business or Copilot Enterprise subscription, your organization or enterprise must grant members the ability to switch to a different model. See Managing policies for Copilot in your organization or Managing policies and features for Copilot in your enterprise.

Note

  • If you use Copilot Extensions, they may override the model you select.
  • Experimental pre-release versions of the models may not interact with all filters correctly, including the duplication detection filter.
  1. To open the chat view, click the chat icon in the activity bar or press Control+Command+i (Mac) / Ctrl+Alt+i (Windows/Linux).
  2. In the bottom right of the chat view, select the CURRENT-MODEL dropdown menu, then click the AI model of your choice.