Cerebras Ultra Fast Inference For GitHub Copilot
di Cerebras
Make GitHub Copilot run 10x Faster - Powered by Cerebras
Make GitHub Copilot run 10x faster - with the World’s Fastest Inference API.
Cerebras Inference powers the world’s top coding models at 2,000 tokens/sec, making code generation instant and enabling super-fast agentic flows.
Simply Pay-As-You-Go for Cerebras Cloud and start building on the world’s fastest AI infrastructure for as little as $10 — no contracts, no friction, just add your credit card and go.
-or-
Get your free API key and get started today.
Have questions? Click the "Contact me" button above to connect with Cerebras.
------------------------------------------------------------------------------------------
API Key Setup
Here's how you can use Cerebras models in VS Code:
- Get your free API key from Cerebras Cloud.
- Install the Cerebras VS Code extension.
- Set up GitHub Copilot if you haven't already done so.
- In the GitHub Copilot chat interface, select Manage Models and choose Cerebras.
- Paste in your API key when prompted.
- Choose which models to enable.
- You're all set! Happy coding 🎉
Note: Bring-your-own-key is not supported for GitHub Copilot Enterprise subscriptions at this time.
Supported Models
This extension provides support for GLM 4.7 in agent mode, as well as the following models in chat mode:
Advanced Tips
Here's how you can accomplish more with Cerebras:
- Get higher rate limits on GLM 4.7 with our Cerebras Code plans, starting at $50/month.
- Generate code at top speed with Cerebras by installing the Cerebras Code MCP server.
- Read our developer documentation to turbocharge your own AI products using Cerebras' Inference API.