🗂️ Navigation

Cerebras-GPT

A Family of Open, Compute-efficient, Large Language Models.

Visit Website →

Overview

Cerebras-GPT is a family of seven open-source GPT-based large language models released by Cerebras Systems. The models range in size from 111 million to 13 billion parameters and were trained using the Chinchilla formula to be compute-optimal, providing high accuracy for a given compute budget. They are released under the Apache 2.0 license, making them available for research and commercial use.

✨ Key Features

  • Family of seven models with sizes from 111M to 13B parameters
  • Trained to be compute-efficient using the Chinchilla formula
  • Open-source under the Apache 2.0 license
  • Trained on the Cerebras Andromeda AI supercomputer

🎯 Key Differentiators

  • Trained to be compute-optimal at every model size
  • Demonstrates the capabilities of Cerebras's AI hardware

Unique Value: Provides a suite of compute-efficient, open-source GPT models for research and commercial use.

🎯 Use Cases (4)

Natural language processing research Text generation Fine-tuning for specific tasks Commercial applications

💡 Check With Vendor

Verify these considerations match your specific requirements:

  • Human-facing dialogue applications without fine-tuning

🏆 Alternatives

Pythia OPT

Offers a clear scaling law and high training efficiency compared to other open-source models.

💻 Platforms

Self-hosted

✅ Offline Mode Available

🔌 Integrations

Hugging Face Transformers

💰 Pricing

Contact for pricing
Free Tier Available

Free tier: Free for research and commercial use under the Apache 2.0 license.

Visit Cerebras-GPT Website →