Professional Documents
Culture Documents
DBRX by Databricks: The Future of FineGrained MoE Architecture in AI
DBRX by Databricks: The Future of FineGrained MoE Architecture in AI
com/
Introduction
What is DBRX?
source - https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm
DBRX boasts several unique features that set it apart from other models:
DBRX offers unique capabilities and benefits that make it a valuable tool
in various applications:
source - https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm
source - https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm
Table above shows Closed Model Comparisons , DBRX Instruct not only
exceeds GPT-3.5 but also rivals Gemini 1.0 Pro and Mistral Medium. It
demonstrates superior performance in general knowledge (MMLU),
commonsense reasoning (HellaSwag and WinoGrande), and excels in
programming and mathematical reasoning (HumanEval and GSM8k).
DBRX Instruct contends with Gemini 1.0 Pro and Mistral Medium,
The weights of the base model, known as DBRX Base, and the fine
tuned model, referred to as DBRX Instruct, are hosted on Hugging Face.
This platform is renowned for its extensive collection of pre-trained
models, making it a go-to resource for machine learning enthusiasts and
professionals alike.
Conclusion
Source
Blog Website: https://www.databricks.com/blog/introducing-dbrx-new-state-art-open-llm
Weights: https://huggingface.co/databricks/dbrx-base
GitHub Repo: https://github.com/databricks/dbrx