G2 takes pride in showing unbiased reviews on user satisfaction in our ratings and reports. We do not allow paid placements in any of our ratings, rankings, or reports. Learn about our scoring methodologies.
StableLM is a suite of open-source large language models (LLMs) developed by Stability AI, designed to deliver high-performance natural language processing capabilities. These models are trained on ex
Mistral-7B-v0.1 is a small, yet powerful model adaptable to many use-cases. Mistral 7B is better than Llama 2 13B on all benchmarks, has natural coding abilities, and 8k sequence length. It’s released
BLOOM-560m is a transformer-based language model developed by BigScience, designed to facilitate research in large language models (LLMs). It serves as a pre-trained base model capable of generating h
Granite-3.1-3B-A800M-Base is a state-of-the-art language model developed by IBM, designed to handle complex natural language processing tasks with high efficiency. This model employs a sparse Mixture
Microsoft Azure’s Phi 3 model redefining large-scale language model capabilities in the cloud.
Athene-70B is an advanced open-weight language model developed by Nexusflow, built upon Meta's Llama-3-70B-Instruct architecture. Utilizing Reinforcement Learning from Human Feedback , Athene-70B achi
BLOOM-1b1 is a multilingual language model developed by the BigScience Workshop, designed to generate human-like text across 48 languages. As a transformer-based model, it utilizes a decoder-only arch
BLOOM-1b7 is a transformer-based language model developed by the BigScience Workshop, designed to generate human-like text across 48 languages. As a scaled-down variant of the larger BLOOM model, it o
BLOOM-3B is a 3-billion parameter multilingual language model developed by the BigScience initiative. As a scaled-down version of the larger BLOOM model, it maintains the same architecture and trainin
BLOOM-7B1 is a multilingual language model developed by BigScience, designed to generate human-like text across 48 languages. With over 7 billion parameters, it leverages a transformer-based architect
Gemma 3 270M is a compact, text-only model within the Gemma family of generative AI models, designed to perform a variety of text generation tasks such as question answering, summarization, and reason
Gemma 3 270M is a compact, text-only model within the Gemma family of generative AI models, designed to perform a variety of text generation tasks such as question answering, summarization, and reason
Gemma 3 270M is a compact, text-only model within the Gemma family of generative AI models, designed to perform a variety of text generation tasks such as question answering, summarization, and reason
Gemma 3n is a generative AI model optimized for deployment on everyday devices such as smartphones, laptops, and tablets. It introduces innovations in parameter-efficient processing, including Per-Lay
Gemma 3n is a generative AI model optimized for deployment on everyday devices such as smartphones, laptops, and tablets. It introduces innovations in parameter-efficient processing, including Per-Lay