2026 Best Software Awards are here!See the list

Best Small Language Models (SLMs) - Page 3

Jeffrey Lin
JL
Researched and written by Jeffrey Lin

Small language models (SLM) are artificial intelligence (AI) language models that are optimized for efficiency, specialization, and deployment in resource-constrained and compute-limited environments. Similar to large language models (LLMs), SLMs are also engineered to understand, interpret, and generate human-like outputs from a wide array of inputs. Leveraging efficient machine learning (ML) techniques, streamlined architectures, and specialized datasets, these models are often repurposed to perform a select array of tasks to maximize resource efficiency. SLMs can be essential for organizations requiring cost-effective and fast deployment of AI models.

Due to their optimized architectures, SLMs can be deployed on edge devices, mobile platforms, and offline systems, facilitating accessible AI deployment. SLMs differ from LLMs, which focus on comprehensive, general-purpose language models that handle complex, diverse tasks across multiple domains. SLMs are designed to be retrained to maximize specialization and resource efficiency, focusing on targeted applications rather than broad intelligence.

A key difference between SLMs and LLMs is their parameter size, which is a direct indicator of their knowledge base and reasoning potential. SLM parameter sizes typically range from a few million to over 10 billion. Whereas LLMs have parameter sizes ranging from 10 billion to trillions of parameters. In practice, some SLMs are also derived from LLMs through methods like quantization or distillation, which reduce model size for efficiency but do not change the original training data. SLMs differ from AI chatbots, which provide the user-facing platform, rather than the foundational models themselves.

To qualify for inclusion in the Small Language Models (SLM) category, a product must:

Offer a compact language model that is optimized for resource efficiency and specialized tasks and capable of comprehending and generating human-like outputs
Contain 10 billion parameters or fewer, whereas LLMs exceed this threshold of 10 billion parameters
Provide deployment flexibility for resource-constrained environments, such as edge devices, mobile platforms, or computing hardware
Be designed for task-specific optimization through fine-tuning, domain specialization, or targeted training for specific business applications
Maintain computational efficiency with fast inference times, reduced memory requirements, and lower energy consumption compared to LLMs
Show More
Show Less

G2 takes pride in showing unbiased reviews on user satisfaction in our ratings and reports. We do not allow paid placements in any of our ratings, rankings, or reports. Learn about our scoring methodologies.

No filters applied
40 Listings in Small Language Models (SLMs) Available
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    NVIDIA Nemotron-Nano-9B-v2 is a compact, open-source language model designed to deliver high-performance reasoning and agentic capabilities. Utilizing a hybrid Mamba-Transformer architecture, it effic

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    NVIDIA
    Year Founded
    1993
    HQ Location
    Santa Clara, CA
    Twitter
    @nvidia
    2,455,322 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    46,062 employees on LinkedIn®
    Ownership
    NVDA
Product Description
How are these determined?Information
This description is provided by the seller.

NVIDIA Nemotron-Nano-9B-v2 is a compact, open-source language model designed to deliver high-performance reasoning and agentic capabilities. Utilizing a hybrid Mamba-Transformer architecture, it effic

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
NVIDIA
Year Founded
1993
HQ Location
Santa Clara, CA
Twitter
@nvidia
2,455,322 Twitter followers
LinkedIn® Page
www.linkedin.com
46,062 employees on LinkedIn®
Ownership
NVDA
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    Phi-3.5-mini is a lightweight, state-of-the-art language model developed by Microsoft, designed to deliver high-quality reasoning capabilities within a compact architecture. Building upon the datasets

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Microsoft
    Year Founded
    1975
    HQ Location
    Redmond, Washington
    Twitter
    @microsoft
    13,088,873 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    226,132 employees on LinkedIn®
    Ownership
    MSFT
Product Description
How are these determined?Information
This description is provided by the seller.

Phi-3.5-mini is a lightweight, state-of-the-art language model developed by Microsoft, designed to deliver high-quality reasoning capabilities within a compact architecture. Building upon the datasets

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Microsoft
Year Founded
1975
HQ Location
Redmond, Washington
Twitter
@microsoft
13,088,873 Twitter followers
LinkedIn® Page
www.linkedin.com
226,132 employees on LinkedIn®
Ownership
MSFT
G2 Advertising
Sponsored
G2 Advertising
Get 2x conversion than Google Ads with G2 Advertising!
G2 Advertising places your product in premium positions on high-traffic pages and on targeted competitor pages to reach buyers at key comparison moments.
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    The Phi-3 Mini-4K-Instruct is a lightweight, state-of-the-art language model developed by Microsoft, featuring 3.8 billion parameters. It is part of the Phi-3 model family and is designed to support a

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Microsoft
    Year Founded
    1975
    HQ Location
    Redmond, Washington
    Twitter
    @microsoft
    13,088,873 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    226,132 employees on LinkedIn®
    Ownership
    MSFT
Product Description
How are these determined?Information
This description is provided by the seller.

The Phi-3 Mini-4K-Instruct is a lightweight, state-of-the-art language model developed by Microsoft, featuring 3.8 billion parameters. It is part of the Phi-3 model family and is designed to support a

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Microsoft
Year Founded
1975
HQ Location
Redmond, Washington
Twitter
@microsoft
13,088,873 Twitter followers
LinkedIn® Page
www.linkedin.com
226,132 employees on LinkedIn®
Ownership
MSFT
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    The Phi-3-Small-128K-Instruct is a 7-billion-parameter, state-of-the-art language model developed by Microsoft. It is part of the Phi-3 family and is designed to handle a context length of up to 128,0

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Microsoft
    Year Founded
    1975
    HQ Location
    Redmond, Washington
    Twitter
    @microsoft
    13,088,873 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    226,132 employees on LinkedIn®
    Ownership
    MSFT
Product Description
How are these determined?Information
This description is provided by the seller.

The Phi-3-Small-128K-Instruct is a 7-billion-parameter, state-of-the-art language model developed by Microsoft. It is part of the Phi-3 family and is designed to handle a context length of up to 128,0

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Microsoft
Year Founded
1975
HQ Location
Redmond, Washington
Twitter
@microsoft
13,088,873 Twitter followers
LinkedIn® Page
www.linkedin.com
226,132 employees on LinkedIn®
Ownership
MSFT
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    Smaller Phi-3 model variant with extended 8k token context and instruction capabilities.

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Microsoft
    Year Founded
    1975
    HQ Location
    Redmond, Washington
    Twitter
    @microsoft
    13,088,873 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    226,132 employees on LinkedIn®
    Ownership
    MSFT
Product Description
How are these determined?Information
This description is provided by the seller.

Smaller Phi-3 model variant with extended 8k token context and instruction capabilities.

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Microsoft
Year Founded
1975
HQ Location
Redmond, Washington
Twitter
@microsoft
13,088,873 Twitter followers
LinkedIn® Page
www.linkedin.com
226,132 employees on LinkedIn®
Ownership
MSFT
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    The Phi-3 Mini-4K-Instruct is a lightweight, state-of-the-art language model developed by Microsoft, featuring 3.8 billion parameters. It is part of the Phi-3 model family and is designed to support a

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Microsoft
    Year Founded
    1975
    HQ Location
    Redmond, Washington
    Twitter
    @microsoft
    13,088,873 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    226,132 employees on LinkedIn®
    Ownership
    MSFT
Product Description
How are these determined?Information
This description is provided by the seller.

The Phi-3 Mini-4K-Instruct is a lightweight, state-of-the-art language model developed by Microsoft, featuring 3.8 billion parameters. It is part of the Phi-3 model family and is designed to support a

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Microsoft
Year Founded
1975
HQ Location
Redmond, Washington
Twitter
@microsoft
13,088,873 Twitter followers
LinkedIn® Page
www.linkedin.com
226,132 employees on LinkedIn®
Ownership
MSFT
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    Phi-4-mini-reasoning is a compact, transformer-based language model developed by Microsoft, specifically optimized for mathematical reasoning tasks. With 3.8 billion parameters and support for a 128K

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Microsoft
    Year Founded
    1975
    HQ Location
    Redmond, Washington
    Twitter
    @microsoft
    13,088,873 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    226,132 employees on LinkedIn®
    Ownership
    MSFT
Product Description
How are these determined?Information
This description is provided by the seller.

Phi-4-mini-reasoning is a compact, transformer-based language model developed by Microsoft, specifically optimized for mathematical reasoning tasks. With 3.8 billion parameters and support for a 128K

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Microsoft
Year Founded
1975
HQ Location
Redmond, Washington
Twitter
@microsoft
13,088,873 Twitter followers
LinkedIn® Page
www.linkedin.com
226,132 employees on LinkedIn®
Ownership
MSFT
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    StableLM 2 1.6B is a 1.6 billion parameter decoder-only language model developed by Stability AI. It is pre-trained on 2 trillion tokens from diverse multilingual and code datasets over two epochs. Th

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    HQ Location
    London
    Twitter
    @StabilityAI
    251,245 Twitter followers
    LinkedIn® Page
    www.linkedin.com
    188 employees on LinkedIn®
Product Description
How are these determined?Information
This description is provided by the seller.

StableLM 2 1.6B is a 1.6 billion parameter decoder-only language model developed by Stability AI. It is pre-trained on 2 trillion tokens from diverse multilingual and code datasets over two epochs. Th

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
HQ Location
London
Twitter
@StabilityAI
251,245 Twitter followers
LinkedIn® Page
www.linkedin.com
188 employees on LinkedIn®
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    Step-1 8k is a large-scale language model developed by StepFun, designed to understand and generate natural language text across various domains. With a context length of 8,000 tokens, it can process

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    StepFun
    HQ Location
    N/A
    LinkedIn® Page
    www.linkedin.com
    35 employees on LinkedIn®
Product Description
How are these determined?Information
This description is provided by the seller.

Step-1 8k is a large-scale language model developed by StepFun, designed to understand and generate natural language text across various domains. With a context length of 8,000 tokens, it can process

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
StepFun
HQ Location
N/A
LinkedIn® Page
www.linkedin.com
35 employees on LinkedIn®
  • Overview
    Expand/Collapse Overview
  • Product Description
    How are these determined?Information
    This description is provided by the seller.

    Multilingual Mixture-of-Experts model supporting 50+ languages with better MMLU performance and reduced hallucinations using online knowledge.

    We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
    Industries
    No information available
    Market Segment
    No information available
  • Seller Details
    Expand/Collapse Seller Details
  • Seller Details
    Seller
    Two AI
    Year Founded
    2021
    HQ Location
    Silicon Valley, US
    LinkedIn® Page
    www.linkedin.com
    49 employees on LinkedIn®
Product Description
How are these determined?Information
This description is provided by the seller.

Multilingual Mixture-of-Experts model supporting 50+ languages with better MMLU performance and reduced hallucinations using online knowledge.

We don't have enough data from reviews to share who uses this product. Leave a review to contribute, or learn more about review generation.
Industries
No information available
Market Segment
No information available
Seller Details
Seller
Two AI
Year Founded
2021
HQ Location
Silicon Valley, US
LinkedIn® Page
www.linkedin.com
49 employees on LinkedIn®