Large language models (LLMs) have become the bedrock of modern artificial intelligence development. They initiated and now define the generative AI era, from straightforward chatbot applications to agentic engineering and other complex automated workflows driven by AI agents. Their advent has marked a fundamental turning point in the history of machine learning.
As the technology matures, new LLMs continue to proliferate. Leading AI developers, new start-ups and established enterprise powerhouses alike are perpetually releasing and refining new models. Meanwhile, the open source community is constantly fine-tuning open source LLMs, merging and modifying existing models on custom datasets to create endless variants. As such, no list of LLMs could reasonably hope to be exhaustive—and even the most “exhaustive” list wouldn’t remain so for very long.
What follows is a list of some of the most prominent and performant LLMs available today. Here are some things to note:
For practical purposes, LLMs can generally be divided into 2 categories: closed source LLMs, available solely as commercial offerings through the model developer, and open models, which are made freely available at no cost.
A closed source model, or proprietary model, can only be accessed directly on the model developer’s platform, other platforms to which they have licensed their model or through the model provider’s proprietary API.
Because closed model developers generally treat their technical details as closely guarded trade secrets, it’s typically impossible to know with certainty the specifics of a closed model’s size, neural network architecture or training process. Some details can be inferred—for instance, by comparing a closed model’s inference speed, GPU memory usage and benchmark performance to that of open models whose details are publicly disclosed—but rarely, if ever, confirmed.
Since at least since roughly 2022, most state-of-the-art frontier models at any given time have been closed models—but that’s largely a reflection of the real-world historical circumstances of the industry, rather than any inherent superiority of closed models to open models. What follows are some of the most notable closed model series, ordered alphabetically.
Anthropic’s Claude language models are among the world’s most performant. Originally founded as an AI safety research lab in 2021 by former OpenAI employees, Anthropic’s approach to model development is built around the unique concept of Constitutional AI. Claude’s “Constitution” is a document that serves not only to guide the conduct of Anthropic employees, but the conduct (and creation of synthetic training data) of Claude models themselves.
Since Claude 3, successive generations of Claude have featured multimodal models in 3 different sizes:
Claude Haiku, Sonnet and Opus can all process text, audio and image inputs, and output text or audio (as text-to-speech). Historically, unlike most of their closed model competitors, they (and the Claude platform that they power) were not capable of image generation—but as of March 12, 2026, Claude can now generate images. When accessing the models through the Claude API, users can set the “effort level” of Sonnet or Opus’s reasoning process to “max,” “high,” “medium,” “low” or “adaptive.”
Gemini is Google’s closed language model series, developed by its subsidiary Google DeepMind and first launched in December 2023. It’s worth noting that Google Brain (which was merged with DeepMind to form Google DeepMind in 2023) is responsible for the creation of the transformer model architecture that enabled the first LLMs, having published the landmark “Attention is All You Need” research paper in 2017.
Since early 2025, Google has released each generation of Gemini models with 3 different sizes, all of which are reasoning models. When accessed through the Gemini API, users can select one of multiple “thinking levels” to customize the amount of tokens and time the model will spend before generating a final output.
Gemini Pro, Flash and Flash-Lite models are natively multimodal: they can process text, audio, image or video inputs and generate text outputs. When accessed through the Gemini platform, multimodal outputs can be generated through Gemini’s separate, specialized models for image generation, video generation or music generation.
Since the release of Gemini 2.5 Pro in March 2025, which achieved then-industry best performance across most academic benchmarks, Gemini models have contended with Claude and OpenAI’s GPT series as the world’s most performant LLMs. Generally speaking, the status of “top” model changes hands each time a new frontier model in one those three series is released.
Grok is a family of proprietary LLMs produced by xAI, first launched in beta preview as a chatbot on X (formerly Twitter) in November 2023. In April 2025, xAI launched API access for Grok 3, which was then its newest, flagship model.
Grok’s model lineup has continued to change over successive generations of model releases.
As of Grok 4, Grok models can process text, image and speech inputs. Though the Grok LLMs cannot provide multimodal outputs, image and video outputs can be generated by xAI’s Aurora model through its Grok Imagine platform.
Unrelated to its raw performance, much of Grok’s history (and particularly that of the Grok chatbot) has been marked by controversy, such as accusations of spreading election misinformation, inserting polarizing viewpoints into unrelated conversations and perpetuating harmful stereotypes.
In public statements, xAI CEO Elon Musk has said that “our general approach is that we will open source the last version when the next version is fully out.”1
xAI open-sourced Grok 1 under Apache 2.0 license in March 2024. Though Grok 3 was released in February 2025, the next open-source release of a Grok model was not until August of 2025. Confusingly, xAI (and Musk) announced that they had open-sourced “Grok 2.5,” 2 though no model had been named nor announced as such prior to that statement. The model’s own Hugging Face model card even refers to the model as “Grok-2.”
In that August 2025 announcement, Musk indicated that Grok 3 would be likewise open-sourced in “about 6 months.” As of 8 months later, said open source release date is yet to be announced.
OpenAI’s GPT series—short for Generative Pretrained Transformer—is largely credited with initiating the current era of generative AI, particularly following the 2022 launch of ChatGPT with their GPT-3.5 model.
OpenAI’s for model naming and variant conventions has changed significantly since 2022, often in a confusing way. For instance, GPT-4.1 was released after GPT-4.5, and the o4 reasoning model was available at the same time as the multimodal non-reasoning model GPT-4o, which was an entirely distinct from their “o4” reasoning model, whose performance was inferior to that of “o3.” In early 2025, OpenAI CEO Sam Altman acknowledged that “We realize how complicated our model and product offerings have gotten.”
Since the release of GPT-5 in August 2025, the company’s consolidated LLM offerings now comprise:
OpenAI has also release 2 open weight GPT models, which are detailed in the “Open models” section of this article.
Mistral AI, a France-based company founded by former employees of Meta AI and Google DeepMind, was originally dedicated entirely to open source models upon the launch of its first model (Mistral 7B) in September 2023. Since then, Mistral has transitioned to a mixed model in which many of its offerings have open releases but select frontier models remain closed source.
As of March 2026, Mistral AI’s flagship proprietary LLMs include:
Mistral’s open weight model offerings are detailed later in this article.
In machine learning, open source is often used colloquially to refer to AI tools whose source code is made available free of charge, but the term is actually a formal designation stewarded by the Open Source Initiative (OSI). The OSI only certifies a given software license as “Open Source Initiative approved” if it deems said license to meet the ten requirements listed in the official Open Source Definition (OSD).
Most “open source” models do not meet all of those requirements. That being the case, the term open model (or open weight model) more accurately refers to any freely distributed LLM. Within the spectrum of open models is a great deal of variability. An open weight (but not open source) model can be used to run inference and can even be fine-tuned—but if its full source code is not provided, it can’t be modified beyond changes to the values of its weights through fine-tuning. Its license might prohibit the model’s use in some scenarios (such as commercial settings) or place other specific stipulations on its application.
A true open source model released with training code and a description of its training procedures, conversely, can be fully modified in any way and used without restriction. The most common and standardized open source licenses are the Apache 2.0 license and MIT license. It’s worth noting, however, that unless an open source model’s developer provides the details of its training data, the model itself is not fully reproducible.
Open source releases are integral to the continued development and improvement of LLMs, and are largely responsible for enabling their invention in the first place. Open models can typically be accessed through their model developer or through popular open source platforms such as GitHub or Hugging Face. What follows is a list of notable open model series, organized alphabetically.
Cohere, a Canada-based company whose founders include one of the authors of “Attention is All You Need,” was launched in 2019. Though the company releases detailed technical reports for each LLM and ostensibly releases them as open weight models, Cohere licenses their open releases under a modified version of the Creative Commons 4.0 license that prohibits commercial use.
Command is Cohere’s flagship foundation model series, designed for enterprise use cases.
In a March 2026 Reddit comment, Cohere CEO Aidan Gomez indicated that the company was actively developing the next generation of Command, and that they would be the organization’s first mixture of experts (MoE) models.
Aya is Cohere’s multilingual-focused model series, first launched in February 2024 with Aya 101—which, as its name suggests, was “capable of following instructions in 101 languages.”
DeepSeek is an integral player in the open source ecosystem, contributing a number of innovations to LLM architectures and training processes. At times, its models’ performance have rivaled that of top closed models. Their LLMs—both weights and code—are open sourced under a standard MIT license. DeepSeek also frequently releases technical papers detailing their findings and techniques.
Despite periodic rumors of an impending DeepSeek-V4 (or “DeepSeek-R2”), their releases are yet to materialize.
The Falcon series of LLMs are developed by the UAE’s Technology Innovation Institute (TTI). Though TII’s first generational of models in 2023 was perhaps most notable for Falcon-180B, which at the time was one of the largest open source models available, TII have since focused on smaller models. Falcon2 had 11B parameters and Falcon3, TII’s first multimodal models (released in December 2024), ranged from 1B to 10B.
The most recent generations of Falcon models have focused on hybrid Mamba-Transformer models.
Falcon models are released under a proprietary Falcon license that is inspired by, but adds notable stipulations and constraints to, the Apache 2.0 framework.
Gemma is Google’s family of open models. According to Google, Gemma models are “built from the same technology that powers [their] Gemini models.”7
Gemma models are released under the Gemma license, whose usage terms are similar to those of the Apache 2.0 license but are governed by the Gemma Prohibited Use Policy.
GLM is a family of LLMs from Beijing-based Z.ai (also called Zhipu AI) that aim for state-of-the-art performance. The company achieved a breakthrough with GLM-4.5, which upon its initial release in late July 2025 ostensibly rivaled the world’s top open models, including the flagship models from DeepSeek and Qwen, across academic benchmarks.
IBM Granite is a series of open source LLMs optimized for enterprise use cases, focused primarily on small, practical and efficient models. First launched in September 2023, Granite rose to prominence upon the release of Granite 3.0 in October 2024, which saw the Granite series reach performance rivaling that of leading open models of comparable size.
Granite 4, launched in October 2025, introduced a new hybrid Mamba2-Transformer architecture for superior speed and memory efficiency, particularly under large workloads, compared to conventional transformer models.
All Granite models are open sourced under a standard Apache 2.0 license and trained on enterprise-safe data. In October 2025, the Granite series became the first major open model family to receive ISO-42001 certification.
GPT-OSS are OpenAI’s open weight language models, released in August 2025 under a standard Apache 2.0 license. They’re the company’s first open LLMs since the release of GPT-2 in 2019.
Both GPT-OSS models were trained with 4-bit quantization of their model weights, significantly increasing their speed and reducing their memory requirements relative to those of conventional models of similar size.
Kimi is a series of open models developed by Beijing-based Moonshot AI.
Kimi models are released under a modified MIT License, requiring users to “prominently display ‘Kimi K2’ on the user interface” of any product with over 100 million monthly active users or more than USD 20M in monthly revenue.
Meta’s Llama models (original stylized as LLaMA, short for “Large Language model Meta AI), have been an integral part of the history of open LLMs. Early Llama releases help democratize LLM methodologies, informing and strongly influencing many standard conventions of LLM development, from training to architecture and sizing variations.
Though Meta often uses the term “open source,” Llama models are released under a custom Llama license that places constraints on usage, attribution and access. The Open Source Initiative has therefore criticized Meta’s use of the term.
Shanghai-based MiniMax Group released their first eponymous LLM, MiniMax-Text-01, and a companion VLM, MiniMax-VL-01, in January 2025. They have since risen to prominence as one of the premier LLM developers in China, prioritizing large-scale models and long context windows.
MiniMax models are offered under a modified MIT License.
Alongside its closed-source offerings, Mistral AI offers a variety of well-regarded open models. Most (but not all) of Mistral’s open models are released under standard Apache 2.0 license.
Preeminent hardware manufacturer NVIDIA’s open LLM series are well regarded for their performance, research literature and architectural innovations.
Olmo, developed by the Allen Institute for AI (“Ai2”), are among the most truly “open” of all open source models: Ai2 typically releases all code, weights, training checkpoints and associated datasets alongside a standard Apache 2.0 release.
Phi is Microsoft’s open model line, historically focused on small models. They’re released under standard MIT License.
The Qwen series of LLMs, developed by Alibaba, have become among the most popular open models in the industry. The model family offers a wide variety of model sizes, architecture and capabilities intended to suit a variety of developer needs.
Accelerate software delivery with Bob, your AI partner for secure, intent-aware development.
Easily design scalable AI assistants and agents, automate repetitive tasks and simplify complex processes with IBM watsonx Orchestrate.
Put AI to work in your business with IBM’s industry-leading AI expertise and portfolio of solutions at your side.
1. “Elon Musk reins in Grok AI bot to stop election misinformation,” The Register, 28 August 2024
2. “Musk’s xAI chatbot Grok keeps randomly responding about ‘white genocide’ in South Africa,” CNBC, 14 May 2025
3. “Elon Musk’s AI chatbot, Grok, started calling itself ‘MechaHitler’,” NPR, 9 July 2025
4. @MarioNawfal tweet, X (formerly Twitter), 18 February 2025
5. “GPT-5.4 pro”, OpenAI, API docs accessed 12 March 2026
6. “Announcing Codestral 25.08 and the Complete Mistral Coding Stack for Enterprise,” Mistral AI, 30 July 2025
7. Gemma, Google DeepMind, accessed 12 March 2026
8. “Alibaba-backed Moonshot releases new Kimi AI model that beats ChatGPT, Claude in coding — and it costs less,” CNBC, 14 July 2025
9. “5 Thoughts on Kimi K2 Thinking,” Interconnects, 6 November 2025
10. Meta Llama: models page (sorted by “Most Downloads”), Hugging Face, accessed 11 March 2026
11. “MiniMax M2.7: Early Echoes of Self-Evolution,” MiniMax, 18 March 2026
12. “Introducing Mistral 3,” Mistral AI, 2 December 2025
13. Mistral AI: models page (sorted by “Most Downloads”), Hugging Face, accessed 11 March 2026