From Wikipedia - Reading time: 6 min
Open-source artificial intelligence is the application of open-source practices to the development of artificial intelligence resources.
Many open-source artificial intelligence products are variations of other existing tools and technologies which have been shared as open-source software by large companies.[1]
Companies often develop closed products in an attempt to keep a competitive advantage in the marketplace.[2] A journalist for Wired explored the idea that open-source AI tools have a development advantage over closed products, and could overtake them in the marketplace.[2]
Popular open-source artificial intelligence project categories include large language models, machine translation tools, and chatbots.[3]
For software developers to produce open-source artificial intelligence resources, they must trust the various other open-source software components they use in its development.[4][5]
This section needs to be updated. (November 2024) |
LLaMA is a family of large language models released by Meta AI starting in February 2023.[6] Meta claims these models are open-source software, but the Open Source Initiative disputes this claim, arguing that "Meta's license for the LLaMa models and code does not meet this standard; specifically, it puts restrictions on commercial use for some users (paragraph 2) and also restricts the use of the model and software for certain purposes (the Acceptable Use Policy)."[7]
| Model | Developer | Parameter count | Context window | Licensing | Ref. |
|---|---|---|---|---|---|
| LLaMA | Meta AI | 7B, 13B, 33B, 65B | 2048 | [6] | |
| Llama 2 | Meta AI | 7B, 13B, 70B | 4k | Custom Meta license | [8][9] |
| Llama 3.1 | Meta AI | 8B, 70B, 405B | 128K | Meta Llama 3 Community License | [10][11][12] |
| Llama 3.2 | Meta AI | 1B to 405B | Research-only | [13] | |
| Mistral 7B | Mistral AI | 7 billion | 8k | Apache 2.0 | [14][15] |
| Mixtral 8x22B | Mistral AI | 8×22B | Apache 2.0 | [13] | |
| GPT-J | EleutherAI | 6 billion | 2048 | Apache 2.0 | [16] |
| GPT-NeoX | EleutherAI | 20B | MIT License | [13] | |
| Pythia | EleutherAI | 70 million - 12 billion | Apache 2.0 (Pythia-6.9B only) | [17][18] | |
| T5 | Google AI | 60 million to 11 billion | Apache 2.0 | [13] | |
| Gemma 2 | Google DeepMind | 2B, 9B, 27B | Apache 2.0 | [13] | |
| OLMo | Allen Institute for AI | Various | Apache 2.0 | [13] | |
| BLOOM | BigScience | 176 billion | OpenRAIL-M | [13] | |
| StarCoder2 | BigCode | Various | Apache 2.0 | [13] | |
| Falcon | Technology Innovation Institute | 7B, 40B | Apache 2.0 | [13] | |
| Jamba Series | AI21 Labs | Mini to Large | Custom | [13] | |
| Sea-Lion | AI Singapore | 7B | Custom | [13] | |
| Qwen Series | Alibaba Group | 7B | Custom | [13] | |
| Dolly 2.0 | Databricks | 12B | CC BY-SA 3.0 | [13] | |
| Granite Series | IBM | 3B, 8B | Apache 2.0 | [13] | |
| Phi-3 Series | Microsoft | Mini to Medium | MIT License | [13] | |
| NVLM 1.0 Family | Nvidia | 72B | CC BY-SA 3.0 | [13] | |
| RakutenAI Series | Rakuten | 7B | Custom | [13] | |
| Grok-1 | xAI | 314B | Apache 2.0 | [13] |
| Model | Developer | Parameter count | Licensing | Ref. |
|---|---|---|---|---|
| Stable Diffusion 3.5 | Stability AI | 2.5B to 8B | OpenRAIL-M | [13] |
| IF | DeepFloyd | 400M to 4.3B | Custom | [13] |
| Model | Developer | Parameter count | Licensing | Ref. |
|---|---|---|---|---|
| SAM 2.1 | Meta | 38.9M to 224.4M | Apache 2.0 | [13] |
| DeepLab | Not disclosed | Apache 2.0 | [13] | |
| Florence | Microsoft | 0.23B, 0.77B | MIT License | [13] |
| CLIP | OpenAI | 400M | MIT License | [13] |