A large language model built by the Technology Innovation Institute (TII) for use in summarization, text generation, and chat bots.
63.9K Pulls Updated 14 months ago
Updated 14 months ago
14 months ago
e2bc879d7cee · 101GB
Readme
Falcon is a family of high-performing large language models model built by the Technology Innovation Institute (TII), a research center part of Abu Dhabi government’s advanced technology research council overseeing technology research.
CLI
ollama run falcon "Why is the sky blue?"
API
curl -X POST http://localhost:11434/api/generate -d '{
"model": "falcon",
"prompt": "Why is the sky blue?"
}'
Parameter counts
Parameter Count | Recommended memory | ||
---|---|---|---|
7 billion | 8GB | View | ollama run falcon:7b |
40 billion | 32GB | View | ollama run falcon:40b |
180 billion | 192GB | View | ollama run falcon:180b |
Variations
chat |
Chat models are fine-tuned on chat and instruction datasets with a mix of several large-scale conversational datasets. |
instruct |
Instruct models follow instructions and are fine-tuned on the baize instructional dataset. |
text |
Text models are the base foundation model without any fine-tuning for conversations, and are best used for simple text completion. |
Falcon 180B
As of September 2023, the 180 billion parameter model, Falcon 180B, is the best-performing openly released LLM. It sits somewhere in between OpenAI’s GPT 3.5 and GPT 4. For running Falcon 180B, a powerful system is recommended with at least 192GB of total memory.
Note: Falcon 180B is released under a different license than its smaller siblings that restricts commercial use under certain conditions. See the model details and license for more information.