Skip to content

tiiuae/Falcon-H1

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

72 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸ¦… Falcon-H Chat | πŸ€— Hugging Face | πŸ“„ Paper | πŸ“° Blog | πŸ“„ Documentation | πŸ–₯️ Hugging Face Demo | πŸ’¬ Discord

News


πŸš€ Introduction

We are excited to introduce Falcon-H1, the latest evolution in the Falcon family of large language models. Built upon an advanced hybrid architectureβ€”where each block integrates both State Space Models (SSMs) and Attention Mechanisms, these models span a wide range of scales, from 500 million to 34 billion parameters, making them suitable for both lightweight inference on edge devices and large-scale deployments in data centers.

Falcon-H1 was initially trained with support for 18 core languages, with scalability to 100+ languages, achieving state-of-the-art multilingual and reasoning performances in instruction following, maths, coding, and multilingual tasks.


✨ Key Highlights

Built by the Technology Innovation Institute (TII) in Abu Dhabi, Falcon-H1 is the latest step in pushing the frontier of hybrid transformer design:

🧩 Hybrid Architecture

Each transformer block processes all channels through both SSM and Attention in parallel, then sums the outputs. This allows the model to benefit from both long-range memory (via SSMs) and local/global attention simultaneously.

πŸ“ Scalable Sizes

Models available at multiple scales or variants: 500M, 1.5B, 1.5B-Deep, 3B, 7B, and 34B parameters.

🧠 Efficient Reasoning

The hybrid structure enhances reasoning and task generalization.

🌐 Multilingual by Design

Native training in 18 languages, with scalability to 100+ languages thanks to our multilingual tokenizer trained on diverse language datasets, with strong zero-shot translation and instruction-following abilities.

πŸ€– Instruction-Following and Agent Capabilities

Tuned for instruction following, multi-turn conversations, and already integrated with major inference engines such as vLLM, Hugging Face Transformers, and llama.cpp β€” with more coming soon.


🧭 Where to Start?

We provide the following documentation and resources to begin working with Falcon-H1:

  • πŸ’¬ Quick Deploy: Try Falcon-H1 instantly using our hosted Chat Interface or the Live Demo from Hugging Face
  • πŸ› οΈ Inference Toolkits: Compatible out-of-the-box with vLLM, Transformers, and llama.cpp. πŸ‘‰ Deployment Instructions. Other runtimes are in progress.
  • βš™οΈ Fine-tuning: Compatible with most frameworks based on Hugging Face Transformers library, out-of-the-box with OUMI, Llama-Factory, etc. πŸ‘‰ Fine-Tuning Guidelines. More framework support coming soon!
  • πŸ’» Local Setup: Full GGUF and HF formats available. Run it efficiently on both GPU and CPU.
  • πŸ”¬ Research: Learn more about our novel hybrid design in the Falcon-H1 technical report (Coming soon).

⚑ Inference

Make sure to install the latest version of transformers or vllm, eventually install these packages from source:

pip install git+https://github.com/huggingface/transformers.git

Refer to the official vLLM documentation for more details on building vLLM from source.

πŸ€— Transformers

Transformers is a library of pretrained natural language processing for inference and training. Refer to the snippet below to run H1 models using πŸ€— transformers:

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

# Load the model
model_id = "tiiuae/Falcon-H1-1B-Base"
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype=torch.bfloat16,
    device_map="auto"
)

# Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_id)

# Perform text generation below

πŸš„ vLLM

vLLM is a high-throughput and memory-efficient inference and serving engine for LLMs. To run Falcon-H1 models, you can refer to the following command:

# pip install vllm
vllm serve tiiuae/Falcon-H1-1B-Instruct --tensor-parallel-size 2 --data-parallel-size 1

πŸ”§ llama.cpp

Falcon-H1 is now natively supported into llama.cpp !

All official GGUF files can be found on our official Hugging Face collection.


1. Prerequisites

  • CMake β‰₯ 3.16
  • A C++17-compatible compiler (e.g., gcc, clang)
  • make or ninja build tool
  • (Optional) Docker, for OpenWebUI integration

2. Clone & Build

# Clone the Falcon-H1 llama.cpp fork
git clone https://github.com/ggml-org/llama.cpp
cd llama.cpp

# Create a build directory and compile
mkdir build && cd build
cmake ..         # Configure the project
make -j$(nproc)  # Build the binaries

Tip: For GPU acceleration, refer to the llama.cpp GPU guide.


3. Download a GGUF Model

Fetch the desired Falcon-H1 checkpoint from Hugging Face’s collection:

# Example: download the 1B Instruct model
wget https://huggingface.co/tiiuae/Falcon-H1-1.5B-Instruct-GGUF/resolve/main/Falcon-H1-1.5B-Instruct-Q5_K.gguf \
     -P models/

All available GGUF files: https://huggingface.co/collections/tiiuae/falcon-h1-6819f2795bc406da60fab8df


4. Run the llama-server

Start the HTTP server for inference:

./build/bin/llama-server \
  -m models/Falcon-H1-1B-Instruct-Q5_0.gguf \
  -c 4096 \
  -ngl 512 \
  --temp 0.1 \
  --host 0.0.0.0 \
  --port 11434

5. Web UI via OpenWebUI

Use the popular OpenWebUI frontend to chat in your browser:

docker run -d \
  --name openwebui-test \
  -e OPENAI_API_BASE_URL="http://host.docker.internal:11434/v1" \
  -p 8888:8888 \
  ghcr.io/open-webui/open-webui:main
  1. Open your browser at http://localhost:8888
  2. Select Falcon-H1-1B-Instruct-Q5_0 from the model list
  3. Start chatting!

For advanced tuning and custom flags, see the full llama.cpp documentation: https://github.com/ggerganov/llama.cpp

Demo Hardware: MacBook M4 Max Chip Model:Falcon-H1-1B-Q6_K

Falcon-H1-1B-Q6_K.mp4

πŸ“Š Performance and Throughput

A detailed dynamic evaluation report is provided in our blogpost and technical report:

  1. πŸ† We compare the performance of each Falcon-H1 model against the strongest models not only with the same size but also twice their size.
  2. πŸ“ˆ We show that Falcon-H1 models achieve state-of-the-art performance in most benchmarks (reasoning, maths, coding, in-context learning, and more), outperforming some closed source models like gpt-4o-mini in coding, reasoning and instruction following related tasks.

The blog post also features a dedicated section comparing Falcon-H1's inference speed to leading attention-based models, across a wide range of sequence lengths, prefillinng and generation scenarios.


πŸ“¦ Falcon-H1 Features at a Glance

  • πŸ”„ Parallel Hybrid Blocks: Attention + SSM in every layer.
  • 🌍 100+ Languages Supported: Multilingual instruction, chat, and translation.
  • πŸ“ Scalable Sizes: From 0.5B to 34B.
  • 🧩 Full Ecosystem Integration: Runs on widely used inference stacks and supports common file formats (HF, GGUF).
  • πŸ”‹ Quantized + Fine-tune Friendly: Models available in 8-bit, 4-bit, and standard FP16.

πŸ‘₯ Join the Community

Got feedback or want to build with Falcon-H1?

Join the conversation on Discord, follow us on Hugging Face, visit our official website, or check out our roadmap and open issues on GitHub.

Citation

Feel free to cite our work if you find it useful for your projects:

@article{falconh1,
    title={Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance},
    author={Jingwei Zuo and Maksim Velikanov and Ilyas Chahed and Younes Belkada and Dhia Eddine Rhayem and Guillaume Kunsch and Hakim Hacid and Hamza Yous and Brahim Farhat and Ibrahim Khadraoui and Mugariya Farooq and Giulia Campesan and Ruxandra Cojocaru and Yasser Djilali and Shi Hu and Iheb Chaabane and Puneesh Khanna and Mohamed El Amine Seddik and Ngoc Dung Huynh and Phuc Le Khac and Leen AlQadi and Billel Mokeddem and Mohamed Chami and Abdalgader Abubaker and Mikhail Lubinets and Kacper Piskorski and Slim Frikha},
    journal = {arXiv preprint arXiv:2507.22448},
    year={2025}
}

License GitHub stars Made with love at TII

About

All information and news with respect to Falcon-H1 series

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published