π§ LFM2
Collection
LFM2 is a new generation of hybrid models, designed for on-device deployment. β’ 28 items β’ Updated β’ 154
winget install llama.cpp
# Start a local OpenAI-compatible server with a web UI:
llama-server -hf LiquidAI/LFM2-1.2B-GGUF:# Run inference directly in the terminal:
llama-cli -hf LiquidAI/LFM2-1.2B-GGUF:# Download pre-built binary from:
# https://github.com/ggerganov/llama.cpp/releases# Start a local OpenAI-compatible server with a web UI:
./llama-server -hf LiquidAI/LFM2-1.2B-GGUF:# Run inference directly in the terminal:
./llama-cli -hf LiquidAI/LFM2-1.2B-GGUF:git clone https://github.com/ggerganov/llama.cpp.git
cd llama.cpp
cmake -B build
cmake --build build -j --target llama-server llama-cli# Start a local OpenAI-compatible server with a web UI:
./build/bin/llama-server -hf LiquidAI/LFM2-1.2B-GGUF:# Run inference directly in the terminal:
./build/bin/llama-cli -hf LiquidAI/LFM2-1.2B-GGUF:docker model run hf.co/LiquidAI/LFM2-1.2B-GGUF:
LFM2 is a new generation of hybrid models developed by Liquid AI, specifically designed for edge AI and on-device deployment. It sets a new standard in terms of quality, speed, and memory efficiency.
Find more details in the original model card: https://huggingface.co/LiquidAI/LFM2-1.2B
Example usage with llama.cpp:
llama-cli -hf LiquidAI/LFM2-1.2B-GGUF
4-bit
5-bit
6-bit
8-bit
16-bit
Base model
LiquidAI/LFM2-1.2B
Install from brew
# Start a local OpenAI-compatible server with a web UI: llama-server -hf LiquidAI/LFM2-1.2B-GGUF:# Run inference directly in the terminal: llama-cli -hf LiquidAI/LFM2-1.2B-GGUF: