π§ LamoFast-1.0 (GGUF)
This repository contains the GGUF release of LamoFast-1.0, a lightweight and fast open-source language model based on Qwen2.5-0.5B, with support for English and Hebrew.
The GGUF format is optimized for efficient local inference, making this model ideal for tools such as llama.cpp, Ollama, and LM Studio.
β¨ Key Features
- π Fast & lightweight β ~0.5B parameters, runs well on CPU and low-memory systems
- π§ General-purpose LLM with additional focus on astronomy & science topics
- π Bilingual β English & Hebrew support
- π¦ GGUF format β optimized for low memory usage and fast loading
- π Open license β Apache 2.0
π¦ Available Files
The GGUF release may one file:
LomaFast_Tiny_v1.gguf
Choose a quantization based on your hardware:
- Lower quantization (Q2βQ4) β faster, lower memory usage
- Higher quantization (Q5βQ8) β better quality, more memory usage
π Usage Examples
llama.cpp
./main -m LamoFast-1.0.Q4_K_M.gguf -p "Explain the Big Bang theory in simple terms." -n 200
Ollama
Create a Modelfile:
FROM LamoFast-1.0.Q4_K_M.gguf
Then run:
ollama create lamofast -f Modelfile
ollama run lamofast
LM Studio
- Open LM Studio
- Import the GGUF file
- Select the model and start chatting
π§ͺ Prompt Format
The model follows a chat-style prompt format compatible with Qwen-style templates.
Example:
<|user|>
Explain black holes in simple terms.
<|assistant|>
β οΈ Notes
- This is a small model by design β it prioritizes speed and efficiency over raw reasoning power
- Best results are achieved with clear, concise prompts
- Works especially well for educational, scientific, and lightweight assistant tasks
π License
Licensed under the Apache License 2.0.
You are free to use, modify, and distribute this model, including for commercial purposes.
π Credits
- Base model: Qwen2.5-0.5B
- Fine-tuning & GGUF release: Raziel1234
If you use LamoFast-1.0, a mention or β on Hugging Face is always appreciated!
- Downloads last month
- 42