SmolVLM2-2.2B-Instruct-GGUF (Q4_K_M)

Q4_K_M quantized GGUF conversion of SmolVLM2-2.2B-Instruct.

Files

File Size Description
SmolVLM2-2.2B-Instruct-Q4_K_M.gguf 1.0 GB Main model (Q4_K_M quantized)
mmproj-SmolVLM2-2.2B-Instruct-f16.gguf 832 MB Vision projector (F16)

Usage

With Ollama

FROM ./SmolVLM2-2.2B-Instruct-Q4_K_M.gguf
PROJECTOR ./mmproj-SmolVLM2-2.2B-Instruct-f16.gguf
PARAMETER num_ctx 4096
PARAMETER temperature 0.1

With llama.cpp

./llama-mtmd-cli -m SmolVLM2-2.2B-Instruct-Q4_K_M.gguf --mmproj mmproj-SmolVLM2-2.2B-Instruct-f16.gguf --image screenshot.png -p "Click the Submit button"

License

Apache 2.0

Downloads last month
35
GGUF
Model size
2B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for pierretokns/SmolVLM2-2.2B-Instruct-GGUF