Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

We also support SmolVLM series which delivers light-speed response thanks to its mini size!

This is perfect for real-time home video surveillance system. That's one of the ideas for my next hobby project!

    llama-server -hf ggml-org/SmolVLM-Instruct-GGUF
    llama-server -hf ggml-org/SmolVLM-256M-Instruct-GGUF
    llama-server -hf ggml-org/SmolVLM-500M-Instruct-GGUF
    llama-server -hf ggml-org/SmolVLM2-2.2B-Instruct-GGUF
    llama-server -hf ggml-org/SmolVLM2-256M-Video-Instruct-GGUF
    llama-server -hf ggml-org/SmolVLM2-500M-Video-Instruct-GGUF


I've been noticing your commits as I skim the latest git commit notes whenever I periodically pull and rebuild. Thank you for all your work on this (and llama.cpp in general)!


Thanks for landing the mtmd functionality in the server. Like the other commenter I kept poring over commits in anticipation.


Ok but what's the quality of the high speed response? Can the sub-2.2B ones output a coherent sentence?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: