#llama_cpp — Public Fediverse posts
Live and recent posts from across the Fediverse tagged #llama_cpp, aggregated by home.social.
-
第905回 新年度特別企画 llama.cppでコマンドラインベースのローカルLLM入門[VRAM容量別編]
https://gihyo.jp/admin/serial/01/ubuntu-recipe/0905?utm_source=feed#gihyo #技術評論社 #gihyo_jp #技術動向 #技術解説 #業界動向 #OS #アプリケーション #ハードウェア製品 #Ubuntu #生成AI #llama_cpp #Intel_ARC_B580
-
第904回 ミドルレンジのグラフィックボードで生成AI入門[Intel編]
https://gihyo.jp/admin/serial/01/ubuntu-recipe/0904?utm_source=feed#gihyo #技術評論社 #gihyo_jp #技術動向 #技術解説 #業界動向 #OS #アプリケーション #お役立ち情報 #Ubuntu #生成AI #llama_cpp #Intel_ARC_B580
-
Tối ưu hóa tốc độ Qwen3 Next được hợp nhập vào llama.cpp (#llama_cpp #TríTuệNhânTạo #AI). Cải thiện hiệu năng mô hình, PR #17996 đã được duyệt. #HocMay #MachineLearning #LMA #MôHìnhNhânTạo
https://www.reddit.com/r/LocalLLaMA/comments/1pnz9xu/qwen3_next_speed_optimization_has_been_merged/
-
#MistralSmall24B-Instruct is a really nice model to run locally for Coding Advice, Summarizing or Creative Writing.
With a recent #llama_cpp on a #GeForce #RTX4090 at Q8, the 24GB VRAM is tightly maxed out and I am seeing text generation at 7-9 token/s.
https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501