LogoFollow AI builders
  • 首页
  • 功能特性
  • 构建者
  • 提交构建者
LogoFollow AI builders

AI Builder 动态聚合平台|不追网红,只追创造者

TwitterX (Twitter)Email
公司
  • 联系我们
法律
  • Cookie政策
  • 隐私政策
  • 服务条款
© 2026 Follow AI builders All Rights Reserved.
GG

Georgi Gerganov

0 位关注者
27 条内容
4最近 7 天 条

简介

24th at the Electrica puzzle challenge | https://t.co/baTQS2bdia

平台

𝕏Georgi Gerganov

内容历史

GG
Georgi Gerganov
⚡github•3 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•3 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•3 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•6 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•8 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•8 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•8 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
⚡github•8 days ago

Activity on repository

ggerganov pushed ggterm

ggerganov pushed ggterm

View on GitHub
GG
Georgi Gerganov
𝕏x•9 days ago

Excited about this!

@Maxime Labonne

LFM2.5-Audio-1.5B > Real-time text-to-speech and ASR > Running locally on a CPU with llama.cpp > Interleave speech and text It's super elegant, I'm bullish on local audio models

View quoted post
View on X
GG
Georgi Gerganov
𝕏x•9 days ago

Recent contributions by NVIDIA engineers and llama.cpp collaborators resulting in significant performance gains for local AI

Recent contributions by NVIDIA engineers and llama.cpp collaborators resulting in significant performance gains for local AI
View on X
GG
Georgi Gerganov
𝕏x•about 1 month ago

Some neat QoL improvements coming to llama.cpp thanks to Johannes Gäßler https://github.com/ggml-org/llama.cpp/discussions/18049

View on X
GG
Georgi Gerganov
𝕏x•about 1 month ago

> llama-cli -hf org/model

> llama-cli -hf org/model
View on X
GG
Georgi Gerganov
𝕏x•about 1 month ago
Retweeted from @Xuan

RT Xuan-Son Nguyen Introducing: the new llama-cli 🦙🦙 > Clean looking interface > Multimodal support > Conversation control via commands > Speculative decoding support > Jinja fully supported Original tweet: https://x.com/ngxson/status/1998763208098853332

Your browser does not support the video tag.
View on X
GG
Georgi Gerganov
𝕏x•about 1 month ago

We joined forces with NVIDIA to unlock high-speed AI inference on RTX AI PCs and DGX Spark using llama.cpp. The latest Ministral-3B models reach 385+ tok/s on @NVIDIA_AI_PC GeForce RTX 5090 systems. Blog: https://developer.nvidia.com/blog/nvidia-accelerated-mistral-3-open-models-deliver-efficiency-accuracy-at-any-scale/

View on X
GG
Georgi Gerganov
𝕏x•about 1 month ago

The new Mistral 3 models in llama.cpp

The new Mistral 3 models in llama.cpp
View on X
GG
Georgi Gerganov
𝕏x•about 1 month ago
Retweeted from @Lysandre

RT Lysandre Transformers v5's first release candidate is out 🔥 The biggest release of my life. It's been five years since the last major (v4). From 20 architectures to 400, 20k daily downloads to 3 million. The release is huge, w/ tokenization (no slow tokenizers!), modeling & processing. Original tweet: https://x.com/LysandreJik/status/1995558230567878975

RT Lysandre
Transformers v5's first release candidate is out 🔥 The biggest release of my life.

It's been five years since the last major (v4). From 20 architectures to 400, 20k daily downloads to...
View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @Jeff

RT Jeff Geerling Just tried out the new built-in WebUI feature of llama.cpp and it couldn't be easier. Just start llama-server with a host and port, and voila!

RT Jeff Geerling: Just tried out the new built-in WebUI feature of llama.cpp and it couldn't be easier. Just start llama-server with a host and port, ...
View on X
GG
Georgi Gerganov
𝕏x•2 months ago

Initial M5 Neural Accelerators support in llama.cpp Enjoy faster TTFT in all ggml-based software (requires macOS Tahoe 26) https://github.com/ggml-org/llama.cpp/pull/16634

View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @Georgi

RT Georgi Gerganov Initial M5 Neural Accelerators support in llama.cpp Enjoy faster TTFT in all ggml-based software (requires macOS Tahoe 26) https://github.com/ggml-org/llama.cpp/pull/16634

View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @Emanuil

RT Emanuil Rusev Re @fishright @ggerganov Just pushed a fix for this — this is what first launch is going to look like in the next version.

RT Emanuil Rusev: Re @fishright @ggerganov Just pushed a fix for this — this is what first launch is going to look like in the next version.
View on X
GG
Georgi Gerganov
𝕏x•2 months ago

LlamaBarn v0.10.0 (beta) is out - feedback appreciated

LlamaBarn v0.10.0 (beta) is out - feedback appreciated
View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @clem

RT clem 🤗 When you run AI on your device, it is more efficient and less big brother and free! So it's very cool to see the new llama.cpp UI, a chatgpt-like app that fully runs on your laptop without needing wifi or sending any data external to any API. It supports: - 150,000+ GGUF models - Drop in PDFs, images, or text documents - Branch and edit conversations anytime - Parallel chats and image processing - Math and code rendering - Constrained generation with JSON schema supported Well done @ggerganov and team!

RT clem 🤗: When you run AI on your device, it is more efficient and less big brother and free! So it's very cool to see the new llama.cpp UI, a cha...
View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @Georgi

RT Georgi Gerganov A detailed look into the new WebUI of llama.cpp

RT Georgi Gerganov: A detailed look into the new WebUI of llama.cpp
View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @yags

RT yags llama.cpp developers and community came together in a really impressive way to implement Qwen3-VL models. Check out the PRs, it’s so cool to see the collaboration that went into getting this done. Standard formats like GGUF, combined with mainline llama.cpp support ensures the models you download will work anywhere you choose to run them. This protects you from getting unwittingly locked into niche providers’ custom implementations that won’t run outside their platforms.Qwen: 🎉 Qwen3-VL is now available on llama.cpp! Run this powerful vision-language model directly on your personal devices—fully supported on CPU, CUDA, Metal, Vulkan, and other backends. We’ve also released GGUF weights for all variants—from 2B up to 235B. Download and enjoy! 🚀 🤗 Link: https://x.com/Alibaba_Qwen/status/1984634293004747252

View on X
GG
Georgi Gerganov
𝕏x•2 months ago
Retweeted from @Qwen

RT Qwen 🎉 Qwen3-VL is now available on llama.cpp! Run this powerful vision-language model directly on your personal devices—fully supported on CPU, CUDA, Metal, Vulkan, and other backends. We’ve also released GGUF weights for all variants—from 2B up to 235B. Download and enjoy! 🚀 🤗 Hugging Face: https://huggingface.co/collections/Qwen/qwen3-vl 🤖 ModelScope: https://modelscope.cn/collections/Qwen3-VL-5c7a94c8cb144b 📌 PR: https://github.com/ggerganov/llama.cpp/pull/16780

View on X
GG
Georgi Gerganov
𝕏x•3 months ago
Retweeted from @Vaibhav

RT Vaibhav (VB) Srivastav BOOM: We've just re-launched HuggingChat v2 💬 - 115 open source models in a single interface is stronger than ChatGPT 🔥 Introducing: HuggingChat Omni 💫 > Select the best model for every prompt automatically 🚀 > Automatic model selection for your queries > 115 models available across 15 providers including @GroqInc, @CerebrasSystems, @togethercompute, @novita_labs, and more Powered by HF Inference Providers — access hundreds of AI models using only world-class inference providers Omni uses a policy-based approach to model selection (after experimenting with different methods). Credits to @katanemo_ for their small routing model: katanemo/Arch-Router-1.5B Coming next: • MCP support with web search • File support • Omni routing selection improvements • Customizable policies Try it out today at hf[dot] co/chat 🤗

Your browser does not support the video tag.
View on X
GG
Georgi Gerganov
𝕏x•3 months ago

simpleDavid Finsterwalder | eu/acc: Important info. The issue in that benchmark seems to be ollama. Native llama.cpp works much better. Not sure how ollama can fail so hard to wrap llama.cpp. The lesson: Don’t use ollama. Espacially not for benchmarks. Link: https://x.com/DFinsterwalder/status/1978372050239516989

simple
simple
View on X