Clem Delangue
简介
Co-founder & CEO @HuggingFace 🤗, the open and collaborative platform for AI builders
平台
内容历史
RT Sam Hogan 🇺🇸 Trust me when I say Egocentric is a fantastic choice of name for an open source datasetEddy Xu: you can download Egocentric-10K on an apache 2.0 license here: https://huggingface.co/datasets/builddotai/Egocentric-10K Link: https://x.com/eddybuild/status/1987951628700578190
Cool to see transformers number 4 top open-source project by contributors in 2025 in that latest @github octoverse! With vLLM, Codex, Ollama, llama.cpp all deeply integrated with transformers, it's an AI take-over of github this year!
You can just open-source things! If you think you don’t have enough experience, support or ressources to release useful open-source, remind yourself that @eddybuild is 18 and released the currently number one trending dataset on Hugging Face, ahead of Nvidia and Meta (respectively valued at 5 and 1.6 trillion dollars). Let's go!Eddy Xu: today, we’re open sourcing the largest egocentric dataset in history. - 10,000 hours - 2,153 factory workers - 1,080,000,000 frames the era of data scaling in robotics is here. (thread) Link: https://x.com/eddybuild/status/1987951619804414416
RT merve I need to follow more people, please send your @huggingface usernames 🤝 (also flexing the ✨)
RT Jeff Boudier 🤗 🔥 New in @Dell Enterprise Hub - @nvidia RTX Pro 6000! The XE7745 PowerEdge systems by Dell Technologies are dream machines for on-premise AI. With 96GB DDR7 vram packed along a Blackwell GPU, you have more than enough to deploy @OpenAI GPT-OSS 120GB... But making that easy wasn't easy, and that's what the Dell Enterprise Hub offers - an optimized, secured and tested container ready to deploy in one line of Docker or via Kubernetes using the CLI of Dell AI SDK (link below). 🦾 Optimized - fixes vLLM issues for Blackwell deployment 🔒 Secured - CVEs in vanilla deployment removed 🧪 Tested - by Dell teams on the same systems customers use Congrats team, great work! 👏 @alvarobartt @juanjucm Thibault Goehringer @PagezySimon Bala Rajendran Colin Byrne Bagus Hanindhito cc @ClementDelangue @MichaelDell
RT Jack Wotherspoon Welcome to the party @huggingface 🤗 Access the Hugging Face Hub directly from Gemini CLI with this new Gemini CLI extension. 🔍 - Search models, datasets and papers 📈 - Find trending models or datasets 🤗 - Learn how to fine-tune models and more! Big thanks to @evalstate and @reach_vb for making this happen!
RT Andi Marafioti Sneak peek of Reachy Mini's conversation capabilities! He can chat fluidly and change languages naturally. You can interrupt him and he can see and react to his environment. They start shipping soon, I'm excited to see what people start building with him!
RT Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) I admit I was skeptical about Alexander's small model obsession, but this is a powerful flex.Alexander Doria: Breaking: we release a fully synthetic generalist dataset for pretraining, SYNTH and two new SOTA reasoning models exclusively trained on it. Despite having seen only 200 billion tokens, Baguettotron is currently best-in-class in its size range. Link: https://x.com/Dorialexander/status/1987930819021635964
RT Beff – e/acc Training robotics foundation models on this dataset could make the US more competitive in manufacturingEddy Xu: today, we’re open sourcing the largest egocentric dataset in history. - 10,000 hours - 2,153 factory workers - 1,080,000,000 frames the era of data scaling in robotics is here. (thread) Link: https://x.com/eddybuild/status/1987951619804414416
RT Alexandra Botez Met Eddy when he was 17, working on AI glasses for chess. Six months later he’s a CEO, open-sourcing the largest first-person robotics dataset ever. The new generation moves fast. Props to the team!Eddy Xu: today, we’re open sourcing the largest egocentric dataset in history. - 10,000 hours - 2,153 factory workers - 1,080,000,000 frames the era of data scaling in robotics is here. (thread) Link: https://x.com/eddybuild/status/1987951619804414416
RT Roy holy shitEddy Xu: today, we’re open sourcing the largest egocentric dataset in history. - 10,000 hours - 2,153 factory workers - 1,080,000,000 frames the era of data scaling in robotics is here. (thread) Link: https://x.com/eddybuild/status/1987951619804414416
RT will brown one of the most exciting data / pretraining releases in quite a while this is a roadmap towards the “cognitive core” the x-axis here is log-scale (!!)Alexander Doria: Breaking: we release a fully synthetic generalist dataset for pretraining, SYNTH and two new SOTA reasoning models exclusively trained on it. Despite having seen only 200 billion tokens, Baguettotron is currently best-in-class in its size range. Link: https://x.com/Dorialexander/status/1987930819021635964
RT Eddy Xu today, we’re open sourcing the largest egocentric dataset in history. - 10,000 hours - 2,153 factory workers - 1,080,000,000 frames the era of data scaling in robotics is here. (thread)
RT Alexander Doria Breaking: we release a fully synthetic generalist dataset for pretraining, SYNTH and two new SOTA reasoning models exclusively trained on it. Despite having seen only 200 billion tokens, Baguettotron is currently best-in-class in its size range.
RT Crystal Starts in 10 minutes! Kimi’s first AMA https://www.reddit.com/r/LocalLLaMA/comments/1oth5pw/ama_with_moonshot_ai_the_opensource_frontier_lab/
Reachy Mini heard about @REKrobot and hired an AI builder coach. We're not ready!
RT elvis Kimi K2 Thinking is a bigger deal than I thought! I just ran a quick eval on a deep agent I built for customer support. It's on par with GPT-5; no other LLM has reached this level of agentic, orchestration, and reasoning capabilities. Huge for agentic and reasoning tasks.
RT Jeremy Howard Side effect of blocking Chinese firms from buying the best NVIDIA cards: top models are now explicitly being trained to work well on older/cheaper GPUs. The new SoTA model from @Kimi_Moonshot uses plain old BF16 ops (after dequant from INT4); no need for expensive FP4 support.Zhihu Frontier: 🚀 "Quantization is not a compromise — it's the next paradigm." After K2-Thinking's release, many developers have been curious about its native INT4 quantization format. 刘少伟, infra engineer at @Kimi_Moonshot and Zhihu contributor, shares an insider's view on why this choice Link: https://x.com/ZhihuFrontier/status/1987125624599970218
RT Matt Rouif Having fun building the @huggingface @ReachyMiniSol robot 🤖 . It’s very accessible! ♥️@ClementDelangue @julien_c @Thom_Wolf
RT Alex Finn Give me 1 reason why I shouldn't buy this top of the line Mac Studio, download Kimi K2 Thinking (best AI model in the world right now), and let it control the computer autonomously 24/7 A full employee working for me year round Would anyone want to this live streamed?