Discover real AI creators shaping the future. Track their latest blogs, X posts, YouTube videos, WeChat Official Account posts, and GitHub commits — all in one place.
RT Steve Ruiz I got the live performance of this last night from @jxnlco . That's dev rel. if you want to announce anything to me, my dms are open Original tweet: https://x.com/steveruizok/status/2030379369806376999
We’re launching Codex for Open Source to support the contributors who keep open-source software running. Maintainers can use Codex to review code, understand large codebases, and strengthen security coverage without taking on even more invisible work. http://developers.openai.com/codex/community/codex-for-oss
View quoted postActivity on rasbt/LLMs-from-scratch
rasbt contributed to rasbt/LLMs-from-scratch
View on GitHubI packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. https://github.com/karpathy/autoresearch Part code, part sci-fi, and a pinch of psychosis :)(I still have the bigger cousin running on prod nanochat, working a bigger model and on 8XH100, which looks like this now. I'll just leave this running for a while...)