document OCR + workflows @llama_index. cofounder/CEO Careers: https://t.co/EUnMNmb4DZ Enterprise: https://t.co/Ht5jwxRU13
SF is back but so is NYC Big thanks to @tuanacelik and our partners (@GoogleDeepMind , @cerebral_valley ) for running an awesome hackathon!
Hello from the @GoogleDeepMind Gemini hackathon from NYC! With @cerebral_valley and @temporal_xyz Been a while since I’ve done one of these, the @llama_index workshop is about to start, come learn all about document agents with LlamaParse and LlamaAgents 🫶
View quoted postSF is back but so is NYC Big thanks to @tuanacelik and our partners (@GoogleDeepMind , @cerebral_valley , @temporal_xyz ) for running an awesome hackathon!
Hello from the @GoogleDeepMind Gemini hackathon from NYC! With @cerebral_valley and @temporal_xyz Been a while since I’ve done one of these, the @llama_index workshop is about to start, come learn all about document agents with LlamaParse and LlamaAgents 🫶
View quoted postdeep learning in 2013-~2020 was the OG vibe coding 1. read some arxiv papers, have no idea what works and what doesn't 2. so you just try out a bunch of random shit, with some inkling of what might work ("i'm just gonna sprinkle in some...batch norm, half the LR, skip connections 👩🍳") 3. kick off like 20 runs across 500 gpus, wait a few hours/days 4. one of these things magically works, you justify it as some huge breakthrough, dress it up with some numbers, publish it to cvpr/neurips
Realize why I love agentic coding so much now It makes software engineering feel like ML research - kick off a bunch of agents (experiments) - monitor their trajectory (loss curves) - kill some & double down on others - async & orchestration
View quoted postRT Tuana Hello from the @GoogleDeepMind Gemini hackathon from NYC! With @cerebral_valley and @temporal_xyz Been a while since I’ve done one of these, the @llama_index workshop is about to start, come learn all about document agents with LlamaParse and LlamaAgents 🫶 Original tweet: https://x.com/tuanacelik/status/2027780520570921178
RT Simon Suo Realize why I love agentic coding so much now It makes software engineering feel like ML research - kick off a bunch of agents (experiments) - monitor their trajectory (loss curves) - kill some & double down on others - async & orchestration Original tweet: https://x.com/disiok/status/2027666939234193687
what's the point of netflix when i have this app
Tonight, we reached an agreement with the Department of War to deploy our models in their classified network. In all of our interactions, the DoW displayed a deep respect for safety and a desire to partner to achieve the best possible outcome. AI safety and wide distribution of
View quoted postRT Clelia Bertelli (🦙/acc) New blog post alert!🦙✨ At @llama_index we shipped LlamaAgents Builder to make it the fastest way to build a document agent only with natural language. Yesterday, we published a full walkthrough of how to use LlamaAgents Builder end-to-end with a real-world use case, covering prompt and context engineering best practices, workflow visualization tricks and tips on how to best iterate with the builder over multiple turns. Read the blog: https://www.llamaindex.ai/blog/creating-a-deal-sourcing-agent-with-llamaagents-builder Original tweet: https://x.com/itsclelia/status/2027498003851055531
RT Tuana Since joining @llama_index, my focus has shifted from 'everything agents' to 'document agents' : agents that can handle work over all manner of complex documents. So, I tried out the latest chart parsing capabilities of LlamaParse. Charts in PDFs are notoriously painful to work with. You can see the data ) bars, axes, labels) but actually getting it into a format you can analyze means is a different matter. I tried out parsing a U.S. Treasury executive summary PDF, pulling a grouped bar chart showing Budget Deficit vs. Net Operating Cost for fiscal years 2020–2024, and turning it into a pandas DataFrame you can run analysis on (although really you can then do whatever, provide it for downstream tasks to an agent..) Once parsed, the chart's underlying data comes back as a table in the items tree for that page. From there: grab the rows, construct a DataFrame, etc. In the example, I'm computing year-over-year changes in both metrics, measuring the gap between them across the five-year window, and just to be sure, I reproduced a bar chart that mirrors the original PDF visualization. You can try it our here: https://colab.research.google.com/drive/1wr_b7JQIiBk998qiaUNxpFa6RvxwZlws?usp=sharing Original tweet: https://x.com/tuanacelik/status/2027488010640765042
RT LlamaIndex 🦙 Turn your PDF charts into pandas DataFrames with specialized chart parsing in LlamaParse! This tutorial walks you through extracting structured data from charts and graphs in PDFs, then running data analysis with pandas - no manual data entry required. 📊 Enable specialized chart parsing to convert visual charts into structured table data 🐼 Extract table rows directly from parsed PDF pages and load them into DataFrames 📈 Perform year-over-year analysis, calculate gaps between metrics, and create visualizations ⚡ Use the items view to get per-page structured data including tables and figures We demonstrate this using a 2024 Executive Summary PDF, extracting a fiscal year chart showing Budget Deficit vs Net Operating Cost data spanning 2020-2024, and reproducing the key financial insights. Check out the full tutorial: https://developers.llamaindex.ai/python/cloud/llamaparse/tutorials/parse_charts_pandas/?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2027429029834531120
We’ve made an in-depth tutorial on building a PE deal-sourcing 🤝 document workflow, by simply typing a natural language prompt There’s often a set of documents coming inbound to let analysts understand what deals to invest in - teasers, CIMs, emails, presentations With our LlamaAgents builder, you can describe what documents are coming in, and the information you want to extract from each document. We perform high-quality document OCR to extract out all the key details from these documents, and give you back an application where you can process these docs at scale. Thanks to @itsclelia for this tutorial. If you’re interested, come check it out: https://cloud.llamaindex.ai/?utm_campaign=parse&utm_medium=jl_socials If you have these applications at scale, come talk to us: https://www.llamaindex.ai/contact?utm_campaign=parse&utm_medium=jl_socials
Build a private equity deal sourcing agent that automatically classifies investment opportunities and extracts key financial metrics using our LlamaAgents Builder. This step-by-step guide shows you how to create an agent that processes deal files like teasers and financial
RT Simon Suo should we rebrand to SF model harness company Original tweet: https://x.com/disiok/status/2027073086055764035
The Model Harness is Everything We are already living in a world of incredible frontier models and incredible agent tools (Claude Code, OpenClaw). But the biggest barrier to getting value from AI is your own ability to context and workflow engineer the models. This is
The Model Harness is Everything We are already living in a world of incredible frontier models and incredible agent tools (Claude Code, OpenClaw). But the biggest barrier to getting value from AI is your own ability to context and workflow engineer the models. This is *especially* true the more horizontal the tool that you’re using. If you’re using a very generic tool like ChatGPT and Claude Code, you need to spend a lot of work clearly articulating your requirements and specifications so that the agent can actually solve the task relative to your specifications. Today that looks like being extremely thoughtful about the tools that you select, and writing English very precisely in a http://skills.md file to articulate the agent these requirements. Some of the work around defining the business workflow is inherently time consuming. Think about any document SOP - simply writing the English can take hours to refine, iterate, and optimize. This is where more vertically focused agents come in; they handle the burden of equipping the agents with relevant prompts to solve a given workflow, so that you can just go in and use the application directly. Another approach is to be specialized services that offer *context* to these agents. This is the space that we (@llama_index) are operating in. We are providing the infrastructure to parse the most complex documents into agent-ready context. For other companies it could be offering web data, sales data, documentation, or codebases as a service. At a high-level any AI startup should provide context or workflows on top of these agents. We’re excited about building enduring tech even as the agent landscape evolves. If you’re specifically excited to unlock the billions of context stored within your documents, come talk to us! https://www.llamaindex.ai/contact
OmniDocBench is getting saturated VLMs are getting increasingly better at document understanding, from OSS (DeepSeek-OCR2, GLM-OCR), to frontier (Gemini 3, Kimi 5.2, GPT-5.2). A popular benchmark to measure document understanding progress has been OmniDocBench. But we're quickly approaching the point where we need a new benchmark. 1. The latest models are pushing ~95% on OmniDocBench and are already overfitting the benchmark, while still having real gaps on document capabilities. 2. The evaluation metrics of OmniDocBench depend completely on exact match and not on semantic correctness. The latter is much more important, especially in today's world where LLMs can reason over text tokens regardless of non-important formatting differences. *(see example below; our own service llamaparse is penalized even though our parsing is 100% semantically correct on scientific notation parsing)* There are so many real-world documents that haven't yet been solved by even the latest models. We'd love to welcome discussion on advancing document understanding benchmarks that are more diverse and properly score models on semantic correctness. Blog: https://www.llamaindex.ai/blog/omnidocbench-is-saturated-what-s-next-for-ocr-benchmarks?utm_source=socials&utm_medium=li_social
Document OCR benchmarks are hitting a ceiling - and that's a problem for real-world AI applications. Our latest analysis reveals why OmniDocBench, the go-to standard for document parsing evaluation, is becoming inadequate as models like GLM-OCR @Zai_org achieve 94.6% accuracy
RT LlamaIndex 🦙 Document OCR benchmarks are hitting a ceiling - and that's a problem for real-world AI applications. Our latest analysis reveals why OmniDocBench, the go-to standard for document parsing evaluation, is becoming inadequate as models like GLM-OCR @Zai_org achieve 94.6% accuracy while still failing on complex real-world documents. 📊 Models are saturating OmniDocBench scores but still struggle with complex financial reports, legal filings, and domain-specific documents 🎯 Rigid exact-match evaluation penalizes semantically correct outputs that differ in formatting (HTML vs markdown, spacing, etc.) ⚡ AI agents need semantic correctness, not perfect formatting matches - current benchmarks miss this critical distinction 🔬 The benchmark's 1,355 pages can't capture the full complexity of production document processing needs The document parsing challenge isn't solved just because benchmark scores look impressive. We need evaluation methods that reward semantic understanding over exact formatting, especially as AI agents become the primary consumers of parsed content. We're building parsing models focused on semantic correctness for complex visual documents. If you're scaling OCR workloads in production, LlamaParse handles the edge cases that benchmarks miss. Read our full analysis: https://www.llamaindex.ai/blog/omnidocbench-is-saturated-what-s-next-for-ocr-benchmarks?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2026342120236396844
RT Clelia Bertelli (🦙/acc) Hey Rustacean friends!🦀 This Friday I'll be talking about how we migrated to Qdrant Edge for the storage layer of Semtools, @llama_index's Rust-powered CLI toolkit for local document intelligence🦙 The @qdrant_engine team and I will discuss lessons learnt when migrating to Qdrant Edge as a backend for on-disk vector storage in Rust, as well as the performance gains and common pitfalls to avoid👩💻 📍The live will start at 4PM CET and you can join us here: https://discord.gg/PcgpnXa2?event=1473577301470875820 If you're building AI systems in Rust, don't miss out!🦀 Original tweet: https://x.com/itsclelia/status/2026227327370469457
Gemini 3.1 Pro <> LlamaParse Use smaller VLMs for high-accuracy document understanding, use frontier reasoning models to synthesize insights over your documents. Check out this e2e expense analysis example over receipts here!
🚀 Big drop from @GoogleDeepMind: Gemini 3.1 Pro is here, and we built a hands-on demo powered by LlamaCloud to put it to work and turn your receipt photos into real financial insights! Using our Agent Workflows, the app: 📸 Parses receipt images with LlamaParse (Agentic tier)
View quoted postWe built an AI agent that lets you vibe-code document extraction - high accuracy and citations over the most complex documents. Our latest release lets you upload documents as context. All you then have to do is describe what you want extracted in natural language. 💡 Our agent will then read the document with file tools to infer the right schema, validation rules, and other pre/postprocessing logic. ✅ It will give you back a workflow that can extract over thousands/millions of documents at scale. You can still of course review and edit every output before approving. Stop handling paperwork manually; just upload files, describe your task, and let our agent handle the rest. Our vision for LlamaAgents is to provide the most advanced and easy-to-use way for you to orchestrate document work. Walkthrough: https://youtu.be/5Nk6KZhBDbQ Check it out: https://cloud.llamaindex.ai/ If you’re interested in reducing the operational burden of document extraction (invoices, claims, onboarding forms), come talk to us! https://www.llamaindex.ai/contact
🚀 LlamaAgents Builder just leveled up: File uploads are here! Our natural language interface for building agentic document workflows now supports file uploads. You can provide example documents as context, and the agent will use them as a starting point to design and tailor
View quoted postRT Clelia Bertelli (🦙/acc) We’ve been cooking at @llama_index🍳 Almost a month ago, we launched LlamaAgents Builder, a natural language interface for building agentic document workflows📁 Today, we’re excited to announce file upload support: add your PDFs, and the agent will use them as context to build your use case🚀 The more representative your examples, the more accurate the generated application will be📈 🎥 Watch the full walkthrough: https://youtu.be/5Nk6KZhBDbQ 🦙 Get started with LlamaCloud: https://cloud.llamaindex.ai/signup Original tweet: https://x.com/itsclelia/status/2025989350216081758
RT LlamaIndex 🦙 🚀 LlamaAgents Builder just leveled up: File uploads are here! Our natural language interface for building agentic document workflows now supports file uploads. You can provide example documents as context, and the agent will use them as a starting point to design and tailor your workflow. The result? Applications that better match your real-world use case. The more representative your sample files, the more accurate your final app. 🎥 Watch the full walkthrough: https://youtu.be/5Nk6KZhBDbQ 🦙 Get started with LlamaCloud: https://cloud.llamaindex.ai/signup Original tweet: https://x.com/llama_index/status/2025978751172096324
it's always fun when twitter hype makes a dent in the real world tbh apple should've just bought them to capture the mindshare of personalized digital assistants. 🍎🦞
Bought a new Mac mini to properly tinker with claws over the weekend. The apple store person told me they are selling like hotcakes and everyone is confused :) I'm definitely a bit sus'd to run OpenClaw specifically - giving my private data/keys to 400K lines of vibe coded
View quoted postThe second highest category is backoffice automation, but imo it's underrated by the AI community. RPA is truly dead, and agentic workflows are taking its place. A lot of backoffice work depends on routine operations over unstructured documents (invoices, claims packets, loan files). The best interface to automate these operations is enabling users to create deterministic workflows at scale, instead of solving ad-hoc tasks through chat. We are starting to build an agentic layer within our own document processing product, LlamaCloud, that lets users "vibe-code" these workflows through natural language. Come check it out: https://cloud.llamaindex.ai/
Software engineering makes up ~50% of agentic tool calls on our API, but we see emerging use in other industries. As the frontier of risk and autonomy expands, post-deployment monitoring becomes essential. We encourage other model developers to extend this research.
Agentic extraction from SCOTUS opinions 🧑⚖️ Today SCOTUS struck down Trump’s tariffs in a 6-3 vote, and one of the most interesting points was Gorsuch’s concurrence, where he devotes a decent portion calling out every justice by name - especially around everyone in the dissent (Kavanaugh, Thomas, Alito by extension). I was curious about this, so I used LlamaExtract on the official court opinion to generate a schema that not only gives me an overall summary, but extracts particular disagreements Gorsuch has in each line item. * We extract the specific disagreements with Kavanaugh and Thomas line by line * For each key, you can also trace back via bounding boxes to the source document so you can read the text! Brief: https://www.supremecourt.gov/opinions/25pdf/24-1287_4gcj.pdf Whether it’s court opinions, legal briefs, or any other complex doc, if you need to do document extraction at scale come check out LlamaCloud (particularly our “extract” feature): https://cloud.llamaindex.ai/
We built a vibe-coding tool to help you accurately extract out information from any document type, simply by describing the workflow in natural language. In the latest walkthrough, @tuanacelik shows you how to split a package of resumes and extract out candidate information from every single resume, simply by describing the task in English. Our agent builder is built on top of LlamaParse. It’s available today - come check it out! https://cloud.llamaindex.ai/
I filmed a walkthrough of LlamaAgent Builder, our new tool for building document agents by just describing what you want @llama_index I revisited my old demo: I took a resume book from NYU (resumes mixed with cover pages and curriculum pages) and just told the agent builder:
View quoted postRT LlamaIndex 🦙 🚀 Big drop from @GoogleDeepMind: Gemini 3.1 Pro is here, and we built a hands-on demo powered by LlamaCloud to put it to work and turn your receipt photos into real financial insights! Using our Agent Workflows, the app: 📸 Parses receipt images with LlamaParse (Agentic tier) 🗂 Stores everything locally in an SQLite database 📊 Aggregates your spending monthly 🧠 Uses Gemini 3.1 Pro to analyze trends and generate actionable tips to improve your finances Check out the demo below!👇 👩💻 GitHub repo: http://github.com/run-llama/receipts-analyzer 🦙 Get started with LlamaCloud: http://cloud.llamaindex.ai/signup Original tweet: https://x.com/llama_index/status/2024892621911679102
Coding agents are fundamentally changing software engineering in terms of velocity, role, and org structure. We published a memo to our internal engineering team detailing our growing expectations in terms of role/scope. 🟠 Before, the tasks of prioritization, engineering planning, and implementation were divided between EMs, PMs, senior ICs, and junior ICs 🟢 Now, ICs are expected to handle *all* of product prioritization, product speccing, and implementation This is due to a few trends 📈: - Coding agents have brought implementation costs down to ~0. The role of engineers is writing prompts - LLMs and sub-agents have reduced the PM work of synthesizing feedback down to ~0 too The main job of any “engineer” is to be an e2e product owner: being able to translate requirements into specifications, and delegate tasks to various subagents for implementation. Every engineer is told to offload as much as possible to their favorite tools, whether it’s Claude Code, Cursor, Devin, Codex, regular ChatGPT and more. We celebrate and share learnings around burning tokens, as long as it helps drive additional productivity!
Increased thinking doesn't correlate with increased document understanding Models are getting much better at reasoning which translates well for math/coding/general intelligence tasks. We did some experiments with GPT-5.2 on different thinking modes to see if it actually helps improves scores on OmniDocBench. Result: it doesn't, and in fact the higher the thinking, the more the output structure might deviate from the existing input. Shoutout to Boyang on the @llama_index team for this blog. Come check out our post! https://www.llamaindex.ai/blog/the-cost-of-overthinking-why-reasoning-models-fail-at-document-parsing
More reasoning doesn't always mean better results - especially for document parsing. We tested GPT-5.2 at four reasoning levels on complex documents and found that higher reasoning actually hurt performance while dramatically increasing costs and latency. 🧠 Reasoning models
RT LlamaIndex 🦙 More reasoning doesn't always mean better results - especially for document parsing. We tested GPT-5.2 at four reasoning levels on complex documents and found that higher reasoning actually hurt performance while dramatically increasing costs and latency. 🧠 Reasoning models hallucinate content that isn't there, filling in "missing" table cells with inferred values 📊 They split single tables into multiple sections by overthinking structural boundaries ⚡ Processing time increased 5x with xHigh reasoning (241s vs 47s) while accuracy stayed flat at ~0.79 💰 Our LlamaParse Agentic outperformed all reasoning levels at 18x lower cost and 13x faster speed You can't reason past what you can't see. Vision encoders lose pixel-level information before reasoning even starts, and no amount of thinking tokens can recover that lost detail. Our solution uses a pipeline approach - specialized OCR extracts text at native resolution, then LLMs structure what's already been accurately read. Each component plays to its strengths instead of forcing one model to handle everything. Read the full analysis: https://www.llamaindex.ai/blog/the-cost-of-overthinking-why-reasoning-models-fail-at-document-parsing?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2024529937462706517
RT Tuana I filmed a walkthrough of LlamaAgent Builder, our new tool for building document agents by just describing what you want @llama_index I revisited my old demo: I took a resume book from NYU (resumes mixed with cover pages and curriculum pages) and just told the agent builder: split this into individual resumes, ignore the rest, extract graduation year, work experience, etc. The agent builder figured out it needed Split and Extract, configured both, built the workflow, and deployed it: API + UI, code in my GitHub the whole point is you're describing the problem, not building the pipeline. the agent builder decides the architecture BUT with the caveat that coding agents aren't perfect and the code is yours to edit and perfect! Give it a try and let me know what you build. We're running a contest for the most difficult document workflow you can throw at it. Full video here: https://youtu.be/0Zhf5z2Onjs?si=Em6uUZPxQz6YNlyY Original tweet: https://x.com/tuanacelik/status/2024500524851298764
We’re on a mission to parse the world’s hardest PDFs, and we’d love your help There are so many document types that introduce a million edge cases for current VLMs / OCR: handwritten forms, badly scanned/rotated pages, charts, diagrams, and more. We are running a contest right now for you to try to extract the hardest PDFs you can find. Come sign up on our agent builder, describe what you want to extract through natural language, upload your document, and show the results. If our platform doesn’t work, even better; this is great feedback for us to improve our service. Either way submit your project and we’d love to get your feedback! Check out LlamaCloud here: https://cloud.llamaindex.ai/
RT LlamaIndex 🦙 🏆 We're running a LlamaAgents contest right now. Throw your hardest documents at our agent builder, and tell us how it goes. Want help getting started? We have a new walkthrough for the LlamaAgent Builder by @tuanacelik 💬 Describe a document workflow in natural language, and it builds a full agent for you. In this video, the prompt was basically: "split a resume book into individual resumes, ignore cover pages and curriculum pages, extract resume work and education related fields..." 🛠️ From that, the agent builder reasons about which LlamaCloud tools to use, lands on LlamaSplit + LlamaExtract, configures both, iterates on the workflow structure, and gives you a deployable agent with an API and UI. No dragging boxes around. No writing workflow code (unless you want to). Just describe the problem and let it figure out the architecture. You own the code, it pushes to your GitHub. Clone it, open in Cursor, customize whatever you need. https://www.youtube.com/watch?v=0Zhf5z2Onjs Original tweet: https://x.com/llama_index/status/2024176418767429826
Extracting large-scale structured information from complex PDFs is really hard, even with LLMs. 80% accuracy isn’t good enough; a lot of business applications require 98%+. You need specialized capabilities that can not only parse complex tables and charts, but also trace back to the source elements with confidence scores and citations. We’ve created a best in class document extraction service with LlamaExtract, and you can see for yourself in ~5 mins! Simply log on to LlamaCloud and click on our templates to take a look. Check out LlamaCloud here: https://cloud.llamaindex.ai/ If you’re looking to productize document extraction, come talk to us: https://www.llamaindex.ai/contact
A lot of documents are extremely dense and repetitive in information: stapled together resumes, invoices, insurance claims, loan applications 📑 One of the main promises of AI is being able to automatically extract structured information from massive amounts of unstructured context. But in order to do this accurately over these massive documents, you need: ✅ Page attribution ✅ Bounding boxes linking back to the source text ✅ No hallucinations or dropped outputs, even if there are *hundreds* of structured fields ✅ Calibrated confidence scores to map to human review. This is exactly what we're building with our Extract feature in LlamaCloud. Our page-level extraction lets you extract massive amounts of dense information at scale from all these documents with all the auditability and citations you need, letting humans review documents much faster than before. https://www.llamaindex.ai/blog/beyond-full-text-extraction-why-page-level-granularity-matters
"It's somewhere in the PDF" is not a citation. Page-level extraction in LlamaExtract gives you: ✓ Data mapped to specific pages ✓ Bounding boxes showing exact locations ✓ Audit-ready citations Turn 200-page docs into skimmable, structured insights 👇 https://www.llamaindex.ai/blog/beyond-full-text-extraction-why-page-level-granularity-matters
View quoted postRT Ankur Goyal We sent this note to our customers to let them know that Braintrust has raised a new round of funding, and thank them for their support. While the money is exciting, our focus hasn't changed: we're building Braintrust to help our customers ship quality AI products. In 2026, AI is moving to production but teams have never had less conviction about what will fail next. Our customers are building AI products that serve millions and simply need to work. If Braintrust makes their lives easier and their products better, I know we are doing our job. Thank you to @ICONIQCapital for leading our Series B, and to @a16z, @GreylockVC, @basecasevc, and @eladgil for doubling down. Thank you to the Braintrust team for all the incredible work you've done over the past year. And thank you to our customers, who have made this growth possible. Original tweet: https://x.com/ankrgyl/status/2023810273598128588
RT LlamaIndex 🦙 "It's somewhere in the PDF" is not a citation. Page-level extraction in LlamaExtract gives you: ✓ Data mapped to specific pages ✓ Bounding boxes showing exact locations ✓ Audit-ready citations Turn 200-page docs into skimmable, structured insights 👇 https://www.llamaindex.ai/blog/beyond-full-text-extraction-why-page-level-granularity-matters Original tweet: https://x.com/llama_index/status/2023804723875508302
RT Tuana Most document AI is either rigid extraction pipelines or "upload a PDF and chat with it." Both useful. But documents that we want agents to do "work" on are ever changing. Example, for research: you gather sources, draft, get comments, revise, do more research, then the source material changes and you start again. That doesn't fit into a single prompt-response cycle. So, the next step isn't making agents faster at responding to prompts. It's making them less dependent on prompts. AI agents that react to events: · A trigger: new files, comments, updated source docs · The continuous response: is to manage their own task queue instead of waiting for you to tell them what to do next. Your memo isn't a one-time deliverable, it's a living artifact that rebuilds when inputs change. This blog on long-horizon document agents is definitely a good read to understand how we're thinking about the future of document agents: https://www.llamaindex.ai/blog/long-horizon-document-agents Original tweet: https://x.com/tuanacelik/status/2023789925465027005
Our document parsing is really good You can see for yourself with our new feature💫: convert complex PDFs with tables, charts, multi-column layouts to clean markdown/JSON representations through our new clickable templates! 1. We convert complex tables with gaps into clean markdown structures 2. We parse chart + line graphs into interpretable 2d tables Check out LlamaCloud: https://cloud.llamaindex.ai/?utm_campaign=parse&utm_medium=jl_socials Come talk to us: https://www.llamaindex.ai/contact
RT LlamaIndex 🦙 What if an AI agent could review every invoice against your contracts — and flag what doesn't match? That's exactly what our Invoice Reconciler demo does. Here's how it works: 📄 Upload your contracts and invoices → LlamaParse converts them into clean, LLM-readable Markdown 📂 Everything gets indexed in LlamaCloud — searchable and ready for RAG 🔍 Define your reconciliation rules (unit price match, correct math, line item match, etc.) 🤖 A LlamaAgent workflow analyzes each invoice against your contracts and rules — then approves or rejects with confidence scores and detailed reasoning You can even chat with your invoices and contracts directly — ask "what have we bought?" or "what contracts do we have in place?" and get cited answers instantly. The whole thing is powered by LlamaCloud: LlamaParse for document ingestion, LlamaCloud indexes for retrieval, and LlamaAgent Workflows for orchestration. 🎥 Watch the full walkthrough: https://www.youtube.com/watch?v=DHFAYWYIxuA Original tweet: https://x.com/llama_index/status/2023443263651451294
we are all jestermaxxing today i'm so sorry my brain cells are dying
Just in case Gen Z is trying to understand what happened today: Claude was mogging OpenAI for weeks. Then this gymcel dev ships Clawdbot which was the fastest growing OSS thing ever, absolute looksmax for the whole ecosystem. Anthropic tries to dairygoon him with legal. Dev
View quoted postRT Simon Suo honestly an Anthropic fumble Original tweet: https://x.com/disiok/status/2023153271305908497
Peter Steinberger is joining OpenAI to drive the next generation of personal agents. He is a genius with a lot of amazing ideas about the future of very smart agents interacting with each other to do very useful things for people. We expect this will quickly become core to our
View quoted postI parsed OpenAI’s tax filings for fun 📋 I used the “Extract” capability in LlamaCloud to automatically extract out all relevant fields into structured JSON outputs. You can see their mission statement “ensure that artificial general intelligence benefits all of humanity” with the corresponding bounding boxes. (Inspired by @simonw’s post on this recently: https://simonwillison.net/2026/Feb/13/openai-mission-statement/) The extraction is powered by our core OCR engine, LlamaParse. LlamaParse is able to reconstruct this complex form PDF into markdown tables that captures each cell with ~100% accuracy. Check it out: https://cloud.llamaindex.ai/
RT ODSC (Open Data Science Conference) AI In this episode of the ODSC Ai X Podcast, we speak with @jerryjliu0, @llama_index, about the shift from early “RAG frameworks” to document AI workflows. 🎧Listen to the full episode here - https://hubs.li/Q042_7Wq0 Original tweet: https://x.com/_odsc/status/2022838519886926130
We have a native integration with @posthog! Parse, extract, and reason from your docs, and watch it show up in the analytics dashboard. Check out the example below 👇
🚀 The @posthog team has just rolled out LlamaIndex support for their LLM Analytics, and we built a demo to showcase what’s possible. Using LlamaIndex, LlamaParse, and OpenAI, our Agent Workflow compares product specifications and matches users with the most suitable option for
View quoted postRT LlamaIndex 🦙 🚀 The @posthog team has just rolled out LlamaIndex support for their LLM Analytics, and we built a demo to showcase what’s possible. Using LlamaIndex, LlamaParse, and OpenAI, our Agent Workflow compares product specifications and matches users with the most suitable option for their use case 🛠️ 🦔 Thanks to PostHog’s observability integration, the demo automatically tracks OpenAI usage, including: •Token consumption •Cost breakdown •Latency metrics 🎥 Check out the video below to see it in action 👇 👩💻 GitHub: https://github.com/run-llama/product-specs-comparison 📚 Docs: https://posthog.com/docs/llm-analytics/installation/llamaindex 🦙 LlamaCloud: https://cloud.llamaindex.ai/signup Original tweet: https://x.com/llama_index/status/2022355660504207766
RT LlamaIndex 🦙 Next week at @DeveloperWeek, fuel up for hackathon success with free breakfast at the same bakery that fed the Superbowl champions! 🥐 We're teaming up with @kilocode, @MiniMax_AI, and @withmartian to host breakfast for all @DeveloperWeek hackathon attendees on February 19th in San Jose. Look forward to: 🥞 Special edition cookies, breakfast burritos, and and plenty of coffee to start your day right 🚶 Just a 5-minute walk from the convention center 🤝 Zero networking pressure, just good food and good company with fellow developers ⚡ Get energized before you dive into building something amazing Sign up for free breakfast: https://luma.com/devchampions Original tweet: https://x.com/llama_index/status/2022083315302547738
Extracting complex patent documents 💡📄 with natural language Patent documents are long and have varied sections: claims, prior art references, figures, and more. Our agent builder lets you express what you want to extract through natural language, and builds a workflow that lets you extract 1M+ patent documents accurately at scale! Check out the video below. Define the overall extraction in the chat builder, which deploys an agent app that can extract over any document and flag uncertain fields. If you have complex documents that you want to analyze, come check out our agent builder in LlamaCloud. You don’t even have to write a single line of code! https://cloud.llamaindex.ai/
This is AGI
I built a Claude Code notification system that uses Warcraft III Peon voice lines. It's probably the stupidest thing I've ever shipped. And according to everybody that has used it, it's also incredibly useful. (sound on)
View quoted postExisting AI agents are largely short-horizon (e.g. chat) or constrained (e.g. agentic process automation). @sequoia predicted that 2026 is the year of long-horizon agents, and long-horizon agents ~= AGI We’re already seeing this with coding agents. But I’m specifically excited about how this translates to non-technical knowledge work over documents (e.g. legal, finance, insurance, back office) that humans typically perform for hours on end. I wrote a blog post that articulates the broader class of use cases this would unlock, and the ideal architecture/UX of an “agent inbox” that allows agents to run autonomously without being blocked on human input. Unlike a chat interface. Some examples: 1. An agent that can continuously create a living FAQ from your Sharepoint, Slack, call transcripts 2. An agent that can e2e interface with lawyers to do contract redlining This fundamentally requires building agents that can monitor event triggers beyond human chat inputs, and can use core capabilities around parsing, extracting, creating, and editing documents of various formats. Check out the blog! https://www.llamaindex.ai/blog/long-horizon-document-agents If you’re interested in the building blocks around parsing, extraction, with more coming soon, come check out LlamaCloud: https://cloud.llamaindex.ai/
RT LlamaIndex 🦙 2026 is the year of long-horizon agents. @sequoia predicts that this year, agents will be able to tackle long-horizon tasks and work autonomously for hours to solve ambiguous tasks. We're excited about how this translates to knowledge work automation, particularly over documents. Let's take a look at "Long Horizon Document Agents" 🕰️ Agents are evolving to work autonomously over weeks, not just minutes, handling complex document tasks end-to-end. 🔄 These agents can continuously monitor events like document changes, comments, and deadlines - not just respond to chat prompts 📝 They maintain persistent task backlogs and can collaborate iteratively on living documents like FAQs, PRDs, and legal contracts 🎯 The interface shifts from chat boxes to "agent inboxes" that manage ongoing document tasks with clear status and context ⚡ This enables true automation of multi-step knowledge work - from due diligence memo updates to contract redline collaboration loops 2026 is shaping up to be the year agents evolve from "workflows" to "employees" - and we're building the document processing infrastructure to make this possible. Read @jerryjliu0's full blog on long horizon document agents: https://www.llamaindex.ai/blog/long-horizon-document-agents?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2021992214038241477
We love parsing diagrams. Anthropic’s recent report on coding trends has a nice diagram on the evolution from single-agent to hierarchical multi-agent architectures With our latest VLM-enabled document parsing, we’re able to one-shot this diagram into a `mermaid` plaintext representation! Check out the results below. This capability lets you convert even the most complex diagrams within PDFs/Powerpoints into digestible graph representations that LLMs can understand. This lets you use AI to understand complex docs at scale; VLMs either can’t understand these diagrams out of the box, or you also end up burning unnecessary vision tokens. The report itself is an interesting overview of multi-agents, check it out: https://resources.anthropic.com/hubfs/2026%20Agentic%20Coding%20Trends%20Report.pdf?hsLang=en For diagram parsing, sign up to LlamaCloud: https://cloud.llamaindex.ai/ If you’re interested in chatting more about this, come talk to us: https://www.llamaindex.ai/contact
who needs Clay for outbound personalization when you have this
Coding agents are not great for open-source software, and it needs to adapt This is a really nice article from @LoganMarkewich that draws from our experience maintaining our own OSS library in the past 3 years. 1. Open-source used to be nice for human knowledge sharing. Now you really need to optimize it so that coding agents can pick it up. 2. Light abstractions (wrappers around LLMs, vector databases, etc.) are dead 3. There is a massive slopocalypse of ai-generated PRs, which breaks the fun community feeling of OSS. If you want to lean in and accept this, you need to document patterns clearly so coding agents can pick it up https://www.llamaindex.ai/blog/on-the-incoming-slopocalypse-and-the-death-of-open-source?utm_source=socials&utm_medium=li_social
The rise of coding agents is fundamentally changing open source - Our head of OSS @LoganMarkewich breaks down how LLM-powered coding agents are impacting core pillars of open source: 👥 Community interaction, which is getting complicated by low-quality, massive AI-generated PRs
RT LlamaIndex 🦙 The rise of coding agents is fundamentally changing open source - Our head of OSS @LoganMarkewich breaks down how LLM-powered coding agents are impacting core pillars of open source: 👥 Community interaction, which is getting complicated by low-quality, massive AI-generated PRs 💪 Personal skill development suffers when developers rely too heavily on AI assistance 🧠 Knowledge sharing is shifting as LLMs become the frontend for learning But open source isn't dead - it's evolving. We're shifting toward hackable reference implementations, community-driven knowledge sharing, and agent-friendly codebases that work with AI tools rather than against them. Read the full blog by Logan on how he views this evolution of open source projects: https://www.llamaindex.ai/blog/on-the-incoming-slopocalypse-and-the-death-of-open-source?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2021631328236802318
Scaling Document Ingestion for AI Agents We're excited to partner with StackAI on this webinar. We'll show you how to build a modern agent stack that can automate knowledge work over millions of documents, across finance, legal, insurance use cases and more. Come check it out! https://lnkd.in/gzrWt33y
Are you trying to solve high-quality document ingestion for your product? Gain lessons from the field on how @stackai uses LlamaCloud to power high-accuracy document ingestion & retrieval across PDFs, images, spreadsheets & more — at enterprise scale. ➡️ Register now:
Parsing PDFs at scale with LLMs is cost prohibitive. Newer models (e.g. gemini 3) are good at reading pdfs, but you burn unnecessary vision tokens even when the page is text heavy. We’ve built in a “cost-optimizer” within LlamaParse that will dynamically route pages to fast/cheap parsing depending on its complexity. Complex pages (e.g. those with tables/charts/diagrams) will still get routed to our VLM-enabled modes. This will let you save anywhere from 50-90% of parsing costs, at much higher accuracy compared to the comparable mode of feeding screenshots into VLMs. Check it out! https://cloud.llamaindex.ai/
RT LlamaIndex 🦙 Are you trying to solve high-quality document ingestion for your product? Gain lessons from the field on how @stackai uses LlamaCloud to power high-accuracy document ingestion & retrieval across PDFs, images, spreadsheets & more — at enterprise scale. ➡️ Register now: https://landing.llamaindex.ai/webinar-stackai-and-llamaindex-scaling-document-ingestion Original tweet: https://x.com/llama_index/status/2021259881823834552
RT ODSC (Open Data Science Conference) AI In this episode of the ODSC Ai X Podcast, we speak with @jerryjliu0, @llama_index, about the shift from early “RAG frameworks” to document AI workflows. 🎧Listen to the full episode here - https://hubs.li/Q042hjhv0 Original tweet: https://x.com/_odsc/status/2021026586779111700
We built LobsterX 🦞, an @openclaw specialized for document work on your computer. It uses high-accuracy document parsing, extraction, classification through LlamaCloud, meaning it can comb through complicated PDFs (with scans, tables, diagrams) and extract out 100% accurate context! It can run as a Telegram bot and is built on top of agentfs (@tursodatabase) as a file system. Big shoutout to @itsclelia. This is a fun project inspired by @openclaw’s success, and besides being a fun tool to use, it can be a great reference for building your own generalized coding agents! Readme: https://github.com/AstraBert/workflows-acp/blob/main/packages/lobsterx/README.md LlamaCloud: https://cloud.llamaindex.ai/signup
The tech world went crazy for @openclaw, so I decided to build a similar crustacean agent called LobsterX🦞, with a focus on document-processing tasks. I wrote a blog about it: http://clelia.dev/2026-02-09-the-anatomy-of-a-document-processing-agent, but, as a tl;dr: - LobsterX uses @llama_index cloud products to parse, extract
View quoted postRT Clelia Bertelli (🦙/acc) The tech world went crazy for @openclaw, so I decided to build a similar crustacean agent called LobsterX🦞, with a focus on document-processing tasks. I wrote a blog about it: http://clelia.dev/2026-02-09-the-anatomy-of-a-document-processing-agent, but, as a tl;dr: - LobsterX uses @llama_index cloud products to parse, extract structured data and classify files - The agent has only access to a virtual filesystem (AgentFS by @tursodatabase) to avoid damaging your real one, and cannot execute arbitrary bash commands, preventing it from performing dangerous or security-critical operations🔒 - The agent is self-hostable and can be used both as a @Docker image and as a uv tool, and is available as a @telegram bot💬 📦 Install: 𝘶𝘷 𝘵𝘰𝘰𝘭 𝘪𝘯𝘴𝘵𝘢𝘭𝘭 𝘭𝘰𝘣𝘴𝘵𝘦𝘳𝘹 --𝘱𝘳𝘦𝘳𝘦𝘭𝘦𝘢𝘴𝘦=𝘢𝘭𝘭𝘰𝘸 📚 Read more: http://github.com/AstraBert/workflows-acp/blob/main/packages/lobsterx/README.md 📝 Read the blog: http://clelia.dev/2026-02-09-the-anatomy-of-a-document-processing-agent Original tweet: https://x.com/itsclelia/status/2020913910686355709
RT LlamaIndex 🦙 Everybody’s talking about @openclaw, so @itsclelia decided to build her own crustacean AI assistant for document workflows: LobsterX 🦞 LobsterX is built on top of our Agent Workflows and leverages LlamaCloud for document parsing, structured data extraction, and classification via powerful, modular tools🛠️ Designed with a safety-first mindset, the agent runs on AgentFS (by @tursodatabase) to protect your real filesystem and intentionally avoids full shell access to prevent security-critical or destructive operations 🔒 It’s fully self-hostable, can be run as a uv tool or @Docker container, and works out of the box as a @telegram bot 💬 📦 Install: 𝘶𝘷 𝘵𝘰𝘰𝘭 𝘪𝘯𝘴𝘵𝘢𝘭𝘭 𝘭𝘰𝘣𝘴𝘵𝘦𝘳𝘹 --𝘱𝘳𝘦𝘳𝘦𝘭𝘦𝘢𝘴𝘦=𝘢𝘭𝘭𝘰𝘸 📚 Read more: https://github.com/AstraBert/workflows-acp/blob/main/packages/lobsterx/README.md 🦙 Sign up to LlamaCloud: https://cloud.llamaindex.ai/signup Original tweet: https://x.com/llama_index/status/2020906615323623642
Yes he is on Singles Inferno but also we would unironically hire him (Samuel if you can’t find love you can always find AI)
In hindsight, a quant trader going on a Netflix dating show was definitely the top signal…
potrero hill in C tier is 100% rage bait
San Francisco Neighborhood Ranking S Tier: Pac Heights, Cole Valley, North Beach A Tier: Marina, Noe valley, Richmond, Hayes valley, Nob hills B Tier: Russian Hill, Haight Ashbury, Rincon Hill, Mission, Castro C Tier: Potrero Hill, Bernal Heights, Sunset D tier : SoMa,Dog patch
View quoted postParsing line charts is a hard task for VLMs VLMs are generally fine at coarse visual understanding, but they have a hard time reasoning about precise coordinates. Ask most VLMs, even though tuned to chart understanding, to parse a line chart to a table and they will struggle. We tested over a few samples. Docling’s new granite-vision model, gemini 3 flash, gpt 5.2 pro, and a v0.1 of our own chart parsing (which is in beta and rapidly evolving). Out of these, most models fail, and sometimes miss the entire chart correctly. gpt 5.2 pro is closest but spends an absurd number of tokens reasoning through each point. Our own parsing is actually quite good, though of course, there’s still some things we need to do to get to 100% accuracy. If you want to parse complex documents with diagrams/charts, come check out LlamaCloud! https://cloud.llamaindex.ai/
Extracting structured outputs with LLMs is easy. But doing large-scale extraction with precise citations and bounding boxes back to the source documents is way harder. With our latest release in LlamaExtract, we extract citation bounding boxes along with every single key and value within a document. You can see this in the UI. Hover over any k:v pair and you’ll be able to see the corresponding highlights in the source doc. If you’re a human reviewing a million docs (resumes, IDs, invoices, claims, contracts), this will help you 5x your ability to verify values and make sure things are correct. Check out these new extraction upgrades in LlamaCloud: https://cloud.llamaindex.ai/
LlamaExtract citations just got an upgrade: we now show you exactly where extracted data comes from in your documents with new citation bounding boxes 🎯 This citations upgrade gives you visual proof of where each field originates: 📍 Precise bounding boxes highlight the exact
View quoted postRT LlamaIndex 🦙 LlamaExtract citations just got an upgrade: we now show you exactly where extracted data comes from in your documents with new citation bounding boxes 🎯 This citations upgrade gives you visual proof of where each field originates: 📍 Precise bounding boxes highlight the exact location of extracted data in your source documents 🔍 Full citation transparency so you can verify and trust your extraction results 🚀 Also available through our API for seamless integration into your applications 📄 Perfect for compliance, auditing, and quality assurance workflows where traceability matters This makes LlamaExtract even more reliable for production document processing where you need to show your work and validate results. Try it out through the cloud UI or the API and see the difference visual citations make for your document extraction pipeline. Sign up to LlamaCloud to get started: https://cloud.llamaindex.ai?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2019823118794330180
If you are around SF First Thursdays today, we are giving away gold ⚱️(like 1g of 24k gold for each qualified registrant*) Our office is right by 2nd and Mission Street overlooking the party. See the QR code on the right window! The main requirements are that you have use cases around PDF parsing, and you’re willing to see a demo of our agentic OCR technology within LlamaParse. First-come, first-served. We have 50g of gold reserves. 🙂 Outside of this, let us know if you ever want to come say hi!
this is an unhinged tweet lol
the sexual tension between anthropic and openai employees in SF must be insane
View quoted postwith opus 4.6, anthropic is expanding into general intelligence from its stronghold of coding with codex 5.3, openai is expanding into coding from its stronghold in general intelligence maybe this was always obvious, but there was a moment last year where i thought frontier labs would specialize/segment a bit more. but nope it's an arms race towards the same thing
We are running a contest on document understanding 🥇. 1️⃣ Find the *hardest* document you’ve seen - whether it’s a scanned form with barely legible handwriting, or a financial presentation with multiple line charts on the same page 2️⃣ Describe what you want to do over the document with natural language, in our agents builder. Deploy the workflow within LlamaCloud. 3️⃣ Share what you’re doing with us! We’re giving out $200 to the top “document agents” built over the next 3 weeks. Bonus points if the input document is super hard, the use case is very relevant, and/or the workflow is complex. Check it out here: https://cloud.llamaindex.ai/?utm_source=socials&utm_medium=li_social
Kicking off the Document Agent Olympics. Build a document agent - Win $200 🥇 Document agents turn messy PDFs, invoices, and filings into structured data you can actually use. Think: extracting financials from SEC filings, reconciling invoices against contracts, or processing a
View quoted postwith opus 4.5, anthropic is expanding into general intelligence from its stronghold of coding with codex 5.3, openai is expanding into coding from its stronghold in general intelligence maybe this was always obvious, but there was a moment last year where i thought frontier labs would specialize/segment a bit more. but nope it's an arms race towards the same thing
Introducing Claude Opus 4.6. Our smartest model got an upgrade. Opus 4.6 plans more carefully, sustains agentic tasks for longer, operates reliably in massive codebases, and catches its own mistakes. It’s also our first Opus-class model with 1M token context in beta.
View quoted postRT LlamaIndex 🦙 Kicking off the Document Agent Olympics. Build a document agent - Win $200 🥇 Document agents turn messy PDFs, invoices, and filings into structured data you can actually use. Think: extracting financials from SEC filings, reconciling invoices against contracts, or processing a stack of resumes to surface top candidates. We're giving away three $200 prizes for the best agents built over the next 3 weeks. To enter: 💛 Deploy the agent to LlamaCloud 🤝 Make sure the agent repository is public 🚀 Explain what your document agent is solving 🔥 Bonus points for a good readme and demo video And most importantly, let us know what you think about our new LlamaAgents Builder! Ready to participate? Join the contest and start building your winning agent! Signup to LlamaCloud to get started https://cloud.llamaindex.ai?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2019457047415472237
2025 superbowl: kendrick 2026 superbowl: claude
I cannot comment on inter-company disagreements. What I can say is that Codex™ is now available, has 500,000 downloads, and is statistically likely to increase your builder productivity by a non-zero amount. Would you like me to help you get started?
First, the good part of the Anthropic ads: they are funny, and I laughed. But I wonder why Anthropic would go for something so clearly dishonest. Our most important principle for ads says that we won’t do exactly this; we would obviously never run ads in the way Anthropic
View quoted postRT LlamaIndex 🦙 Ready to master production-grade multi-agent AI systems in one intensive day? 🚀 We're partnering with @AWS and leading AI companies for the AWS AI Builder Lab with @aicampai in San Francisco on Feb 13th - a hands-on competition where you'll build sophisticated agentic workflows, not basic chatbots. 🤖 Design and orchestrate powerful multi-agent workflows using cutting-edge tools 🏆 Compete in real-time challenges with live leaderboards and prizes for top performers 🔧 Get production-ready patterns you can implement immediately, plus enterprise tool trials Coordinate intelligent systems to tackle real-world challenges. Limited to 250 qualified developers. Register here: https://www.aicamp.ai/event/eventdetails/W2026021308 Original tweet: https://x.com/llama_index/status/2019139663462850628
Our parsing models are able to parse this massive diagram into mermaid format with 100% accuracy. Original doc on the left, parsed mermaid on the right. This is the “agentic plus” mode in LlamaParse, powered by state-of-the-art VLMs / agentic reasoning that can interpret complex relationships within any document page. If you have docs with a lot of flowcharts and diagrams, come check us out! https://cloud.llamaindex.ai/
We’re excited to partner with @NTTData to help them scale out their document parsing/RAG pipelines for both their internal use cases to client-facing use cases. This helps provide higher accuracy and allows internal teams to focus on building agents instead of data quality. Big thanks to Manuel for the kind words here. Check it out: https://www.youtube.com/watch?v=9bfkXXfSIkw
How does @NTTDATA scale enterprise AI? They use LlamaIndex to improve document parsing & power RAG applications for faster, more accurate AI systems. Watch the customer story 🎥: https://www.youtube.com/watch?v=9bfkXXfSIkw #AI #RAG #LlamaIndex #EnterpriseAI
View quoted postRT Tuana MCP connects agents to live systems: databases, APIs, external services. It's designed for runtime tool access . But the moment you need to teach your agent how to approach a problem domain, you need something else. Skills aren't about (just) accessing data. They're about embedding knowledge into your agent's reasoning. When your agent needs to understand "here's the right sequence for debugging a data pipeline" or "this is how you validate and process complex documents," skills allow you to bake that knowledge into how the agent thinks. There's then also the whole matter of how they work fundamentally: MCP tools rely on an external connection and API calls. Skills are local.. The issue isn't choosing between them. It's understanding that they kiiinda serve different purposes. MCP extends your agent's capabilities at runtime. Skills shape how your agent reasons about problems. I wrote all about this with @itsclelia in our latest blog: https://www.llamaindex.ai/blog/skills-vs-mcp-tools-for-agents-when-to-use-what Original tweet: https://x.com/tuanacelik/status/2019106807437038029
RT LlamaIndex 🦙 How does @NTTDATA scale enterprise AI? They use LlamaIndex to improve document parsing & power RAG applications for faster, more accurate AI systems. Watch the customer story 🎥: https://www.youtube.com/watch?v=9bfkXXfSIkw #AI #RAG #LlamaIndex #EnterpriseAI Original tweet: https://x.com/llama_index/status/2019094169290223807
no one's mentioned pie punks yet pie punks is extremely underrated
we spent a lot of time figuring out how to context engineer claude code with mcp tools and skills this comparison diagram should help you understand the tradeoffs between the two 👇
Are you choosing between MCP servers and skills for your agent? @tuanacelik and I wrote a blog post about the differences between the two and their PROs and CONs, informed by the lessons learnt while building our own coding agent (LlamaAgents Builder). The main takeaway? Skills
RT Clelia Bertelli (🦙/acc) Are you choosing between MCP servers and skills for your agent? @tuanacelik and I wrote a blog post about the differences between the two and their PROs and CONs, informed by the lessons learnt while building our own coding agent (LlamaAgents Builder). The main takeaway? Skills are easier to set up but tougher to keep up-to-date, whereas MCPs are more dev-facing, but offer a more deterministic toolset with less maintenance overhead. Check out the article 👉 https://www.llamaindex.ai/blog/skills-vs-mcp-tools-for-agents-when-to-use-what Original tweet: https://x.com/itsclelia/status/2018821269752611102
Should you use MCP or skills in your coding agent (?) We gained a lot of insight into this when we built our own coding agent (the LlamaAgents builder) within LlamaCloud. tl;dr Skills are way easier to setup, but unreliable/hard to maintain. Blog: https://www.llamaindex.ai/blog/skills-vs-mcp-tools-for-agents-when-to-use-what?utm_source=socials&utm_medium=li_social
Confused about whether to use Skills or MCP tools for your AI agents? We break down the key differences and when to use each approach. 🔧 MCP tools provides a more deterministic interface with fixed schemas - perfect for precise, predictable operations but require more technical
RT LlamaIndex 🦙 Confused about whether to use Skills or MCP tools for your AI agents? We break down the key differences and when to use each approach. 🔧 MCP tools provides a more deterministic interface with fixed schemas - perfect for precise, predictable operations but require more technical setup and an overload of MCP tools can cause a confusion of choice to an agent 📝 Skills use natural language instructions in markdown files - minimal setup required but open to LLM interpretation variations ⚡ MCP involves network latency while Skills run locally, but MCP offers centralized updates that propagate automatically Read our full analysis with real examples: https://www.llamaindex.ai/blog/skills-vs-mcp-tools-for-agents-when-to-use-what?utm_source=socials&utm_medium=li_social Original tweet: https://x.com/llama_index/status/2018749615907213457
A week after PaddleOCR-VL-1.5 took the top spot on OmniDocBench, *another* 0.9B model dethrones it! GLM-OCR shows SOTA results on doc parsing benchmarks and it's apparently 50-100% faster https://huggingface.co/zai-org/GLM-OCR
Introducing GLM-OCR: SOTA performance, optimized for complex document understanding. With only 0.9B parameters, GLM-OCR delivers state-of-the-art results across major document understanding benchmarks, including formula recognition, table recognition, and information extraction.
Our default document parsing mode is now able to parse a complex research report with multiple embedded charts on a single page. This is the cheapest document OCR model out there that can turn complex visual documents into LLM-ready markdown. This is our agentic mode in LlamaParse. It starts at ~1c per page; if you’re looking to scale consumption we offer volume discounts. Come check it out! Sign up: https://cloud.llamaindex.ai/
RT LlamaIndex 🦙 AI in 60 Seconds ⏱️ @Experian's Head of AI Innovation shares how AI is shaping the future of fintech, from smarter decisions to better financial experiences. Watch now 👇 https://www.youtube.com/shorts/SHAAHecmNeE #AI #Fintech #LlamaIndex Original tweet: https://x.com/llama_index/status/2018374580834804216
I built a simple paralegal agent to detect, classify, and extract information from court filings🧑⚖️- complaints, motions, orders, and more! Wrote a prompt in English, and it encoded it into a deterministic, repeatable workflow that you can use to run on millions of court filings at scale. 1. A new docket entry arrives 2. Detect what type of filing it is (motion/order/complaint/etc) 3. Define a separate schema for each new filing 4. Synthesize results in a UI with human review. Can also plug it in as an API into your workflows This is available to everyone in our LlamaCloud agents builder. The prompt is exactly what you see in the screenshot; you can make it more/less verbose if you’d like. Come check it out! LlamaCloud by @llama_index: https://cloud.llamaindex.ai/
would be incredible if this single-handedly reverses claude code's dominance
@Yuchenj_UW I don’t let Claude Code on my codebase. It’s all codex. Would be too buggy with Opus.
View quoted post