Docling’s cover photo
Docling

Docling

Technology, Information and Internet

Get your documents ready for gen AI

About us

Docling unlocks the information trapped in your PDFs, Office files, images, and more, so you can automate document processing and build AI applications with ease and speed.

Industry
Technology, Information and Internet
Company size
2-10 employees
Type
Nonprofit

Employees at Docling

Updates

  • Docling reposted this

    I’m really happy to see IBM Elite Support for Docling become generally available. Docling has grown through its open-source community, and more teams are now using it in production for document intelligence and RAG pipelines. With IBM Elite Support, Docling users now have the option to pair open-source flexibility with professional, enterprise-grade support — including guidance on architecture, deployment, and production best practices. This is a great step for organizations that want to adopt Docling with confidence, while keeping the project open and community-driven. 👉 Learn more or connect with an IBM seller: https://lnkd.in/euPQK4rn Thanks IBM for supporting the Docling OSS community! Jordan Youngblood Roy Derks Charna Parkey, Ph.D. Ed Anuff Sriram Raghavan Abdel Labbi Bill Higgins Peter W. J. Staar #Docling #OpenSource #IBM #EnterpriseAI #GenAI #RAG #DeveloperTools #AIInfrastructure

  • Docling reposted this

    Sunday Coffee & Code: Successful End-to-End RFP Generation Run (Web UI + Multi-Agent Orchestration) This week I got a clean end-to-end run of my RFP Factory: upload through the Web UI, multi-agent orchestration does the heavy lifting, and a complete draft package comes out the other side. It’s one thing to have individual components working; it feels like a whole other step to have the whole pipeline behave like a real system. What worked (and why it matters) ● Multi-Agent Orchestration (Microsoft Agent Framework - MAF) MAF performed exceptionally well as the orchestration layer - both for offline Model (Ollama) and OpenAI integration with tool execution. The “agent team” approach is now predictable enough that it feels like engineering, not experimentation. ● Local models via Ollama (Qwen3 continues to impress) Even though this run used OpenAI’s API end-to-end, it reinforced that local models are not just “nice demos.” The Qwen3 family in particular performed strongly for this workflow. ● MCP (Docling via MCP: surprisingly frictionless) Docling via MCP has been one of the simplest pieces to integrate: download it, start it, expose it as a tool to agents, and it just works. That’s exactly what “tooling” should feel like - boring, reliable, repeatable. ● Agentic coding (Anthropic Claude Code from VSCode + Agent) This was a major step-change. Not just “faster coding", it changes the way I worked. I spent more time on the specification and acceptance criteria up front. Troubleshooting does shift a bit: being more “distant from the code” can mean it takes longer to reacquire context when something breaks. Experience matters here: knowing where to look (configs, log files) helps. What I’ve learned building this (beyond the RFP use case). This project has been a practical vehicle for exploring a stack of emerging patterns and technologies - and it’s clarified a few things for me: ● Agents have matured dramatically since my first builds in 2024. Reliability, tool use, and orchestration patterns are materially better now. ● Specs are the new leverage. The better the spec, the more “agentic coding” compounds across quality and speed. Not entirely sure what is next yet - this has been a bit of a wander, but in the best way: exploring new tech with a practical application in mind. Likely next steps: ● Improve RAG setup (talked about this a few times). ● Automate startup and deployment. ● Add agents to further improve the workflow. ● Build this out as a solution? Cost snapshot (because this part matters) - today’s run: ● 407,743 tokens for $1.45. ● 16.5 minutes of Amazon Web Services (AWS) server time. ● Total of $1.70 for a draft. If you’re experimenting with agents, MCP, or multi-agent orchestration, I’m happy to compare notes - this space is moving quickly, and the practical details are where the learning is. (Lots of screen shots attached - front-end, log file, agent activity logs, output zip file contents attached.)

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • Docling reposted this

    Goudlokje en de drie taalmodellen IBM vertelt in dit recente filmpje over SLMs, LLMs en FMs (https://lnkd.in/eZTxAsJ5). Het vertelt dat je heus niet het grootste model hoeft te gebruiken voor elke taak. En daar ben ik het helemaal mee eens. Sterker nog: gebruik het kleinste model dat kan volstaan. Dat scheelt kosten en milieubelasting. En als je handig genoeg bent en een klein genoeg model kunt gebruiken, kan het gewoon op je eigen computer! Dat heet dan Edge-AI. En dat is wat #FietsbelAI makkelijk wil maken.  En als die SLMs net niet volstaan? Dan kun je ze een heel klein beetje finetunen, maar dan hot-swappable; je kunt het zomaar aan- en uitzetten en je voegt niets toe aan het gigantische aantal van 2.528.840 taalmodellen op Hugging Face Hoe moet je dit dan ongeveer zien? Nou, zo. Zie onderstaande aardappels. En hier valt een hele hoop over te zeggen, zoals of je 70B wel op workstations moet willen draaien, of hoe hoog de verschillen tussen de pieken precies moeten zijn. Zodra we een schaal gaan kiezen, gaan we dat heel precies doen. Waar het mij om gaat, is dat SLMs op de Edge *vandaag* mogelijk zijn. Sommige superkleine SLMs, zoals Docling, zijn zeker goed genoeg voor één enkele taak. Iets grotere SLMs kun je met een beetje LoRa (bijvoorbeeld via InstructLab) nóg nuttiger en taakspecifieker maken. Misschien nog niet tot het niveau van alles wat LLMs en FMs kunnen, maar genoeg om je achter de oren te krabben of de allergrootsten nog wel per sé nodig zijn. Zeker als je meeneemt, dat niet elke toepassing echt blijvende waarde toevoegt. One more thing: onze computers worden alleen maar krachtiger en kleine taalmodellen alleen maar beter. Dus de SLM aardappels schuiven de komende jaren omhoog!

    • No alternative text description for this image
  • Docling reposted this

    As the workday comes to a close, I find myself reflecting on the significance of community. Being part of the right community can make a substantial difference. 🌟 This week, I had the opportunity to connect at the IBM #DevDay, where I was inspired by tech leaders sharing their experiences with #Bob, agent orchestration, and Docling. 💻✨ I am also honored to be involved in an agentic project utilizing IBM watsonx Orchestrate, and the support there has been incredible. 🙌 I feel fortunate to have discovered my "tech family," where I can share ideas and grow together. Thank you to everyone for being such an amazing team! 🤝❤️ #gratitude #Friday #ibmchampion

  • Docling reposted this

    🌟 LF AI & Data Project Highlight of the Week: Docling Docling has surpassed 50,000 GitHub stars and doubled its daily PyPI downloads! This growth is a direct result of the incredible work from our contributors and community members around the world. Thank you for helping build open, trusted tools for document intelligence. Explore Docling and get involved: https://lnkd.in/eDZagt_i Michele Dolfi, PhD Peter W. J. Staar Cara Delia Carol Chen #LFAIData #Docling #OpenSourceAI #OpenSource #AICommunity #MachineLearning #DataScience #MLOps #GitHub #PyPI #OSS #TechCommunity

    • No alternative text description for this image
  • View organization page for Docling

    3,758 followers

    𝗖𝗮𝗹𝗹𝗶𝗻𝗴 𝗔𝗹𝗹 𝗗𝗼𝗰𝗹𝗶𝗻𝗴 𝗘𝗻𝘁𝗵𝘂𝘀𝗶𝗮𝘀𝘁𝘀! In this next installment of Docling Office hours, we will be hearing from Taylor Agarwal on how RedHat used Docling to migrate over 32,000 legacy vendor contracts, as well as from Thomas Vitale and Eric Deandrea, creators of the Docling-Java project, the official Java client and tooling for Docling.  And as always, plenty of time for questions! 𝗛𝗼𝘄 𝘁𝗼 𝗝𝗼𝗶𝗻: Meeting link: https://lnkd.in/epwvDFiX Public Project Calendar (iCal): https://lnkd.in/gSY4_ZJR 𝗦𝘁𝗮𝘆 𝗖𝗼𝗻𝗻𝗲𝗰𝘁𝗲𝗱: Subscribe to our LF AI announcement list to receive updates about future meetings: https://lnkd.in/g3J-dsrn 𝗛𝗲𝗹𝗽 𝗨𝘀 𝗣𝗿𝗲𝗽𝗮𝗿𝗲: Please take a moment to complete our short survey so we can tailor the content to your interests and needs: https://lnkd.in/gjdZJWrs We look forward to seeing you there!

  • Docling reposted this

    Quarkus Insights #323: What is Docling? Michele Dolfi, Thomas Vitale, and Eric Deandrea join us to discuss Docling, a project which converts documents into structured data for AI models. Learn how it parses various file types, such as PDFs, DOCX, and images, to extract rich information like layout, tables, and reading order, making documents suitable for generative AI applications like Retrieval-Augmented Generation (RAG) systems and how to use it with Quarkus.

    Quarkus Insights #323: What is Docling?

    Quarkus Insights #323: What is Docling?

    www.linkedin.com

  • Docling reposted this

    🔥 Formatted data is the secret sauce powering smarter AI! Plain text removes the formatting and context that give your documents meaning: 🏷️ Headings → section hierarchy 📊 Tables → rows/columns + relationships 📄 Layout → reading order (especially multi-column PDFs) A super common failure is ⚠️ wrong reading order on multi-column pages: the extractor mixes left + right columns (and sometimes headers/footers), so your chunks become scrambled. Retrieval returns “garbage context,” and the model answers confidently from a broken source. Same with tables: when rows/columns collapse into a paragraph, numbers lose their headers and relationships. 🦆 Docling and 🦙 LlamaIndex both have the power to reformat and structure documents effortlessly, turning raw files into AI‑ready knowledge. Together, they transform messy, unstructured content into a clean, formatted foundation that makes retrieval, reasoning, and generation far more reliable and context‑aware for any RAG or document‑intelligence workflow. They truly made my day so much better! 🚀 Excited to share that a live demo is coming soon, stay tuned! #Docling #opensource #RAG #DocumentAI #GenAI #LLMOps #LLM #Llamaindex #GitHub

    • No alternative text description for this image
  • Docling reposted this

    Sunday Coffee & Code: RAG improvements, and a few detours Most of my recent Coffee & Code time has been going into one very specific thing, improving the RAG layer behind my Microsoft Agent Framework–based RFP multi-agent solution. Here’s the path so far. Step 1: Chroma + Docling (via MCP) ● This was the first “real” implementation. Simple vector search over RFP input, fed by Docling. ● It worked well enough to support one live RFP response, which is an MVP bar for me - whether something is more than an experiment. Step 2: LightRAG ● I then went looking for ways to improve retrieval quality and came across LightRAG (https://lnkd.in/gCftxBBr), which combines vector search, knowledge graph, and re-ranking. ● Technically, it was impressive. I got it running, built the knowledge graph, and the architecture made a lot of sense. ● But it also did something unsettling: the knowledge graph started containing things that were never in the source documents (for example, “Noah Carter – World Athletics Championship”). Whether this was a hallucinated artifact or not, it was a reminder to take a closer look. Step 3: OpenRAG Last week I came across IBM's OpenRAG stack - https://www.openr.ag/ - (Langflow + Docling + OpenSearch Project) and it checked a lot of boxes for me: ● LangFlow (where I built my first real agent back in 2024) ● Docling (already central to my document pipeline) ● OpenSearch (scalable, production-grade search) ● Langfuse support (observability - also used it before and it provides valuable insights) The install was quick (Docker, Inc based), though I did hit a startup issue on my Amazon Web Services (AWS) EC2 instance that took some fiddling to resolve (GitHub issue - https://lnkd.in/g38qb8QK - fixed with my PR: https://lnkd.in/g594VTQ6). Once running, though, this stack is genuinely interesting: configurable models, Langfuse tracing, and full access to LangFlow flows sitting right under the RAG layer. This feels much closer to something I could operate, tune, and trust over time (as well as getting support for). Next steps: Now it’s about stabilising the OpenRAG deployment and wiring it cleanly into my RAG Manager so the agent can start using it for real RFP work. As always, this about exploring the latest tools as well as answering a simple consulting question: Can I use this in a real, production, workflow and trust it? More to come.

    • No alternative text description for this image
    • No alternative text description for this image

Similar pages