Auto-selects OCR, VLM, or text extraction based on content.
Ask your documents. Get grounded answers.
The most advanced local document Q&A technology available. Query PDFs, Office documents, HTML, Markdown, and more. Powered by semantic RAG, adaptive layout analysis, and a fully customizable agent with tool calling and memory. 100% on-device.
Full document structure understanding with page elements.
Embedding-based retrieval with intelligent chunking.
Every answer traced to document, page, and passage.
Cloud calls
Traceable
Documents
Semantic RAG meets document layout analysis.
LM-Kit.NET delivers the most advanced local document Q&A technology available. The underlying system combines semantic retrieval-augmented generation with a complete document layout analysis stack. Supports PDF, Office documents (Word, Excel, PowerPoint), HTML, Markdown, images, and more.
The engine is extremely fast, using an adaptive approach that intelligently engages OCR, Vision Language Models, layout processing, or direct text extraction based on content discovery. Each page is analyzed and processed using the optimal strategy automatically.
Built by IDP pioneers: This isn't a wrapper around generic RAG. It's purpose-built document intelligence from a team with 20+ years of experience processing billions of documents in production worldwide.
Adaptive document processing pipeline.
Content-aware processing that automatically selects the optimal extraction strategy for each page based on content discovery.
Step 01
Document import
PDF, Office, HTML, Markdown, images analyzed page-by-page.
Step 02
Auto-detection
Content type determines processing mode.
Step 03
Extraction
Text, OCR, VLM, or layout analysis.
Step 04
Semantic RAG
Embeddings, retrieval, grounded answers.
Intelligent content-aware processing.
The engine automatically selects the optimal strategy for each page based on content analysis, or you can specify your preference.
Default · Auto
Content-driven selection
Analyzes each page and selects the best processing strategy automatically. Uses VLM for image-heavy pages, text extraction for digital content.
- Zero configuration required
- Optimal quality/speed balance
- Handles mixed document types
Mode · TextExtraction
Fast, direct processing
Extracts text directly from PDF structure with OCR fallback for scanned or image-based pages. Maximum speed.
- Fastest processing
- Lower resource usage
- Best for clean digital PDFs
Mode · DocumentUnderstanding
VLM-powered analysis
Vision Language Models analyze pages visually to understand layout, structure, tables, and relationships. Markdown output.
- Best for complex layouts
- Preserves document structure
- Tables, forms, mixed content
Production-ready document intelligence.
Everything you need to build document Q&A applications that actually work.
Capability
Multi-document queries
Load multiple documents and ask questions that span all of them. Compare contracts, cross-reference reports, search collections.
Capability
Source attribution
Every answer includes document names, page numbers, and custom metadata. Full traceability for compliance and audit.
Capability
Intelligent caching
Processed documents cached via IVectorStore. Subsequent loads are instant. Filesystem or custom backends (Qdrant, PostgreSQL).
Capability
Smart context
Small documents included in full. Large documents use semantic passage retrieval. Automatic optimization per document.
Capability
Multi-turn dialogue
Follow-up questions maintain context. Natural conversation flow. Ask clarifying questions without re-explaining.
Capability
100% local
All processing on your infrastructure. Documents never leave. Air-gapped deployments. HIPAA, GDPR, compliance-ready.
Capability
Agentic capabilities
PdfChat is a fully customizable document agent. Connect external tools, maintain conversation memory, and integrate with MCP servers for extended functionality.
Complete demo application.
A fully-featured console application demonstrating all capabilities, ready to run.
Chat with PDF demo
Interactive console app that lets you load PDFs, ask questions, and see the full document Q&A pipeline in action with source references, generation stats, and real-time streaming.
- Model selection with download progress
- Standard or vision processing modes
- Multi-document loading with caching
- Interactive commands (/help, /status, /add)
- Token counts and generation speed metrics
Built for real-world applications.
Document intelligence that solves actual business problems.
Use case
Contract analysis
Query legal agreements for specific clauses, obligations, termination conditions, and payment terms with full source attribution.
Use case
Financial review
Ask questions about revenue, expenses, projections, and risk factors across multiple financial reports and statements.
Use case
Technical documentation
Search manuals and specifications for configuration details, procedures, system requirements, and troubleshooting steps.
Use case
Research & academia
Query research papers for methodology, findings, citations. Cross-reference multiple sources for literature reviews.
Use case
Compliance & audit
Verify policy adherence with traceable source references. Generate audit trails with document and page attribution.
Use case
Customer support
Build knowledge bases from product documentation. Answer customer questions automatically with grounded responses.
Choose your models
LM-Kit.NET supports a wide range of vision-capable chat models, embedding models, and specialized OCR models. Browse our model catalog to find the right combination for your use case, hardware, and accuracy requirements.
Key classes.
The building blocks for document Q&A applications.
Class
PdfChat
High-level document agent for question-answering. Supports tool calling, conversation memory, MCP integration, and full customization.
View documentationClass
DocumentRag
Lower-level document RAG engine with full control over processing modes, chunking, and retrieval parameters.
View documentationInterface
IVectorStore
Interface for embedding storage and caching. Use FileSystemVectorStore or implement custom backends.
View documentationClass
VlmOcr
Vision-based document parser using VLMs. Preserves layout and structure as markdown output.
View documentationChat plus the rest of Document Intelligence.
OCR
When the PDF is a scan, OCR runs transparently. Native engine plus VLM OCR with PaddleOCR-VL, GLM-OCR, LightOnOCR.
Document to Markdown
Universal converter that picks the right strategy per page. The same primitive feeds the chat pipeline.
Document RAG engine
Lower-level control: explicit lifecycle, chunking strategies, custom vector stores, source attribution.
Document summarisation
Sometimes the right answer is a summary. Recursive summarisation handles documents bigger than the context window.
Build it. Read it. Try it.
Working console demos on GitHub, step-by-step how-to guides on the docs site, and the API reference for the classes used on this page.
Chat with PDF
Console demo: drop a PDF, ask questions, get cited answers.
Open on GitHub → How-to guideChat with PDF documents
End-to-end how-to with adaptive layout analysis.
Read the guide → How-to guideBuild a private document Q&A
Build a question-answering surface over your own corpus, on-device.
Read the guide → API referencePdfChat
API reference for the PdfChat class.
Open the reference →Ready to build document intelligence?
The most advanced local document Q&A technology. Semantic RAG and layout analysis. 100% on your infrastructure.