raglite V.1.0 is live
raglite V.1.0 is live
raglite V.1.0 is live

Table of contents

RAGLite V.1.0 is live: Faster ingest, smarter queries, better retrieval
Executive summary: Why RAGLite matters for your organization
What’s new in RAGLite v1.0?
1. Enhancing retrieval quality with improved chunking and multi-vector search
2. Support for DuckDB to improve local search
3. Support for Qwen3 models and reasoning tools
4. Parallel document insertion
5. Built-in benchmarking
6. Smarter query adaptation
Designed for builders, loved by engineers
Try RAGLite now

Table of contents

RAGLite V.1.0 is live: Faster ingest, smarter queries, better retrieval
Executive summary: Why RAGLite matters for your organization
What’s new in RAGLite v1.0?
1. Enhancing retrieval quality with improved chunking and multi-vector search
2. Support for DuckDB to improve local search
3. Support for Qwen3 models and reasoning tools
4. Parallel document insertion
5. Built-in benchmarking
6. Smarter query adaptation
Designed for builders, loved by engineers
Try RAGLite now

Table of contents

RAGLite V.1.0 is live: Faster ingest, smarter queries, better retrieval

RAGLite V.1.0 is live: Faster ingest, smarter queries, better retrieval

RAGLite V.1.0 is live: Faster ingest, smarter queries, better retrieval

14 Aug 2025

14 Aug 2025

Discover RAGLite V.1.0, the first major update to our lightweight open-source Python RAG toolkit.

In the fast-evolving world of Retrieval-Augmented Generation (RAG), where accurate, scalable, and efficient information retrieval defines competitive edge, RAGLite v1.0 arrives as a serious contender. Designed for teams building RAG applications, RAGLite emphasizes simplicity, interoperability, and benchmark-beating performance, all in a minimal footprint.

Whether you're a CTO exploring how to streamline AI infrastructure, an innovation lead building a prototype, or a technical team integrating RAG into your workflows, RAGLite offers a clean and powerful foundation. The new v1.0 release brings significant updates that elevate its retrieval quality, performance, and flexibility, without adding operational complexity.

Executive summary: Why RAGLite matters for your organization

RAGLite is a lightweight, open‑source Python toolkit designed by Superlinear that makes it easier to build Retrieval‑Augmented Generation (RAG) pipelines. It connects document retrieval systems with language models (LLMs), letting users demand in natural language and get answers grounded in external documents, all without heavyweight frameworks. 

Traditional RAG stacks can be heavy, slow, and hard to customize. RAGLite fixes that by offering fast, modular components for each stage of the pipeline, so you can build reliable RAG workflows with minimal overhead. 

Key business advantages:

  • Simplicity: RAGLite eliminates the complexity of traditional RAG stacks, replacing a maze of tools, vector databases, and orchestration frameworks with a single, unified Python toolkit for ingestion, chunking, hybrid search, reranking, generation, and UI.

  • Easy to deploy and maintain: RAGLite reduces deployment and maintenance complexity by using a single configuration to run different components and stacks in our RAG app. It includes ingestion, evaluation, UI, and Claude MCP integration, all powered by one lightweight database file.

  • High-quality retrieval: RAGLite improves response quality through semantic chunking, contextual headings, late chunking options, multi-vector embeddings, hybrid search, reranking, and query adaptation, leading to more relevant and accurate results without requiring external components.

  • Align with user expectations: To address user expectations for more than just basic retrieval, RAGLite incorporates intelligent RAG pipelines powered by adaptive language models. It integrates with Claude via a built-in Model Context Protocol (MCP) server that any MCP client, such as Claude Desktop, can connect to. It also supports Chainlit as an optional, customizable ChatGPT-like frontend for web, Slack, and Teams.

  • Compatibility: Works with any LLM available from that framework: LiteLLM, embedding model, or reranker (supported by AnswerDotAI/rerankers), including open-weight models like Qwen3.

  • Performance: Outperforms LlamaIndex, OpenAI Vector Store, and Azure AI Search on retrieval benchmarks.

  • Evaluation-ready: Includes built-in tools to test and evaluate RAG performance with just two lines of code.

For decision-makers, this means faster iteration cycles, fewer vendor lock-ins, and more control over retrieval quality and governance.

What’s new in RAGLite v1.0?

RAGLite’s first major release introduces several innovations, ranging from backend database improvements to smarter chunking strategies. 

Here’s a breakdown of the most notable upgrades:

1. Enhancing retrieval quality with improved chunking and multi-vector search

This update introduces several improvements aimed at boosting the accuracy and relevance of information retrieval. 

Introducing ‘Chunklets’ for better chunking

A new middle layer between sentences and chunks improves information granularity and alignment with document structure.

  • Each chunklet groups about three statements, where a statement is roughly the length of a typical sentence.

  • Increased default chunk size (from 360 tokens to 512 tokens) aligns with modern LLMs' context windows and improves coherence.

This approach replaces the previous method that optimized chunks based on overlapping sentence windows, simplifying the system. Multi-vector retrieval is a mix between chunklet embeddings and sentence embeddings, which helps improve search quality since chunklets provide richer context.

Adding front matter for context

Document metadata such as filenames or URLs is added as front matter to chunk embeddings. This extra context is especially useful when document structure (like Markdown headings) is missing, helping the search engine understand the content better.

Improved chunk ranking method

The way chunks are ranked in vector search has been refined.

  • Previously, the system ranked chunks based on the total combined similarity of all the sentences inside them (using the L1 norm). So, chunks with more sentences that matched the query tended to rank higher, even if some sentences weren’t very relevant, just because they had more content.

  • After this update, chunks are ranked based on their single most relevant sentence (using the L∞ norm). That means the chunk’s rank depends on the best matching/most relevant part, not the total sum of all sentences. This helps highlight chunks with strong, focused matches rather than just longer chunks with many sentences.

This change helps prioritize chunks based on their strongest match to the query, improving the quality of retrieval results.

Together, these changes result in more precise, context-aware responses from your RAG application.

2. Support for DuckDB to improve local search

RAGLite replaces SQLite with DuckDB as the default database backend to address limitations encountered with SQLite in local retrieval scenarios.

Why the change?

SQLite’s vector search capabilities have limitations: its extensions are restricted on macOS, lack advanced indexing for large datasets, and its full-text search does not support BM25 scoring, which affects search relevance.

Advantages of DuckDB

DuckDB offers easily installable vector search and full-text search extensions, supports accelerated indexing methods, and provides more flexibility for managing and updating indexes.

Key updates

The system now uses DuckDB by default, with updated configurations and code to leverage DuckDB’s native search and indexing features. This includes simplifying search implementations, improving full-text search relevance with BM25, and enhancing development workflows.

DuckDB brings scalability, search accuracy, and maintainability for local data retrieval tasks.  Ideal for teams wanting full control without operational overhead.

3. Support for Qwen3 models and reasoning tools

RAGLite now natively supports Qwen3, one of the best-performing open-weight LLMs and enables improved handling of reasoning before tool use.

  • Model upgrade: Replaces older Llama models with more modern Qwen3 models, which are gaining popularity for their performance and efficiency.

  • Reasoning support: Adds the ability to perform a reasoning step before calling tools, improving flexibility in how queries are processed. This feature can be turned off if needed.

  • Performance improvements: Updates underlying dependencies for better compatibility and speed, and adopts recommended settings for Qwen3 models.

These changes ensure better alignment with emerging open-source model standards and provide more adaptable behavior in structured query handling.

4. Parallel document insertion

This update improves performance and usability by introducing parallel document insertion and integrated benchmarking:

  • Better ingestion: Document insertion is now parallelized, enabling RAGLite to support high-speed, multi-threaded ingestion.

  • Refined internals: Several internal changes streamline document handling and chunking logic.

These changes improve performance and give teams better tools to measure and monitor retrieval quality at scale. This is a critical feature for production environments that need to refresh or update knowledge bases regularly.

5. Built-in benchmarking

Evaluating retrieval quality no longer requires custom scripts or third-party pipelines. A new command-line tool (raglite bench) makes it easy to evaluate performance using standard datasets and metrics. 

This means teams can measure how changes to embeddings, chunking, or rerankers actually impact retrieval quality, before rolling into production.

6. Smarter query adaptation

Retrieval quality isn’t just about what’s indexed, it’s about understanding the user query. RAGLite V.1.0 brings some improvements to the query adapter algorithm, which is responsible for refining how search queries return relevant results.

  • Improved accuracy: Refines how relevant and irrelevant results are separated, leading to more precise rankings.

  • More control: Adds the ability to tune how much the query adapter adjusts results.

These changes are subtle, but help make search behavior more stable, adaptable, and aligned with recent retrieval improvements.

Designed for builders, loved by engineers

From a developer’s perspective, RAGLite provides a highly modular and extensible codebase:

  • One-liner configuration for any LLM, embedding model, or reranker.

  • No vendor lock-in: You can swap models and databases easily.

  • PostgreSQL support: Prefer Postgres over DuckDB? RAGLite has you covered.

  • Minimal dependencies: No heavy frameworks; designed to be lightweight and hackable.

This makes RAGLite a strong fit for any team looking to rapidly prototype and deploy RAG pipelines without overengineering.

Try RAGLite now

RAGLite v1.0 is a thoughtful, efficient, and surprisingly powerful retrieval engine that serves both AI teams looking for high-quality retrieval and business stakeholders needing scalable, controllable AI infrastructure.

By combining strong technical foundations (DuckDB, Qwen3) with thoughtful features (benchmarking, evaluation, chunklets), RAGLite makes it easier to ship smarter, faster, and more reliable RAG applications, from laptop to production.

If you're building a knowledge-driven AI assistant, research tool, or domain-specific LLM product, RAGLite deserves a spot at the center of your stack.

Build serious RAG apps, without the complexity. 
Try
RAGLite v1.0 today and see the difference in retrieval speed, quality, and control.

Author(s):

Contact Us

Ready to tackle your business challenges?

Stay Informed

Subscribe to our newsletter

Get the latest AI insights and be invited to our digital sessions!

Stay Informed

Subscribe to our newsletter

Get the latest AI insights and be invited to our digital sessions!

Stay Informed

Subscribe to our newsletter

Get the latest AI insights and be invited to our digital sessions!

Locations

Brussels HQ

Central Gate

Cantersteen 47



1000 Brussels

Ghent

Planet Group Arena

Ottergemsesteenweg-Zuid 808 b300

9000 Gent

© 2024 Superlinear. All rights reserved.

Locations

Brussels HQ

Central Gate

Cantersteen 47



1000 Brussels

Ghent

Planet Group Arena
Ottergemsesteenweg-Zuid 808 b300
9000 Gent

© 2024 Superlinear. All rights reserved.

Locations

Brussels HQ

Central Gate

Cantersteen 47



1000 Brussels

Ghent

Planet Group Arena
Ottergemsesteenweg-Zuid 808 b300
9000 Gent

© 2024 Superlinear. All rights reserved.