Type something to search...
5 Things I Learned Building Qdrant + RAG That Aren't in the Documentation

5 Things I Learned Building Qdrant + RAG That Aren't in the Documentation

5 Things I Learned Building Qdrant + RAG That Aren’t in the Documentation

You want to design a Qdrant + RAG system. This means taking your documents, breaking them into pieces, converting them to vectors, storing them in a vector database, and pulling them out when needed using “cosine similarity.” So you’re not teaching the system anything - you’re just building your own smarter DB system.

Wait a minute, wouldn’t it be pretty much the same if you just trained an LLM? Yes, exactly… Our learning mechanism isn’t that different either. How many of us question what we’ve learned and go after something better? Or how many of us reject information we’ve already learned? Answer: none of us.

So we need to turn our data into chunks. Why? Why can’t it be in one piece? Are we afraid of increasing vector dimensions? No. By breaking documents into the smallest pieces possible, we’re actually trying to create a data map. The smaller they are, the more points and the more potential for extracting context… Okay then, why don’t we break it down word by word? No, that’s a terrible idea. Because we want what we’re breaking down to be the “SMALLEST MEANINGFUL PIECE.” Not a small meaningless piece.

Okay, now we have a knowledge cloud made from our own data. But what we’re looking for still isn’t in the knowledge cloud. What we find is still just the closest thing to what’s in this cloud in our system. Oops, there’s a problem. The closest thing it finds might have nothing to do with what I’m looking for. Of course. Remember how sometimes LLMs give you answers that have nothing to do with you? Or how a Generative AI generates an image that has nothing to do with what you wanted? It does. All of these are errors resulting from working with these distances and probabilities.

If you integrate a RAG system alongside your LLM system, you gradually design a system that contains the same information and doesn’t have to reach the LLM. Only when it gets an answer below a certain threshold value should it go to the LLM, and it should write the answer result to the vector db - this will both reduce LLM costs and allow you to build your own specialized system without depending on any high-cost fine-tuning process.

The System’s Soft Spot…

Now everything sounds incredibly good and flawless, right? It shouldn’t. Because after a while, this system can start spinning around itself like a dog trying to catch its tail and get stuck on the same things. Classic overfitting won’t let you off the hook here either. Because it always knows the same topic, it will start saying the same things like your boring relative who always talks about the same subject. The solution to this is to ensure it includes information on broader topics, also considering the sub-topics and related topics of its specialized subject.

So doing everything by the book doesn’t mean everything will be perfect :)

Stay Ahead in Tech

Join thousands of developers and tech enthusiasts. Get our top stories delivered safely to your inbox every week.

No spam. Unsubscribe at any time.

Related Posts

2025 AI Recap: Top Trends and Bold Predictions for 2026

2025 AI Recap: Top Trends and Bold Predictions for 2026

If 2025 taught us anything about artificial intelligence, it's that the technology has moved decisively from experimentation to execution. This year marked a turning point where AI transitioned from b

read more
Google’s 2025 AI Research Breakthroughs: Gemini 3, Gemma 3 & More

Google’s 2025 AI Research Breakthroughs: Gemini 3, Gemma 3 & More

Key HighlightsThe Big Picture: Google’s 2025 AI research pushes models from tools to true utilities, with Gemini 3 leading the charge. Technical Edge: Gemini 3 Flash delivers Pro‑grade reasoning at

read more
Weekly AI News Roundup: The 5 Biggest Stories (January 1-7, 2026)

Weekly AI News Roundup: The 5 Biggest Stories (January 1-7, 2026)

Happy New Year, everyone! If you thought 2025 was wild for artificial intelligence, the first week of 2026 just looked at the calendar and said, "Hold my beer." We are only seven days into the year, a

read more
Daily AI News Roundup: 09 Jan 2026

Daily AI News Roundup: 09 Jan 2026

Nous Research's NousCoder-14B is an open-source coding model landing right in the Claude Code moment Nous Research, backed by crypto‑venture firm Paradigm, unveiled the open‑source coding model NousCo

read more
Unleashing Local AI Power with Nexa.ai's Hyperlink

Unleashing Local AI Power with Nexa.ai's Hyperlink

Key HighlightsFaster indexing: Hyperlink on NVIDIA RTX AI PCs delivers up to 3x faster indexing Enhanced LLM inference: 2x faster LLM inference for quicker responses to user queries Private and secure

read more
Activation Functions: The 'Secret Sauce' of Deep Learning

Activation Functions: The 'Secret Sauce' of Deep Learning

Have you ever wondered how a neural network learns to understand complex things like language or images? A big part of the answer lies in a component that acts like a tiny decision-maker inside the ne

read more
Light-Based AI Computing: A New Era of Speed and Efficiency

Light-Based AI Computing: A New Era of Speed and Efficiency

Key HighlightsAalto University researchers develop a light-based method for AI tensor operations This approach promises dramatically faster and more energy-efficient AI systems The technique could be

read more
Adobe Firefly Image 5 Revolutionizes AI Image Generation

Adobe Firefly Image 5 Revolutionizes AI Image Generation

As the AI image generation landscape continues to evolve, Adobe is pushing the boundaries with its latest Firefly Image 5 model. This move reflects broader industry trends, where companies like Canva

read more
Adobe's AI Creative Director

Adobe's AI Creative Director

As the lines between human and artificial intelligence continue to blur, companies like Adobe are pushing the boundaries of what's possible with AI-powered creative tools. This move reflects broader i

read more