AI Engineering, JavaScript Solutions, Competitive programming in JavaScript, MCQ in JS

Wednesday, May 6, 2026

Day 8 - AI Engineering - Why Prompting Alone Is Not Enough (Understanding RAG)

Day 8 - Why Prompting Alone Is Not Enough (Understanding RAG)

AI Engineering — Day by Day
My journey to becoming an AI Engineer






After building my LLM playground and experimenting with prompts, I started noticing something:

No matter how good the prompt is… the model still fails in certain situations.

This made me question something fundamental:

If LLMs are so powerful, why do they still struggle with real-world tasks?

That’s when I started exploring something called:

👉 Retrieval Augmented Generation (RAG)


🧠 The Problem I Ran Into

Before learning RAG, my approach was simple:

Write a better prompt  
→ Get a better answer

But this approach has clear limits:

  • The model doesn’t have real-time knowledge
  • It hallucinates even when it “knows” something
  • It struggles with large or specific documents

Even after improving prompts, these issues didn’t go away.


⚠️ Why Prompting Alone Fails

Prompting works well when:

  • The question is general
  • The model already knows the answer

But it fails when:

  • The data is private (company docs, policies)
  • The data is large (PDFs, knowledge bases)
  • The data is dynamic (frequently changing)

At this point, I realized:

The issue is not just prompting — it’s access to the right information.

🔄 The Mental Shift

Instead of asking:

“How do I write a better prompt?”

I started asking:

“How do I give the model the right information before it answers?”


🚀 What RAG Actually Does

RAG changes the flow completely.

Instead of:

Prompt → LLM → Output

We now do:

User Query  
↓  
Retrieve relevant information  
↓  
Add it to the prompt  
↓  
LLM generates answer  

🧠 Key Insight

The most important realization for me was:

RAG is not about giving more data —
it’s about giving the right data at the right time.

⚠️ Important Observation

I also realized something critical:

If retrieval gives wrong information, the final answer will also be wrong.

This means:

  • The model is not the only component anymore
  • Retrieval quality becomes equally important

🧠 What Changed for Me

Before:

  • I focused only on prompts
  • I blamed the model for bad answers

Now:

  • I think in systems (retrieval + generation)
  • I focus on providing the right context

🚀 What’s Next

Now that I understand why RAG is needed, the next step is:

How does retrieval actually work?

In the next post (Day 9), I’ll explore:

  • What embeddings are
  • How similarity search works

💭 Final Thought

LLMs are powerful… but incomplete on their own.

RAG is what makes them practical.

This is Day 8 of my journey —
and this is where I stopped thinking in prompts…
and started thinking in systems.

No comments:

Post a Comment