Yes. It's a super simple flow of: Embed query and retrieve from Vector Store Build a big prompt with
Yes. It's a super simple flow of:
Embed query and retrieve from Vector Store
Build a big prompt with query + RAG results
Send that to LLM
There is no need to double guardrail it. The prompt is already checked in the embedding step.
And the big query is build from safe-bits (guard before entrying into DB + guarded user query).
So the second guard is just extra latency.
Embed query and retrieve from Vector Store
Build a big prompt with query + RAG results
Send that to LLM
There is no need to double guardrail it. The prompt is already checked in the embedding step.
And the big query is build from safe-bits (guard before entrying into DB + guarded user query).
So the second guard is just extra latency.









