Secure Retrieval-Augmented Generation: Preventing Data Leakage with Provenance and Policy Enforcement

Main Article Content

Narendra Bhargav Boggarapu

Abstract

Retrieval-Augmented Generation (RAG) has become a paradigm architectural design to implement large language models in industries that face some form of regulation, but its probabilistic nature of retrieval creates a high risk of data leaks where access controls are not enforced or are used improperly. Secure RAG must have a comprehensive governance posture covering ingestion, indexing, retrieval, and generation mediated by identity-aware filtering, attribute-based access control, and policy-as-code frameworks that are version-controlled and auditably separate. Threats in this list are prompt injection, entitlement bypass, index poisoning, and generation-stage inferential disclosure, which need different prevention measures at the correct stage of the pipeline. The structural basis of the consistent enforcement is a two-plane architecture between the data plane and the control plane, where Open Policy Agent is the common decision-making and verifiable evidence bundle that meets the traceability considerations of the IEEE P7001 transparency standard. Assessment should be ongoing as opposed to periodic, and in terms of leakage rate, entitlement violation rate, provenance fidelity, and refusal correctness against a security harness that is consistent with the four core governance functions of the NIST AI RMF. The outcome is a deployment model where the enforcement of policy can be measured, provenance can be demonstrated to regulatory inquisitions, and leakage risk is structurally constrained as opposed to the widely held belief that this risk is mitigated.

Article Details

Section
Articles