Academic Knowledge RAG — knowledge orchestration for higher education institutions
Secure RAG assistant centralizing regulations, syllabi and procedures to accelerate document retrieval and reduce compliance errors.
Executive summary
A document RAG prototype designed for higher education institutions. The goal: centralise academic reference materials, regulations and procedures in a secure assistant capable of answering with source citation and confidence level. Designed for a pilot deployment on a campus of 5,000 students.
Business problem
Teaching teams spend an average of 40 minutes per query locating a regulation or syllabus across fragmented systems. The risk of non-compliance during accreditation audits is real and costly. No existing system centralises sources with full response traceability.
Solution
Secure RAG system with an automated ingestion pipeline, role-based access control (RBAC), systematic source citation with confidence scores, and a conversational interface tailored to academic profiles. Every answer is traceable and auditable.
Target KPIs
< 8 sec
P95 response time
70%
Document search time reduction
95%
Source citation precision
0
Target compliance incident
Technical architecture
Modular 5-layer pipeline: document ingestion (AWS Lambda + LangChain), vector storage (Supabase pgvector + HNSW), hybrid retrieval (vector + BM25 + Cohere Rerank v3), generation (Claude Sonnet 4.6 via LiteLLM), observability (Langfuse). OIDC auth via Supabase, audit trail persisted in database.
General architecture
Recommended stack
Competitive advantages
No SaaS product on the market combines confidence-scored citation, audit trail and granular RBAC control adapted to French and Swiss academic compliance requirements. The system is designed for accreditation, not just productivity.
Risks and mitigations
The primary risk is source document quality: poorly structured PDFs degrade retrieval precision. Mitigation: a quality validation pipeline at ingestion. Second risk: user adoption. Mitigation: a simple conversational interface and short onboarding. Third risk: LLM cost at scale. Mitigation: semantic cache and an economic fallback model.
Impact
- Prototype / evaluation in progress.
- Detailed impact data available on request.
Prototype / evaluation in progress.
Project scope
Pilot scope: 1 institution, 3 departments, 5,000 source documents. POC duration: 6 weeks. Environment: AWS eu-west-1 + Supabase Europe. Governance: GDPR, EU-hosted data, no personal data ingested.
Hosting and resilience
Deployment: Vercel (frontend) + AWS Lambda (ingestion) + Supabase (DB + auth). Target availability: 99.5% SLA. Recovery: RTO < 1h, RPO < 24h. Semantic Redis cache (TTL 24h) to absorb load spikes.
Role
Architecture design, data ingestion design, RAG engineering, security review
Next steps
Industrialisation of the document pipeline, extended compliance rule coverage, and campus SSO integration.
Tech stack
Timeline
S1–S2
Ingestion
Document corpus ingestion and indexing
S3–S4
Retrieval
Retrieval, reranking and quality evaluation
S5
Interface
Interface, OIDC auth and audit trail
S6
Pilot
User pilot and KPI measurement