Transform your video library into an interactive, intelligent knowledge base with our stunning Glassmorphism Aurora interface.
๐ Quick Start โข โจ Features โข ๐จ Live Demo โข โ๏ธ Architecture โข ๐ Documentation
Beautiful gradient-based UI with glassmorphism effects and pre-built example prompts
Real-time streaming responses with context-aware answers and video timestamps
|
Production-ready conversational assistant with stunning Glassmorphism UI |
|
FastAPI Backend
Vanilla JS/CSS Frontend
Ollama AI Models
|
Hybrid RAG Pipeline
Cross-Encoder Reranking
Streaming Responses
|
Glassmorphism UI
Aurora Theme
Mobile-First
|
|
Remembers your entire conversation history for coherent, multi-turn dialogues. Ask follow-up questions naturally without repeating context.
ChatGPT-style token-by-token streaming for instant feedback. See responses as they're generated, not after completion.
Hybrid search combining BM25 keyword matching with semantic embeddings, followed by cross-encoder reranking for maximum accuracy.
|
graph TD
Q[๐ User Question] -->|Hybrid Search| S[๐ Retrieval]
S -->|BM25 + Vector| R[๐ฏ Reranking]
R -->|Cross-Encoder| C[๐ Context]
C -->|Prompt| L[๐ค LLM Generation]
L -->|Stream| A[๐ฌ Final Answer]
style Q fill:#7F00FF,color:#fff
style A fill:#00D9FF,color:#fff
Jump directly to exact timestamps in source videos. Every answer includes clickable timestamps that take you to the relevant moment.
Beautiful Aurora theme with purple-cyan gradients, glassmorphism effects, smooth animations, and dark mode optimization.
|
| Hybrid Search | Reranking | First Token | Local & Private |
|
High-Performance Server โข |
Advanced Retrieval System โข Multi-stage retrieval pipeline โข Context-aware generation โข Smart chunk selection โข Prompt engineering โข Source attribution |
Real-Time Responses โข Token-by-token streaming โข Newline-delimited JSON โข Graceful error handling โข Stop generation control โข Progress feedback |
Persistent History โข SQLite-based storage โข Multi-session support โข Conversation history โข Context retention โข Export capabilities |
|
Stunning Visuals โข Purple-cyan gradient palette โข Smooth color transitions โข Dark mode optimized โข Custom CSS variables โข Consistent branding |
Modern Design Language โข Translucent card effects โข Backdrop blur filters โข Elevated shadows โข Frosted glass aesthetics โข Layered depth |
Works Everywhere โข Mobile-first architecture โข Adaptive layouts โข Touch-optimized controls โข Tablet support โข Cross-browser compatible |
Rich User Experience โข Typing animations โข Stop generation button โข Copy to clipboard โข Markdown rendering โข Syntax highlighting |
|
๐ค BM25 Keyword Search
โ ๐งฎ Semantic Vector Search
= โจ Best of Both Worlds |
๐ Initial Retrieval
โฌ๏ธ ๐ฏ Precise Scoring
โฌ๏ธ ๐ Top-K Results
|
๐น Video References
โ ๐ Chunk Metadata
= ๐ Clickable Links |
|
Frontend |
Backend |
LLM Engine |
Vector DB |
Database |
Models: llama3.2:latest โข bge-m3:latest โข nomic-embed-text โข gemma3:4b
โ
Python 3.8 or higher
โ
Ollama installed with models: llama3.2, bge-m3
โ
Modern web browser (Chrome, Firefox, Edge)|
Clone the Repository git clone <repository-url>
cd RAG-Based-AI |
|
|
Install Dependencies cd project/backend
pip install -r requirements.txt |
|
|
Start the Backend Server # From project/backend directory
uvicorn main:app --port 8000 --host 0.0.0.0Or use the convenient startup script: # Windows
.\start_aurora.bat
# Linux/Mac
./start_aurora.shโจ API Documentation: |
|
|
Launch the Frontend Simply open # Or serve it locally
cd project/frontend
python -m http.server 3000๐ Open: |
graph TB
User[๐ค User] -->|HTTPS| Frontend[๐จ Aurora Frontend UI]
Frontend -->|REST API| Backend[โก FastAPI Backend Server]
Backend --> Pipeline[๐ง RAG Pipeline Orchestrator]
Pipeline --> Search[๐ Hybrid Search Engine]
Search --> Vector[๐ Semantic Search<br/>ChromaDB + Embeddings]
Search --> Keyword[๐ Keyword Search<br/>BM25 Index]
Pipeline --> Rerank[๐ฏ Cross-Encoder Reranking<br/>Score & Sort Results]
Pipeline --> Context[๐ Context Builder<br/>Prompt Engineering]
Context --> LLM[๐ฆ Ollama LLM<br/>llama3.2:latest]
LLM -->|Streaming Tokens| Backend
Backend -->|JSON Stream| Frontend
Backend --> DB[(๐พ SQLite Database<br/>Chat History)]
style Frontend fill:#7F00FF,color:#fff
style Backend fill:#00D9FF,color:#fff
style Pipeline fill:#FFD700,color:#000
style LLM fill:#00FF88,color:#000
|
1๏ธโฃ
|
2๏ธโฃ
|
3๏ธโฃ
|
4๏ธโฃ
|
5๏ธโฃ
|
6๏ธโฃ
|
RAG-Based-AI/
โ
โโโ ๐ project/
โ โ
โ โโโ ๐ backend/ # โก FastAPI Backend Application
โ โ โโโ main.py # ๐ API entry point & endpoints
โ โ โโโ rag_pipeline.py # ๐ง Core RAG orchestration logic
โ โ โโโ search.py # ๐ Hybrid search engine
โ โ โโโ models.py # ๐ Pydantic request/response models
โ โ โโโ config.py # โ๏ธ Configuration management
โ โ โโโ prompts.py # ๐ฌ Prompt templates
โ โ โโโ utils.py # ๐ ๏ธ Helper utilities
โ โ โโโ database.py # ๐พ SQLite operations
โ โ โโโ requirements.txt # ๐ฆ Python dependencies
โ โ โโโ embeddings.joblib # ๐ Pre-computed embeddings (7MB)
โ โ โโโ bm25_index.joblib # ๐ BM25 search index (8MB)
โ โ โโโ chat_history.db # ๐ฌ Chat session database
โ โ
โ โโโ ๐ frontend/ # ๐จ Aurora UI (Vanilla JS)
โ โโโ index.html # ๐ Single Page Application
โ โโโ style.css # ๐ Glassmorphism styling
โ โโโ script.js # โก Chat logic & API calls
โ โโโ ๐ assets/ # ๐ผ๏ธ Images & resources
โ
โโโ ๐ legacy/ # ๐ฆ Legacy/Archived Scripts
โ โโโ app.py # Old Streamlit prototype
โ โโโ video_to_mp3.py # Video ingestion pipeline
โ โโโ preprocess_json.py # Data preprocessing
โ โโโ ...
โ
โโโ ๐ jsons/ # ๐ Transcription JSON files
โ โโโ 01_Installing VS Code.json
โ โโโ 02_Your First HTML.json
โ โโโ ... (18 course transcripts)
โ
โโโ ๐ tests/ # ๐งช Testing & Evaluation
โ โโโ evaluate.py # RAG evaluation metrics
โ โโโ eval_dataset.json # Test dataset
โ
โโโ README.md # ๐ This file
โโโ .env # ๐ Environment variables
โโโ start_aurora.bat # ๐ช Windows startup script
โโโ start_aurora.sh # ๐ง Linux/Mac startup script
|
|
| Method | Endpoint | Description |
|---|---|---|
GET |
/ |
API information and health status |
POST |
/chat |
Send a query and receive streaming response |
GET |
/sessions |
List all chat sessions |
GET |
/sessions/{id} |
Get specific session history |
GET |
/models |
List available Ollama models |
GET |
/health |
Backend health check |
๐ Full API Docs: Visit http://localhost:8000/docs after starting the backend
|
๐ Search Speed
|
๐ฏ Reranking
|
๐ค First Token
|
๐ Index Size
|
Tested on: Intel i7-12700K, 32GB RAM, RTX 3080
1. Fork the repository
2. Create your feature branch (git checkout -b feature/AmazingFeature)
3. Commit your changes (git commit -m 'Add some AmazingFeature')
4. Push to the branch (git push origin feature/AmazingFeature)
5. Open a Pull Request
Read CONTRIBUTING.md for detailed guidelines
Follow our CODE_OF_CONDUCT.md
This project is licensed under the MIT License - see the LICENSE file for details.
Built with amazing open-source tools:
FastAPI โข Ollama โข ChromaDB โข Sentence-Transformers โข BM25

