Ollama-Profile is a Retrieval-Augmented Generation (RAG) system built to answer questions about me using structured personal data instead of hallucinated responses.
The system uses:
- JSON-based personal knowledge
- Vector embeddings stored in ChromaDB
- Local LLM inference using Ollama (Phi-3)
- Keyword-based routing for targeted retrieval
This ensures responses are context-aware, fast, and accurate.
-
📄 Structured Personal Knowledge Base
- Personal data stored as JSON (projects, skills, education, achievements, etc.)
-
🧠 Vector Search with ChromaDB
- Data is chunked, embedded, and stored for semantic retrieval
-
🔀 Query Routing Logic
- Keyword-based routing restricts retrieval to relevant chunks
- Example: project-related queries search only project vectors
-
🤖 Local LLM via Ollama
- Uses
phi3model for lightweight, fast inference - No cloud APIs, fully local execution
- Uses
-
❌ No Hallucinations
- Model answers strictly from retrieved context
- If data doesn’t exist, it says so
User Query
↓
Keyword Router
↓
Relevant Vector Collection (ChromaDB)
↓
Context Retrieval
↓
Ollama (Phi-3)
↓
Final Answer
Ollama-Profile/
│
├── data/
│ └── profile.json # Structured personal data
│
├── vector_store/
│ └── chroma/ # ChromaDB persistent storage
│
├── embedding.py # Chunking & vectorization logic
├── retriever.py # Context retrieval + routing
├── llm.py # Ollama model interface
├── main.py # Entry point
│
├── requirements.txt
└── README.md
Python
Ollama
Phi-3 LLM
ChromaDB
Sentence Transformers / Embeddings
JSON-based knowledge representation
Personal data is written in structured JSON format
Data is chunked and converted into embeddings
Embeddings are stored in ChromaDB
User query is analyzed using keyword routing
Only relevant chunks are retrieved
Ollama’s Phi-3 generates answers using retrieved context
# Clone the repo
git clone https://github.com/rahulkumarparida/Ollama-Profile.git
cd Ollama-Profile
# Create virtual environment
python -m venv venv
source venv/bin/activate # Windows: venv\Scripts\activate
# Install dependencies
pip install -r requirements.txt
# Pull Phi-3 model
ollama pull phi3
# Run the app
python main.py
📌 Example Queries
“What projects has he built?”
“What technologies does he know?”
“Tell me about his education”
“What are his achievements?”
Each query retrieves only the relevant vector space, not the entire dataset.
Demonstrates real RAG implementation, not just API calls
Shows understanding of:
Embeddings
Vector databases
Context grounding
Query routing
Fully local, privacy-preserving AI system