Skip to main content

Nebula AI Onboarding Assistant

An agentic RAG system using LangGraph and Gemini to help new hires navigate company policies and org structure with multi-step reasoning.

Visit Repoistory
  • LangGraph Agent Development
  • RAG Architecture
  • Full-Stack Development
  • CI/CD & Docker Orchestration
Nebula AI dashboard showing real-time agent reasoning and chat interface

The Challenge

New employees spend hours searching scattered documentation for basic questions about policies, team structure, and role requirements. Traditional keyword search fails for multi-step queries like "Who is Sarah Chen's manager, and what tools do they need?"

Simple RAG systems can retrieve documents but can't reason across multiple sources or explain their process. Nebula solves this with an agentic architecture that thinks before it answers.

Agentic RAG Architecture

Nebula uses a ReAct agent powered by LangGraph instead of a traditional retrieve-and-generate pipeline. The agent evaluates each query, selects the right tools, and chains operations to build comprehensive answers.

The backend runs on FastAPI with Server-Sent Events for streaming responses. Incremental vector ingestion uses MD5 hashing to skip unchanged documents, and conversation history persists across sessions via SQLite checkpointing.

System architecture diagram showing LangGraph agent, tools, and data flow

Multi-Step Reasoning

The agent has three specialized tools: search_policies for vector search over company documents, lookup_employee for the org chart, and lookup_role_requirements for role definitions. For complex queries it automatically chains these tools together.

The Streamlit frontend displays the agent's reasoning chain in real-time, showing each tool call and its result. This transparency helps users understand how answers are constructed.

Example queries showing agent reasoning chain and tool orchestration

Technical Implementation

  • CI/CD: GitHub Actions with Ruff linting and 23 unit tests on every commit.
  • Docker: Multi-service docker-compose with health checks and persistent volumes.
  • Smart Caching: Incremental ingestion tracks file hashes to skip unchanged documents.
  • Streaming UX: SSE enables real-time token streaming and tool event updates.

Stack: LangGraph ReAct Agent · Gemini 2.5 Flash · FastAPI · Streamlit · ChromaDB · Docker · pytest