Learn how Umaku uses RAG, vector databases, and sequential AI agents to deliver contextual intelligence across code, tickets, and teams.

In the current landscape of AI, building a “Chat with your PDF” tool is the “Hello World” of Large Language Models (LLMs). But applying AI to Project Management (PM) requires solving a much harder problem: Contextual Intelligence.
A standard LLM can tell you what a piece of code does. But it cannot tell you if that code satisfies the Acceptance Criteria written three weeks ago by a Product Manager, or if it conflicts with the DevOps compliance rules defined in the Project Charter.
At Umaku, we realized that to build a truly “Agentic” PM platform—one that doesn’t just store tickets but evaluates them—we had to move beyond simple keyword search. We had to build a system that understands the relationships between three distinct data planes:
This blog explores the engineering architecture behind Umaku, specifically how we built a Massive Vector Database (VDB) and a Sequential Multi-Agent System to give our AI the “Context” it needs to act as a Project Manager, Tech Lead, and QA Engineer simultaneously.
The core challenge of an AI Project Manager is Data Fragmentation. Your code is in GitHub, your requirements are in the Board, and your discussions are in Slack. To an LLM, these are disconnected silos.
To solve this, Umaku uses a unified Context Ingestion Pipeline. We do not query these services in real-time for every request (which would be slow and context-limited). Instead, we ingest, chunk, and embed this data into a centralized Vector Database.
We treat the project as a living organism where every action is a “memory.”

Figure 1: The Umaku Context Ingestion Pipeline.
Ingesting data is easy; retrieving the right data is hard. If a user asks, “Why is the payment module failing?”, a naive RAG system might retrieve every document containing the word “payment.”
Umaku employs a Hierarchical RAG Strategy.
This allows our agents to “see” that Ticket #305 (Implement Stripe) is linked to src/payments/stripe_adapter.py, bridging the gap between plan and reality.
Most PM tools rely on users to report status. If a developer moves a ticket to “Done,” the tool accepts it as truth. Umaku flips this model. We treat a ticket status as a claim that needs verification.
When a developer works on a ticket, they attach their Git commits or PRs. When the “Evaluate” agent is triggered, it performs a Cross-Modal Verification:
For example, if the ticket requires “User must be able to reset password via email,” but the code only contains a UI button without the backend logic, the Agent flags this discrepancy.

Figure 2: The Cross-Modal Truth Verification Loop.
This architecture serves as the foundation for our most advanced feature: The Automated Sprint Retrospective, where multiple agents work in sequence to audit the entire project.
Ingesting data into a Vector Database is only the first step. The real power of Umaku lies in how we act on that data.
In traditional RAG applications, you have one agent answering one user query. But evaluating a completed Agile Sprint—which contains dozens of tickets, hundreds of commits, and thousands of lines of code—is too complex for a single prompt context window. It requires a Sequential Multi-Agent Architecture.
We designed the Sprint Report mechanism not as a single “analysis” step, but as a relay race between specialized agents. We run these agents sequentially because the output of one agent often provides necessary context for the next.
When a Project Manager clicks “Complete Sprint,” the following pipeline is triggered:

Figure 3: The Sequential Multi-Agent Relay Architecture.
One of the hardest parts of Agile management is deciding what to do with a ticket that is mostly done but not fully done. Usually, this results in the entire ticket being dragged into the next sprint, messing up velocity metrics.
Umaku introduces an AI-Driven Descoping Agent that automates this granularity.
Developers often mark a ticket as “Done” when the coding is finished, even if testing or documentation is missing. If the system accepts this, the Sprint Report becomes inaccurate.
When the agents detect a discrepancy between the User Story Requirements and the delivered Code, the Descoping Agent intervenes. It doesn’t just reject the ticket; it proposes a constructive split.
This ensures that the Velocity Chart reflects actual value delivered, rather than a binary “All or Nothing.”

Figure 4: The AI Descoping Logic Flow.
While code is critical, projects are built by people. A key differentiator of Umaku is its ability to understand the “Human Context” of a project.
If a Project Manager asks, “How is Ahmed performing this month?”, a standard RAG system might just look for tickets assigned to Ahmed. Umaku goes deeper by utilizing a Long-Term Memory module for team evaluation.
While code is deterministic, software development is fundamentally a human activity. A project management tool that understands code but ignores the team is incomplete. Umaku bridges this gap by applying Contextual Intelligence to the people building the software.
Traditional metrics like “lines of code written” or “tickets closed” are notoriously bad at measuring productivity. They encourage quantity over quality.
Umaku employs a Team Evaluation Agent that runs weekly to build a comprehensive profile of each contributor. This isn’t a simple counter; it is a multi-modal analysis engine.
The agent aggregates data from two primary streams:
The agent synthesizes this into a narrative report. Instead of saying “Ahmed closed 5 tickets,” it might report:
“Ahmed focused on high-complexity backend tasks this week (Tickets #201, #204). While his ticket volume was lower, his code contribution involved significant refactoring of the authentication module. He was also highly active on Slack, assisting junior developers with environment setup.”
This allows Project Managers to evaluate performance based on value, not just velocity.

Figure 5: The Multi-Modal Collaborator Profiling Engine.
The interface for all this intelligence is the Umaku Chatbot. But this is not a standard “Chat with your Data” bot. It is Entity-Aware.
When a user asks: “What did Ahmed work on last week, and did he use React Native?”, the bot must perform a complex chain of operations:
This level of detail is only possible because we index the meta-data (Who, When, What) alongside the content (The Code).

Figure 6: The Context-Aware Query Resolution Path.
The shift from standard project management tools to Umaku represents a fundamental change in how we view software delivery. We are moving from Passive Data Entry—where humans have to constantly update the tool—to Active Intelligence, where the tool observes, analyzes, and guides the humans.
By building a system that understands the semantic relationships between a Project Charter, a Sprint Goal, a User Story, and a Git Commit, we have created more than just a dashboard. We have created a Context Engine that ensures every line of code written is a step towards the project’s actual goal.
Umaku doesn’t just manage the project; it understands it.