We Run This Stack Today
Graphiti + FalkorDB (primary), Railway Postgres (mirror), LightRAG, and local Ollama models are all live on our own infrastructure. You are not paying for us to learn on your data.
Production-grade temporal knowledge graphs and RAG pipelines built on top of your existing CRM and ERP — without ripping out what you already have. Graphiti + FalkorDB + LightRAG + Ollama, deployed in 4 to 6 weeks.
Graphiti shipped in late 2024. FalkorDB is a Redis-backed graph database that still feels like a secret weapon. LightRAG is a 2024 research project that's suddenly usable in production. In April 2026, fewer than 1,000 developers worldwide have any of these running in production.
9o4t Inc does. Today. We run Graphiti + FalkorDB as our primary memory layer, mirrored to Railway Postgres, with Ollama (llama3.2, mistral) serving local LLM inference and OpenAI handling structured-output tasks. It's not a pitch deck — it's the stack you'd be hiring us to build for you.
We run 4-to-6 week pilots that answer a specific business question: "Can a knowledge graph over our last seven years of Salesforce notes surface the pattern we've been missing?" At the end of the engagement you get a working system, a decision framework for productionizing it, and candid recommendations on what not to build.
Graphiti + FalkorDB (primary), Railway Postgres (mirror), LightRAG, and local Ollama models are all live on our own infrastructure. You are not paying for us to learn on your data.
Kubernetes consultants in 2015 charged $300–$500/hour because almost no one had production K8s experience. Snowflake consultants hit similar windows. Knowledge graphs are in that window right now — and the first-mover premium goes away inside 18 months.
Most AI consultants can build RAG over a pile of PDFs. We can build RAG over Salesforce notes, NetSuite transactions, and HubSpot email threads, with the domain context of someone who's shipped SuiteScript for a decade. That bridge is rare.
Graphiti extracts entities via OpenAI (structured JSON mode). Embeddings run locally via Ollama's nomic-embed-text. For clients who can't send data to OpenAI, we can swap extraction to local Mistral and keep the entire stack on-prem.