Skip to content
Chris Booth edited this page Feb 22, 2024 · 34 revisions

Welcome to the InferGPT wiki!

North Star of how it will (probably) work:

ArchitectureFlowOverview

Why InferGPT?

Language models are great a predicting the next token - as they are designed to. The issue though, compared to humans, is when one human requests another, we very rarely just spew out a response. Instead, we usually ask a question back. For example, if I ask you for a film recommendation, if you know me well, you would think: "Chris loves Marvel, and I know there's been a recent film released" - so you would ask: "Have you seen the latest Ant-Man?"

Alternatively, if you didn't know me well, you would ask things such as: "What genre of films do you like?"

We believe knowledge graphs are the solution to the above issue; to understand the user's current profile and ask questions based on missing context needed to solve their issue. It can then also store conversations, context and new information as time goes on - always remaining contextually updated.

Why a graph?

Graphs are great at this sort of task. They infer fast and they carry deep context with their edges. Most excitingly they also:

  1. Act as super-vector stores with Neo4j's cypher language, providing better performance vs cosine similarity methods.
  2. Make great recommendation models - graphs could even start to predict what you want to do next!

For a deeper dive I highly recommend the Neo4j Going Meta YouTube series

Clone this wiki locally