If you want us to send your context-augmented prompts directly to OpenAI, add your OpenAI API key to the environment variables. Otherwise, leave it blank, and you can send them on your end.
If you're on the starter plan, this will simply be https://api.onecontext.ai. If you're on the enterprise plan, this will be the URL of your private instance of OneContext.
Create your first knowledge base
A knowledge base is a collection of files. We create our first knowledge base and
upload a file:
We want to chunk and embed the files in our knowledebase but first we need
somewhere to store our vectors. We create a vector index and specify the
embedding model that the vector index should expect:
By specifying the model we create a vector index of appropriate dimensions and
also ensure that we never write embeddings from a different model to this index.
Create an Ingestion Pipeline
We are ready to deploy our first ingestion pipeline.
steps:-step:KnowledgeBaseFilesname:inputstep_args:# specify the source knowledgebases to watchknowledgebase_names:["my_kb"]inputs:[]-step:Preprocessorname:preprocessorstep_args:{}inputs:[input]-step:Chunkername:simple_chunkerstep_args:chunk_size_words:320chunk_overlap:30inputs:[preprocessor]-step:SentenceTransformerEmbeddername:sentence-transformersstep_args:model_name:BAAI/bge-base-en-v1.5inputs:[simple_chunker]-step:ChunkWritername:savestep_args:vector_index_name:my_vector_indexinputs:[sentence-transformers]
Let's break down the steps.
The KnowledgeBaseFiles step tells the pipeline to watch
the "my_kb" knowledge base. When the pipeline is first deployed all files in the
knowledge base will be run through the pipeline. Any subsequent files uploaded to this
knowledge base will trigger the pipeline to run.
The Chunker defines how the files will be split into chunks.
The SentenceTransformerEmbedder step embeds the query
The Retriever step embeds the query and performs a similarity search against
the index we defined earlier. This step has a high recall and is great to
retrieve many candidate vectors.
The Reranker step uses cross-encoder model to further narrow down the results
only to the most relevant chunks.
Run the Query Pipeline
We can run the query pipeline and override any of default the step arguments defined in our pipeline at runtime by passing
a dictionary of the form: