Reputation: 41
We have seen some companies use GPT-3.5 or GPT-4 models to train their own data and provide customized answers. But GPT-3.5 and GPT-4 models are not available for fine-tuning.
I've seen the document from OpenAI about this issue, but I had seen OpenAI only allow fine-tuning davinci
, for example.
How do I customize answers from a GPT-3.5 or GPT-4 model if I can't fine-tune them?
Upvotes: 3
Views: 2378
Reputation: 22920
They don't fine-tune the GPT-3.5 or GPT-4 models.
You have two options.
OPTION 1: Using LlamaIndex or LangChain
What they do is use LlamaIndex (formerly GPT-Index) or LangChain. Both of them enable you to connect OpenAI models with your existing data sources.
OPTION 2: Using the OpenAI Embeddings API endpoint
See my past answer. Also, as @peter_the_oak mentioned, you can use Pinecone to store embedding vectors. Pinecone is designed specifically for handling this type of data.
Upvotes: 3
Reputation: 3710
Besides LlamaIndex, there is just the basic combination of Vector Database and LLM. Have a look at Pinecone: https://www.pinecone.io/learn/vector-database/
A vector database stores pieces of text (or pieces of images or sound or else) togtehter with a numeric vector. The numeric vector contains information about the text. A query can be transformed in a numeric vector, too. Now, having two vectors, there are algorithms for finding the one that matches most (e.g. cosine distance). So you can search the database for text that is highly relevant, according to the vector.
Now you can store your "knowledge" in many text/vector pairs. If you get a query, first read the suitable context from the vector database and put the received text in front of the prompt. Like this, the LLM will always have the right context knowledge, together with the custormer query.
Finetuning is overhead for most cases, but prompting is just a very simple and not so powerful solution. Vector Databases use prompting, but offer a mechanism to find suitable prompts, which is a powerful intermediate solution.
Upvotes: 1