I am experiencing an issue with the knowledge base of your service.
I have several documents that I need to use for an avatar we are developing for a museum. When I upload these documents to a service like GPT-4o or Google’s Notebook LM, these systems correctly respond even to highly specific questions.
However, when querying the same documents (using GPT-4o as the model) within ConvAI, the behavior is highly inconsistent. Often, the responses fail to capture even the most prominent details.
The character ID I am using is 09f85624-eea8-11ef-85b7-42010a7be016, and I have currently uploaded several documents to the knowledge base, of which only two have been activated: one of 1.57KB and one of 648.69KB. The imprecise behavior remains the same even when working with just a single document.
Could you help me understand if I am doing something wrong?
Is it correct to expect that responses using GPT-4o with ConvAI should be the same as those obtained by directly using GPT-4o and the “project” feature as RAG?
Note:
Someone in this forum says that knowledge base supports English language only, but I can’t find any reference anywhere stating that the knowledge base only works in English. Could you please confirm this point?
It seems really strange to me that the knowledge base would only work in English… why?
I have already used the knowledge base in Italian for another project in December 2024, and it worked. Moreover, a RAG that uses embeddings to index documents should be language-independent.
Additionally, the behavior I am observing with this avatar is inconsistent: sometimes it responds correctly… other times it does not.
There is one aspect of your knowledge base document guidelines that I don’t fully understand. You mention using a “single file format.”
We have a PRO account with a 20MB limit for the knowledge base. Is it better to upload a single 20MB file rather than 20 files of 1MB each?
Yeh as Francis_Perron has mentioned, best to set your temperature to very low or maybe 0.
According to the docs it’s recommended to use a single text file, I assume even if that’s not possible then to use a minimal amount would be better. I think it is recommended to group the info you have in the text file(s) into paragraphs, each paragraph being the subject of 1 main thing, and iirc have say no more than 4 or 5 sentences in each paragraph. The docs can confirm most of this I think. Here is a sample paragraph from 1 of my Knowledge Bank files :
In the latest news, Russia’s war in Ukraine is still ongoing. President Trump has entered into negotiations with to President Putin about reaching a peace deal. There was a major rift between Ukraine President Zelensky and President Trump. The 2 men were trying to agree a deal that would see some of Ukraine’s valuable raw mineral resources being exchanged for America’s support of Ukraine’s war effort via money and miliary equipment. Ukraine has agreed to Trumps 30 day cease fire proposal although a cease fire is not yet in place because Russia’s Putin is cagey about a ceasefire.
Edit to add : my Knowledge Bank works really well tbh.
My tests were with french texts within multiple files.
I will try to merge them into one file to see if its better and lower the temperature if needed.
I can confirm that it works in italian.
Anyway it’s still not so precise as “standard” llms are.
It can’t find specific information as GPT-4o (for instance) does.
I see. We have a new feature. Would you be interested in giving it a try. The feature should better understand your document structure for AI response. If you are willing I can enable it for you to try
it could be interesting to try it.
The avatar I working on is almost “in production” (starting on May 3rd) and has been tested and optimized based on current ConvAI behaviour.
Is it possible for instance to try this new feature in a separate avatar?
Anyway we could test it on a differente account (we have more than one account)