Explained: The Voiceflow Knowledge Base (Retrieval Augmented Generation)

Voiceflow
4 Dec 202314:27

Summary

TLDRThis video script introduces 'Retrieval Augmented Generation' (REG), a key feature in AI assistants that allows them to answer questions based on uploaded documents. The script explains how documents are broken into chunks and stored in a vector database, which the AI then uses to find relevant information to answer user queries. It also covers the technical aspects of implementing REG in Voice Flow, including the use of AI models, system prompts, and chunk limits to optimize the accuracy and efficiency of the AI's responses.

Takeaways

  • πŸ“š The script introduces Retrieval-Augmented Generation (RAG), a feature in AI tools that allows AI to answer questions based on uploaded documents.
  • πŸ› οΈ The process involves using a vector database to break documents into chunks and match them with user queries to provide relevant information.
  • πŸ” Voice Flow's knowledge base is a key feature that enables the creation of AI assistants that can access and utilize uploaded documents to answer questions.
  • πŸ“ˆ The script explains the technical process of how documents are broken into chunks, stored in a vector database, and then matched to user queries.
  • πŸ“ Chunks are small snippets of text from the original document, which are used to provide context to the AI model when assembling answers.
  • 🧲 The AI model uses the similarity of concepts within the question to find relevant chunks from the vector database to answer the user's query.
  • πŸ”‘ The script highlights the importance of choosing the right AI model, adjusting settings like temperature, max tokens, and chunk limit for optimal results.
  • πŸ”„ The accuracy of the AI's answers depends on the quality of the chunks provided, and the script suggests testing and refining the knowledge base for better accuracy.
  • πŸ›‘ The script mentions the ability to debug and inspect the process, including viewing the API information and similarity scores of chunks used in forming answers.
  • πŸ”§ The script discusses the importance of optimizing chunk usage and the trade-off between accuracy and token usage when increasing the chunk limit.
  • πŸ”— The script concludes with a mention of Voice Flow's API documentation, which allows developers to build custom solutions and integrate document management with the knowledge base.

Q & A

  • What is Retrieval-Augmented Generation (RAG) and why is it important for AI assistants?

    -Retrieval-Augmented Generation (RAG) is a function that allows AI assistants to upload documents and answer questions based on those documents. It is important because it enables AI to provide valuable and contextually accurate answers by referencing specific information from the uploaded documents.

  • What is the role of a vector database in RAG?

    -A vector database is used in RAG to store and manage the chunks of text from uploaded documents as vectors. These vectors represent the content of the chunks and help the AI model to identify and retrieve the most relevant information when answering questions.

  • How does Voice Flow's knowledge base differ from a traditional AI's base layer of knowledge?

    -Voice Flow's knowledge base allows users to upload specific documents that the AI can then reference for answering questions. Unlike a traditional AI's base layer, which relies on general knowledge from its training data, Voice Flow's knowledge base is tailored to the user's provided documents, offering more customized and specific answers.

  • What happens when a document is uploaded to Voice Flow's knowledge base?

    -When a document is uploaded, Voice Flow breaks it down into smaller sections called chunks. These chunks are then processed and stored in a vector database, where they are turned into vectors that represent the content of the document.

  • How does Voice Flow determine which chunks of information to use when answering a question?

    -Voice Flow uses a model to compare the user's question with the chunks in the vector database, identifying the most similar chunks based on the concepts within the question. It then selects the most relevant chunks to include in the AI model's response.

  • What is the significance of the chunk limit setting in Voice Flow?

    -The chunk limit setting determines how many of the most similar chunks are selected to answer a question. Increasing the chunk limit can improve the accuracy of the answer by providing the AI model with more information, but it also increases the number of tokens used per response.

  • How can the quality of answers from Voice Flow's knowledge base be improved?

    -The quality of answers can be improved by ensuring the chunks in the knowledge base are accurate and relevant. This involves testing the knowledge base, reviewing the chunks used in answers, and updating or removing documents that do not provide correct information.

  • What is the purpose of the system prompt in Voice Flow's knowledge base?

    -The system prompt is used to frame the response from the AI model. It can be customized to influence the length, format, or style of the answer, such as requiring a certain number of sentences or a specific structure like bullet points.

  • How does Voice Flow handle the token usage when multiple chunks are used to answer a question?

    -Token usage in Voice Flow is influenced by both the input (number of chunks) and the output (length of the answer). When more chunks are used, the input token count increases, which can affect the overall token usage for each response.

  • What are some of the developer tools available for Voice Flow's knowledge base?

    -Voice Flow provides APIs for uploading, deleting, and replacing documents in the knowledge base, as well as for retrieving answers. These APIs can be used to build custom services or widgets that integrate with Voice Flow, allowing for automated updates to the knowledge base.

  • How can Voice Flow's knowledge base be used to create custom solutions?

    -Developers can use Voice Flow's knowledge base APIs to create custom solutions, such as widgets for uploading documents directly into the knowledge base or services that automatically update the knowledge base with the latest documentation from a CMS.

Outlines

00:00

πŸ€– Introduction to Retrieval-Augmented Generation (REG)

This paragraph introduces the concept of Retrieval-Augmented Generation (REG), a feature in AI tools that allows users to upload documents and ask questions based on those documents. The video aims to explain how REG works conceptually and visually, using Voice Flow's knowledge base as an example. Voice Flow is a platform that enables the design of AI assistants capable of accessing uploaded documents to answer questions. The explanation includes the process of breaking documents into chunks and storing them in a vector database, which the AI uses to find relevant information when answering questions.

05:02

πŸ“š Behind the Scenes of Voice Flow's Knowledge Base

This section delves into the inner workings of Voice Flow's knowledge base, starting with the process of uploading documents and breaking them into text chunks. It explains how these chunks are turned into vectors and stored in a vector database. The paragraph then describes the process of answering user questions by finding similar chunks in the database and using them to inform the AI model, which assembles an answer. The use of system prompts and the AI model's output are also discussed, along with the settings that can be adjusted for optimization, such as chunk limit and model selection.

10:03

πŸ” Debugging and Customizing the Knowledge Base

The final paragraph focuses on debugging and customizing the knowledge base in Voice Flow. It discusses the importance of chunk quality for accurate answers and provides tips on how to identify and improve the quality of chunks. The paragraph also covers the settings that influence the AI's response, such as temperature, max tokens, and system prompt. Additionally, it touches on the chunk limit's impact on accuracy and token usage, advising on how to balance these factors. The paragraph concludes with a mention of Voice Flow's APIs for developers, which can be used to automate the upload, deletion, and replacement of documents in the knowledge base.

Mindmap

Keywords

πŸ’‘Retrieval Augmented Generation (RAG)

Retrieval Augmented Generation, or RAG, is a feature in AI tools that allows the AI to access and utilize uploaded documents to answer questions. It is central to the video's theme as it forms the basis for designing AI assistants that can provide valuable answers. The script explains that RAG uses the AI's base layer of knowledge and a library of provided documents to assemble answers to user queries.

πŸ’‘Knowledge Base

A knowledge base in the context of the video refers to a system that stores documents and information from which an AI can draw to answer questions. The video emphasizes the importance of the knowledge base in the Voice Flow system, illustrating how documents are uploaded and broken down into chunks for the AI to reference when responding to user inquiries.

πŸ’‘Chunks

In the script, 'chunks' are small, overlapping segments of text extracted from the uploaded documents. These chunks are crucial as they represent the pieces of information that the AI uses to form its responses. The script mentions that a single document can be broken down into numerous chunks, which are then stored in a vector database.

πŸ’‘Vector Database

A vector database is a type of database that stores data points as vectors, which are mathematical representations of the data. In the video, chunks of text are turned into vectors within the vector database, allowing the AI to match user queries with the most relevant chunks of information. The script explains the process of how chunks are turned into vectors and used to retrieve information.

πŸ’‘Concepts

Concepts in the video are the underlying ideas or themes within the user's question that the AI identifies to find relevant chunks in the knowledge base. The script describes how the AI matches concepts like 'pricing' from the user's question to similar concepts within the chunks to provide an accurate response.

πŸ’‘AI Model

The AI model mentioned in the script refers to the artificial intelligence system that processes the chunks of information and the user's question to generate a response. The video discusses how the AI model is selected and configured to optimize the accuracy and relevance of the answers it produces.

πŸ’‘Tokens

Tokens in the context of the video are the basic units of text that the AI model uses to process and generate responses. The script discusses the importance of token limits in controlling the length and depth of the AI's responses, as well as the impact on the efficiency of token usage.

πŸ’‘System Prompt

A system prompt is a set of instructions or a framing mechanism provided to the AI model to guide the formation of its response. The video script explains how the system prompt can influence the structure and content of the AI's answer, such as requiring a response in bullet points or a specific number of sentences.

πŸ’‘Chunk Limit

The chunk limit is a setting that determines how many chunks of information are passed to the AI model when generating a response. The script illustrates how adjusting the chunk limit can affect the accuracy and detail of the AI's answers, as well as the number of tokens used in the process.

πŸ’‘API

API, or Application Programming Interface, is a set of rules and protocols that allows different software applications to communicate with each other. The video script mentions the use of APIs in the context of uploading, deleting, and replacing documents in the knowledge base, as well as retrieving answers from it, providing developers with the tools to integrate and automate these processes.

Highlights

Retrieval Augmented Generation (REG) is a key feature for designing AI assistants that can provide valuable answers based on uploaded documents.

REG uses AI's base layer of knowledge and its understanding of how information relates to assemble answers from provided documents.

A vector database is used in the process, where document chunks are turned into vectors for similarity comparison.

Voice Flow's knowledge base allows uploading documents and designing AI assistants to access and answer questions based on those documents.

Documents are broken into chunks in Voice Flow, which are then stored in a vector database for efficient retrieval.

The AI model determines the most similar chunks from the vector database to answer a user's question.

Concepts within the user's question are matched with chunks to find the most relevant information.

The AI model then uses the selected chunks to formulate a response to the user's question.

Voice Flow's settings allow adjusting the chunk limit, which affects the detail and accuracy of the AI's answer.

Debugging in Voice Flow includes inspecting the network to see the API information and similarity scores used to form answers.

Choosing the right AI model and adjusting settings like temperature and max tokens can optimize the AI's response.

The system prompt in Voice Flow can be customized to influence the structure and depth of the AI's response.

Increasing the chunk limit improves accuracy but also increases token usage per response.

The knowledge base's effectiveness depends on the quality of the chunks provided from the uploaded documents.

Testing the knowledge base in preview mode helps identify which documents are being used to answer questions.

Updating or re-uploading documents can improve the accuracy of the knowledge base by replacing incorrect chunks.

Voice Flow offers APIs for uploading, deleting, and replacing documents in the knowledge base, enabling custom integrations.

Developers can leverage Voice Flow's APIs to build custom solutions for knowledge base management.

The presentation by Daniel, head of grow at Voice Flow, provides a comprehensive overview of the knowledge base functionality.

Transcripts

play00:00

if you've used any AI tool of the past

play00:02

year you'll notice that there's

play00:03

typically a feature where you can upload

play00:05

documents and actually ask the AI

play00:07

questions based on those documents now

play00:09

this is a function called retrieval

play00:11

augmented generation or reg for short

play00:14

and it's incredibly important to know

play00:15

how this works if you're going to be

play00:17

designing AI assistants that can

play00:19

actually provide valuable answers now in

play00:21

this video we're going to go over how it

play00:23

works conceptually but we're also going

play00:24

to show you visually what's actually

play00:26

happening when you go and upload

play00:28

documents into voice flows knowledge

play00:29

base which is our own version of

play00:31

retrieval augmented generation this lets

play00:33

you upload a bunch of documents and then

play00:34

actually design an AI assistant around

play00:36

them to be able to send questions and

play00:38

access those documents and even choose

play00:40

certain files within the knowledge space

play00:41

that you've uploaded to be able to uh

play00:43

answer users questions so at a high

play00:45

level retrieval augmented generation

play00:47

works the following way it uses the ai's

play00:50

base layer of knowledge so this is the

play00:52

information that it's already trained on

play00:54

so in the case of GPT it would be the

play00:56

internet but really what's more

play00:57

important is its understanding of how

play00:59

relationships Co and map to each other

play01:01

then what it does is when you ask it a

play01:03

question it says can I answer this with

play01:05

the knowledge that I have and then let

play01:08

me look through a library of documents

play01:10

that I've been provided or library of

play01:12

information that I've been provided to

play01:14

actually assemble an answer that can

play01:16

answer this question now that process

play01:18

typically involves the use of something

play01:20

called a vector database and vectors and

play01:22

so without further Ado let's actually

play01:24

hop into vo low and get a sense of what

play01:26

happens once you upload your knowledge

play01:29

or your

play01:30

to the knowledge base so the knowledge

play01:32

base is the most impactful important

play01:34

feature we built in voice flow and we've

play01:37

got a lot of questions so the first one

play01:38

is obviously how does the knowledge base

play01:40

actually work so if you're within voice

play01:43

flow if if you're remember and I've got

play01:44

a little retail project up here this is

play01:46

the ril template you can find on the

play01:47

website let me just go ahead and copy

play01:49

these URLs and within the knowledge base

play01:52

when you go to head add data source and

play01:54

you drop these URLs in once they start

play01:57

uploading there's a lot of stuff that's

play01:58

going on behind the scenes so as these

play02:00

process I'm going to walk you through

play02:01

just a little illustration of what this

play02:03

looks

play02:03

like so imagine you've got your document

play02:06

here that you've uploaded right so this

play02:08

can be the URL that I just did maybe

play02:09

it's the returns page on your website so

play02:12

once you upload this into voice flow

play02:14

what's happening is that voice flow

play02:16

behind the scenes is actually breaking

play02:17

this one document so this might be a big

play02:19

chunk of text right so you've got a

play02:20

website it's got a bunch of stuff on it

play02:22

it's basically scraping all the text

play02:24

assembling it and then it breaks it into

play02:26

sections called chunks so chunks are

play02:29

basically like little tiny versions of

play02:31

the document that you've built so let's

play02:33

just go ahead and and illustrate this so

play02:34

I'll make them maybe blue and these are

play02:37

basically just Snippets so these are

play02:39

maybe you know, 2,500 characters of of

play02:42

text that overlap a bit throughout your

play02:44

project and so that one document that

play02:46

you've uploaded uh may actually be

play02:48

broken out into like tens 50s hundreds

play02:51

of different chunks that basically are

play02:53

like little Snippets of the information

play02:55

on the page itself so if you've got a

play02:57

really long page with a lot of

play02:58

information and it's talking about your

play02:59

return policy again we're breaking that

play03:01

out into little pieces so that we can

play03:03

digest them and put them in what we call

play03:05

a vector database so these chunks are

play03:09

stored here so let's take them out so

play03:12

let's just say you know this one

play03:13

document now becomes these many little

play03:16

different chunks and these chunks are

play03:18

stored in again what's called a vector

play03:23

database what this means is that this

play03:27

these chunks are basically turned into

play03:29

vectors and you can imagine a vector is

play03:31

like an arrow and that Arrow generally

play03:33

represents what's in the chunk and the

play03:34

chunk is just a piece of information and

play03:37

so what all we've done by this stage is

play03:38

just break a bunch of text like a big

play03:40

chunk document into little pieces that

play03:42

we can then uh utilize uh within the

play03:44

flow itself or within some of R

play03:46

ml now when a user asks a question and

play03:49

you can imagine here that you know

play03:51

you're not just uploading one document

play03:52

you're uploading many different

play03:54

documents into your project and so

play03:56

you've now got tons of chunks that are

play03:58

in this Vector database that are all

play04:00

kind of waiting to be used so now when a

play04:02

user asks a question what's happening is

play04:04

that we take that question and we look

play04:06

at our Vector database and we say hey

play04:09

what are some chunks that are similar to

play04:12

the question being as asked and so we

play04:14

may return let's just say you know maybe

play04:16

it's this one make it yellow this one

play04:19

here make it yellow again and maybe

play04:22

another one from the same

play04:24

document and so we have a model that's

play04:26

running that's determining that based on

play04:27

the question being asked these are the

play04:29

most similar chunks of information that

play04:32

most likely answer or pertain to that

play04:34

question how this works is it's looking

play04:37

at Concepts that are within the question

play04:40

so if someone asks how much are your

play04:42

products or what is your pricing it's

play04:44

able to take that concept of pricing and

play04:46

compare it to the different chunks here

play04:48

to say are there another similar concept

play04:50

of pricing in these chunks of

play04:51

information that I have if it finds

play04:53

multiple and you can set a max value

play04:55

here so let's say for example the max

play04:57

value is three here it'll pick the three

play04:59

three most similar chunks or chunks that

play05:01

have the most similarity to the question

play05:04

being

play05:05

asked then what it's going to do is it's

play05:07

going to pull out those chunks so let's

play05:08

just go ahead and actually pull them out

play05:10

here and now it's going to take that

play05:15

it's going to package it with a prompt

play05:16

that's just us being able to say hey

play05:19

answer this user's question and that's

play05:23

going to then be sent to a AI

play05:26

model and so we'll call this AI model

play05:31

AI

play05:34

model now this is the information that's

play05:36

going to the AI model and we're

play05:38

basically saying hey AI model this is

play05:40

the question here's the relevant

play05:42

information from this user's database or

play05:44

the information they gave us assemble an

play05:47

answer uh that actually answers the

play05:49

question and it'll be able to come out

play05:52

with uh a model an answer just like that

play05:54

so we'll make this answer green and

play05:56

we'll just say that this is the

play05:58

output

play06:00

so we'll call this output AI

play06:05

output and so we've got here the AI

play06:08

input that's being sent over and then

play06:11

we've got the output which is the actual

play06:12

answer of the question so that's

play06:14

generally what's happening behind the

play06:16

scenes here there's also a lot of

play06:17

smaller optimizations and a lot of other

play06:19

models that we have running that we're

play06:21

constantly learning and trying to

play06:22

improve to ensure that answers are

play06:24

actually they're more accurate than what

play06:26

you would find and so that's how this is

play06:28

working and so now when we look at voice

play06:30

flow and let's go into our retail

play06:32

assistant here and let's go ahead and

play06:34

ask a question like what is your return

play06:37

policy or yeah what is your return

play06:42

policy so if you remember I've already

play06:45

because I've already uploaded all my

play06:46

documents those have already been broken

play06:48

out into chunks it's taking my question

play06:50

it's looking and finding chunks it's

play06:51

passing those to a language model and

play06:52

it's summarizing the answer and you can

play06:54

see here in the preview mode you can

play06:55

actually see the chunks that it took it

play06:57

from and so in this case I've got one

play06:59

chunk that I grabbed it from and then

play07:01

this was the chunk that it passed to the

play07:03

AI model with my question to say come up

play07:05

with a response now in the settings I

play07:08

have the option to actually change some

play07:09

stuff so I can make the chunk limit five

play07:13

for example so let's hit save and let's

play07:15

run this question again now and you can

play07:17

see that the question the answer is

play07:19

actually going to be much different so

play07:20

what is your return

play07:22

policy uh now because I've got five

play07:24

chunks that are being passed in I've got

play07:26

a lot more information to work with it's

play07:27

also going to use more tokens but you

play07:29

you can see that the answer is actually

play07:30

a bit more detailed and when I go I've

play07:32

got 1 2 3 four five different chunks

play07:35

that it was able to actually pull from

play07:37

to inform this

play07:39

answer another thing you can do if you

play07:41

want to go even deeper into debugging is

play07:43

hit inspect and what you're able to do

play07:46

is you can actually see some of the

play07:47

information that's going on behind the

play07:48

scenes here so if you go to network and

play07:50

then let's just rerun this question so

play07:52

what is your return

play07:55

policy you can see that here under

play07:57

knowledge base xhr I can I can actually

play07:59

see the API information on what's being

play08:01

pulled and so this is the information

play08:02

you get when you actually use our API

play08:03

directly but I can see what's being

play08:05

passed to voice flow and so I can see

play08:07

each of the chunks here and you can see

play08:08

that when I mentioned there's a

play08:09

similarity score right so our system

play08:12

determined that these chunks had a

play08:14

different they all have different

play08:15

similarity scores and so it's accounting

play08:16

for that in how it forms the answer as

play08:18

well with the AI model so there's a lot

play08:20

more going on behind the scenes here

play08:22

than just kind of looking up information

play08:23

in your document that's trying to ensure

play08:25

the accuracy of

play08:26

it so going into some of the settings

play08:29

here that are important to know the

play08:31

first one is which model you pick so in

play08:34

our documentation here in our little

play08:35

diagram this the AI out input and the AI

play08:38

output are being determined by a model

play08:40

of your choosing so in voice flow you've

play08:42

got a couple different models you've got

play08:44

from gpt3 to GPT 4 and everything in

play08:46

between these are ranked in order of how

play08:48

expensive they are and how accurate they

play08:50

are you can see here that GPT 4 is

play08:52

incredibly expensive you do not want to

play08:54

use this for knowledge based just

play08:56

general stuff it's going to use up a lot

play08:58

of your tokens you can see the GPT 3.5

play09:00

and clae instant are 1X tokens these are

play09:02

really good to use for knowledge base

play09:04

and if you really need to improve

play09:05

accuracy you can increase it to claw one

play09:07

or two but I would recommend just

play09:08

sticking with these two as your accuracy

play09:10

is going to be pretty

play09:11

decent the second one is temperature so

play09:14

this is an AI model setting most AI

play09:16

models have this this is basically

play09:17

telling the model like how creative it

play09:19

can be with its answer or how closely it

play09:21

should ST stick to the exact text or the

play09:23

exact information that it has Max tokens

play09:26

this is important to know as well so

play09:28

when it comes to your output and so this

play09:30

is your output here the answer that the

play09:32

AI actually forms this slider is what's

play09:35

going to dictate how long or how short

play09:37

that answer is going to be but what's

play09:39

important to note is that this is

play09:40

actually just setting a limit so if I

play09:42

set the limit of 300 tokens it's going

play09:44

to be about like a th characters cuz

play09:46

each token is about three to four

play09:48

characters and so that's the maximum

play09:50

length it can be but what I want to do

play09:52

is I want to play with this and I want

play09:54

to play with the system prompt right so

play09:56

the system prompt here is what's

play09:57

actually going to as part of the prompt

play09:59

to be able to frame the response back

play10:00

from the AI and if you want a shorter

play10:02

answer you want a longer answer you want

play10:03

a formatted answer this is where you'd

play10:05

put it so I I would say something like

play10:07

your sentences your answers must be at

play10:09

least at least you know like three

play10:12

sentences or you can say you must

play10:14

structure in a bullet point form or you

play10:16

can say you know you must say you're an

play10:17

AI assistant this is where you can start

play10:19

to tweak how long or how robust or How

play10:21

deep the response is and then you can

play10:23

use the max token slider to also help

play10:25

account for that finally is the chunk

play10:28

limit so this is now probably the most

play10:30

important concept to think about is that

play10:33

when you increase the chunk limit you're

play10:35

going to have a higher accuracy because

play10:37

the AI model has more information to

play10:38

work with but you are also going to

play10:40

increase the number of tokens you use

play10:43

per response that's because token out

play10:46

like the amount of tokens you use is

play10:48

both input and output we already talked

play10:50

about a slider where you can control the

play10:52

output but the input is controlled by

play10:54

the number of of of of chunks that you

play10:57

have so in here where I I mentioned I've

play10:59

got five chunks you can just imagine

play11:01

that when I do my preview and I do that

play11:03

same question so what is your return

play11:07

policy that all five of those chunks so

play11:09

all of that information is being passed

play11:11

to the large language model and so there

play11:14

are like all of these are counting

play11:16

against your tokens so that's 1 2 3 4 5

play11:18

it's quite a bit and so your token uses

play11:20

is going to go way up so you really have

play11:21

to be cautious around what you want to

play11:23

do if you want to again increase those

play11:24

or decrease those honestly two or three

play11:27

is fine I would really stick with two

play11:29

you can go as low as one if you're

play11:30

really conscious about your token usage

play11:32

but that's how you can start to modify

play11:34

it so now going back to our presentation

play11:36

and I hope that made sense we can look

play11:38

at a couple other really specific

play11:40

scenarios so on average the knowledge

play11:42

base uses between 500 and 2,000 tokens

play11:45

depending on how many chunks it's using

play11:47

so again when you're running a question

play11:49

in the knowledge base the like input and

play11:51

output the input is is determined by the

play11:54

number of chunks and so that's what you

play11:56

really want to pay attention to the

play11:59

quality of the answer depends on the

play12:00

quality of the chunks provided so my

play12:02

recommendation is as you're testing out

play12:03

your knowledge base go ahead and answer

play12:06

question or ask questions in the preview

play12:07

mode because you can start seeing what

play12:10

what documents is actually pulling from

play12:11

to answer your question if you're

play12:12

finding that you are getting bad quality

play12:14

answers it's because you have chunks in

play12:17

your project that are don't have good

play12:19

information and they're being included

play12:20

so you want to find out where those are

play12:22

coming from and you want to either

play12:23

remove that document or you want to

play12:24

update and re-upload that document when

play12:27

you delete a document it removes the

play12:28

chunks when you reupload the document it

play12:30

reads chunks so again what is your

play12:33

return

play12:36

policy you can see here where all the

play12:39

chunks are coming from and so if I look

play12:41

at and say you know I've say let's say

play12:44

this isn't correct let's say it's

play12:45

actually 60 days I can see that okay

play12:47

cool it's pulling most of the

play12:48

information from this returns page if my

play12:51

information in the answer is incorrect

play12:52

it's because there's some answer

play12:54

incorrect in your returns page so you

play12:56

need to go there figure where it is

play12:57

remove it or update it and then

play12:59

re-upload the document so that it's

play13:00

pulling the chunks so that's how you

play13:02

would improve the accuracy of your

play13:03

project and how you would debug where

play13:05

those answers are coming

play13:06

from going back to over here we just

play13:09

walked through how to determine and find

play13:11

out where chunks are coming from we

play13:12

talked about optimizing tunk chunks so

play13:14

with just determin the number that

play13:16

you're using the last one here is

play13:17

knowledge based apis and so over here in

play13:20

our API documentation so developer.

play13:21

vo.com so if you're a developer you

play13:23

should absolutely go check this out we

play13:25

have a ton of apis and so for our

play13:26

knowledge base the most important ones

play13:28

uh that are relevant here are the upload

play13:30

and delete and replace document

play13:33

endpoints this means that you can build

play13:35

a custom service on top of VO those apis

play13:37

or if you are a developer with a react

play13:39

chat you could build a widget so someone

play13:41

could actually upload a document and

play13:42

this would allow people to upload

play13:44

documents right into the knowledge base

play13:45

or remove them or delete them or replace

play13:46

them without needing to actually touch

play13:48

voice flow so if you have a customer or

play13:51

or you are a business where you have all

play13:53

of your documentation in like a CMS for

play13:56

example you can use this API to

play13:58

constantly upload update and replace all

play14:00

the documentation your knowledge base

play14:01

without touching voice so automatically

play14:03

so again something to look at something

play14:05

to keep an eye for if you're a developer

play14:06

we've also got a API to actually get an

play14:08

answer from the knowledge base so this

play14:09

gives you a bit more control as

play14:11

well so that's uh going to be really

play14:14

important you should see some custom

play14:15

Solutions coming out from community

play14:16

members as well uh that are using

play14:19

that so thanks again my name is Daniel

play14:21

I'm the head of grow. voice flow I will

play14:23

see you in Discord along with the rest

play14:25

of my team bye

Rate This
β˜…
β˜…
β˜…
β˜…
β˜…

5.0 / 5 (0 votes)

Related Tags
AI ToolsDocument RetrievalKnowledge BaseVoice FlowAI AssistantVector DatabaseData ChunkingNatural LanguageContent MappingRetrieval Augmented GenerationCustom Solutions