Enhance RAG Chatbot Performance By Refining A Reranking Model

Labelbox
26 Apr 202417:15

Summary

TLDRThis video script outlines a workflow using Labelbox to enhance a retrieval process for a custom chatbot. It demonstrates how to utilize a re-ranker model to improve document retrieval by incorporating human-in-the-loop for context evaluation. The process involves embedding documents into vector space, fine-tuning the model with human annotations, and testing the model's performance. The script highlights the importance of human expertise in refining AI responses, showcasing an effective technique for improving chatbot accuracy.

Takeaways

  • πŸ“š The script demonstrates a workflow for improving a retrieval process using the Labelbox platform.
  • πŸ€– A custom chatbot is utilized to interact with an internal corpus of documents through a retrieval method to find relevant information chunks.
  • πŸ” The process involves using a ranker model to enhance retrieval performance by re-ranking initial search results, which may not always be contextually relevant.
  • πŸ§‘β€πŸ”§ A human loop is introduced to evaluate text chunks against search queries, using human judgment to refine the reranking model.
  • πŸ“ˆ The script shows a baseline experiment for retrieval using a consistent Transformer model to embed documents into vector embeddings.
  • πŸ“‘ An example of embedding the extensive NFL rulebook into a vector index for efficient retrieval is provided.
  • πŸ“ The script details how to create a training dataset for fine-tuning the ranker model by extracting top chunks from initial data sets for human review.
  • πŸ”„ The use of Labelbox SDK to format and submit text queries and response chunks to the Labelbox platform for human review is demonstrated.
  • πŸ“Š The process of generating pre-labels using a model like OpenAI's GP4 to assist annotators in efficiently labeling data is explained.
  • πŸ“ Annotation projects are set up in Labelbox to allow human annotators to review and adjust pre-labels, improving the relevance of retrieved chunks.
  • πŸ”§ After human review, the annotations are exported in JSON line format to fine-tune the ranker model, incorporating human expertise into the model's learning process.
  • πŸ“ˆ The fine-tuned ranker model is then used to re-rank query-response pairs, aiming to improve the context provided to the language model for generating responses.

Q & A

  • What is the primary goal of using Labelbox in the described workflow?

    -The primary goal of using Labelbox in the workflow is to improve the retrieval process by fine-tuning a re-ranker model with human-in-the-loop to enhance the performance of retrieving relevant information from an internal corpus of documents.

  • What is a 'chunk' in the context of this script?

    -In the context of this script, a 'chunk' refers to a segment of text retrieved from the internal corpus of documents, which is used as context for the language model to generate a response.

  • Why is a re-ranker model used in the retrieval process?

    -A re-ranker model is used to improve retrieval performance by re-evaluating the initial top results and potentially promoting more contextually relevant information that might be ranked lower by the initial retrieval algorithm.

  • What role does the human loop play in the fine-tuning process of the re-ranker model?

    -The human loop involves expert annotators who evaluate the relevance of each text chunk to the search query, providing valuable feedback that is used to fine-tune the re-ranker model, ensuring it aligns with human expertise.

  • How does the script suggest improving the baseline retrieval performance?

    -The script suggests improving baseline retrieval performance by using a re-ranker model that has been fine-tuned with human feedback, which helps in retrieving more contextually relevant information.

  • What is the significance of the NFL rules PDF in the provided example?

    -The NFL rules PDF serves as an extensive document corpus for the example, which referees or a chatbot would need to be familiar with to answer queries related to NFL rules effectively.

  • What is the purpose of pre-processing in the context of this script?

    -Pre-processing in this context involves loading documents into a consistent Transformer model to embed them into vector embeddings, which is a foundational step for both experiments and the retrieval process.

  • How does the script handle the limitation of token length in LLMs?

    -The script addresses the token length limitation by selecting only the top two documents as context for the LLM model, which helps in managing the input size while still providing relevant information.

  • What is the process of exporting annotations and ground truth data from Labelbox?

    -The process involves using the export section of the labeling project in Labelbox, where code is provided to export all ground truth data. This data is then formatted into a JSON line format suitable for fine-tuning the re-ranker model.

  • How does the fine-tuning process of the re-ranker model work?

    -The fine-tuning process involves feeding the model with labeled data in JSON line format, which contains positive and negative examples. The model learns to score the relevance of responses, with the goal of improving the ranking of contextually relevant information.

  • What is the final step in the workflow after fine-tuning the re-ranker model?

    -The final step is to replicate the experiment using the same LLM and embedding models but now with the fine-tuned re-ranker model to rank all query-response pairs, aiming to provide improved context for generating responses.

Outlines

00:00

πŸ” Introduction to Labelbox for Retrieval Process Improvement

The video script introduces the use of the Labelbox platform to enhance a retrieval process. It explains the workflow where a custom chatbot uses a retrieval method to search through an internal corpus of documents to find relevant text chunks. These chunks are then used as context for a language model to generate responses. The script outlines the process of using a re-ranker model to improve retrieval performance by incorporating contextual information from documents that may not initially rank high in the search results. The human loop process is introduced as a method to evaluate and fine-tune the re-ranking model using human expertise.

05:00

πŸ“š Setting Up the Labelbox Platform for Human-in-the-Loop Annotation

The script proceeds to demonstrate the setup for a training dataset on the Labelbox platform. It describes the process of extracting top chunks from initial data for human review and the use of the Labelbox SDK to format and submit text queries and response chunks. The video shows how to use the platform's catalog to manage tasks and how to prepare data for human annotators, including the use of model predictions to provide pre-labels and reduce the workload for annotators.

10:01

πŸ€– Utilizing Model Predictions and Annotation for Fine-Tuning

The script explains how to use model predictions to assist human annotators in the labeling process on the Labelbox platform. It details the creation of an annotation project with a specific ontology for relevance assessment of text chunks. The use of pre-labels generated by an AI model to expedite the annotation process is highlighted. The video also covers how to export annotations and ground truth data in JSON line format, which is essential for fine-tuning the re-ranking model.

15:04

πŸ“ˆ Fine-Tuning and Testing the Re-Ranking Model

The final part of the script discusses the fine-tuning process of the re-ranking model using the exported annotations and ground truth data. It illustrates how to select a model, input sample query and response pairs, and utilize the fine-tuned model to rank query-response pairs. The script concludes with a demonstration of how the re-ranking model, when combined with an LLM model, can improve the context provided for generating responses, thereby enhancing the overall retrieval process.

Mindmap

Keywords

πŸ’‘Labelbox

Labelbox is a data labeling platform used for training machine learning models. In the video, it is utilized to improve the retrieval process by fine-tuning a re-ranker model through a human-in-the-loop process. The platform allows for the evaluation of text chunks for relevance to search queries, which is essential for enhancing the accuracy of the retrieval system.

πŸ’‘Retrieval method

The retrieval method refers to the process of searching through a corpus of documents to find the most relevant information in response to a query. In the context of the video, it is the initial step in fetching top chunks of text from an internal document set to provide context for generating responses using a language model.

πŸ’‘LLM (Large Language Model)

LLM stands for Large Language Model, which is an AI model trained on vast amounts of text data to generate human-like responses. In the video, the LLM is used to generate responses based on the context provided by the retrieval method, and its performance is improved by incorporating the results of the re-ranker model.

πŸ’‘Re-ranker model

A re-ranker model is a machine learning model designed to improve the order of search results by re-evaluating their relevance to a query. In the video, the re-ranker model is fine-tuned using human feedback from Labelbox to better identify contextually relevant information from lower-ranked documents.

πŸ’‘Human-in-the-loop

Human-in-the-loop is a concept where human judgment is integrated into the process of data analysis or model training. In the video, this process involves human annotators evaluating text chunks for relevance on the Labelbox platform, which helps to train the re-ranker model to improve retrieval performance.

πŸ’‘Vector embeddings

Vector embeddings are numerical representations of words, phrases, or documents in a multi-dimensional space. In the script, a consistent Transformer model is used to convert documents into vector embeddings, which are essential for the retrieval and ranking processes.

πŸ’‘PDF

PDF, or Portable Document Format, is a file format used to present documents in a manner independent of application software, hardware, and operating systems. In the video, a PDF containing NFL rules is used as the document corpus for the retrieval and ranking processes.

πŸ’‘Chunk size

Chunk size refers to the portion of text or data that is processed as a single unit. In the context of the video, defining an appropriate chunk size is a parameter that can be tuned to improve the efficiency and accuracy of the retrieval process.

πŸ’‘Model Foundry

Model Foundry is a term used in the video to refer to a service or tool that provides pre-trained models for generating predictions. It is used to create pre-labels for the annotators on the Labelbox platform, which helps to streamline the human-in-the-loop process.

πŸ’‘Json Line Format

Json Line Format is a data format where each line of a file is a separate JSON object. In the video, this format is used to structure the data for fine-tuning the re-ranker model, with each line containing information about a query and its associated relevant or irrelevant examples.

πŸ’‘Fine-tuning

Fine-tuning is the process of further training a machine learning model on a specific task, using a dataset that is more specialized than the original training data. In the video, fine-tuning is applied to the re-ranker model using annotations from the human-in-the-loop process to better match the specific retrieval needs.

Highlights

Introduction to the workflow using the Labelbox platform to improve the retrieval process.

Explanation of the custom chatbot and the retrieval method to enhance response generation.

Importance of fine-tuning a reranker model to improve retrieval performance.

Use of a human-in-the-loop process to evaluate and improve the relevance of retrieved text chunks.

Description of the baseline experiment for retrieval using a Transformer model for vector embeddings.

Introduction of an example using an NFL rulebook PDF to demonstrate the retrieval process.

Details on defining chunk size and storing it in a vector database for improved retrieval.

Steps to test the baseline performance of retrieval with no reranking.

Creation of a training dataset for fine-tuning the reranker model.

Explanation of the process to extract top five chunks for each query and submit them to the Labelbox platform.

Overview of the Labelbox platform and the process of importing text queries and response chunks.

Description of the ontology needed for the reranker model, focusing on relevant and non-relevant classifications.

Utilization of the model Foundry to generate pre-labels and assist annotators in the labeling process.

Steps to set up the annotation project in the Labelbox platform.

Details on exporting annotations and ground truth data for fine-tuning the reranker model.

Explanation of fine-tuning the reranker model using a JSON line format and sample queries.

Analysis of the fine-tuned reranker model's performance on sample data.

Replication of the experiment using the same LLM and embedding models with the fine-tuned reranker model.

Comparison of baseline responses with responses improved by the fine-tuned reranker model.

Conclusion highlighting the effectiveness of the human-in-the-loop process in improving retrieval and response quality.

Transcripts

play00:02

hello so let me show you a quick

play00:04

workflow on how you can use the labelbox

play00:07

platform to improve a retrieval process

play00:11

um so typically how it works is the user

play00:13

asks a question it's going to go through

play00:15

this custom

play00:16

chatbot and we're going to use the

play00:18

retrieval method to look through your

play00:20

internal Corpus of documents so that you

play00:23

can retrieve the top few chunks and that

play00:26

chunk is going to be added as context to

play00:29

your llm so that you can generate a

play00:32

response so how we're going to use Label

play00:34

Box is we're going to find tun a re

play00:37

ranker model and a ranker model can help

play00:41

improve your retrieval performance

play00:43

because in many studies it's shown that

play00:47

contextual information might not belong

play00:50

in the top five results of your initial

play00:53

retrieval algorithm it could be stuck in

play00:56

the lower levels of relevance so we're

play01:00

going to use a human Loop process

play01:03

evaluate each chunk of text as it

play01:06

pertains to the search query and then

play01:08

we're going to use that to fine-tune

play01:09

this reranking model so that it can take

play01:13

context from um retrieval documents far

play01:17

below and add that as context to

play01:19

formulate maybe a better response um so

play01:23

let me show you how this works so I'm

play01:25

going to start off

play01:27

by running this pre-processing note book

play01:30

and just show you a baseline experiment

play01:33

for retrieval so I'm just going to load

play01:35

in some documents here and as you can

play01:39

see here I have a consistent Transformer

play01:42

model to embed my documents into Vector

play01:46

embeddings so we're going to keep this

play01:47

consistent for both

play01:50

experiments and for the sake of our

play01:51

example today we're going to upload a

play01:54

PDF containing all of the NFL rules so

play01:58

all the NFL referees probably has to be

play02:01

familiar with this PDF and it's quite

play02:04

extensive as you can see here the rule

play02:07

book consists of uh 245 pages and we're

play02:11

going to formulate our spilling strategy

play02:12

here right so we're going to define a

play02:14

chunk size and in many retrieval use

play02:17

cases right this is another method or

play02:20

another set of parameters that you can

play02:21

tune to improve your

play02:23

retrieval we'll save it as a vector

play02:26

index you can save it into your vector

play02:28

DB as well

play02:30

and we can test this out to see how well

play02:32

it performs right so here's a sample

play02:35

query that might want to ask a chat bot

play02:38

and you can see here that the most

play02:40

relevant

play02:41

result it consists of this chunk of text

play02:44

right

play02:45

here so let's go ahead and evaluate the

play02:47

Baseline performance retrieval with no

play02:50

reranking uh for our NFL documents with

play02:53

this set of embedding model and Baseline

play02:56

llm model so here's our list of uh

play03:00

queries that we want to ask the

play03:02

chatbot and because of token limitations

play03:05

for many LOL models right we're just

play03:07

going to get the top two documents and

play03:10

we're going to use the top two documents

play03:12

as context for our llm model so here's

play03:16

the model name we got it off hugging

play03:19

face and we can use those top two

play03:21

results as context for this model so

play03:26

here's the question sample

play03:27

question and here's the results based

play03:30

off of the

play03:31

context and now we can just run it

play03:33

across all of our list of

play03:36

questions and get a corresponding

play03:38

response based off of the context we'll

play03:41

go ahead and save this to a CSV file uh

play03:44

which we can use later on okay so that's

play03:46

our Baseline experiment let's go ahead

play03:48

and move on to creating the training

play03:50

data set for fine-tuning our ranker

play03:53

model so here let's go ahead and extract

play03:56

the top five chunks for each query

play03:58

within our initial data set and these

play04:01

response chunks are going to be the data

play04:04

that we're going to send over to the

play04:05

labelbox platform for that human the

play04:07

loop review

play04:09

process okay so to send it to Label Box

play04:12

let's go ahead and import the label box

play04:15

SDK let's go ahead and format our pandas

play04:18

data frame a little bit and let's go

play04:20

ahead and create that asset object where

play04:23

we can submit all of our text queries

play04:26

and response chunks into the labelbox

play04:28

platform as a task so there's no errors

play04:33

as you can see here let's head over to

play04:35

the labelbox

play04:37

platform so I'm in the labelbox catalog

play04:41

and as you can see here my job was

play04:42

successful we have

play04:45

354 of our user queries and the top five

play04:49

most relevant chunks all loaded in as

play04:52

text assets within the label box

play04:54

platform so this is going to be that

play04:56

platform where we can do human the loop

play04:58

response

play05:00

so before moving further uh might make

play05:02

sense to show you uh what a ranker model

play05:05

looks like and what types of response it

play05:08

accepts to F tuna model so I'll head

play05:11

over to this coh here

play05:13

website and you can see here we need to

play05:16

upload our data in a Json L format Json

play05:19

line

play05:20

format and this is very relevant to our

play05:23

onology right so for each chunk we're

play05:26

going to have to flag it as relevant or

play05:30

not relevant so that's all we need the

play05:32

ontology is fairly simple two options

play05:35

for each chunk of

play05:37

text so here I'm back in my labelbox

play05:40

platform and before I create that

play05:42

annotation project with the ontology

play05:44

that I just learned let's actually use

play05:47

model Foundry to take a first pass we

play05:49

know that your expert

play05:51

annotators uh are trained for efficiency

play05:54

and accuracy but let's take some weight

play05:56

off their shoulders by generating a set

play05:58

of pre- labels that they can rely on so

play06:01

we'll select all the data within this

play06:03

data set and we'll click on the predict

play06:06

with Foundry option and as you can see

play06:08

here we give you different options for

play06:11

passing your data through initial set of

play06:14

foundation models so you can choose any

play06:16

of the foundation models here so that I

play06:19

can generate that pre-label for you

play06:21

based off of your

play06:23

prompt so if that doesn't make sense let

play06:25

me show you how this works so here I

play06:28

have a open AIG gp4 model that I have

play06:33

created as you can see here I uploaded

play06:36

the ontology right two responses for

play06:38

each

play06:39

chunk and for each chunk here right I

play06:42

have developed the prompt to tell us

play06:46

whether the chunk is relevant to the

play06:47

query or not or if it's negative if it

play06:50

doesn't contain information that need to

play06:52

answer a specific

play06:54

query so by formulating this prompt with

play06:58

the ontology I can now generate some

play07:00

previews from my open AI

play07:03

model after a few seconds the previews

play07:06

has been generated right so now for each

play07:08

query and chunk pairs uh we can now get

play07:12

pre- labels for how open AI determines

play07:15

whether each chunk is not relevant or

play07:19

relevant to this specific question that

play07:22

the user may be

play07:23

asking so if I'm satisfied with this

play07:26

response I can now submit the job and

play07:29

apply these classifications throughout

play07:31

my entire data set so now my annotators

play07:34

has some pre- labels to work

play07:36

with once the model job is finished I

play07:40

can now click on each row within my data

play07:42

set back in the

play07:43

catalog and determine and check out the

play07:47

predictions right so now I have pre-

play07:50

labels for every chunk within each row

play07:54

within my data

play07:55

set and once I send this over to my

play07:58

annotators they're going to save a lot

play08:00

of time so let me show you how to set up

play08:02

that annotation

play08:04

project so let's set that up so we'll

play08:06

head over to the annotate component of

play08:09

the labelbox

play08:10

platform and let's go ahead and create a

play08:12

new project we're working with text so

play08:15

we'll select the text modality we'll

play08:18

give this a name maybe demo with today's

play08:22

date and let's go ahead and confirm the

play08:25

project so now all we need to do to

play08:28

finalize the configuration

play08:30

is to add the data and set up the

play08:32

ontology and labeling

play08:33

experience let's set the ontology up

play08:35

first so we'll head over to

play08:38

settings set edit the standard labeling

play08:41

editor and as you can see here we can

play08:44

choose the ontology that we want to use

play08:46

in our case we want the retrieval

play08:48

ontology

play08:50

here um this is going to give you a

play08:52

preview of what that onology looks like

play08:54

what are the guard rails as to how your

play08:56

annotators can annotate each data row

play09:00

and as you can see here we have five

play09:03

Global classifications right one for

play09:05

each chunk and based off the guideline

play09:08

that cooh here has provided us earlier

play09:10

right we want two options whether that's

play09:13

relevant or not relevant for each

play09:15

retrieved chunk so this is that human

play09:18

Loop process that we have to find let's

play09:21

save

play09:22

it and all we need to do now is to add

play09:25

the data so we'll head over to catalog

play09:30

we'll choose all the data rows within

play09:32

this data set that we uploaded earlier

play09:34

select all of them it we'll add it to

play09:37

this annotation

play09:39

project we'll choose the project that we

play09:42

have just created we want to include the

play09:44

model

play09:45

predictions from our gp4 model to give

play09:48

our annotators a head

play09:50

start and we could set a

play09:54

priority once all this has been

play09:56

configured let's go ahead and submit

play09:58

this list to data rows to our annotation

play10:01

project as a

play10:04

job so now we can head over back into

play10:07

our annotation

play10:09

project confirm that the data row has

play10:11

indeed been

play10:13

added you can see here we have a data

play10:16

row and now our expert team of

play10:20

annotators that are familiar with the

play10:22

rules of the NFL can start the labeling

play10:25

process and because we added pre-label

play10:28

into our Label Box workflow right you

play10:30

can see here that the results of the gp4

play10:33

model has been automatically populated

play10:36

into the labeling editor so as an

play10:39

annotator this is going to save me a lot

play10:41

of time instead of starting from scratch

play10:44

I can simply review and change the

play10:46

results of each answer of these pre-

play10:50

labels I can use these hot keys to speed

play10:52

up my labeling task and hit submit to

play10:55

move on to the next data

play10:56

row and once my labeling ation is

play10:59

finished that human Loop process is

play11:01

finished let's go ahead and Export all

play11:04

of our annotations and ground truth data

play11:06

to F tune that model and if you remember

play11:09

earlier right we're going to need it in

play11:11

a Json line

play11:13

format so if we head over to the export

play11:15

section of our labeling project uh it

play11:18

gives us the code that we need to export

play11:20

all of our ground truth data so let me

play11:22

show you how that

play11:25

works so

play11:27

now we can go ahead and paste that chunk

play11:29

of code with our project ID initialize

play11:33

the labelbox

play11:35

client and now we can just go ahead and

play11:37

break down the Json and convert our

play11:40

labels into a data frame format for

play11:43

easier

play11:45

visualization we can also save as a

play11:48

CSV and convert it into a jsonline

play11:51

format that open source libraries or

play11:54

coher uses to fine-tune a reranking

play11:57

model what one line of that Json looks

play12:01

like you got the query your positive

play12:03

examples as a list and your negative

play12:06

examples as a

play12:07

list we'll write this to a Jon online

play12:10

file format and we're ready to start the

play12:13

reranking fine-tuning

play12:16

process so we'll head over to the next

play12:19

notebook and the first step is to select

play12:22

the model so you can use coh here or in

play12:25

this case we're using a open source

play12:26

model that we found on hugging face

play12:30

and from the documentation of this model

play12:31

right we can pass in some sample query

play12:34

and response pairs right so you know two

play12:38

of the same queries and two different

play12:40

answers and it's going to tell us which

play12:43

one is more relevant so the Nega the

play12:46

lower the score the less relevant the

play12:48

response is so that's how this model

play12:52

works and let's go ahead and feed in our

play12:54

Json line format containing our labels

play12:59

to fine-tune this ranker

play13:02

model as you can see here from the log

play13:05

files are across a few epochs our loss

play13:09

starts to

play13:11

decrease and our learning rate starts to

play13:13

decrease as

play13:16

well and now we can test our model on

play13:18

some sample data right so here's a quick

play13:21

query how many points is a touchdown

play13:24

worth under some conditions and here's

play13:27

all the positive and negative examples

play13:30

and from our human Loop

play13:33

process right we can see here that we

play13:36

have provided this data row uh three

play13:40

positive

play13:41

examples and two negative examples so

play13:45

the key thing to note here is what

play13:47

matters is the relative order of the

play13:49

scores not the absolute value so as we

play13:52

can expect there is three positive

play13:55

examples right they all have relatively

play13:57

High values with two negative examples

play14:01

with very low values right this shows us

play14:03

that our model is fine tuned to the

play14:07

examples that we

play14:08

provided so the next step is fairly

play14:11

simple we want to replicate our

play14:13

experiment the same llms the same

play14:16

embedding models to rank all of our

play14:19

query response pairs with this ranker

play14:22

model that we have fin toe okay so we're

play14:26

going to load in our queries again with

play14:29

all of labels and

play14:32

responses and we can go ahead and take a

play14:34

look at the score distributions for

play14:36

Relevant versus non-relevant

play14:38

labels and as you can see here the

play14:41

non-relevant ones are typically negative

play14:44

and the histogram of relevant ones are

play14:46

typically uh positive and

play14:49

higher okay so let's put it all

play14:52

together we'll use the reranking method

play14:55

we'll retrieve the top 20 documents now

play14:58

for each specific query and then we'll

play15:01

use our ftuned reranking model to get

play15:03

the top two results as context for the

play15:06

same llm model that we're going to

play15:09

use okay so we'll initialize that same

play15:12

embedding model from hugging face now

play15:15

we're going to retrieve the top 20

play15:17

responses for each

play15:19

query so now for each query we got a

play15:21

list of 20

play15:23

responses and now we're going to use

play15:25

that fine to ranker model to go ahead

play15:28

and only extract the top

play15:30

two so in this case let's go ahead and

play15:33

bring in our llm model which is the same

play15:36

Baseline model as

play15:38

well and for each

play15:40

query we can now feed in the top context

play15:43

from the results of our fine to ranker

play15:47

model we'll save this as a CSV

play15:50

file and then we can just do a simple

play15:53

inner join to compare the responses for

play15:56

each question

play15:59

the updated response right from our fine

play16:01

tun ranker model in our initial

play16:05

response and just taking a look at

play16:07

initial pass of our

play16:09

responses uh you can see here that some

play16:12

responses are better right you know this

play16:16

ranker method is certainly not a Magic

play16:19

Bullet uh some results are definitely

play16:22

better uh others are a little bit

play16:25

worse uh but as you can see overall this

play16:27

ranker method does work to match the

play16:30

pattern of your data he uses the human

play16:32

Loop process to incorporate real human

play16:35

expertise so that you can't get a better

play16:37

answer such as this one below here right

play16:41

uh how many Feats in bounce to

play16:43

constitute a catch in the NFL right it

play16:45

everyone who watches NFL knows that is

play16:48

two feet not

play16:51

20 so hopefully this video shows you how

play16:54

you can use the label boox platform to

play16:57

incorporate human loot process to

play16:59

fine-tune a ranker model as part of your

play17:01

retrieval process while this is not a

play17:04

Magic Bullet this is definitely a

play17:06

technique which can help improve your

play17:09

customized chat bots in other retrieval

play17:11

processes thanks

Rate This
β˜…
β˜…
β˜…
β˜…
β˜…

5.0 / 5 (0 votes)

Related Tags
Chatbot OptimizationRetrieval ProcessLabelbox PlatformHuman-in-the-LoopRanking ModelDocument AnalysisNFL RulesEmbedding VectorsContextual AIFine-TuningData Annotation