How RAG Turns AI Chatbots Into Something Practical

bycloud
24 Aug 202410:17

Summary

TLDRThe video explains the concept of Retrieval-Augmented Generation (RAG), a method that improves AI performance by retrieving accurate information from external documents instead of relying solely on pre-trained neural networks. It outlines the stages of RAG, including indexing, retrieval, and generation, and discusses its practical applications in tools like ThinkBuddy. RAG boosts AI's accuracy and cost-effectiveness without needing expensive fine-tuning. While considered a short-term workaround, its evolving complexity positions it as a significant field in AI research. The video also highlights tools and resources to implement RAG effectively.

Takeaways

  • 🤖 Current AI chatbots are powerful but can be impractical in work settings due to hallucinations and inconsistencies.
  • 📈 A short-term workaround called Retrieval-Augmented Generation (RAG) significantly improves chatbot accuracy and performance.
  • 📚 RAG retrieves accurate information from external documents, allowing more cost-effective and reliable outputs without expensive AI training.
  • 🧩 RAG is broken down into three stages: indexing documents, retrieving relevant data, and generating responses.
  • 🧠 Vector databases and transformer models are used in RAG to encode and retrieve the most semantically relevant information.
  • 🔍 Graph RAG is a promising new method that organizes information into knowledge graphs for better explainability and traceability.
  • ⚙️ Hybrid search and input query rewriting help refine the retrieval process by improving accuracy and reducing irrelevant results.
  • 💡 RAG pipelines can include re-ranking models and autocut functions to prioritize the most relevant responses and prevent hallucinations.
  • 🚀 Think Buddy, an AI tool for Mac OS, leverages RAG principles to improve workflow with local storage and access to multiple models.
  • 🛠️ Think Buddy offers deep integrations for screen capture, PDF processing, and code analysis, with voice input powered by Whisper.

Q & A

  • What is the main limitation of current AI chatbots according to the transcript?

    -The main limitation of current AI chatbots is that they can hallucinate or generate incorrect information, making them unreliable for consistent use in professional settings.

  • What is RAG and how does it help mitigate the limitations of current AI models?

    -RAG stands for Retrieval Augmented Generation. It improves AI performance by retrieving accurate information from external sources, such as a collection of documents, instead of relying solely on the AI's neural network. This reduces hallucination and enhances the usability of AI in tasks requiring specific and accurate data.

  • What are the three main stages of a naive RAG pipeline?

    -The three main stages of a naive RAG pipeline are: 1) Indexing, where documents are divided into chunks and stored in a searchable vector database; 2) Retrieval, where relevant information is retrieved based on a query; and 3) Generation, where the AI generates a response using the retrieved content.

  • Why is RAG considered a 'short-term solution'?

    -RAG is considered a short-term solution because it introduces additional moving parts like indexing, retrieval, and blending of data, which can introduce points of failure. It is seen as a workaround to the limitations of current AI models, rather than a long-term architectural solution.

  • What are some challenges associated with the RAG pipeline?

    -Some challenges of the RAG pipeline include managing multiple components like document indexing, retrieval accuracy, and the AI's ability to blend and generate relevant responses. Any issue in these components can lead to poor output quality.

  • What is a 'knowledge graph' and how does it improve RAG?

    -A knowledge graph is a structured representation of entities, relationships, and key claims extracted from documents. In RAG, it helps organize data more effectively, making retrieval more accurate and context-aware, and it improves the traceability and auditability of the AI's responses.

  • What role does 'reranking' play in the RAG pipeline?

    -Reranking involves retrieving multiple top results and then passing them through a model to determine which results are the most relevant. This ensures that the most contextually appropriate response is used, reducing the chance of inaccurate results.

  • How can RAG help prevent AI hallucination?

    -RAG can prevent hallucination by using a reranking model and auto-cut functions to remove unrelated retrieved results. If the AI cannot find a relevant match, it can be forced to admit that it does not know the answer, instead of generating an incorrect response.

  • What are some tools and libraries mentioned for building RAG systems?

    -The transcript mentions several tools and libraries for building RAG systems, including Llama Index for general frameworks, Hugging Face for embedding models, Microsoft's GitHub for graph RAG implementation, and RAG-assess for evaluating RAG pipelines.

  • What is Think Buddy, and how does it enhance productivity for developers?

    -Think Buddy is a macOS AI lab designed for developers. It combines strengths from multiple AI models like GPT-4, Claude, and Gemini Pro to generate responses by remixing the best parts of each. It integrates deeply with macOS, supports hotkeys for instant AI help, and handles various file formats like PDF and docx, making it a powerful tool for enhancing workflow.

Outlines

00:00

🤖 Limitations and Solutions for Current AI Chatbots

The current generation of AI chatbots excels in many areas but struggles with practicality in consistent work environments due to their tendency to hallucinate incorrect information. As companies take time to develop more advanced AI, a temporary solution called RAG (Retrieval-Augmented Generation) has emerged. RAG enhances chatbot performance by retrieving accurate information from uncompressed documents rather than relying solely on pre-trained data. This reduces costs and improves accuracy without requiring expensive fine-tuning. Additionally, the browsing function in some chatbots is an extension of RAG, useful when large amounts of reference data exceed the context window of language models (LLMs).

05:02

🔍 The Process of RAG: Indexing, Retrieval, and Generation

RAG operates in three key stages: indexing, retrieval, and generation. During the indexing phase, documents are divided into chunks and stored in a vector database. The retrieval stage involves analyzing user queries to pull relevant information from this database using semantic matching, often with a BERT model to find the most meaningful data. The final generation stage leverages this retrieved content to generate coherent, contextually appropriate responses. However, RAG’s complexity introduces points of failure, and it remains a short-term solution to overcome limitations in current AI systems while more refined architectures are developed.

10:05

📊 The Evolving Meta of RAG

RAG's methodology continues to evolve with new variations and techniques. One approach involves using trainable embedding models to improve document retrieval by encoding them into vectors, making semantic retrieval more precise. For example, in coding contexts, models fine-tuned for coding text can better understand code structure, like indentation. A promising development is 'Graph RAG,' which uses knowledge graphs to extract and organize key relationships and claims, making results more traceable and explainable. This structured approach enhances the accuracy and contextual relevance of retrievals, preventing the generation of irrelevant responses.

🧠 Refining Queries and Hybrid Search Techniques

To ensure accurate information retrieval, the input query is also encoded for search. However, irrelevant parts like greetings or notations should be excluded. Query rewriting models help streamline queries to focus on key information. Hybrid search techniques, combining nearest neighbor searches with word frequency analysis, improve retrieval accuracy. In some cases, web search integration ensures that time-sensitive information remains current. APIs can be integrated at this stage, providing flexibility for additional data sources.

🔄 Reranking and Improving Output Quality

RAG has introduced methods like reranking, where multiple results are retrieved and then ranked for relevance. Domain-specific fine-tuning can further improve accuracy. Additionally, methods like autocut remove irrelevant results, and reranking models set thresholds to prevent irrelevant or incorrect data from being returned. This approach helps prevent the hallucination problem common in LLMs, ensuring higher quality and more reliable responses.

🔧 Tools and Resources to Build RAG Systems

For those looking to implement RAG, there are several tools and models available. Llama Index and Llama Pars are popular frameworks for organizing documents. Many fine-tuned embedding models can be found on Hugging Face, and Cohere provides models optimized for RAG. Microsoft's GitHub has code for Graph RAG, while Llama Index also offers implementations. Additionally, tools like RAG-as-a-service help evaluate and optimize RAG pipelines. These resources enable developers to create tailored RAG systems for specific needs.

💡 ThinkBuddy: A Productivity-Boosting AI Lab for Developers

ThinkBuddy is a powerful AI tool designed for developers, offering seamless integration with macOS and access to over 10 leading models, such as GPT-4 and Claude. Its unique 'AI Remix' feature combines the strengths of different models to produce more accurate responses. ThinkBuddy supports various file formats (PDF, DOCX, XLSX), includes screen capture capabilities, and offers voice input powered by Whisper, supporting over 80 languages. Its local data storage ensures privacy, and it's working towards integrating local models for even faster, secure processing. ThinkBuddy offers both a free basic tier and a discounted lifetime deal for users.

🙏 Gratitude and Further Engagement

The speaker extends thanks to their supporters on Patreon and YouTube and encourages viewers to follow them on Twitter for more updates. They also invite viewers to subscribe to their newsletter for insights on the latest AI research papers, providing an additional way to stay informed about cutting-edge developments in the field.

Mindmap

Keywords

💡RAG (Retrieval Augmented Generation)

RAG refers to a method where an AI model retrieves information from an external set of documents or data, then uses this information to generate responses. This approach enhances the accuracy and practicality of AI systems by addressing one of their main issues, hallucination (the creation of false information). RAG retrieves accurate data from stored sources, making it more reliable for tasks like writing emails or responding to specific queries, which is a key theme in the video.

💡Indexing

Indexing in the context of RAG is the process of organizing documents into searchable chunks and converting them into vectors that the AI can efficiently retrieve. This step is foundational in the RAG system because it allows the AI to quickly access relevant data without repeatedly going through the entire document set. The video highlights how indexing is a crucial first step for making AI practical in retrieving accurate information.

💡Vector Database

A vector database stores documents in a mathematical form, known as vectors, allowing AI systems to search and retrieve data based on semantic relevance rather than simple word matching. In the video, the importance of vector databases is underscored in enabling more accurate and efficient retrieval of information. This database makes it easier for AI models to pull the most contextually relevant information when generating responses.

💡Semantic Similarity

Semantic similarity is the process of measuring how closely related two pieces of information are in meaning, rather than just word matching. The video emphasizes the use of semantic similarity in RAG to ensure that the AI retrieves information that not only contains relevant words but is also contextually meaningful to the user’s query. It is achieved through models like BERT (Bidirectional Encoder Representations from Transformers), which are mentioned as a solution for better accuracy.

💡Embedding Model

An embedding model converts text into vectors for the AI to understand and compare. The video discusses embedding models, particularly those that can be fine-tuned for specific tasks (like coding or natural language). For example, a specialized embedding model for coding would prioritize code syntax and indentation when converting text, improving the AI's ability to retrieve meaningful information in technical contexts.

💡Knowledge Graph

A knowledge graph is a structured representation of knowledge, showing relationships between entities, such as people, places, or concepts. The video introduces 'Graph RAG,' where a knowledge graph is used to organize and visualize information more clearly. This approach helps the AI avoid irrelevant answers by making the retrieval process more traceable and auditable, improving the overall reliability of the generated content.

💡Re-ranking

Re-ranking is the process of sorting multiple retrieved results by relevance before generating a final response. In the video, re-ranking is mentioned as a way to pass the top results through a model to determine which ones are most relevant. This technique prevents the AI from using irrelevant or low-quality information, thus improving the quality of responses, especially in domain-specific tasks.

💡Hybrid Search

Hybrid search combines different search techniques, such as nearest neighbor search and word frequency matching, to increase the accuracy of information retrieval. The video discusses hybrid search as a method used within the RAG framework to improve the retrieval process by balancing speed and relevance. This technique is particularly useful in ensuring that the AI finds the most relevant information, especially when dealing with large datasets.

💡Hallucination

Hallucination refers to the problem where AI models generate false or misleading information. This is a key issue the video addresses, as hallucination makes AI unreliable in professional or practical settings. The introduction of RAG, by retrieving accurate information from documents, aims to reduce the frequency of hallucinations, making AI more practical for everyday use, like writing emails or generating reports.

💡Fine-tuning

Fine-tuning refers to the process of adjusting a pre-trained AI model for a specific task or dataset to improve its performance. In the video, fine-tuning is mentioned several times, especially in the context of embedding models and re-ranking systems. By fine-tuning, the AI becomes more specialized and accurate in handling specific types of data, such as coding or domain-specific language, which is crucial for making RAG more effective.

Highlights

RAG (Retrieval-Augmented Generation) drastically improves the performance and usability of LLMs by retrieving accurate information from uncompressed documents instead of relying solely on the LLM's neural network.

RAG serves as a short-term solution to overcome LLM hallucinations by introducing document retrieval mechanisms that are cost-effective and accurate without needing to fine-tune the model.

RAG's process involves three key stages: indexing, retrieval, and generation, which together enhance LLM performance using external documents.

Indexing is the first stage in RAG, where documents are divided into meaningful chunks and stored as vectors in a database for easy retrieval.

In the retrieval stage, a bi-encoder model is used to capture semantic similarities between the input query and the indexed documents by measuring vector distances.

The final generation stage uses the LLM to combine the input and retrieved content to produce coherent, contextually relevant responses.

RAG can fail due to several variables, including how information is indexed, retrieved, and integrated, which introduces instability in the pipeline.

Graph RAG is an emerging method that uses knowledge graphs to organize and retrieve information, offering better traceability and explainability compared to traditional vector databases.

Hybrid retrieval techniques, like combining nearest-neighbor search with word frequency matching, can improve the accuracy of retrieved information.

Query rewriting with an LLM is used in RAG to remove irrelevant tokens like greetings, ensuring the input query is focused on key information before retrieval.

Reranking retrieved results using a domain-specific model helps ensure the most relevant documents are used in response generation.

Autocut is a technique to remove unrelated retrieved results based on similarity thresholds, preventing LLM hallucinations and forcing it to admit when it doesn't know the answer.

LlamaIndex is a popular RAG framework for general use, offering tools for organizing documents and retrieving information efficiently.

Graph RAG is supported by Microsoft, with open-source implementations available on GitHub for further experimentation and application.

ThinkBuddy is an AI-powered tool for MacOS that integrates with RAG, providing access to multiple LLMs and allowing users to combine strengths from different models into one response.

Transcripts

play00:00

the current AI chap Bots are good at

play00:01

everything except for being practical

play00:04

and it's really frustrating that it

play00:05

cannot be easily utilized consistently

play00:08

in work settings because it'll just

play00:09

hallucinate the most out- of pocket

play00:11

thing you can imagine and since we don't

play00:13

have the patience to wait for the Mega

play00:15

corpse to train an even more powerful AI

play00:17

we need ways to utilize the chat Bots we

play00:19

have right now to get ahead of the curve

play00:21

before AGI replaces us to write email

play00:24

SLO so there's this short-term

play00:25

workaround called rag which stands for

play00:27

retrieval augmented generation in all of

play00:30

the latest research and even in services

play00:31

like CH gbt or claw rag has drastically

play00:34

improved the performance and usability

play00:36

for these L because instead of spawning

play00:39

the key information in a neural network

play00:41

that has compressed the data with rag it

play00:43

retrieves accurate information from a

play00:45

collection of uncompressed documents

play00:46

separately stored that even the LM may

play00:49

not have been trained on this method

play00:51

would provide results that are both cost

play00:53

effective and accurate without needing

play00:54

to train or fine-tune the L that would

play00:56

have cost tens of thousands and remember

play00:59

the function we browsing that a lot of

play01:00

chat Bots have this function is also an

play01:02

extension of rag which makes rag perfect

play01:05

when you need to use a large amount of

play01:06

documents as reference that cannot fit

play01:08

within the context window of an llm so

play01:10

with these insane benefits why would it

play01:12

be a short-term solution well before we

play01:14

can answer that we first need to

play01:16

understand how it generally Works let's

play01:18

take a naive rag as an example oh yeah

play01:20

keep in mind the field is still a bit

play01:22

new so the process can be differentiated

play01:23

a bit differently from place to place

play01:25

but I would decompose them into three

play01:28

main stages index ret retrieval and

play01:30

generation in the index stage you will

play01:32

be indexing your documents for the AI to

play01:34

easily retrieve it later this index

play01:36

process can vary but the most common way

play01:38

is to divide the documents into

play01:40

meaningful chunks and store it in a

play01:42

vector form that can be searched easily

play01:44

this stage usually doesn't repeat once

play01:46

all your documents has been indexed and

play01:48

stored within a vector database this

play01:50

then brings us to the retrieval stage

play01:52

where we will need to retrieve the

play01:53

information for the LM to use to know

play01:55

what information the LM needs to

play01:57

retrieve we need to First Look at the

play01:59

user input to see what the query is

play02:00

about so we can bring out the most

play02:02

relevant data from the vector database

play02:04

for the LM to work with while there is

play02:06

the classic word frequency matching

play02:08

between the input query and the

play02:10

documents to retrieve the ideal

play02:12

information it still can capture the

play02:14

semantic information between the words

play02:16

so a bir model that is an encoder only

play02:18

Transformer is used to encode and

play02:20

provide measurements for semantic

play02:22

similarities between the documents and

play02:24

the input query so by measuring their

play02:25

Vector distances we can find the most

play02:27

semantic relevant information from the

play02:29

doc doent and provide it to the llm for

play02:31

further processing which brings us to

play02:33

the last stage the generation stage this

play02:35

stage relies on D llm to utilize the

play02:37

retrieve content and the input content

play02:39

to formulate coherent and contextually

play02:41

relevant responses it needs to strike a

play02:44

balance between following the

play02:45

information on the reference document

play02:47

and transforming into a response that

play02:48

answers the input query so without any

play02:50

fine-tuning the llm would be able to

play02:53

respond to questions about your

play02:54

documents with Rag and now you would

play02:56

realize that with that many components

play02:59

rag could have have a single point of

play03:00

failure there are now various moving

play03:02

Parts like how you index information how

play03:04

you retrieve them and how good the model

play03:06

is at blending and presenting the output

play03:08

that have the power to affect the

play03:10

quality of rag so it's kind of

play03:12

reasonable to call it a short-term

play03:14

solution because it is certainly a hacky

play03:16

way to bypass L's limitations by

play03:18

introducing more unstable variables but

play03:20

being hacky about it kind of transforms

play03:23

rack into a whole new field of research

play03:25

with a more applied mentality rather

play03:27

than a theoretical one as a better AR

play03:29

chitectural model would be more of a

play03:31

long-term solution so naturally this

play03:33

simple pipeline has evolved into

play03:35

something even more complex which brings

play03:37

us to the million dooll question what is

play03:39

the current meta for rag okay there are

play03:42

way too many variants right now and

play03:44

sometimes it boils down to what works

play03:46

the best with your own data but here's

play03:47

how The Meta roughly looks like starting

play03:49

from the indexing stage other than

play03:51

chunking the document semantically using

play03:53

LM to better organize the information

play03:56

retrieved a trainable embedding model

play03:58

which is to convert text to vect vectors

play04:00

can be used to better connect the input

play04:02

query and the relevant documents when

play04:03

storing and comparing them within a

play04:05

vector database so for things like

play04:07

indentation where it holds a more

play04:09

significant meaning in coding compared

play04:10

to your typical writing an embedding

play04:12

model that is fine-tuned on coding would

play04:15

be much more mindful about this detail

play04:17

when converting the text into Vector

play04:19

encoding so later on when the AI is

play04:21

retrieving the influence of indentations

play04:23

is much more respected another really

play04:25

new and promising method is something

play04:27

called graph rag this technique uses a

play04:29

knowledge know graph and utilizes L to

play04:31

extract entities relationships and key

play04:34

claims from your document then the

play04:35

hierarchical clustering with the lien

play04:37

technique is used to organize the graph

play04:39

which can be easily visualized for

play04:41

better traceability and is much more

play04:43

explainable and auditable than looking

play04:45

at a vector database which brings us to

play04:47

the retrieval stage where the model now

play04:49

can not only retrieve the most relevant

play04:51

information with the input query but

play04:54

also obtain the context of the retrieved

play04:56

content thanks to the knowledge graph by

play04:58

using the structured data previous ly

play05:00

generated this makes mistakes much more

play05:01

traceable and preventable as answers

play05:04

that might be contextually irrelevant to

play05:05

the input query can be ignored but for

play05:07

the model to retrieve information

play05:09

accurately the previously mentioned

play05:11

embedding model would need to encode the

play05:13

input query too right however not all

play05:16

the input query is needed for search

play05:18

like the greetings in the inputs the

play05:20

next line notation or end of sentence

play05:22

token should be completely left out so

play05:24

an input query rewriting LM would be

play05:26

here to help to condense or even

play05:28

transform the query to its key

play05:29

information that is then encoded into

play05:32

Vector form to be compared and searched

play05:34

within the vector database or the

play05:35

knowledge graph to retrieve more

play05:37

accurate information additionally a

play05:39

hybrid search can be used in the

play05:41

meantime like the FIS SS nearest

play05:43

neighbor plus word frequency to increase

play05:45

the chance of getting the desired

play05:47

retrieval optionally web search can be

play05:49

done at this stage too which is really

play05:51

useful to ensure any time sensitive

play05:53

information or citations are correct and

play05:55

this is also the part where you can

play05:57

literally insert any apis then in the

play05:59

final stage which is Generation

play06:01

something called reranking is often used

play06:04

now where instead of retrieving only

play06:05

once you would instead retrieve topk

play06:08

results in the retrieving stage and pass

play06:10

the topk results into a reranking model

play06:13

to see which results are actually the

play06:15

most relevant that would be able to

play06:17

enter the input query and the rerank

play06:19

model can also be fine-tuned to be

play06:21

domain specific another function code

play06:23

autocut would also be used to remove

play06:25

unrelated retrieved results based on

play06:27

similarity distance gaps and sometimes

play06:30

the content relevant score measured by

play06:32

the rerank model would have a threshold

play06:34

in place so if the retrieved information

play06:36

is not as relevant it'll Force the model

play06:38

to say they don't know anything about

play06:40

the input query instead of hallucinating

play06:43

or providing bad results so yeah that's

play06:45

roughly the current mattera of rag but

play06:47

since I've only been talking about the

play06:49

conceptual ideas here are some relevant

play06:51

sources you can use to build your own

play06:53

rag for a more General rag framework

play06:56

llama index is a more popular one it

play06:58

also has Library like llama pars which

play07:01

is good for organizing your documents

play07:02

for retrievals for the embedding models

play07:05

there are a lot of fine-tuned ones on

play07:06

hugging face which are free download so

play07:08

pick one at your own cost for rack model

play07:12

commend R models from cooh here are some

play07:14

of the best rack optimized models and

play07:16

they also offer some really easy to ous

play07:18

rerank and embedding models but of

play07:20

course they're not free for graph rag

play07:22

you can check out Microsoft's official

play07:24

GitHub and yoink their codes from there

play07:26

and I think llama index also has an

play07:28

implementation so you can check that out

play07:30

you can also check out rag ass which is

play07:32

a framework that helps you evaluate your

play07:34

rag pipelines so on the topic of rag

play07:36

locally I want to talk about an

play07:37

incredible application that can boost

play07:39

your workflow that is think buddy it's

play07:41

not just another AI app it's a

play07:43

full-fledged Mac OS AI lab for 10x

play07:46

deaths like you guys what sets think

play07:48

Buddy apart as its LM remix feature

play07:50

imagine combining the strengths of gbd4

play07:52

CLA Sona 3.5 and Gemini Pro in one

play07:55

response by collecting the best parts of

play07:57

each answer and that's exactly what

play07:59

think Buddy does you get access to 10

play08:01

plus leading models without any extra

play08:03

subscriptions but the Deep Mac OS

play08:06

integration is the actual game changer

play08:08

you can capture your screen and ask

play08:09

questions to AI use customizable hot

play08:12

keys to utilize prompt templates and

play08:14

basically get AI to help you instantly

play08:16

anytime for example assume that I have a

play08:18

hotkey set up to analyze code just

play08:20

select the text use the shortcut and

play08:22

think Buddy provide suggestions on how

play08:24

to improve it the voice input powered by

play08:26

opening eyes whisper lets you dictate

play08:28

while multitask it even supports 880

play08:30

plus languages with great accuracy and

play08:33

they even fix whisper output by using

play08:34

gp4s Advan reasoning and for youth

play08:37

developers and researchers think Buddy

play08:38

handles PDF docx xlsx and many other

play08:42

file types you can ask questions to LM

play08:44

by sourcing documents and you can see

play08:46

responses of each model and a remix

play08:47

answers in under 10 seconds and don't

play08:49

worry about privacy all Chad storage is

play08:52

local plus they're working on

play08:53

integrating local LM soon which makes

play08:55

processing even faster and more secure

play08:57

now here's the deal that you don't want

play08:59

to miss think Buddy is offering an

play09:00

exclusive lifetime deal but it's only

play09:02

available for August and it's closing

play09:04

very soon the regular lifetime deal sale

play09:06

is over but you can use the code bcloud

play09:08

with the link down in the description to

play09:09

get 30% off bringing the price down to

play09:11

just 130 bucks which is basically around

play09:13

the same price if you have chbt and

play09:15

Claud for over 3 months and this code is

play09:17

only limited to the first 50 sales and

play09:19

if you're still on the fans they offer a

play09:21

30-day no questions asked refund policy

play09:23

and think Buddy also has a free basic

play09:25

tier for you to try out and their basic

play09:27

version is only 10 bucks a month still

play09:29

cheaper than subscribing to multiple AI

play09:31

Services separately I've been also

play09:33

trying out thank buddy for a while now

play09:34

and so far I found it extremely fitting

play09:36

for my workflow so if you want to

play09:37

experience this AI Powerhouse check them

play09:39

out using the link down in the

play09:40

description and thank you think buddy. a

play09:42

for sponsoring this video if you like

play09:43

what I talked about today you should

play09:45

definitely check out my newsletter on

play09:47

there I will be breaking down the latest

play09:48

hottest research papers coming out left

play09:50

and right for you on a weekly basis so

play09:53

even if I am late to the news or don't

play09:54

have the chance to talk about it in a

play09:56

video you would 100% cat the most juicy

play09:58

stuff on there but anyways a big shout

play10:01

out to andul lelas chrisad do Alex J

play10:05

Alex Marice migam Dean Fel robbers aasa

play10:12

and many others that support me through

play10:13

patreon or YouTube follow my Twitter if

play10:15

you having and I'll see you all in the

play10:16

next one

Rate This

5.0 / 5 (0 votes)

相关标签
AI toolsRAG technologyworkflow boostAI optimizationdocument retrievalvector searchLLM efficiencyknowledge graphhybrid searchtech innovation
您是否需要英文摘要?