A Practical Introduction to Large Language Models (LLMs)

Shaw Talebi
22 Jul 202314:57

Summary

TLDRIn this data science series, Shah introduces large language models (LLMs), emphasizing their vast parameters and emergent properties like zero-shot learning. He explains the shift from supervised to self-supervised learning, highlighting next word prediction as the core task. Shah outlines three practical levels of LLM use: prompt engineering, model fine-tuning, and building your own LLM. The series aims to make LLMs accessible, with future videos covering APIs, open-source solutions, and practical applications.

Takeaways

  • ๐Ÿ˜€ Shah introduces a new data science series focused on large language models (LLMs) and their practical applications.
  • ๐Ÿ” The series will cover beginner-friendly introductions to LLMs, practical aspects, APIs, open-source solutions, fine-tuning, and building LLMs from scratch.
  • ๐Ÿ—ฃ Large language models, like Chat GPT, are advanced chatbots that can generate human-like responses to queries.
  • ๐Ÿ“ 'Large' in LLM refers to the vast number of model parameters, ranging from tens to hundreds of billions, which define the model's functionality.
  • ๐ŸŒŸ A key qualitative feature of LLMs is 'emergent properties', such as zero-shot learning, which allows models to perform tasks without explicit training for those tasks.
  • ๐Ÿ”„ The shift from supervised learning to self-supervised learning in LLMs has been significant, with self-supervised learning relying on the structure within the data itself.
  • ๐Ÿ”ฎ The core task of LLMs is next word prediction, which they learn through exposure to massive amounts of text data, allowing them to understand word associations and context.
  • ๐Ÿ›  Three levels of working with LLMs are discussed: prompt engineering (using LLMs out of the box), model fine-tuning (adjusting model parameters for specific tasks), and building your own LLM.
  • ๐Ÿ’ป Prompt engineering can be done through user interfaces like Chat GPT or programmatically via APIs and libraries like OpenAI or Hugging Face Transformers.
  • ๐Ÿ”ง Model fine-tuning involves taking a pre-trained LLM and updating its parameters using task-specific examples, often resulting in better performance for specific use cases.
  • ๐Ÿ— For organizations with specific needs, building a custom LLM may be necessary, involving data collection, pre-processing, model training, and deployment.

Q & A

  • What is the main focus of Shah's new data science series?

    -The main focus of Shah's new data science series is to discuss large language models (LLMs) and their practical applications.

  • What is the difference between a large language model and a smaller one?

    -Large language models differ from smaller ones in two main aspects: quantitatively, they have many more model parameters, often tens to hundreds of billions; qualitatively, they exhibit emergent properties like zero-shot learning that smaller models do not.

  • What is zero-shot learning in the context of large language models?

    -Zero-shot learning refers to the capability of a machine learning model to complete a task it was not explicitly trained to do, showcasing an emergent property of large language models.

  • How does self-supervised learning differ from supervised learning in the context of large language models?

    -In self-supervised learning, models are trained on a large corpus of data without manual labeling, using the inherent structure of the data to define labels. This contrasts with supervised learning, which requires manually labeled examples for training.

  • What is the core task that large language models are trained to do?

    -The core task that large language models are trained to do is next word prediction, where they predict the probability distribution of the next word given the previous words.

  • What are the three levels of working with large language models mentioned by Shah?

    -The three levels of working with large language models mentioned by Shah are: 1) Prompt Engineering, 2) Model Fine-tuning, and 3) Building your own Large Language Model.

  • What is meant by prompt engineering in the context of large language models?

    -Prompt engineering refers to using a large language model out of the box, without altering any model parameters, and crafting prompts to elicit desired responses.

  • How does model fine-tuning in large language models work?

    -Model fine-tuning involves adjusting at least one internal model parameter of a pre-trained large language model to optimize its performance for a specific task using task-specific examples.

  • Why might an organization choose to build its own large language model?

    -An organization might choose to build its own large language model for security reasons, to customize training data, or to have full ownership and control over the model for commercial use.

  • What resources does Shah recommend for further exploration of large language models?

    -Shah recommends the blog in Towards Data Science and a GitHub repository for more details, example code, and further exploration of large language models.

Outlines

00:00

๐Ÿš€ Introduction to Large Language Models

Shah introduces a new data science series focused on large language models (LLMs), emphasizing their practical applications. The video aims to provide a beginner-friendly introduction to LLMs, explaining their significance and how they differ from traditional language models. Shah highlights the impressive capabilities of LLMs, such as Chat GPT, which can generate human-like responses. The key distinguishing features of LLMs are their vast number of parameters and emergent properties like zero-shot learning, which allows them to perform tasks without explicit training. The video sets the stage for future discussions on practical aspects, including using APIs, open-source solutions, fine-tuning, and building LLMs from scratch.

05:00

๐Ÿ” Deep Dive into Large Language Models

This paragraph delves deeper into the workings of large language models, contrasting them with traditional machine learning models. It explains the self-supervised learning paradigm used to train LLMs, which involves predicting the next word in a sequence based on the context. The process is described as an auto-regression task, where the model learns to predict the probability distribution of the next word given the previous words. Shah emphasizes the importance of context in language modeling and how a single word change can significantly alter the model's output. The paragraph also outlines the three levels of working with LLMs: prompt engineering, model fine-tuning, and building your own LLM. Each level requires varying degrees of technical expertise and computational resources.

10:00

๐Ÿ›  Practical Applications of Large Language Models

Shah discusses the practical applications of large language models, focusing on three levels of engagement: prompt engineering, model fine-tuning, and building your own LLM. Prompt engineering involves using LLMs without altering their parameters, either through intuitive interfaces like Chat GPT or programmatically via APIs and libraries. Model fine-tuning adjusts the model parameters for specific tasks, building on the pre-trained model's capabilities. Shah mentions techniques like low-rank adaptation and reinforcement learning with human feedback. For organizations with specific needs and security concerns, building a custom LLM might be the best approach. The paragraph concludes with a call to action for viewers to subscribe for more content and engage with the series through comments and suggestions.

Mindmap

Keywords

๐Ÿ’กLarge Language Models (LLMs)

Large Language Models, or LLMs, refer to advanced AI systems that have been trained on vast amounts of text data and can generate human-like text. They are characterized by having billions of parameters, which are the variables that the model learns to adjust during training. In the context of the video, LLMs are the central theme, with the speaker discussing their capabilities, applications, and the distinction between large and smaller language models. An example from the script is the mention of ChatGPT, which is an LLM that can respond to user prompts in a conversational manner.

๐Ÿ’กParameters

In the context of machine learning, parameters are the weights and biases that a model learns during training. These are the numbers that define how the model processes input data to generate an output. The video script emphasizes that large language models have many more parameters than traditional models, which contributes to their advanced capabilities. The speaker mentions that these parameters are in the range of tens to hundreds of billions, highlighting the scale of complexity in LLMs.

๐Ÿ’กZero-Shot Learning

Zero-shot learning is a capability where a machine learning model can perform a task without being explicitly trained for it. This is considered an emergent property of large language models, meaning it is a feature that arises from the scale and complexity of the model rather than being directly programmed. The script uses zero-shot learning as an example to illustrate the advanced capabilities of LLMs, suggesting that these models can understand and respond to prompts in ways that smaller models cannot.

๐Ÿ’กSelf-Supervised Learning

Self-supervised learning is a machine learning paradigm where models learn from input data without requiring manual labeling. In the video, the speaker contrasts this with supervised learning, where each training example is manually labeled. Large language models are often trained using self-supervised learning, such as predicting the next word in a sequence, which allows them to learn from the structure of the data itself without the need for human-generated labels.

๐Ÿ’กPrompt Engineering

Prompt engineering involves crafting input prompts to elicit desired responses from a language model. It is described in the video as the most accessible way to use LLMs, where users can interact with the model without needing to alter its underlying parameters. The script mentions using platforms like ChatGPT or APIs to provide prompts and receive model-generated text, which can be useful for a variety of applications without needing deep technical expertise.

๐Ÿ’กFine-Tuning

Fine-tuning refers to the process of adjusting a pre-trained model to perform better on a specific task. In the video, the speaker explains that fine-tuning involves updating the parameters of a large language model using task-specific examples. This is a step beyond simply using a model out of the box and can result in improved performance for particular use cases. The script mentions that models like ChatGPT are actually fine-tuned versions of larger pre-trained models.

๐Ÿ’กAutoregression

Autoregression is a statistical method where a model predicts a value based on its own previous values. In the context of LLMs, it is used to describe the task of predicting the next word in a sequence given the previous words. The video script provides an example of autoregression with the notation P(W_n | W_1^(n-1)), which represents the probability of the nth word given the previous n-1 words. This is a fundamental task that LLMs are trained on, and it underlies their ability to generate coherent text.

๐Ÿ’กEmergent Properties

Emergent properties are characteristics that arise from complex systems that are not present in the individual components of the system. In the video, the speaker discusses how certain capabilities, like zero-shot learning, emerge in large language models due to their size and complexity. These properties are not explicitly programmed but appear as a result of the model's training and the vast amount of data it has been exposed to.

๐Ÿ’กChatGPT

ChatGPT is an example of a large language model that is designed to interact with users in a conversational manner. It is mentioned in the script as a tool that can be used to demonstrate the capabilities of LLMs. Users can ask ChatGPT questions, and it generates responses based on its training data. ChatGPT is an instance of an LLM being used for practical applications, showcasing the technology's potential for natural language understanding and generation.

๐Ÿ’กHugging Face Transformers

The Hugging Face Transformers library is an open-source library that provides access to pre-trained large language models and tools for working with them. In the video, the speaker mentions this library as an alternative to using proprietary APIs like OpenAI's, allowing users to work with LLMs locally and without the need for an internet connection. It is an example of how the technology behind LLMs can be made accessible to developers and researchers.

Highlights

Introduction to a new data science series focusing on large language models (LLMs).

Description of three levels of working with LLMs: prompt engineering, model fine-tuning, and building a custom LLM.

Definition of a large language model and its distinguishing properties.

Quantitative aspect of LLMs: the vast number of model parameters.

Qualitative aspect of LLMs: emergent properties like zero-shot learning.

Comparison between supervised learning and self-supervised learning in LLMs.

Explanation of next word prediction as the core task of LLMs.

Importance of context in language modeling demonstrated through example.

Introduction to prompt engineering as the most accessible way to use LLMs.

Discussion on using LLMs through user interfaces like chat GPT.

Mention of programmatic access to LLMs via APIs and open-source libraries.

Overview of model fine-tuning as a way to customize LLMs for specific tasks.

Explanation of the process of fine-tuning a pre-trained LLM for a specific use case.

Introduction to the concept of building a custom LLM for organizations with specific needs.

Discussion on the potential of LLMs for various applications and the importance of understanding the technology.

Invitation to subscribe for future videos in the series and to engage with the content through comments and suggestions.

Transcripts

play00:00

everyone I'm Shah and I'm back with a

play00:02

new data science Series in this new

play00:04

series I'm going to be talking about

play00:06

large language models and how to use

play00:08

them in practice

play00:13

in this video I will give a beginner

play00:15

friendly introduction to large language

play00:18

models and describe three levels of

play00:21

working with them in practice future

play00:23

videos in this series will discuss

play00:25

various practical aspects of large

play00:27

language models things like using open

play00:29

ai's python API using open source

play00:32

Solutions like the hugging face

play00:34

Transformers Library how to fine-tune

play00:36

large language models and of course how

play00:39

to build a large language model from

play00:40

scratch if you enjoyed this content

play00:42

please be sure to like subscribe and

play00:45

share with others and if you have any

play00:46

suggestions for me to include in this

play00:49

series please share those in the

play00:51

comments section below and so with that

play00:52

let's get into the video so to kick off

play00:54

the video series in this video I'm going

play00:57

to be giving a practical introduction to

play00:59

large language models and this is meant

play01:01

to be very beginner friendly and high

play01:03

level and I'll leave more technical

play01:05

details and example code for future

play01:08

videos and blogs in this series so a

play01:11

natural place to start is what is a

play01:14

large language model or llm for short so

play01:18

I'm sure most people are familiar with

play01:20

chat GPT however if you are enlightened

play01:23

enough to not keep up with new cycles

play01:25

and Tech hype and all this kind of stuff

play01:27

chat GPT is essentially a very

play01:30

impressive and advanced chat bot so if

play01:34

you go to the chat GPT website you can

play01:36

ask it questions like what's a large

play01:39

language model and it will generate a

play01:41

response very quickly like the one that

play01:45

we are seeing here and that is really

play01:47

impressive like if you were ever on AOL

play01:50

Instant Messenger also called aim you

play01:53

know back in early 2000s or in the early

play01:56

days of the internet there were chat

play01:58

Bots then there have been chat Bots for

play02:00

a long time but this one feels different

play02:02

like the text is very impressive and it

play02:04

almost feels human-like a question you

play02:07

might have when you hear the term large

play02:08

language model is what makes it large

play02:11

what's the difference between a large

play02:12

language model and a not large language

play02:15

model and this was exactly the question

play02:16

I had when I first heard the term and so

play02:18

one way we can put it is that large

play02:20

language models are a special type of

play02:23

language model but what makes them so

play02:25

special and I'm sure there's a lot that

play02:26

can be said about large language models

play02:28

but to keep things simple I'm going to

play02:29

talk about two distinguishing properties

play02:31

the first quantitative and the second

play02:33

qualitative so first quantitatively

play02:36

large language models are large they

play02:38

have many many more model parameters

play02:41

than past language models and so these

play02:43

days this is anywhere from tens to

play02:45

hundreds of billions of parameters the

play02:48

model parameters are numbers that Define

play02:51

how the model will take an input and

play02:54

generate the output so it's essentially

play02:57

the numbers that Define the model itself

play02:59

okay so that's a quantitative

play03:01

perspective of what distinguishes large

play03:03

language models from not large language

play03:05

models but there's also this qualitative

play03:07

perspective and these so-called emergent

play03:10

properties that start to show up when

play03:12

Lang language models become large and so

play03:15

emergent properties is the language used

play03:17

in this paper cited below a survey of

play03:19

large language models available in the

play03:20

archive really great beginner's guide I

play03:22

recommend it but essentially what this

play03:24

term means is there are properties in

play03:27

large language models that do not appear

play03:30

in smaller language models and so one

play03:33

example of this is zero shot learning

play03:35

one definition of zero shot learning is

play03:38

the capability of a machine learning

play03:40

model to complete a task it was not

play03:43

explicitly trained to do so while this

play03:46

may not sound super impressive to us

play03:48

very smart and sophisticated humans this

play03:51

is actually a major innovation in how

play03:53

these state-of-the-art machine learning

play03:55

models are developed so to see this we

play03:57

can compare the old state-of-the-art

play03:59

Paradigm to this new state-of-the-art

play04:01

paradigm the old way and not too long

play04:04

ago we can say like about five ten years

play04:06

ago the way the high performing best

play04:09

machine learning models were developed

play04:10

was strictly through supervised learning

play04:13

what this would typically look like is

play04:15

you would train a model on thousands if

play04:18

not millions of labeled examples and so

play04:21

what this might have looked like is you

play04:23

have some input text like hello Ola

play04:26

how's it going nastabian so on and so

play04:28

forth and you take all these examples

play04:31

and you manually assign a label to each

play04:35

example here we're labeling the language

play04:37

so English Spanish so on and so you can

play04:40

imagine that this would take a

play04:41

tremendous amount of human effort to get

play04:45

thousands if not millions of high

play04:47

quality examples so let's compare this

play04:49

to the more recent Innovation with large

play04:51

language models who use a different

play04:52

Paradigm they use so-called

play04:54

self-supervised learning so what that

play04:56

looks like in the context of large

play04:57

language models is you train a very

play05:00

large model on a very large Corpus of

play05:03

data and so what this can look like is

play05:06

if you're trying to build a model that

play05:08

can do language classification instead

play05:10

of painstakingly generating this labeled

play05:13

data set you can just take a corpus of

play05:16

English text and a corpus of Spanish

play05:18

text and train a model in a

play05:21

self-supervised way so in contrast to

play05:23

supervised learning self-supervised

play05:25

learning does not require manual

play05:27

labeling of each example in your data

play05:30

set the so-called labels or targets for

play05:33

the model are actually defined from the

play05:36

inherent structure of the data or in

play05:38

this context of the text so you might be

play05:40

thinking to yourself how does this

play05:41

self-supervised learning actually work

play05:43

and so one of the most popular ways that

play05:45

this is done is the next word prediction

play05:48

Paradigm so suppose we have this text

play05:50

listen to your and we want to predict

play05:53

what the next word would be but clearly

play05:56

there's not just one word that can go

play05:58

after the string of words there are

play06:00

actually many words you can put after

play06:02

this text and it would make sense in

play06:04

this next word prediction Paradigm what

play06:06

the language model is trying to do is to

play06:09

predict the probability distribution of

play06:12

the neck next word given the previous

play06:14

words what this might look like is

play06:16

listen to your heart might be the most

play06:19

probable next word but another likely

play06:21

word could be gut or listen to your body

play06:23

or listen to your parents and listen to

play06:25

your grandma and so this is essentially

play06:27

the core task that these large language

play06:29

models are trained to do and the way the

play06:32

large language model will learn these

play06:34

probabilities is that it'll see so many

play06:36

examples in this massive Corpus of text

play06:40

that is trained on and it has a massive

play06:43

number of internal parameters so it can

play06:45

efficiently represent all the different

play06:48

statistical associations with different

play06:50

words and an important Point here is

play06:52

that context matters if we simply added

play06:55

the word don't to the front of this

play06:56

string here and it changed it to don't

play06:59

listen to your then this probability

play07:01

distribution could look entirely

play07:03

different because just by adding one

play07:05

word before this sentence we completely

play07:07

change the meaning of the sentence and

play07:09

so to put this a bit more mathematically

play07:11

and I promise this is the most technical

play07:13

thing in this video this is an example

play07:16

of a auto regression task so Auto

play07:18

meaning self regression meaning you're

play07:21

trying to predict something so what this

play07:22

notation means is what is the

play07:24

probability of the nth text or more

play07:27

technically the nth token given the

play07:30

preceding M token so n minus 1 and minus

play07:33

two and minus three and so on and so

play07:34

forth and so if you really want to boil

play07:36

everything down this is the core task

play07:39

most large language models are doing and

play07:42

somehow through this very simple task of

play07:45

predict the next word we get this

play07:47

incredible performance from tools like

play07:48

chat GPT and other large language models

play07:51

so now with that Foundation said

play07:53

hopefully you have a decent

play07:54

understanding of what large language

play07:56

models are and how they work and a

play07:57

broader context for them now let's talk

play07:59

about how we can use these in practice

play08:02

here I will talk about three levels in

play08:05

which we can use large language models

play08:07

these three levels are ordered by the

play08:09

technical expertise and computational

play08:11

resources required the most accessible

play08:13

way to use large language models is

play08:15

prompt engineering next we have model

play08:17

fine tuning and then finally we have

play08:20

build your own large language model so

play08:21

starting from level one prompt

play08:23

engineering here I have a pretty broad

play08:25

definition of prompt engineering here I

play08:27

Define it as just using an llm out of

play08:30

the box so more specifically not

play08:32

touching any of the model parameters so

play08:35

of these tens of billions or hundreds of

play08:37

billions of parameters that Define the

play08:39

model we're not going to touch any of

play08:40

them we're just going to leave them as

play08:42

is here I'll talk about two ways we can

play08:43

do this one is the easy way and I'm sure

play08:46

is the way that most people in the world

play08:48

have interacted with large language

play08:50

models which is using things like chat

play08:52

GPT these are like intuitive user

play08:55

interfaces they don't require any code

play08:57

and they're completely free anyone can

play08:59

just go to the Chad gbt website type in

play09:01

a prompt and it'll spit out a response

play09:03

so while this is definitely the easiest

play09:05

way to do it it is a bit restrictive in

play09:08

that you have to go to their website

play09:09

this doesn't really scale well if you're

play09:11

trying to build a product or service

play09:13

around it but for a lot of use cases

play09:15

this is actually super helpful so for

play09:17

applications where the easy way doesn't

play09:19

cut it there is the less easy way which

play09:21

is using things like the open AI API or

play09:24

the hugging phase Transformers library

play09:26

and these tools provide ways to interact

play09:29

with large language models

play09:30

programmatically so essentially using

play09:32

python in The Case of the openai API

play09:35

instead of typing your request in the

play09:37

chat GPT user interface you can send it

play09:40

over to openai using Python and their

play09:42

API and then you will get a response

play09:45

back of course their API is not free so

play09:48

you have to pay per API call another way

play09:50

we can do this is via open source

play09:52

Solutions one of which is the hugging

play09:54

phase Transformers Library which gives

play09:55

you easy access to open source large

play09:58

language models so it's free and you can

play10:00

run these models locally so no need to

play10:03

send your potentially proprietary or

play10:05

confidential information to a third

play10:07

party and open AI so future videos of

play10:09

the series we'll dive into all these

play10:11

different aspects I'll talk about the

play10:13

openai API what it is how it works share

play10:16

example code I'll dive into the hugging

play10:18

face Transformers Library same situation

play10:20

what the heck is it how does it work and

play10:22

then sharing some python example code

play10:24

there I'll also do a video talking about

play10:26

prompt engineering more generally how

play10:27

can we create prompts to get good

play10:29

responses from large language models and

play10:31

so while prompt engineering is the most

play10:33

accessible way to work with large

play10:35

language models just working with a

play10:37

model out of the box may give you

play10:39

sub-optimal performance on a specific

play10:42

task or use case or the model has really

play10:45

good performance but it's massive it has

play10:47

like a hundred billion parameters so

play10:48

question might be is there a way we can

play10:50

use a smaller model but kind of tweak it

play10:52

in a way to have good performance on our

play10:55

very narrow and specific use case and so

play10:57

this brings us to level two which is

play10:59

model fine tuning which here I Define as

play11:02

adjusting at least one internal model

play11:05

parameter for a particular task and so

play11:08

here there are just generally two steps

play11:10

one you get a pre-trained large language

play11:12

model maybe from open AI or maybe an

play11:15

open source model from the hugging phase

play11:17

Transformers library and then you update

play11:19

the model parameters given task specific

play11:22

examples kind of going back to the

play11:24

supervised learning versus

play11:25

self-supervised learning the pre-trained

play11:27

model is going to be a self-supervised

play11:29

model so it will be trained on this

play11:31

simple word prediction task but in step

play11:34

two here's where we're going to do

play11:36

supervised learning or even

play11:37

reinforcement learning to tweak the

play11:40

model parameters for a specific use case

play11:42

and so this turns out to work very well

play11:45

models like Chachi BT you're not working

play11:47

with the raw pre-trained model the model

play11:50

that you are interacting with in chat

play11:51

GPT is actually a fine-tuned model

play11:53

developed using reinforcement learning

play11:55

and so a reason why this might work is

play11:57

that in doing this self-supervised task

play12:00

and doing the word prediction the base

play12:03

model this pre-trained large language

play12:04

model is learning useful representations

play12:07

for a wide variety of tasks so in a

play12:09

future video I will dive in more deeply

play12:11

into fine tuning techniques popular one

play12:14

is low rank adaptation or low raw for

play12:17

short and then another popular one is

play12:19

reinforcement learning with human

play12:21

feedback or rlhf and of course there is

play12:23

a third step here you'll deploy your

play12:25

fine-tuned large language model to do

play12:27

some kind of service or you know use it

play12:29

in your day-to-day life and you'll

play12:30

profit somehow and so my sense is

play12:33

between prompt engineering and model

play12:35

fine tuning you can probably handle 99

play12:39

of large language model use cases and

play12:42

applications however if you're a large

play12:44

organization large Enterprise and

play12:46

security is a big concern so you don't

play12:48

want to use open source models or you

play12:50

don't want to send data to a third party

play12:52

via an API and maybe you want your large

play12:55

language model to be very good at a

play12:59

relatively specific set of tasks you

play13:01

want to customize the training data in a

play13:03

very specific way and you want to own

play13:05

all the rides have it for commercial use

play13:07

all this kind of stuff then it can make

play13:08

sense to go one step further Beyond

play13:11

monofine tuning and build your own large

play13:14

language model and so here I Define it

play13:16

as just coming up with all the model

play13:17

parameters so I'll just talk about how

play13:20

to do this at a very high level here and

play13:22

I'll leave technical details for a

play13:23

future video in the series first we need

play13:25

to get our data and so what this might

play13:27

look like is you'll get a book Corpus a

play13:30

Wikipedia Corpus and a python Corpus and

play13:32

so this is billions of tokens of text

play13:34

and then you will take that and

play13:36

pre-process it refine it into your

play13:39

training data set and then you can take

play13:41

the training data set and do the model

play13:43

training through self-supervised

play13:45

learning and then out of that comes the

play13:47

pre-trained large language model so you

play13:49

can take this as your starting point for

play13:51

level two and go from there and so if

play13:53

you enjoyed this video and you want to

play13:54

read more be sure to check out the blog

play13:57

in towards data science there I share

play13:59

some more details that I may have missed

play14:01

in this video this series is both a

play14:04

video and blog Series so each video will

play14:07

have an Associated blog and there will

play14:09

also be tons of example code on the

play14:11

GitHub repository story The goal of the

play14:13

series is to really just make

play14:15

information about large language models

play14:16

much more accessible I really do think

play14:18

this is the technological innovation of

play14:21

our time and there's so many

play14:22

opportunities for potential use cases

play14:25

applications products services that can

play14:28

come out of large language models and

play14:30

that's something that I want to support

play14:31

I think we'll be better off if more

play14:33

people understand this technology and

play14:35

are applying it to solving problems so

play14:37

with that be sure to hit the Subscribe

play14:39

button to keep up with future videos in

play14:41

the series if you have any questions or

play14:43

suggestions for other topics I should

play14:45

cover in this series please drop those

play14:47

in the comments section below and as

play14:49

always thank you so much for your time

play14:50

and thanks for watching

Rate This
โ˜…
โ˜…
โ˜…
โ˜…
โ˜…

5.0 / 5 (0 votes)

Related Tags
Large Language ModelsData ScienceMachine LearningChat GPTPrompt EngineeringModel Fine-TuningSelf-Supervised LearningZero-Shot LearningHugging FaceOpenAI API