Conversation w/ Victoria Albrecht (Springbok.ai) - How To Build Your Own Internal ChatGPT

Techsylvania
3 Apr 202429:44

Summary

TLDRVictoria Albrecht, co-founder of Springbok AI, discusses the complexities and considerations of building and utilizing large language models (LLMs) in business. She emphasizes the high costs and expertise required for developing proprietary LLMs, suggesting that for most companies, fine-tuning existing models or using prompt architecture may be more practical. Albrecht outlines a framework for decision-making regarding LLM integration, highlighting the importance of aligning AI strategy with business goals rather than blindly following tech giants.

Takeaways

  • ๐Ÿ‘‹ Victoria Albrecht, co-founder and managing director of Springbok AI, leads a team of engineers and a commercial team on various projects focusing on AI development.
  • ๐ŸŒŸ Victoria's previous experience includes scaling a food tech business and working with Rasa, a conversational AI framework, which aligns with her session's focus on building internal chat GPTs.
  • ๐Ÿค– The session aims to provide insights on whether building your own large language model (LLM) is the right approach for a company or if there are alternative solutions.
  • ๐Ÿš€ Building an LLM is resource-intensive, with costs reaching into billions of dollars and requiring significant expertise and time investment, making it a strategy suitable primarily for tech giants aiming for global domination.
  • ๐Ÿ”ฎ A common misconception is that building an LLM will automatically lead to a 'money printing machine,' but the reality is much more complex and costly.
  • ๐Ÿ›  For companies not seeking global dominance or lacking the resources to build an LLM, fine-tuning an existing LLM can be a viable option, provided they have a substantial dataset and the need for on-premise hosting.
  • ๐Ÿ”’ Fine-tuning an LLM can offer domain-specific information, data security, and compliance, but it also comes with challenges such as the black-box nature of the model and the high costs associated with retraining.
  • ๐Ÿ’ผ Many companies express a desire for a bespoke AI solution to automate and streamline processes, improve decision-making, and maintain data security, but often they do not require building or fine-tuning an LLM to achieve this.
  • ๐Ÿข Prompt architecture is introduced as a scalable and controllable method for leveraging LLMs, allowing for high data security and low risk without the need for developing or fine-tuning an LLM.
  • ๐Ÿ”‘ Prompt architecture involves context-based text enhancement and response accuracy checks, providing a software layer for control and steerability, which can be adapted based on company needs and data.
  • ๐ŸŒ Springbok AI has developed an enterprise platform utilizing prompt architecture, enabling clients to upload documents and query them effectively, showcasing the practical application of LLMs in business processes.

Q & A

  • Who is Victoria Albrecht and what is her role at Springbok AI?

    -Victoria Albrecht is the co-founder and managing director of Springbok AI, where she leads a team of 40 engineers and a commercial team on multiple projects, driving the further development of the business.

  • What was the topic of Victoria's session at the event?

    -Victoria's session was about sharing insights on how to build your own internal chat GPT, which is a conversational AI framework.

  • What is the significance of the story about Andreas and Vlad in Victoria's talk?

    -The story about Andreas and Vlad illustrates the early support and trust that helped Victoria and Springbok AI grow, highlighting the importance of networking and connections in the tech industry.

  • Why did Victoria share her experiences in Japan during her presentation?

    -Victoria shared her experiences in Japan to emphasize the global reach and recognition of AI technologies, showing that the interest in AI extends beyond tech circles and has a widespread impact.

  • What is the main challenge companies face when considering building their own large language model (LLM)?

    -The main challenge is the significant investment in resources, expertise, and time required to develop an LLM, which may not be justifiable for companies that do not aim for global domination in the tech industry.

  • Why did the executive of a major toy company consider developing their own LLM?

    -The executive considered developing their own LLM as part of their AI strategy, possibly influenced by the actions of big tech companies, without necessarily considering whether it was the most suitable path for their business.

  • What is the cost implication of developing a large language model like Chat GPT?

    -Developing a large language model like Chat GPT involves a substantial financial investment, with OpenAI alone having spent roughly two billion dollars on its development.

  • What is the alternative to building or fine-tuning an LLM for companies that do not have the resources or need for such extensive AI models?

    -The alternative is prompt architecture, which allows companies to leverage existing LLMs like Chat GPT through a software layer that provides high control, data security, and low risk.

  • How does prompt architecture help companies utilize LLMs for their specific needs?

    -Prompt architecture enables companies to input contextual information and instructions to tailor the LLM's responses to their specific requirements, enhancing control and ensuring data security.

  • What are some examples of use cases where companies might not need their own LLM but can benefit from prompt architecture?

    -Examples include sales leaders wanting to automate the generation of sales contracts, HR departments providing 24/7 access to company policies, and investors querying internal databases for startup information.

Outlines

00:00

๐ŸŽค Welcoming Victoria Albrecht to Techsylvania

Victoria Albrecht, co-founder and managing director of Springbok AI, is introduced as the next speaker at Techsylvania. With a team of 40 engineers and a commercial team, she is leading multiple projects and business development. Her past experience includes scaling a food tech business and working with Rasa, a conversational AI framework. Victoria will share insights on building an internal chat GPT, a topic relevant to many attendees. She also shares a personal story about her previous engagements at Techsylvania and her appreciation for the connections she has made there, including with Vlad and Andreas, who have supported her journey in the AI industry.

05:01

๐ŸŒ The Global Reach of AI and Building Your Own Language Model

Victoria discusses the global impact of AI, sharing anecdotes from her travels in Japan where she found that even in remote areas, people were aware of AI advancements. She then transitions into the topic of building one's own large language model (LLM), noting that while it's a popular idea, it doesn't make sense for every company. She emphasizes the massive investment and expertise required to develop an LLM, citing the example of OpenAI's investment of around two billion dollars and the complexity of training such models. Victoria also highlights the importance of considering whether building an LLM aligns with a company's goals and resources.

10:01

๐Ÿค– The Risks and Considerations of Developing a Large Language Model

In this section, Victoria outlines the risks and considerations involved in developing a large language model. She mentions the need for significant resources, expertise, and the ability to set up state-of-the-art architecture. Victoria also points out that building an LLM is a massive undertaking, with OpenAI's model having 175 billion parameters and 96 layers, which took months to develop with uncertain outcomes. She advises that unless a company is aiming for global domination or has specific needs, it may not be worth the investment and risk.

15:03

๐Ÿ”„ The Alternative to Building an LLM: Fine-Tuning

Victoria introduces the concept of fine-tuning an existing large language model as an alternative to building one from scratch. She explains that fine-tuning can be beneficial for companies that do not seek global domination but want to dominate their specific space. The process involves using a pre-trained LLM and training a small portion of it with the company's data. However, she also points out the challenges, such as the need for a large dataset, the black-box nature of the model, and the high costs involved.

20:06

๐Ÿ“š Leveraging Documents with Prompt Architecture

The speaker explores the use of prompt architecture as a solution for companies that want to leverage their existing documents without needing to build or fine-tune an LLM. This approach involves using an LLM to generate responses based on contextual information and specific instructions. Victoria explains that prompt architecture can provide high control, data security, and low risk, making it a scalable solution for software development based on LLMs. She also discusses the process of context-based text enhancement and response accuracy checks to ensure the generated responses are appropriate and accurate.

25:07

๐Ÿ› ๏ธ Transforming Business with Large Language Models

In the final paragraph, Victoria summarizes the key points of her presentation and invites the audience to consider how they will use large language models to transform their businesses. She emphasizes the importance of understanding the options available, such as building an LLM, fine-tuning, or using prompt architecture, and making informed decisions based on a company's specific needs and resources. Victoria also expresses her hope to have provided clarity on the topic of large language models and invites questions from the audience.

Mindmap

Keywords

๐Ÿ’กSpringbok AI

Springbok AI is the company co-founded and managed by Victoria Albrecht, the speaker in the video. It is a team of engineers and a commercial team working on multiple projects, focusing on the development of AI technologies. The company's role is central to the video's theme of discussing AI strategies and the building of internal chat GPTs.

๐Ÿ’กConversational AI

Conversational AI refers to artificial intelligence systems that can interact with humans through natural language processing. In the video, Victoria Albrecht discusses her previous experience with 'why food' and 'Rasa', a conversational AI framework, which is directly related to the session's focus on building internal chat GPTs.

๐Ÿ’กLarge Language Model (LLM)

A Large Language Model is an AI system that can understand and generate human-like text based on vast amounts of data. The video discusses the considerations and challenges of building one's own LLM versus using existing models, which is a central topic in the presentation.

๐Ÿ’กFine-tuning

Fine-tuning in the context of AI refers to the process of retraining a model on a specific dataset to adapt to a particular task or domain. The video explores the option of fine-tuning an existing LLM to suit a company's specific needs instead of building a new model from scratch.

๐Ÿ’กData Security

Data security is a critical aspect of AI deployment, especially when dealing with sensitive or proprietary information. The video mentions the desire for a bespoke AI that keeps data secure, highlighting the importance of data protection in AI strategy.

๐Ÿ’กOn-Premise

On-premise refers to software or services that are hosted and operated within a company's own infrastructure. The video discusses the need for on-premise solutions for companies with strict data handling policies, which relates to the decision-making process for AI deployment.

๐Ÿ’กCloud Hosted Services

Cloud hosted services are applications or platforms delivered over the internet and hosted on remote servers. The video touches on the use of cloud services as an alternative to on-premise solutions for AI deployment, offering flexibility and scalability.

๐Ÿ’กPrompt Architecture

Prompt architecture is a method of interacting with AI models by providing them with specific instructions or 'prompts' to guide their responses. The video presents prompt architecture as a solution for companies seeking control and high data security without building their own LLM.

๐Ÿ’กReinforcement Learning

Reinforcement learning is a type of machine learning where an agent learns to make decisions by performing actions in an environment to maximize a reward. The video mentions the integration of reinforcement learning from human feedback as a feature that makes certain AI models impressive.

๐Ÿ’กTech Slovenia

Tech Slovenia is an event where the video's speaker, Victoria Albrecht, has been invited to share her insights. It serves as a backdrop for the discussion on AI strategies and is a platform for networking and learning about the latest in technology.

๐Ÿ’กChat GPT

Chat GPT is a specific instance of a conversational AI developed by OpenAI. The video discusses building an 'internal chat GPT', indicating the desire to create customized AI solutions for internal company use, which is a key point in the discussion about AI deployment.

Highlights

Victoria Albrecht, co-founder and managing director of Springbok AI, discusses the development and application of AI in business strategies.

Albrecht shares her experience scaling a food tech business and her involvement with Rasa, a conversational AI framework.

The importance of building your own internal chat GPT and how it can be a challenge many businesses are grappling with.

Albrecht's story about Techsylvania and the growth of AI, highlighting the global impact and adoption of AI technologies.

The observation that executives often suggest building their own large language model (LLM), despite the complexity and cost.

The reality of developing an LLM, with OpenAI spending around two billion dollars and a decade of research.

Factors to consider when deciding to build your own LLM, such as resources, expertise, and the potential for global domination.

The challenges of fine-tuning an LLM, including the need for a large dataset and the black-box nature of the process.

The high costs and uncertain outcomes associated with fine-tuning an LLM, potentially reaching seven hundred thousand dollars.

The alternative to building or fine-tuning an LLM: prompt architecture, offering high control and data security.

Prompt architecture as the future of scalable and LLM-based software, with examples of its application in various industries.

How prompt architecture allows for full control and steerability of AI responses, enhancing accuracy and security.

The practicality of prompt architecture for companies looking to leverage their existing documents and streamline processes.

Examples of user stories that illustrate the potential of prompt architecture in sales, HR, and investment sectors.

Springbok AI's development of an Enterprise platform utilizing prompt architecture for document uploading and querying.

Albrecht's call to action for businesses to consider how they will use large language models to transform their operations.

The conclusion emphasizing the importance of understanding the options available when it comes to integrating AI into business strategies.

Transcripts

play00:00

[Music]

play00:02

foreign

play00:08

we have our next speaker so please all

play00:12

join me in welcoming Victoria Albrecht

play00:14

the co-founder and managing director of

play00:16

Springbok AI where she leads a team of

play00:19

40 engineers and Commercial team on

play00:23

multiple projects and is driving the

play00:26

further development of that that

play00:27

business her previous experience

play00:30

includes scaling a food Tech business

play00:31

why food and eating sales at rasa a

play00:34

conversational AI framework which is

play00:36

exactly in line with the session that

play00:39

we're going to have today where she'll

play00:40

share her insights and how to build your

play00:43

own internal chat GPT which I think many

play00:47

of us have been grappling with so I'm

play00:49

sure that all of you are going to learn

play00:50

a lot so please join me in giving a warm

play00:52

welcome to Victoria

play00:55

[Music]

play01:02

thank you

play01:08

while the text being figured out um I'll

play01:10

tell you guys a a story about tech

play01:13

Sylvania because this is not my first

play01:16

time speaking at texylvania or one of

play01:18

vlad's events in fact it's my third time

play01:20

speaking here

play01:22

um and there's a very nice man I don't

play01:24

know if he's in the audience already or

play01:27

not his name is Andreas and a few years

play01:29

ago when Springbok was just

play01:33

you know nine months old

play01:35

um Andreas decided to introduce me to

play01:37

Vlad and said hey I met this great

play01:39

founder you should speak to her you know

play01:41

she might be cool for your events

play01:43

and

play01:45

um you know back five five-ish years ago

play01:48

AI wasn't as big as it is now and we

play01:51

were one of the only three AI

play01:53

consultancies in all of London

play01:55

um and we had like very little

play01:57

reputation to go off of as a company but

play02:00

Vlad was actually one of the people the

play02:02

early people who believed in in me and

play02:04

believed in us and sort of took a punt

play02:07

and said you know what why don't you

play02:08

come come speak we'll see how it goes

play02:11

um so that was the first event and then

play02:13

we got introduced to flutter

play02:16

um in here in in plush I'm at the

play02:18

managing director we had some really

play02:20

nice conversations he ended up

play02:22

introducing us to uh flutter in London

play02:24

and we ended up working with flutter for

play02:27

many years

play02:28

so you know if I guess if this is your

play02:30

first time to Tech Slovenia then welcome

play02:35

um I hope that you find great

play02:36

connections as I've found some great

play02:38

connections

play02:39

um and yeah since we have a couple of

play02:41

minutes to to fill just wanted to take a

play02:43

moment to say thank you to Vlad and his

play02:46

team for putting this on and also

play02:48

everybody that's sort of volunteering

play02:50

and you know behind the scenes

play02:52

making a great like Philip as well who

play02:55

I've met several years ago

play02:56

it was uh nice to be back

play02:59

yeah how are we doing

play03:01

oh yeah Round of Applause for black

play03:03

definitely

play03:06

how are we doing on the tech

play03:12

no worries it's a great reminder that

play03:15

nowadays building like almost super

play03:17

human AIS easier than connecting a

play03:19

laptop to a screen

play03:20

absolutely we could have built a whole

play03:22

chapter plug-in in the meantime

play03:26

yeah I was um I was traveling in Japan a

play03:29

few weeks ago in March and I thought it

play03:32

was so incredible like obviously you

play03:34

know we're in our in our Tech bubbles

play03:35

you guys in inclusion wherever you

play03:37

you've come from

play03:39

um and it hadn't really occurred to me

play03:42

that that news had traveled as far as

play03:43

Japan for Chachi BT but I was sitting in

play03:45

a cooking class and this cooking teacher

play03:47

just a local Kyoto guy uh you know asked

play03:50

what do you do for work and I kept it

play03:52

quite high level and I said I work in Ai

play03:54

and he goes oh Chachi PT

play03:57

I was like okay I guess it's made it

play04:00

here okay and then I went really deep

play04:02

into the countryside to a town called uh

play04:05

kinosaki Onsen I don't know does that

play04:07

that ring a bell for anyone in the room

play04:10

no okay it's really a tiny tiny tiny

play04:12

town in Japan and so I was um sitting in

play04:15

the in the Onsen these little hot

play04:17

springs uh next to this lady and again

play04:20

like similar conversation you know what

play04:21

do you do what brings you here

play04:24

um and I said I work in Ai and she goes

play04:26

Chad gbt

play04:28

I just thought wow okay like really

play04:32

um this is taken off news has traveled

play04:34

fast and I guess the the growth Spike

play04:36

the growth curve that you've seen in

play04:38

terms of adoption is not just you know

play04:40

in our own Tech circles um but is is

play04:43

really Global so

play04:45

um so those are some some really great

play04:47

little reminders by how far it's

play04:49

actually spread

play04:51

um and I'll maybe just start with my my

play04:54

presentation already

play04:56

um and we can dial in in a second

play04:58

um and it actually starts with a with

play05:00

another little anecdote which is that um

play05:04

well actually more of an observation so

play05:07

um something that I found really

play05:09

interesting over the past

play05:11

pretty much two months specifically

play05:13

um has been that when I've gone into

play05:16

meetings with Executives and sort of

play05:18

board level to talk about an AI strategy

play05:21

and discuss you know where they want to

play05:22

take the business

play05:23

there was very often at least one person

play05:26

in the room who suggested that it might

play05:29

be a good idea to build their own large

play05:31

language model

play05:32

just a an idea that was floated

play05:35

and while for some companies that's a

play05:37

really understandable Avenue to want to

play05:39

take there's really not very few for

play05:41

whom that actually makes sense and

play05:43

applies

play05:44

um and it was when the executive of um

play05:48

one of the world's biggest toy companies

play05:50

a toy that you've a company whose toys

play05:53

you've definitely played with in your

play05:54

childhood said you know we're also

play05:56

considering developing our own llm that

play05:59

I thought okay this is now Beyond

play06:03

um

play06:04

you know just an idea this has now

play06:06

become a bit of like a buzz theme if

play06:08

that's if that's a word

play06:10

um

play06:10

and I was asking into the into the room

play06:13

with the speakers uh last night we had a

play06:15

small get together and I was asking them

play06:17

um who who of them have heard

play06:19

um you know the the idea of an llm being

play06:22

built for a portfolio company of theirs

play06:24

or in their own company or their own

play06:26

startup building uh their own llm being

play06:28

floated and I was uh pretty Amazed by

play06:31

the response I got so I can't see you

play06:32

all super well because the light's

play06:34

pretty bright from here but um I'd love

play06:37

to hear it see by a show of hands

play06:39

um if you've been privy to a

play06:41

conversation where somebody said let's

play06:43

just build our own llm can I get a small

play06:45

show of hands

play06:49

all right that's probably about eight or

play06:50

nine of you in the room yeah not bad not

play06:54

bad at all yeah so it's it's been it's

play06:56

been a big topic you know a lot of

play06:57

companies have been interested in

play06:58

building their own llm and it to be hon

play07:02

yeah as I said to be honest it doesn't

play07:03

always it doesn't always make sense

play07:05

to do that

play07:07

um but of course we all look up to the

play07:09

likes of um Google and um meta Alibaba

play07:13

you know the big tech company so we look

play07:15

at what are they doing and is there

play07:17

anything that they're doing that I

play07:18

should be doing

play07:20

um so it's easy to think you know okay

play07:22

that's that's their plan

play07:24

um I must imitate but of course

play07:26

um the important thing to remember is

play07:27

those guys are you know cross

play07:30

um industry players

play07:32

um their entire play

play07:33

their entire play is in data right

play07:37

um and they've built their their they've

play07:39

built their entire

play07:41

um their entire systems and their entire

play07:44

product to basically sell the outcomes

play07:46

of the data that they that they process

play07:49

um now for them it makes sense to sort

play07:51

of further grow their their Global World

play07:53

World Domination by developing their own

play07:55

large language model and partially

play07:57

that's because I guess the the sexy

play08:00

Narrative of why this is interesting is

play08:02

right the idea is you build it you build

play08:03

the large language model once and then

play08:05

you have a money printing machine

play08:07

um the reality is is that open AI

play08:10

um you know has been working together

play08:11

with Google in in research

play08:14

for probably about 10 years just open AI

play08:17

alone have spent roughly two billion

play08:20

dollars on developing what is now what

play08:23

you're using um chat2pt

play08:25

so it's really not

play08:27

how are we doing

play08:29

okay I know

play08:31

um so it's really not the easiest feat

play08:33

to just come in and say you know I'm

play08:34

gonna come and compete with you here

play08:38

um but what I'm hoping to

play08:40

um go through with you guys in my talk

play08:42

today is just really a little bit of um

play08:44

of a framework for how to think about

play08:46

those conversations that inevitably will

play08:48

be coming up again and again and will be

play08:49

happening more and more about whether

play08:51

building your own llm is the right

play08:53

approach for you for your company for

play08:55

your portfolio company and if it's not

play08:57

building your own llm then you know what

play08:59

is the alternative solution

play09:02

um and uh one way to start thinking

play09:05

about

play09:06

um you know whether you want your own

play09:07

alarm is on the one hand you know do you

play09:08

want um Global and world uh domination

play09:11

is that is that what you want to achieve

play09:12

but the other big part of it as I

play09:14

mentioned is of course the resource

play09:15

aspect of it you know do you have one to

play09:18

two billion to blow do you have a

play09:20

timeline of two to three years that

play09:22

could set you back because you really

play09:24

want to you know own and grow that thing

play09:27

um do you have the expertise in-house or

play09:29

are you willing to partner with

play09:30

universities are you willing to partner

play09:32

with

play09:33

um Recent research institutions are you

play09:36

willing to set up your own team

play09:37

internally to to run this and do you

play09:41

have the ability to set up

play09:42

state-of-the-art architecture

play09:45

um uh for uh for a large language model

play09:47

and I'm sure all of you that are product

play09:49

Builders here that have seen

play09:51

architecture models you know what they

play09:53

look like you know quite quite cute

play09:54

quite simple

play09:55

quite straightforward but the thing

play09:58

that's important to keep in mind is that

play10:01

an actual architecture for a large

play10:03

language model in open ai's case has 175

play10:06

billion parameters and 96 layers that's

play10:11

not something that you just whip up

play10:12

that's something that you you try you

play10:16

you run you rerun and you do that over

play10:18

months and months and months and you

play10:19

don't actually know if the outcome that

play10:21

you're going to get is what you expect

play10:22

so what I'm saying is that it's a huge

play10:26

risk to try and build your own LM the

play10:28

cost is massive and unless you want

play10:30

world domination or nothing it's

play10:32

probably not worth going for

play10:35

um

play10:36

so that's on the on the topic of

play10:40

excuse me on the topic of building your

play10:41

own llm

play10:48

do you have a clicker

play10:50

of some sort

play10:54

otherwise I just click on my laptop

play10:57

it's fine

play10:59

those are my slides

play11:05

um so this is the framework that I'm

play11:07

gonna I'm gonna walk you through

play11:08

um and it'll all make sense towards the

play11:11

end

play11:12

and here's what we were just talking

play11:13

about regarding world domination

play11:17

um and then we've already talked about

play11:18

what it entails to build your large

play11:20

language model

play11:21

um yeah so on the data side the other

play11:23

thing that I forgot to mention

play11:25

um handy to have slides right

play11:27

um is that you need to also look for uh

play11:30

integrating the reinforcement learning

play11:32

from Human feedback and that's actually

play11:33

something that's made chat topt

play11:35

um really really impressive is it's not

play11:37

just training the data but it's also

play11:38

trained on all the training and

play11:40

reinforcement learning that's been going

play11:42

on that that openai has paid

play11:44

basically some Kenyan labelers for about

play11:47

six months to do full time

play11:49

and they just turn this around in Cycles

play11:53

um

play11:54

my slides are gone again just

play11:56

as a heads up

play12:01

no

play12:04

all right

play12:07

um

play12:08

so where we got to it's pretty much here

play12:11

which is that building your own llm is

play12:13

really only worth it if you're in oil

play12:15

and gas defense a financial institution

play12:18

or an aspiring Tech Giant because the

play12:21

other thing you want to keep in mind is

play12:22

that if your solution if your large

play12:24

language model has to be on premise and

play12:26

you really have a ton of data to train

play12:28

it on and you want to build your own

play12:29

then you know it's also also worth a

play12:32

consideration

play12:34

um

play12:35

so the other thing that might that you

play12:37

might think

play12:39

um in your in your consideration here of

play12:40

whether you know your company or a

play12:42

portfolio company is is cut out for

play12:46

building your own large language model

play12:47

and you get to the point place where

play12:49

you're thinking about you know do we

play12:50

have the resources the expertise do we

play12:51

have the risk appetite to really build

play12:54

our own here and build us from scratch

play12:55

and your answer is no then there is

play12:58

still the option to fine-tune an llm

play13:01

which is what I'm going to go into next

play13:04

thank you um excuse me just wondering

play13:07

would it be possible to have it on there

play13:10

so sorry

play13:11

yeah great thank you this is probably

play13:14

the most chaotic presentation I've ever

play13:16

done thank you for bearing with me

play13:19

um where were we

play13:21

all right so the thing that that um that

play13:24

turns out is that when we actually ask

play13:26

companies you know why do you want your

play13:29

own large language model

play13:30

something that we really often hear is

play13:33

actually this

play13:34

we want to get a bespoke chapter BT for

play13:37

a company that keeps our data secure

play13:38

that is tailored to our information and

play13:41

that supercharges our processes in

play13:42

decision making so basically they're

play13:44

saying we want to take everything that

play13:46

we've learned

play13:47

and then we want to apply it to

play13:49

everything that we're doing in the

play13:49

future to be smarter to make better

play13:51

decisions to be more efficient right

play13:53

that's essentially what they're saying

play13:56

so the question is should you be

play13:57

fine-tuning instead

play14:00

um so here's why fine tuning might make

play14:02

sense for you

play14:04

um you're not hungry for world

play14:05

domination again

play14:07

um you obviously still hungry for

play14:09

dominating your space no one's

play14:11

questioning that

play14:12

um but if you just want to be the

play14:14

biggest the best Healthcare company for

play14:15

example Health tech company then you

play14:17

know you're probably fine

play14:19

um the next question is is everything in

play14:21

your company hosted on premise

play14:24

and if the answer is yes your company is

play14:26

super strict with everything being

play14:27

hosted on premise and you have super

play14:30

huge data sets available then probably

play14:32

yes fine-tuning your own llm is the way

play14:34

to go

play14:35

so what does that look like what is the

play14:37

promised land so the dream on the

play14:39

marketing brochure is basically this

play14:42

you skip the harpit and you fine-tune

play14:44

someone's someone else's work right and

play14:46

in practice if we just look at this

play14:48

diagram what it would look like is

play14:51

um in the case if we if we choose open

play14:52

AI um open ai's model you use the

play14:55

pre-trained existing llm

play14:58

um which is obviously a black box but

play15:00

you use theirs their hard work you

play15:02

manually manually collect the data that

play15:04

you have available

play15:05

in your company and you just train a

play15:08

small retrain a small portion of the llm

play15:10

that sounds easy enough right so what

play15:13

are the suppose benefits or the promise

play15:15

benefits of this well

play15:17

um supposedly domain specific

play15:19

information

play15:21

um

play15:21

you have data security and compliance of

play15:24

course if it's if you can then run it on

play15:26

premise and approved accuracy and

play15:29

relevance because the ideas that you've

play15:30

trained in on your data right

play15:32

well it's not quite as easy

play15:34

unfortunately and I'll start with the

play15:36

with the top right

play15:38

namely the manual collection of the

play15:40

domain specific data set what that

play15:42

basically means is that you need a huge

play15:44

data set in your company to be able to

play15:46

train against or rather alongside

play15:50

model and

play15:54

what was really interesting

play15:55

and I think it was the the hearing in

play15:57

Congress

play15:58

uh with Sam Altman and he said

play16:01

you know even if I wanted to I couldn't

play16:03

really explain how this um how Chachi BT

play16:06

Works

play16:07

um how exactly the large language model

play16:09

is is built what the weights are and why

play16:11

it behaves the way it behaves we're

play16:12

still figuring that out they feed it a

play16:14

bunch of data and then they sort of have

play16:16

to see how that changes the outcome but

play16:18

every time you run it you're running it

play16:20

it takes you about a month

play16:22

and then you need to see what that does

play16:23

to the outcome but if we think about the

play16:26

fact that Chachi PT is built on 45

play16:28

terabytes of data

play16:29

and you don't have 45 terabytes of data

play16:34

but maybe only

play16:35

500 megabytes of data to train it

play16:38

against

play16:38

and I know that doesn't sound like a lot

play16:40

but just for context

play16:44

um uh 45 terabytes of data is 292

play16:48

million documents and that's what Chad

play16:51

gbt or open AI has used to train train

play16:53

their model on

play16:54

so if you're not coming anywhere close

play16:57

to the data set that they have like you

play16:58

really don't have massive amounts of

play17:00

data it'll be really hard to make a

play17:01

difference and even if you're able to to

play17:04

bring a huge data set because this is a

play17:06

black box it's so hard to control what

play17:09

actually happens so you're putting in

play17:11

you're putting in a bunch of you're

play17:12

spending a bunch of money you're

play17:14

retraining and retraining and you don't

play17:15

know why it does the things that it does

play17:18

which is not that great of a proposition

play17:20

right

play17:22

um and yeah I mean retraining the the

play17:24

llm costs a ton of money

play17:27

I think you can probably expect some

play17:29

pretty uncertain technical outcomes and

play17:32

a spend of at least seven hundred

play17:34

thousand uh seven hundred thousand

play17:35

dollars

play17:37

um

play17:38

of course if you do need it need

play17:40

something hosted on premise then this is

play17:42

probably still your best option

play17:45

but if fine tuning isn't the answer then

play17:48

you know what is the answer

play17:51

well let's go through

play17:53

um actually I'm interested uh raise your

play17:56

hand if you're using anything on the

play17:59

cloud

play18:03

raise your hand if you're using a hybrid

play18:04

of on-premise and Cloud

play18:09

okay I'm

play18:11

I'm gonna ask actually another question

play18:13

raise your hand if your company does

play18:15

everything on premise

play18:18

right okay so for my first two questions

play18:20

all of your hands whatever should have

play18:22

gone up right

play18:24

um which just means that that you are

play18:26

part of the the 94 who wouldn't

play18:29

necessarily need to to build their own

play18:31

language model unless we have a new

play18:34

Microsoft or a new uh Apple in in the

play18:37

room

play18:39

um so okay so we go down the decision

play18:41

tree no we can use cloud Hostess service

play18:43

services and we can use apis

play18:46

and the other question is

play18:48

um

play18:49

do you need a high level of control over

play18:51

the model output well we just looked at

play18:54

or we just explored you know the fact

play18:56

that if you have a data set and you

play18:58

train it together with or you you try to

play19:00

fine-tune open ai's data set the guarant

play19:03

the output isn't necessarily guaranteed

play19:04

right

play19:06

um so if the answer to that question is

play19:08

yes as well and you don't have it

play19:09

necessarily have a huge data set

play19:10

available but you have maybe hundreds

play19:12

maybe thousands maybe tens of thousands

play19:14

of documents that you'd like to train it

play19:16

against

play19:17

then probably

play19:19

um there's another solution

play19:21

so again what do most companies want

play19:23

when they say they want an llm well

play19:27

um just for for recap most of them say

play19:29

we want a bespoke to rgbt for a company

play19:32

that keeps our data secure so they want

play19:33

control

play19:34

is tailored to information and helps us

play19:37

make better decisions and be a more

play19:38

efficient company and automate our

play19:40

processes

play19:42

um and then when we drill a little bit

play19:43

deeper we find that

play19:46

um roughly 80 percent of companies

play19:48

actually want this which is to leverage

play19:52

our existing documents to create a

play19:54

process that automates or streamlines

play19:56

querying them

play20:00

way so

play20:05

um

play20:08

I'm just going to walk through a few

play20:09

examples here

play20:11

um and you know maybe think about as I

play20:13

do which of these might sound familiar

play20:15

to you

play20:16

of something that you've thought of or

play20:18

that would be would be helpful

play20:20

to you or to a startup that you work

play20:22

with

play20:23

um so the first one as a sales leader

play20:25

these are user stories I'm sure you guys

play20:27

are all familiar with how this works as

play20:29

a sales leader I want my team to be able

play20:31

to automatically generate sales

play20:32

contracts based on our company best

play20:34

practices

play20:36

so that we don't need to make them from

play20:37

scratch every time and so that I don't

play20:39

need to review all of my sales reps

play20:41

contracts myself

play20:43

right

play20:46

um or as a head of HR I want our

play20:49

employees to have access to rhr 24 7. to

play20:52

able to ask any question based on our

play20:54

company policies so that my team is

play20:56

freed up to deal with more complex or

play20:59

interesting HR topics for example

play21:01

implementing a four-day work week

play21:06

um or as an investor

play21:08

and this is actually based on Jenny from

play21:11

eqt I was speaking to her yesterday on

play21:13

the bus and asked her to give me a use

play21:15

case and she was saying I would love to

play21:18

as an investor I would love to be able

play21:19

to query our internal database startup

play21:22

pitch decks and our notes about startups

play21:24

so that I don't need to give this to my

play21:26

associate to do manually

play21:28

uh when we're looking for the types of

play21:29

startups that we've already looked at

play21:31

the types of trends that are occurring

play21:32

in the industry

play21:33

that we've already observed and so on

play21:36

and so forth

play21:37

and you won't be surprised to hear me

play21:39

say that actually none of these use

play21:41

cases oopsies require you to use your

play21:43

own large language model

play21:45

or to fine-tune one

play21:47

so what does that leave us with well the

play21:50

concept is called

play21:52

prompt architecture

play21:54

so for the 94 of you

play21:56

prompt architecture will likely be the

play21:58

answer

play22:00

and I really see and we at Springbok

play22:03

really see prompt architecture

play22:05

as the future of building scalable and

play22:08

llm-based software with high control

play22:10

High data security and low risk

play22:14

and I'll explain to you what I mean

play22:18

so here's how this works

play22:20

um and this is really more of a

play22:23

of a framework than you know

play22:26

specifically our architecture or

play22:27

anything

play22:29

um so the the user inputs text so for

play22:31

example in this HR case

play22:34

um how many days of annual leave

play22:35

do I get per year

play22:38

next

play22:40

we have context-based text enhancement

play22:42

so you have in this case three types of

play22:44

contextual information that's added

play22:46

further so

play22:48

the user information would be based on

play22:52

potentially your your employee employee

play22:55

database

play22:56

that tells you or that tells the system

play22:58

Jessica's in a probation period please

play23:00

answer all these questions with this in

play23:02

mind then you have the Persona

play23:05

instructions so who is this

play23:08

um this this bot who's the system this

play23:10

this expert system representing so here

play23:13

the prompt is something along the lines

play23:15

of you're an HR expert you must be

play23:17

helpful polite and professional and then

play23:19

we have contextual information so the

play23:22

company policy so this is relying on

play23:24

company internal documents company

play23:26

handbooks Etc maybe your notion

play23:29

um so company policy is that you're only

play23:31

permitted to three days of holiday

play23:32

during your probation period so those

play23:34

are the contextual inputs in this

play23:35

example

play23:38

and then the large language model

play23:39

generates a response that doesn't get

play23:42

sent to the user right away no no

play23:44

we first have the response accuracy

play23:46

check

play23:47

so we check it for uh you check it for

play23:51

offensive language factual correctness

play23:53

turn a voice response response length

play23:55

semantic similarity

play23:58

um things like that just to make sure

play24:00

that it actually reflects you know the

play24:02

information that you've already been

play24:03

been given it

play24:06

um and then your text is enhanced with a

play24:08

request to amend the factual

play24:10

incorrectness if there was anything that

play24:11

was wrong

play24:12

and that is then sent as a message to

play24:14

the user the message of the user is then

play24:16

added to the conversation history and

play24:18

this is what the diagram looks in its

play24:20

entirety

play24:22

now this is really what I see as this is

play24:26

a very simplified version but this is

play24:28

really what I see as the state of the

play24:30

art of the next generation of software

play24:33

that's going to be developed based on

play24:36

large language models with as an example

play24:39

chat GPT

play24:40

as an underlying framework

play24:43

um

play24:44

and what this enables what this

play24:46

architecture enables is for you to have

play24:48

full control using a software layer so

play24:51

you can control exactly what you know in

play24:53

the example that we used chat topt watch

play24:56

does

play24:59

uh it provides accuracy so you can

play25:01

directly provide the information that

play25:03

Chad gbt uses right you can upload your

play25:05

own documents and you know that it's

play25:07

only giving you information based on the

play25:09

information that you have provided it

play25:11

it's not going to somewhere in the

play25:13

internet that you don't know but it's

play25:15

from a trusted source which is your data

play25:17

that's the important bit and the

play25:19

steerability so you can give it very

play25:21

clear instructions for the type of

play25:22

persona it's supposed to adopt the tone

play25:24

of voice you can make that different

play25:25

based on different users there's like a

play25:27

lot of control that you have

play25:30

um and of course depending on how you're

play25:31

building with it you can make that

play25:33

Solution on premise or you can make that

play25:34

solution hosted in the cloud right that

play25:36

flexibility still very much exists

play25:39

um the one thing that you do need to do

play25:41

of course

play25:42

um here is work with the API so in this

play25:45

case the chatgpt or the gpt4 currently

play25:50

3.5 and student four uh to lgbt4 API

play25:54

um

play25:55

so yeah that's what what

play25:58

um what you'll be able to to do with

play26:00

that with that type of architecture and

play26:02

so when we have

play26:04

um you know the the customer is saying

play26:07

we want a bespoke chat to BT for the

play26:08

company that does these things

play26:11

um and we turned it into

play26:13

um or we've we then get to the point

play26:15

where they where they realized that

play26:16

actually what they want is to turn and

play26:18

leverage their existing documents to be

play26:19

able to query them for any topic that

play26:22

they're specifically interested in

play26:24

um and you know like theme specific uh

play26:26

channels is the way that we we kind of

play26:29

think about that

play26:31

um then you know that's something that

play26:33

we've at Springbok we've been thinking

play26:35

about it for for a really long time and

play26:39

um we've actually been in the in the

play26:41

large language model space for I mean we

play26:44

started Spring Walk In 2017. and we

play26:47

started working with large language

play26:48

models about three and a half uh three

play26:51

and a half years ago

play26:52

um and

play26:54

through just a you know a ton of uh

play26:57

customer interviews a ton of customer

play26:58

interactions

play27:00

um we found that this just is what comes

play27:02

up again and again

play27:03

but this is not unique to uh you know

play27:06

the way that we're thinking about

play27:07

product this is the way that products in

play27:09

general will be developing so I hope

play27:11

that the

play27:13

um the framework here that I provided is

play27:16

maybe a little bit of a glimpse into the

play27:17

future of what that might look like for

play27:19

your guys's uh for your as a software

play27:21

development as well

play27:23

and so what we've done

play27:25

um recently we've launched this with a

play27:27

few clients now has actually set this up

play27:29

into into an Enterprise platform that

play27:32

our clients can use so the idea is that

play27:33

you just need to be uploading your

play27:35

documents and you can start start

play27:36

querying it and you have a sort of all

play27:38

the Enterprise Suite features

play27:41

um that you would like

play27:44

um cool so what I hope what I hope that

play27:47

I've been able to achieve today despite

play27:49

all of the technical difficulties and

play27:52

thank you for bearing bearing with us is

play27:55

to provide you a little bit more clarity

play27:58

in this mumbo-jumbo of the large

play28:01

language models that's currently

play28:02

floating around as everyone's figuring

play28:04

out how to Grapple with a topic

play28:07

um and to let you get a little bit

play28:08

clearer of an understanding of you know

play28:11

how to think about it next time you're

play28:12

in the room with someone and you're

play28:14

deciding or they're deciding whether

play28:16

it's worth you know developing your own

play28:18

large language model whether it's worth

play28:21

exploring fine tuning as an option and

play28:24

if those two things aren't an option for

play28:26

whatever the reasons

play28:28

um you know on the on the diagram where

play28:30

that we walked through together

play28:32

um you know you have a Third Avenue to

play28:35

go down to think about building your

play28:36

future products

play28:38

so yeah I guess my question to you is

play28:41

how you will lose use large language

play28:43

models to transform your business

play28:46

and um yeah feel free to I mean I'm

play28:49

happy to take any questions if there's

play28:51

if there's time I appreciate there's

play28:53

probably not that much time left but

play28:55

feel free to connect with me on LinkedIn

play28:56

and find me afterwards and also

play29:00

um Ryan my colleague is just reminding

play29:02

me of a thing that I always forget

play29:05

which I hope we have one more minute for

play29:06

that's um

play29:09

thank you

play29:10

which is to take a photo with you guys

play29:12

in the background

play29:13

you cool with that

play29:15

awesome

play29:16

okay

play29:20

all right

play29:22

let's get in there cheers

play29:28

cool all right I think we got it

play29:31

Philip thank you so much Victorious

play29:37

foreign

Rate This
โ˜…
โ˜…
โ˜…
โ˜…
โ˜…

5.0 / 5 (0 votes)

Related Tags
AI StrategyLanguage ModelsBusiness InsightsTech TransformationData SecurityConversational AICustom SolutionsOn-Premise HostingCloud ServicesPrompt Architecture