Unpopular Opinion: LLMs Didn't Add Value to AIs—They Just Made Them More Accessible

Project A Ventures
22 Nov 202347:02

Summary

TLDRこのトークは、LMS(Language Models)が技術的に新しいテーブルを持ち出しているか、それとも単に技術をよりアクセス可能にし、利用を拡大したにすぎないかという議論の中心に立脚しています。パネルは、LMSがビジネスや一般大衆の理解にどのように影響を与えたか、そしてこれらのモデルが将来どのように進化し、どのよう活用されるかについて洞察を提供します。彼らは、LMSが特定の業界でどのように使用され、その利点と課題をどのようにバランスを取るかについても議論します。

Takeaways

  • 🤖 AIは、人間のように行動するコンピュータシステムの総称であり、特に機械学習がその大きな分野の中で中心的な役割を果たしています。
  • 📊 LMS(Language Models)は技術的には新しいものではなく、RNNやCNNなどの既存技術に基つていますが、より高いアクセス性と広範な利用をもたらしました。
  • 🚀 Transformerアーキテクチャは、より大きなテキストや文脈を処理できるようになり、自然言語処理の分野で大きな進歩を遂げました。
  • 🧠 LMSの進化は技術的な進歩だけでなく、ビジネス価値の発見と適用面での革新も重要です。
  • 💡 LMSは現在、単純なタスクを自動化し、生産性を向上させることができていますが、複雑な意思決定やビジネスプロセスを完全に takeover することはまだ達成されていません。
  • 📈 LMSの成功事例は、医療や法律などのテキストに基づく分野で特に高いポテンシャルがあり、専門用語や規則に基づいたタスクを効率的に処理できるようになりました。
  • 🔍 LMSの信頼性については、まだ完璧ではありません。実際の専門家と比較して、LMSは試験で高いスコアを獲得しても、実際の業務では不十分な場合があります。
  • 🌐 LMSが将来的にシステムとの最初の接触点となる可能性があるとの視点があるが、効率的なユーザーインタラクションを実現するために、様々な方法が検討され続けています。
  • 🔗 LMSの発展は、多様な言語や文化を考慮に入れた包括的なアプローチが求められ、これにより技術の普及と利用が促進されます。
  • 📚 LMSの今後の進化は、応答の正確さと信頼性を向上させることで、より多くのビジネス分野で実装され、より広範に受け入れられるようになるでしょう。

Q & A

  • LMSが既存のAIシステムにどのような価値を提供したと思いますか?

    -LMSは、技術的には新しい価値を提供していないものの、よりアクセスしやすいインターフェースを提供し、既存のAIシステムをより广く利用できるようにしました。

  • LMSが技術的にどの分野で進化を遂げたと感じますか?

    -LMSは、Transformersアーキテクチャを基にしており、より大きなテキストデータセットを扱えるようになり、自然言語処理のコンテキストを大幅に改善しました。

  • LMSがビジネス価値を提供するための課題は何だと思いますか?

    -LMSがビジネス価値を提供するためには、技術的な課題だけでなく、適用層でのイノベーションも重要です。特定のビジネスニーズに合ったLMSの活用方法を見つけることが鍵となります。

  • LMSが医療や法務などの規制された分野でどのように役立つでしょうか?

    -LMSは、医療や法務などの分野で、豊富なテキストデータの処理やルールベースの分析を行い、専門家の仕事をサポートすることができます。ただし、LMSは最終的な決定を下すわけではなく、専門家の判断を補完する役割を果たします。

  • LMSの応用において、どのような技術的な課題があると感じますか?

    -LMSの応用においては、技術的な正確性と信頼性の向上、そして実際の現場での適用可能性の確認が課題となっています。また、LMSが誤った情報を提供するリスクにも注意が必要です。

  • LMSが自然言語処理の分野でどのような進化を見せましたか?

    -LMSは自然言語処理の分野で、より長いテキストや複雑な文脈を理解し、応答する能力を向上させました。これにより、より自然な人机交互が実現し、さまざまな業界で応用が可能となりました。

  • LMSを導入する際、どのようなビジネスのニーズを考慮すべきですか?

    -LMSを導入する際には、ビジネスプロセスの効率化、カスタマーサービスの改善、データ分析の支援など、企業が抱える課題やニーズに応じた適切な活用方法を検討する必要があります。

  • LMSが提供する自然言語処理能力を最大限に活用するためには、どのようなアプローチが効果的ですか?

    -LMSを効果的に活用するためには、適切なデータセットで事前学習を行い、特定のタスクや業界に関連するデータをファインチューンすることが重要です。また、ユーザーとのインタラクションを通じてLMSを徐々に改善していくことも効果的です。

  • LMSの開発において、今後どのような進化が期待されています?

    -LMSの開発において、今後期待される進化には、より高精度な自然言語理解能力の向上、多様なデータタイプとの連携能力の強化、そして倫理的な問題に対する対処などが含まれます。

  • LMSが誤った情報を提供するリスクがあることはどのように対処すべきですか?

    -LMSが誤った情報を提供するリスクに対しては、適切なデータセットで学習させ、定期的な評価とファインチューニングを行い、また、重要な決定においては人間の審査や介入を確保することが重要です。

  • LMSを導入する際のユーザーエクスペリエンスの改善について、どのようなポイントが重要ですか?

    -LMSを導入する際には、ユーザーが自然で使いやすいインターフェースを提供し、不明瞭なプロンプトや質問を最小限に抑えることが重要です。また、フィードバックメカニズムを設けることで、ユーザーの声を取り入れながらLMSを改善していくことも効果的です。

Outlines

00:00

🎤 イベントの開会とLMSの議論

イベントの開会とLMSの価値についての議論が行われています。議論の中心は、LMSが既存のAIシステムに新たな価値を提供したのか、単に技術をよりアクセスしやすくしたのかということです。また、LMSが技術的な進歩をもたらしたのか、または単にインターフェースとして機能したかについても話し及んでいます。

05:01

🌟 LMSの進化と技術的な革新

LMSの進化と技術的な革新について、3人のパネルリストが見解を共有しています。彼らはLMSが持続的な技術の進歩であると思わないという意見も示しており、LMSが現在注目されている理由は、技術的な進歩だけでなく、その技術が一般大衆にアクセス可能になったことにあると考えています。また、トランスフォーマーアーキテクチャとその前身の技術がどのように進化してきましたか、そしてLMSがどのようにスケールとコンテキストを扱う能力を向上させたかについても議論されています。

10:03

💡 LMSのビジネス価値と産業への影響

LMSがビジネスと産業に与える影響について、パネルリストが意見を述べています。LMSは企業にとって新しいビジネスケースを開き、AI技術をより使いやすくする重要な要素となりました。また、LMSは特定の業界(医療や法律など)において、ルールベースの分野でのパターン認識を改善し、効率を向上させる可能性があると指摘されています。

15:05

🚀 LMSの課題と今後の方向性

LMSの現在の課題と今後の方向性について、パネルリストが議論しています。LMSの応用が広まる中で、重要なのはモデルの正当性と信頼性の確保です。LMSは確率的なモデルであり、実際には期待通りに機能しない場合もあります。そのため、LMSの適切な訓練、フィードバック、改善が重要です。また、LMSが単に応答を生成するだけでなく、ユーザーとインタラクションを行って学び、改善する双向的なシステムとなることが期待されています。

20:07

🌐 LMSの多様性と言語の役割

LMSの多様性と言語の役割について、パネルリストが見解を共有しています。LMSは多様なモダリティ(テキスト、画像、動画など)を扱うことができ、将来的には基盤としてさらに発展する可能性があるとされています。また、言語は人間の間だけでなく、機械間のインターフェースとしても役立つとされ、LMSは自然言語でのインタラクションを容易にするでしょう。

25:08

📈 LMSの成功と信頼性のしやすさ

LMSの成功と信頼性のしやすさについて、パネルリストが意見を述べています。LMSは完璧ではないため、適切な成功率を設定することが重要です。簡単なタスクではLMSが自動化できるものの、複雑な決定には人間の参加が必要であるとされています。また、LMSは応答を生成するだけでなく、ユーザーのフィードバックを学び、改善することが望ましい双向的な学習システムとなるべきです。

Mindmap

Keywords

💡LMS (Language Models)

LMSは、自然言語処理技術の一種で、人間のような文章生成や理解を可能にします。このトークでは、LMSが既存のAIシステムにどのような価値を提供しているか、また技術的な進歩としてどの程度の意義を持つかが議論されています。

💡AI (Artificial Intelligence)

AIは、コンピュータシステムが人間の知的行為や行動を模倣する技術のことを指します。このトークでは、AIがどの程度進化し、LMSがその進歩の中でどのように役立っているかがテーマとなっています。

💡Transformers

Transformersは、自然言語処理における重要なアルゴリズムであり、平行処理が可能で、より大きな文脈を扱えるようになりました。このトークでは、TransformersがLMSの発展においてどのような役割を果たし、どのような技術的な進歩をもたらしたかが説明されています。

💡Fine-tuning

Fine-tuningは、機械学習において、既存の訓練済みモデルを特定のタスクに適応させる技術です。このトークでは、LMSを特定の業界やアプリケーションに適応させるためにFine-tuningがどのように使用されるかが議論されています。

💡Accessibility

Accessibilityは、技術やサービスが利用者にとって使いやすいかどうかを指す概念です。このトークでは、LMSが自然言語を通じてAIの技術をより広範なユーザーに利用可能にし、その意義が強調されています。

💡Probabilistic models

Probabilistic modelsとは、予測を行うために確率論を用いたモデルのことを指します。このトークでは、LMSがどのような確率的な予測を提供し、その限界についても議論されています。

💡Human in the loop

Human in the loopは、自動化システムにおいて人間の判断や参画が必要である場合に使用される概念です。このトークでは、LMSが特定のタスクを適切に実行するために、人間の介入が必要であることが強調されています。

💡Semantic understanding

Semantic understandingは、コンピュータが文脈や意味を理解し、適切な応答を行う能力を指します。このトークでは、LMSが語義的な理解を向上させることで、より高品質な応答を提供できることが説明されています。

💡Stochastic parrots

Stochastic parrotsは、LMSが確率的に応答を生成することを指す表現で、正確さと不確実性の問題を示唆しています。このトークでは、LMSがどのようにして確率的な予測を行うか、そしてその限界について議論されています。

💡Multimodality

Multimodalityは、複数の感覚やデータタイプ(例えば、テキスト、画像、音声)を組み合わせて情報処理を行うことを指します。このトークでは、LMSが多様なデータタイプを扱う能力が、技術的な進歩の一部として強調されています。

💡Knowledge base

Knowledge baseは、組織が蓄積した情報や知識を整理し、検索や分析のために利用できるようにしたデータベースです。このトークでは、LMSが知識ベースを構築し、効率的に活用することの重要性が強調されています。

Highlights

The panel discussion revolves around the value addition of Language Models (LMs) and their impact on Artificial Intelligence (AI) systems.

The panelists include experts from tech consulting, machine learning project management, and Google Cloud, providing diverse perspectives on LMs.

The discussion explores whether LMs have brought new technical advancements or simply increased accessibility to existing AI technologies.

Panelists share their definitions of AI, highlighting the importance of machine learning and the ability of systems to mimic human intelligence.

The evolution of AI is described as a continuum, with LMs like ChatGPT and DALL-E relying on transformer architectures and previous technological advancements.

The accessibility factor of LMs is emphasized, as they have made AI more approachable and fun for the general public.

The role of LMs in transforming user interaction with systems is discussed, with a focus on their potential as the first point of contact in various industries.

The importance of fine-tuning LMs for specific industry use cases, such as medical and legal applications, is highlighted to improve their accuracy and relevance.

The challenge of validating LMs is addressed, as their probabilistic nature makes it difficult to ensure their reliability in practical applications.

The potential of LMs to assist human experts in decision-making processes is discussed, with the emphasis on their current role as supportive tools rather than replacements.

The discussion touches on the need for LMs to understand context and handle complex queries, such as distinguishing between different types of quarters in a financial query.

The future of LMs is envisioned as part of the infrastructure, integrating with various data types and systems to form a comprehensive knowledge base.

The importance of inclusivity in LM interactions is stressed, with the need for systems to cater to diverse language abilities and literacy levels.

The potential of LMs to learn from human feedback, gradually improving their performance and ability to automate tasks, is highlighted as a promising area of development.

The debate on LMs as 'stochastic parrots' is addressed, acknowledging their probabilistic nature while also considering their potential for reasoning and understanding.

The discussion concludes with thoughts on the threshold of success rate for LM implementation and the need for a human-in-the-loop for sensitive use cases.

Transcripts

play00:03

okay welcome uh to our panel the second

play00:05

but last talk um on you can see the

play00:09

topic on public opinion LMS didn't add

play00:12

any value to our existing AIS um they

play00:14

just made more accessible what we

play00:16

already had and uh when I was scoping

play00:18

the topic two or three months ago I

play00:20

couldn't imagine that we're g to have

play00:21

that many talks on LMS uh but here we

play00:24

are and uh even if you have attended

play00:26

former talks I'm sure you're going to

play00:28

find this one uh valuable as well so I'm

play00:30

chrisan I joined project a this year in

play00:32

the tech Consulting

play00:33

team and I'm super curious to find out

play00:36

what our panelists have to say um I

play00:39

already read out the topic um we're

play00:41

going to talk about um if LMS brought

play00:44

anything to new to the table technically

play00:46

um or if they were just like acting as

play00:49

an interface or a font technology and

play00:52

making thereby LMS more accessible So

play00:55

today we're going to have the goal of um

play00:58

giving you some context to better

play00:59

understand the the real technical

play01:01

advancement behind behind LMS and also

play01:04

to give you some food for thought um how

play01:06

LMS could develop further and um how

play01:11

they could be utilized in a way that

play01:12

they provide real undeniable value with

play01:16

me today are three lovely panelists um

play01:19

we all like you all had have paneled in

play01:22

different um

play01:24

situations recently at the I think J

play01:27

breakfast for marantic so um to start

play01:30

off Hannah Hannah is um the team lead

play01:34

for machine learning project management

play01:36

at metics momentum marantic momentum for

play01:39

those of you who don't know um

play01:40

identifies use cases for II and ml um

play01:44

with their customers and also then

play01:47

develops and um um also operates those

play01:51

um those solutions to to the use cases

play01:54

then yakob you have seen yakob on stage

play01:56

this morning already um yakob is a

play01:59

customer engineer here at Google cloud

play02:01

and uh is basically together with their

play02:04

clients developing and um optimizing the

play02:08

data infrastructure so um for all Cloud

play02:11

native gcp clients um I would say I

play02:14

would call you a an ML and II Enthusiast

play02:17

and you're appearing like recently

play02:19

appearing on many talks panels

play02:22

hackathons um yeah and last but not

play02:25

least is my dear colleague saman who's a

play02:27

data scientist and engineer here project

play02:30

a um together with our ventes he works

play02:32

on data infrastructure machine learning

play02:34

and also uh data

play02:37

culture so um yeah I'm super happy to

play02:40

have you here and um to kick things off

play02:43

I'd like to uh get some interaction and

play02:46

um ask you the audience on your opinion

play02:50

um which on which side um you see

play02:52

yourself also on which side on the

play02:54

Spectrum so I would like to raise your

play02:56

hand if you you are of the opinion that

play02:58

LMS didn't add any value to our existing

play03:02

a

play03:03

technically is there anyone who's of the

play03:05

opinion that llms didn't add any

play03:11

value okay counter check if you're just

play03:14

lazy raising your hand or because no one

play03:16

is of the opinion who's of the opinion

play03:18

that they add any that they did add any

play03:20

value technically

play03:23

okay so then we I think we're going to

play03:25

have a interesting discussion um and to

play03:28

kick things off and go right in I would

play03:30

like to ask you three um to give us a

play03:34

one- sentence definition of AI because

play03:36

of course we know everyone defines a

play03:38

differently uh and then so we can can

play03:41

put your opinion um into perspective

play03:44

maybe to start off with Hannah

play03:48

sure no it was on before it was on

play03:51

before sorry yeah happy to kick it off

play03:54

so for me AI is really a collective term

play03:56

for any kind of computer systems that in

play03:58

some way simulate

play04:00

yeah Behavior that's otherwise

play04:02

identified or associated with human

play04:04

behavior or intelligent behavior and I

play04:07

think for me very pragmatically most

play04:09

importantly is really this machine learn

play04:11

topic of machine learning um as the

play04:13

biggest cluster within the huge field of

play04:19

AI yeah pretty I can I'm pretty much in

play04:22

line with what he said but maybe to put

play04:24

it into one sentence I think AI is a

play04:27

collection of statistical and

play04:28

mathematical

play04:30

that sort of mimic human decision making

play04:33

in some

play04:35

sense I'm also in line with all the

play04:37

other panels have said but I think I

play04:38

would just go a bit broader and move

play04:39

away from the statistical part of it and

play04:41

just say any system that in one way or

play04:43

the other um mimics aspect of human

play04:46

intelligence whether that might be

play04:47

Vision speech text writing that I was

play04:51

ciz as as

play04:52

AI Co thanks a lot so um to let's get uh

play04:56

so let's get technical for a minute um

play04:58

at project a we value knowledge sharing

play05:00

really a lot so that's why for example

play05:02

our conference tomorrow is called

play05:03

project a knowledge conference and also

play05:06

um at project a we have regular brown

play05:08

bag lunches uh where a or colleagues can

play05:11

share recent project outcomes or also

play05:13

like any topic they find relevant to

play05:15

them or they care of and recently saman

play05:18

shared or did a brownb lunch on AI in

play05:21

general and had a pretty I would say

play05:23

critical view on this um beyond the bus

play05:27

I would say and that's why I like to ask

play05:29

you um is the Advent of LMS rather a

play05:32

Continuum or a prival technical

play05:35

advancement um in the ai's journey and

play05:37

why or why

play05:38

not um I think technically I wouldn't

play05:41

say it's pivotal actually um I think if

play05:44

you look at it from a technical

play05:46

background current implementation of AI

play05:48

which I think today we talk about these

play05:49

geni aspects like chat GPT um Dar ma

play05:53

Journey what else you have um rely on

play05:55

this transform architecture which

play05:56

basically then relies on lsdm which then

play05:58

relies on RNN CNN so I think you have a

play06:00

clear Evolution and and history of you

play06:04

might even say incremental additions and

play06:06

changes that have let to this point and

play06:08

therefore arguing that now suddenly we

play06:10

have these these these gen

play06:13

uh applications are suddenly changing

play06:16

would be a bit well exaggerating I would

play06:18

say and I think if there was something

play06:20

like shim who were in here he would now

play06:22

say oh no I did all of this and I think

play06:24

I would maybe partly agree with this as

play06:27

on a technical level it's a clear

play06:28

Evolution what

play06:30

however makes us such a buzz in my

play06:31

opinion that's why we have all of these

play06:32

panels and we talk about this so much is

play06:34

then um that somebody like my 67y old

play06:38

dad now has a chat GPT app on his phone

play06:40

and uses his because it's funny and cool

play06:43

and that's essentially I think the

play06:44

accessibility factor of it that previous

play06:45

implementations were not able to do if

play06:48

you talk to somebody about an lsdm is

play06:49

they're like bro I don't know

play06:51

but then you can show them hey give me a

play06:53

cooking list for whatever I have in my

play06:54

fridge and they can use it and they're

play06:56

like oh this is insane and this

play06:58

is a bit where I would sort of of argue

play07:00

it's definitely not a pivotal change

play07:02

technically um might be pivotal in sense

play07:05

of how it's being perceived and

play07:07

applied dig a bit into deeper into this

play07:10

um what especially so you mentioned for

play07:12

example Transformer architecture um is

play07:14

this is this what you would you would

play07:16

consider the technical advancement for

play07:18

example that you now can um really

play07:20

process it in parallel or can um get can

play07:24

input like larger contents um or

play07:26

contexts and it can understand it or

play07:29

where do you see the technical

play07:30

advancement because otherwise I would

play07:32

say um there is there is nothing that

play07:34

that we can really like um yeah hold on

play07:38

to that that says okay it's technical

play07:40

technically um I suppose and and I think

play07:43

that's also so funny when we talk about

play07:44

Transformers because they came out in

play07:46

2017 I think was your former co-workers

play07:49

paper when it was published um uh the

play07:53

the main benefit I would say of

play07:55

Transformers compared to say current

play07:56

applications or for applications was

play07:58

then you were able to um which goes into

play08:01

this paration effort is you are now able

play08:03

to work with much much much larger

play08:05

corpuses Cory of body that were not able

play08:08

to do so before right um if you take a

play08:11

look at again it's lsdm which we might

play08:12

have had before you were may be able to

play08:15

go back as far as a sentence or two

play08:17

because then you've got into these

play08:17

diminishing returns and you're like oh I

play08:19

don't even remember what you said and

play08:21

now you know chat GPT can write you

play08:23

books of content because this sort of

play08:25

shortcoming has been alleviated with SC

play08:27

implementation which yeah can be

play08:29

attributed to a Transformer sure um

play08:32

applications of it then um of what say

play08:34

open AI does what BART does I think are

play08:37

then more that on

play08:42

scale maybe to add on to that I I

play08:45

absolutely agree with what you've said

play08:47

um I think what's I think what's very

play08:49

much a Continuum is also when you think

play08:51

about language we are always or the the

play08:54

whole field of natural language

play08:55

processing is kind of concerned about

play08:57

capturing context and for the longest

play09:00

time this was incredibly hard um as you

play09:02

said like we would didn't really get

play09:04

past sentences of context that were able

play09:07

to be processed and then 2017 this

play09:10

Transformer comes along of course also

play09:11

building on Tech that was there

play09:13

previously but this basically sparked

play09:14

this explosion of yeah exponential

play09:17

growth in what you can process in terms

play09:19

of context whole whole books whole like

play09:22

even libraries maybe at this stage so um

play09:25

this this is really incredible and I

play09:28

would also agree that the change we're

play09:30

seeing now what's quite pivotal is um

play09:32

definitely that we can now as an end

play09:35

user also shape more how tool behaves um

play09:39

without having to train it without

play09:40

having to be super technical about it

play09:43

but I do think this is partially also a

play09:44

technical um an invention for for sure

play09:47

um coming from the way these models are

play09:50

pre- chained and now allow us to for

play09:52

example um yeah give examples give

play09:54

context provide intense and say dear

play09:57

model please behave as if you were XY Z

play09:59

or please write in the style of this um

play10:03

and and this is quite quite pivotal

play10:05

because this definitely makes some

play10:07

things way easier and we can discuss

play10:09

later if this is uh if this this a

play10:11

Continuum or if this is really

play10:13

groundbreaking yeah I will come back to

play10:15

this definitely um yob why why did your

play10:17

colleagues um did take so much time from

play10:20

2017 until now um until they until

play10:22

transform architecture really um yeah

play10:26

really developed itself to to a point

play10:28

where could could be utilized in a way

play10:31

that that we can now really um see at

play10:34

least a business value of of

play10:36

course hard for me to make a statement

play10:39

for for the researchers at this point

play10:41

from from my perspective obviously but

play10:44

yeah interesting question I think it's

play10:46

really in line with the fact that in my

play10:49

opinion technology usually doesn't make

play10:52

pivotal changes and that's sometimes a

play10:54

little bit of a tough truth because I

play10:56

think us as humans we're also a little

play10:58

bit wired in the sense that we love we

play11:01

love the story aspect about these

play11:03

individual Geniuses that make an

play11:05

invention and then everything changes

play11:07

and then we have suddenly this blackbox

play11:09

solution that just solves all our

play11:10

problems based on one genius's idea but

play11:13

usually usually in technology and

play11:14

research it doesn't work like that it's

play11:16

rather one researcher Builds on the work

play11:18

of many others we're all sort of working

play11:21

on shoulders of giants in the end we

play11:23

never invent everything from scratch and

play11:26

that's also fine like that right that's

play11:28

that really Mak the research uh culture

play11:31

and the research Community right that

play11:32

they collaborate and build on top of

play11:34

each other's ideas and really the

play11:36

pivotal changes in my opinion happen on

play11:38

the application layer similar to what

play11:40

someone said right suddenly there's a

play11:42

break breakthrough someone um had the

play11:45

idea of hey why don't we bring together

play11:47

this technology with a certain business

play11:49

use case where we see a large large need

play11:52

this could solve this specific use case

play11:54

and then these two perspectives come

play11:56

together and when this B both both fits

play12:00

right there's a technology that has

play12:02

certain strength and then on the other

play12:03

hand you have a business use case that

play12:05

can be solved um and that makes sense

play12:08

together and if this actually has value

play12:10

then it really explodes then you have

play12:11

your P to change but again that's more

play12:13

rather on the application layer than on

play12:14

a tech

play12:17

technological in my opinion yeah so it's

play12:20

it's partly a interface because it's

play12:22

sticking together to Technologies and

play12:24

then there would be an interface um like

play12:26

the application acting as an interface

play12:28

and making

play12:29

front end whatever you want to call it

play12:30

and making it then

play12:33

accessible in a sense yeah of course of

play12:35

course of course I mean in the end you

play12:36

pluck together multiple Technologies and

play12:38

then you try to get certain output right

play12:41

yeah absolutely okay and then um to to

play12:43

briefly talk about like the business

play12:45

value um of llms Hannah you work closely

play12:48

together with clients across Industries

play12:51

um what would what would you say um the

play12:54

introduction of llms um has it or like

play12:57

where has it led to a significant change

play12:58

in the perception but also in the

play13:00

understanding um a among the masses and

play13:02

B um in in business

play13:07

General yeah so I think and I I would

play13:10

expect we all agree with that and you

play13:12

already mentioned that what what's

play13:13

really this amazing Trend we are seeing

play13:15

is that um everyone is getting hands on

play13:18

and trying to work with with these tools

play13:21

so I don't really know anyone who hasn't

play13:23

at least tried to generate some greeting

play13:25

card text or something like that with

play13:26

llms which is amazing to see but to to

play13:29

talk about industry I think maybe two

play13:31

interesting things happen so first of

play13:33

all we can we can feel the formo uh so

play13:36

if there was anyone before not not

play13:38

thinking about okay how can I use these

play13:41

these how can I use AI essentially to

play13:43

accelerate my bris business is now for

play13:45

sure thinking about it at least if not

play13:48

acting on it and and secondly um of

play13:51

course it also unlocks certain use cases

play13:54

or certain thoughts and I would also say

play13:57

of course sometimes the the

play13:58

conversations we have it seems like chbt

play14:02

is is seen as like the solution to

play14:05

everything which is clearly not the case

play14:06

and we are also seeing that for example

play14:08

in Industry we we have a lot of clients

play14:10

in Industry um the main data sources

play14:13

that we really have there are time

play14:15

series and tabular data and this should

play14:18

not be underestimated so everything that

play14:19

was always said before like build up

play14:22

your data um warehouses build up your

play14:24

lake houses build up your um make sure

play14:27

you you build the Foundation to utilize

play14:30

this value this still holds true um up

play14:32

to this point and there are a ton of

play14:35

methods that should not be

play14:37

underestimated um and then on top of

play14:39

that language comes in as a very

play14:42

interesting interface to access this

play14:44

data in a new way and I think that's

play14:46

that's really cool to see um but we also

play14:50

we always have to make a bit of this

play14:52

trade-off to see where is really where

play14:54

is it language that solves the use case

play14:55

and where is language that's the cool

play14:57

addition to have a new way to interact

play15:00

can I just react on this so generally

play15:02

100% get your get your basic straight

play15:05

before you jump into the more Advan St

play15:07

stuff couldn't agree more but where I

play15:09

would disagree with you is really since

play15:12

we in the point you said um that you see

play15:16

most with customers you see mostly

play15:18

structured data in Time series yes they

play15:21

do have a lot of that but they have a

play15:24

lot of that because they didn't pay

play15:26

attention to all the textual data or

play15:28

potentially the data that could

play15:30

potentially Ed be used for um for NLP

play15:32

use cases I would say so obviously

play15:35

there's a lot of structured data and

play15:36

time series and obviously companies

play15:38

still struggle with the basics in

play15:40

getting really the infrastructure state

play15:42

but um I would say there is even more

play15:44

textual data they just didn't unlock

play15:46

that yet as data that has potential to

play15:49

be used for AI so much absolutely yeah I

play15:51

think I can I can only agree with that

play15:52

and I think maybe to to add one one

play15:54

aspect there um um is that I think in

play15:57

every company us have functional units

play16:00

that have a lot of text being that the

play16:01

legal space being that um yeah HR

play16:05

internal knowledge bases way where you

play16:07

have a lot of data that you can unlock

play16:09

um in in new ways um nowadays and then

play16:12

in your core business it really depends

play16:14

on what you're doing if that's also

play16:15

textual or if that's um different data

play16:17

but I agree um in that sense yes I think

play16:20

every company has to some extent

play16:22

potential to look at your text

play16:25

Data thanks a lot for for for the rewind

play16:29

um to come to like the current current

play16:31

scenario um we briefly talked Hana you

play16:35

mentioned it um about like the the the

play16:38

way you WR prompts um also there's F

play16:42

short learning um you mentioned it

play16:43

before in your talk in your talk as well

play16:46

to to um utilize for example giving a

play16:48

demonstration of um of different use

play16:52

case of um the things that the AL is

play16:55

supposed to do um to like fine tune it

play16:58

in a way

play16:59

um is there um any are there any

play17:03

examples could you share any example

play17:04

where where for example these

play17:08

um the these two types of fine-tuning in

play17:11

LM a task osting llm um has led to any

play17:16

impact in any

play17:18

industry um by for example adding this

play17:21

context and thereby for example

play17:23

unlocking longtail use cases that

play17:26

formerly couldn't couldn't even Sol

play17:28

because there's simply no return invest

play17:30

sure so generally what I'm seeing in in

play17:34

my experience from working with LMS is

play17:36

that find tuning multi things so

play17:37

generally supervised learning in the

play17:39

sense what we already know fromal

play17:41

machine learning makes a lot of sense

play17:43

and it increases the quality of your

play17:45

answers by a huge margin um so I believe

play17:49

most use cases will go in the direction

play17:51

of using some sort of multi pumping fine

play17:53

tuning Etc um that's the basic basic

play17:56

layer to as as an answer of your

play17:58

question in terms of specific industry

play18:00

use cases yes I think there are some

play18:02

industries that have high potential for

play18:05

the use of LMS for example Med medical

play18:07

industry legal use cases um some others

play18:11

right um but you can already see that

play18:13

because these are very text based at the

play18:16

same time text based usually um but at

play18:19

the same time very rule based in some

play18:21

sense right it's sort of pattern

play18:23

recognition in both both of these cases

play18:25

a lot of cases but at the same time

play18:27

they're also highly regulated and you

play18:29

work with highly sensitive data and

play18:31

specifically in the medical case you

play18:33

work with data and you work with

play18:34

decisions that you should not get

play18:37

wrong right um even more in the medical

play18:40

space because it's sometimes about lives

play18:42

right in legal space it's a lot it's

play18:44

about a lot of money usually so there is

play18:47

an inherit value for llms the task is um

play18:51

makes sense to solve but we have

play18:53

specific challenges why anms need to get

play18:55

these points right and in these cases we

play18:58

probably should think about how to train

play19:00

L&M that are specific for the industry

play19:02

great example for this is met Palm so

play19:04

our foundation model is Palm right and

play19:07

we're one of the research projects that

play19:09

is in preview at the moment actually is

play19:12

metp for example right our research

play19:14

trained Palm fine tuned it on medical

play19:17

data um fine tun it for this specific

play19:20

use case so essentially in the back

play19:23

combine factual knowledge that the llm

play19:26

needs to get right in any case with the

play19:28

generational knowledge of of Pama the

play19:32

generational capabilities right and this

play19:34

works quite well I mean p is not public

play19:36

yet but it already outperforms any other

play19:38

model in the medical um certification

play19:41

tests in the US for example um similar

play19:43

for the legal knowledge I don't I'm not

play19:46

aware at least for legal llm purely

play19:48

legal find you LM probably it's out

play19:50

there already but I can imagine very

play19:51

well that will fine tune or train

play19:54

specifically in llm on for example a

play19:55

certain legal codex um to to answer

play19:59

specific questions and compare claims

play20:01

between parties um and have the legal

play20:04

system in the rules of that yeah thanks

play20:07

so some we at project a and and you and

play20:09

your work we we don't have the capacity

play20:11

to train whole models ourselves to find

play20:13

you find you them um have you seen any

play20:16

recent either in our portfolio or like

play20:18

um on on blogs or have you read about

play20:20

recent implementation um success stories

play20:24

also challenges uh with

play20:26

llms um

play20:28

I think I would go in the direction that

play20:30

that that yakob just mentioned um

play20:32

because smart struggle with LMS I don't

play20:35

want to talk too much of the skeptic is

play20:36

in this aspect um there is a possibility

play20:39

to find trun LM based on these

play20:41

foundational models given a certain

play20:43

distinct situation or data set and make

play20:44

them perform very very well on those you

play20:46

mentioned Health you mentioned legal HR

play20:48

possible whatever um however what I then

play20:53

very much strug given that these are

play20:54

strongly probabilistic models is the

play20:57

possibility

play20:58

that these actually are not doing what

play21:01

they're supposed to do because um

play21:04

validating these llms is incredibly

play21:06

difficult um so there are numerous

play21:08

articles where I've seen that while they

play21:10

might perform very very well on these um

play21:12

legal bar exams or these Health exams

play21:15

whatever that might be putting them to

play21:17

practice they always are being

play21:19

outperformed easily by an actual lawyer

play21:21

by an actual doctor um which makes this

play21:24

this um intelligence aspect which is

play21:26

being touted as this the strength of

play21:28

these llms kind of ambiguous because

play21:31

what does it actually mean if you

play21:32

perform well on the test but then in

play21:33

practice you're actually quite shite um

play21:34

I don't know I'm not the one to answer

play21:36

this question but this is why I sort of

play21:38

struggle with this notion that when you

play21:40

have some sort of AI model in an LM

play21:41

model it's opposed to um I don't want to

play21:45

say replace but it's actually I think

play21:46

just there's an assistant and therefore

play21:49

I I I'm bit skeptical to see it's going

play21:52

to have this massive massive impact that

play21:54

a much much simpler models have had so

play21:57

far so um if if you ask me if there's

play22:00

one thing that is the biggest or had the

play22:02

biggest value to anything it would be

play22:03

linear regression but this came out what

play22:05

150 years ago or something right U but

play22:07

it still works and it works very

play22:09

very well um and if LM I think there's

play22:12

this risk to it that that might not be a

play22:14

more deterministic approach that that

play22:15

integration might offer

play22:18

um so I think for me and and I think

play22:21

it's going bit divert a bit from your

play22:22

original question um thinking of

play22:24

industry and application my struggle

play22:26

still seems to be a validation and

play22:28

making sure that these things work the

play22:29

way they're supposed to do no as on like

play22:33

old man get get clout but this is sort

play22:35

of the approach that I take when we

play22:37

think about applying these things to

play22:38

Industries because no in our portfolio

play22:40

it's not really that big um because

play22:43

there I think we rely also much more on

play22:45

hanus approach mentioned earlier like

play22:46

get your Basics right get your linear

play22:49

regression right before you can talk

play22:50

about um llms etc etc etc because that's

play22:54

too far away and most of the time might

play22:56

not have impact you actually desire

play23:00

might you add on this I would love to I

play23:03

think I would like to share one example

play23:05

which I think illustrates quite well how

play23:08

in this case the legal space Also

play23:10

benefits from the recent advantages and

play23:13

also how we at momentum kind of handle

play23:15

this this risk because yes I absolutely

play23:17

agree um we have probabilistic models

play23:20

here we have models that hallucinate um

play23:22

we have these risks and it's very

play23:24

important for us to make that very very

play23:25

transparent and handle that but I do

play23:28

think that assisting work um while an

play23:32

expert a human expert has the final say

play23:34

is still something that can have huge

play23:36

benefits um huge impact so one case i'

play23:40

I'd like to share from our project um

play23:42

portfolio is that we are working with a

play23:45

with a legal partner um and this one is

play23:47

about um yeah working and processing

play23:50

relatively standardized

play23:52

um legal regulation um think basically

play23:56

ndas or something like that um and

play23:59

what's very interesting is that I think

play24:02

handling something in the legal space

play24:04

benefits a lot from recent advantages so

play24:06

one thing for sure is models being

play24:08

fine-tuned on legal data in general so

play24:10

being able to process this kind of

play24:12

language and being able to process

play24:13

longer documents with a lot of context

play24:16

but I what I find very interesting and

play24:17

this wasn't basically possible um 10

play24:21

months ago is that in the legal space

play24:24

you also always have not only one fixed

play24:26

set of laws but you also have a

play24:28

continuous flow of cases of reference

play24:32

cases you need to consider if you're

play24:33

making a specific decision and um this

play24:36

new way of context specific learning

play24:39

essentially allows you to feed that into

play24:41

your model without having to train it

play24:43

but just you can give examples you can

play24:45

give reference cases and you can extract

play24:48

Knowledge from all of that and at least

play24:49

like provide a summary of um what the

play24:52

models then train to um detect as

play24:55

relevant passages um give that to an

play24:58

expert and it's already helping you and

play25:00

giving these pointers and giving

play25:01

references and yes it needs a lot of

play25:03

constraint so that it will give this

play25:05

references but it allows you to combine

play25:08

internal and external sources and I

play25:09

think that's very very um cool to see

play25:12

and definitely something that's that's

play25:13

possible and that we are currently

play25:15

working on for example I wouldn't AG I

play25:17

wouldn't disagree with this I think I

play25:18

would actually sort of agree with this

play25:20

um because I think the main benefit of

play25:21

this as you said is assistance that I

play25:23

would say complement existing work so

play25:25

you always read about um companies that

play25:27

use AI had an productivity benefit of

play25:29

uplift of what 20% and I think that's

play25:31

that 20% that a lawyer would spend

play25:33

reading this long boring documents for

play25:35

being P to make sense of what's going on

play25:36

there and and that's essentially when we

play25:39

go to a bit more of an established

play25:41

approach which is sematic understanding

play25:43

right I think translation of documents

play25:45

was more less a solved issue in in 2010

play25:47

whatever right and and there we had a

play25:50

clear Benchmark right um I don't know

play25:52

what it's called but it's essentially

play25:53

when you have a machine translated text

play25:55

versus human translated text how Sim

play25:57

arties to and that essentially defines

play25:59

how well the transation is going and

play26:00

that's how we also measured how well

play26:02

Transformers perform actually whenever

play26:03

launched um I think this you this this

play26:06

one Benchmark is the what's missing to

play26:09

then rely on what's a legal um what say

play26:13

illegal AI might do and I ask a gen

play26:16

based on legal text hey is this bad am I

play26:19

being suit now I wouldn't trust it

play26:21

advice but I would trust it hey what

play26:23

does it tell me essentially this um this

play26:26

transfer knowledge that we're having

play26:28

from what is been generated and is it

play26:29

actually applicable can I do something

play26:31

with it I think that bridge is still

play26:32

missing but the first step that's need

play26:34

before understanding what's going on

play26:36

yeah that's amazing right that

play26:37

works quite well I would

play26:39

say oh thanks um then now looking

play26:42

forward you you already mentioned that

play26:44

there's like there's like a gap until

play26:46

until LMS like until you can fully trust

play26:48

LMS or if you want if you want to add

play26:51

this this maybe the first question if

play26:52

you want to fully trust alms um but but

play26:56

more generally speaking um

play26:58

what like yob to you the question um

play27:00

what what further advancements do you

play27:01

see to undergo for example to get rid of

play27:04

this the symptom that you would like

play27:06

only um consult lmms and then make yeah

play27:11

make your own decision um to for example

play27:14

being LMS being um able to decide for

play27:18

themselves and and act on

play27:22

this difficult question I think um that

play27:26

we're quite far are still from a place

play27:29

where llms do everything and take over

play27:31

all the decision processes and I think

play27:33

you have to differentiate a little bit

play27:35

what's the difficulty of a certain task

play27:37

difficulty of a certain decision right

play27:39

for simple tasks that don't require 100%

play27:43

accuracy all the time we're already

play27:45

there they can already take over quite a

play27:47

lot in my opinion for example like chbd

play27:49

plugins insta card just giv me like a a

play27:52

recipe and then order it on inst card

play27:54

right right yeah yeah simple

play27:55

interactions right um there's also maybe

play27:58

not that much of the line we can already

play28:00

do that I would say um to get to a point

play28:03

where llms do much more complex

play28:06

decisions uh or take them over

play28:08

completely I don't think we're going to

play28:10

get there anywhere soon because also

play28:13

traditional machine learning doesn't

play28:15

most of the time do that I would say

play28:18

right I think in some cases very simple

play28:21

case it in cases it does and we need to

play28:23

scale up a lot of very simple decisions

play28:26

yes I see it in a similar sense maybe

play28:28

less quantitatively based but more text

play28:30

based decision that are that are being

play28:31

taken um I think especially in in the

play28:34

interaction space right really having

play28:35

this as a sort of first interaction

play28:39

layer with a user with a customer

play28:41

internal user external user I think

play28:43

that's where LMS will take over a lot of

play28:46

decisions very independently but really

play28:48

making pivotal business decisions for

play28:50

example um I don't think we're going to

play28:52

get there yet I think it's really

play28:54

similar to machine learning as it does

play28:56

already taking over over the 80% of the

play28:59

manual work and then still leaving the

play29:01

20% and the final decision- making to

play29:05

the to the human in many cases yeah well

play29:09

thanks so so so you rather say okay if

play29:11

your question if you really want it want

play29:13

them to to make decisions um rather

play29:16

being like the first point of contact um

play29:19

yeah exactly exactly yeah um okay so

play29:23

really interesting because I recently um

play29:25

watched the talk from the Nvidia CEO and

play29:27

um he envisions a future or predicts a

play29:29

future where LMS will be um the first

play29:32

point of contact for every interaction

play29:34

with the system um Hannah would you go

play29:37

uh with his uh his prediction of the

play29:39

future and maybe also um what do we need

play29:42

to undergo in terms of uix um until we

play29:45

reach maybe such a

play29:48

point well I love this question it's a

play29:50

very steep hypothesis uh

play29:53

so I think I have a slightly different

play29:56

take there um to be honest I don't think

play29:58

language is always the most efficient

play30:00

way to communicate with a system if I'm

play30:02

very honest I think sometimes a simple

play30:04

drop down and hit and go actually does

play30:06

the job quite well so I I think we

play30:09

should keep that in mind like what what

play30:11

is really what is the good experience

play30:14

for the user to interact with a system

play30:16

if there if you have for example an FAQ

play30:19

um space with only five questions adding

play30:21

a chap but on top can make it even like

play30:24

actually quite confusing because you

play30:25

would have to kind of guess what this

play30:27

chatbot can help you with instead of

play30:28

just like browsing quickly having a

play30:30

glance at these five

play30:32

questions but in general I I do agree

play30:34

what's an very interesting trend is to

play30:36

see language being more and more an

play30:38

interface between um humans and machines

play30:41

even between machine and machines in the

play30:43

sense that we see this trend of

play30:44

multimodal models involving where you

play30:47

have models trained on C different

play30:49

modalities like for example image and

play30:51

text or so video and text so you can now

play30:54

ask questions about an image for example

play30:56

and get the output read to you for

play30:59

example by an AI voice so there's so

play31:02

many ways how language kind of becomes

play31:04

this interface and I think this is an

play31:06

incredibly interesting Trend so yes I do

play31:08

think that we will see a further rise of

play31:11

language based uh interactions but I'm

play31:14

not quite convinced that prompting and

play31:16

like iterating on prompts to get the

play31:18

right answer is necessarily the solution

play31:20

for for every

play31:23

system think you just described clippy

play31:26

basically um so I generally do not think

play31:30

language is anywhere close to being the

play31:32

most efficient way to to to with any

play31:34

sort of system I think and that's

play31:36

essentially also what I attribute to the

play31:37

success of chat2 is the the tendency of

play31:41

humans to feel much more comfortable

play31:44

with anthropomorphic systems like chat

play31:46

talks and reads like a human being and

play31:48

that's what makes it so impressive

play31:49

whereas a complicated I mean we all Tech

play31:53

space here CLI or whatever interface you

play31:55

might have is then sort of scary and

play31:56

spooky

play31:57

but I think most of us in this room

play31:59

might actually prefer a CLI because

play32:01

that's the most efficient way to maybe

play32:02

communicate with whatever is that you

play32:03

want to do do I think text is efficient

play32:06

absolutely not it takes a long time to

play32:08

write takes a long to read a drop down

play32:10

is two clicks so I hope that we do not

play32:12

get into this direction because I think

play32:14

we go in circular we can see that uh the

play32:17

understanding of systems is increasing

play32:19

you know as I said my dad is quite well

play32:21

as iPhone these days you know um so we

play32:23

don't need these systems from like

play32:26

Clippy which were designed to in the

play32:28

'90s to make people who do not

play32:30

understand what you know the trash can

play32:31

means on your desktop to explain it to

play32:34

them but but one point where why I

play32:38

partly disagree let's say is that sure I

play32:42

agree with an educ with your point for

play32:45

an educated bubble of very tech savvy

play32:48

people but now that's not what Humanity

play32:51

looks like right I think the really

play32:54

average and the majority of folks that

play32:56

used to technology use most of the

play32:58

products they are rather uncen towards

play33:01

technology right for them is technology

play33:03

is something that they use as a user but

play33:05

they don't think about what's the most

play33:07

efficient way in communicating with it

play33:09

and for them natural language is the

play33:12

most natural way for sure and they don't

play33:14

even think about whether it's the most

play33:16

efficient or not for them it's the first

play33:17

way that someone would communicate so in

play33:20

that light I would say surely it's a

play33:22

bold statement to say it's every single

play33:25

time the first um layer of communication

play33:28

but I would say we're going to hit I

play33:30

know if you look at all the system inter

play33:32

system interaction probably over all the

play33:35

users probably 90% will be hitting a

play33:39

natural language layer first and then um

play33:43

access functionality and I mean a good

play33:45

example of that is um who would have

play33:48

thought before search engines came out

play33:50

that some sort of keyword search uh

play33:53

weird query is the way of interact with

play33:56

online content right is this the most

play33:58

efficient way to interact with it

play33:59

probably not right but it's the way that

play34:02

really that everyone can use right that

play34:04

everyone used to be able to use and

play34:06

similarly I think the way of querying

play34:08

documents quering knowledge is going to

play34:10

change quite a bit more towards the

play34:12

semantic way and more use more towards

play34:15

using natural language but similar to

play34:18

search I think we're going to have

play34:21

similar manner that search bars work now

play34:23

going have more agents or search bars

play34:26

that act as agents to interact as a

play34:28

first layer that's for

play34:30

sure maybe to just add one thought

play34:32

because I recently read an interesting

play34:34

Comon from the niss Norman Foundation

play34:37

which you might know it's like a famous

play34:38

ux research organization so one thing

play34:42

they said about prompting is that you

play34:43

also have to keep in mind that it

play34:45

requires a pretty high degree of

play34:47

literacy to to frame your thoughts in a

play34:49

way that you get the answer you want at

play34:51

least at the moment because you yeah we

play34:53

all know this experience of prompting

play34:55

and then not exactly getting what we

play34:58

wanted and then refining the prompts so

play34:59

so I think one interesting aspect there

play35:01

is also to make to make sure this is

play35:03

very inclusive if language becomes more

play35:05

and more an interface so um thinking

play35:08

about also the developments in research

play35:11

towards integrating like more longtail

play35:13

languages for example um allowing

play35:16

interactions not only in English but

play35:18

like in a diverse set of um languages

play35:20

it's just one strand of research of

play35:22

things we are seeing that still need to

play35:24

happen in the space of NLP next to

play35:26

getting rid of hallucinations or aspects

play35:29

like that but but I think language and

play35:32

extending to to these more long TR

play35:34

languages um is very

play35:36

important thought there um completely

play35:38

agree that obviously prompting in the

play35:40

back right when you're hitting the llms

play35:42

is complex and need to follow a certain

play35:44

structure but on the other hand I'm not

play35:47

expecting from my user to write these

play35:49

prompts right if I'm building an

play35:51

building an application and I'm

play35:52

integrating LMS then then it's on me to

play35:56

make sure the user can use whatever

play35:57

natural language they um used to and

play36:01

then I need to write the prompt that

play36:02

talks to the to the llm directly and

play36:04

make sure that it includes um or makes

play36:08

the llm understand whatever query the

play36:10

user word in the natural language so I

play36:12

don't think that or literacy in terms of

play36:15

languages for sure obviously needs to

play36:17

inclusiv needs to beur but I actually

play36:20

think that lmms have the huge advantage

play36:23

that different types of language

play36:25

literacy in a sense can be considered

play36:28

someone with very complex language can

play36:30

write a switch query and it can be

play36:32

similarly understood to someone with

play36:34

very easy language because the llm can

play36:36

make can actually understand what they

play36:37

want so I'm not so sure about it though

play36:40

I think that's I think that's what we

play36:42

disagree U because I think language is

play36:45

and I'm not just talking about this

play36:46

bubble that we in it's just not the most

play36:47

efficient way to transort information um

play36:51

I can't recall into how many arguments

play36:53

I've gotten on WhatsApp because a

play36:55

message I've written was receiv in a

play36:57

wrong way and if any sentence you say

play36:59

and you put the emphasis on something on

play37:01

a different place it has a completely

play37:03

different meaning I told you that's a

play37:05

nice shirt you might be offended whereas

play37:08

I might actually as a genuine compliment

play37:09

because I also like wuang you know um

play37:12

and and I think this is then the issue

play37:13

that you have when you then interact

play37:14

with these systems where if you say hey

play37:16

I want um last quarter's

play37:19

Revenue what quot are we talking is it

play37:21

the financial year is it being you know

play37:23

the calendar year and and these context

play37:26

is what then need to figure out from

play37:27

person Business site and then also say

play37:30

if you talk about the 90% of

play37:32

consumers what does it mean to them we

play37:35

talked about um it being very accessible

play37:37

and I think this sort of reminds me of

play37:39

um when iOS launched they had this this

play37:41

heavily geomorphic design so your

play37:43

contacts were like like a roller de

play37:45

because we had to sort of accustom

play37:47

people that swiping means you go through

play37:49

something you had to pinch to zoom

play37:51

because well I don't know what this

play37:52

occurs naturally but we had all of these

play37:54

metaphors for very familiar objects and

play37:57

we've moved away from this because now

play37:59

we have more phones and people on this

play38:00

planet and we sort of understand this is

play38:02

the Gen the overall gesture to zoom into

play38:05

something and this means up and down and

play38:07

I think with more technical literacy

play38:09

that we have as as a human race we're

play38:12

also going to move away from these

play38:13

abstractions that might make things

play38:14

actually a bit more difficult my opinion

play38:17

because prompting and getting the right

play38:18

thing is actually not as straightforward

play38:19

as it likely to

play38:21

be abolutely actually it's funny that

play38:23

you took that you um chose the Q3

play38:25

Revenue last quue example because that's

play38:28

exactly the one of the examples I had in

play38:29

my talk earlier right but I believe that

play38:32

with a good semantic understand layer

play38:37

that understands semantic meaning you'll

play38:39

be able to cover that with an L because

play38:41

an LM can either ask you question ask

play38:43

you back hey there are multiple quarters

play38:45

right which one do you mean or quarter X

play38:48

quarter y quarter z um and that's

play38:50

exactly the major advantage in my

play38:53

opinion because you can use the natural

play38:56

language and it can make these

play38:57

connections put them into context maybe

play39:00

see what's available in my database in

play39:02

my knowledge base and um draw

play39:05

conclusions across essentially so yeah

play39:08

completely see the challenge 100% um but

play39:11

that that's really on the prompting

play39:12

don't give the answer directly but um

play39:15

consider all the options or however

play39:18

right okay I I have the feeling that the

play39:21

conversation could go could go on for

play39:23

like U plenty more minutes I have a lot

play39:26

of questions as well but I want to give

play39:28

you as the audience the chance to ask

play39:30

questions if there are any um to our

play39:34

three

play39:35

panelists otherwise as I said I have I

play39:37

have plenty more so um but we are happy

play39:41

to assist you with a microphone and

play39:43

then you can ask

play39:46

him if not uh then I will ask my

play39:49

questions um so we now talked about like

play39:53

the basic interaction with the system

play39:55

but now coming back to M um what do you

play39:58

think will the the role be of llms in

play40:00

this whole uh sphere of interaction with

play40:03

the with the system where will the the

play40:05

role in the future be um will it be for

play40:08

example for every textual interaction

play40:10

will it be an llm or uh where do you see

play40:17

it I think the main benefit of llms at

play40:20

the moment are are the multimodality

play40:22

right they can work with texts but they

play40:24

can also maybe work with tables you

play40:25

canook up to your dat datase they can

play40:27

work images and um in a in a in a

play40:31

scenario maybe 10 20 years from now we

play40:34

could speak that would be is that this

play40:36

sits basically a part of infrastructure

play40:39

it's part of your office suite as part

play40:41

of all the tables that you have and from

play40:43

that generates your overall knowledge

play40:45

base and having essentially your your

play40:47

your your Google llm your marantic llm

play40:49

your llm which essentially is then your

play40:51

go-to person for questions you have and

play40:53

okay listen where are these slides that

play40:55

are built for four years ago and they're

play40:57

oh here they go oh thank you um this is

play41:00

what I see essentially this main core

play41:02

productivity benefit that an llm can

play41:04

offer on a productive scale as part of

play41:07

any organization really I think that's a

play41:08

bit about the accessibility here this

play41:10

can be part of any literally any company

play41:11

it does matter how boring or fashioned

play41:13

you are this can be embedded anywhere

play41:16

right now I think we're not there yet

play41:18

absolutely not because LS are then

play41:20

always constricted to some sort of

play41:21

microcosm that we want to explore maybe

play41:23

that's your own your database for your

play41:25

SES Maybe it's only you know some

play41:27

textual data as part of your HR

play41:28

department whatever um but I think this

play41:31

is missing the Integrity of integration

play41:33

across all departments which is where

play41:35

going to get the main benefit

play41:38

overall and I think textual information

play41:42

is the start of it but then really the

play41:44

underlying value it's always thinking a

play41:46

little bit what are the patterns that

play41:48

where again um technology strength of

play41:51

the technology come together with the

play41:53

business value right what the common

play41:54

patterns you can kind of reuse use use

play41:56

in every single company and I think this

play42:00

um building a semantic understanding for

play42:02

whatever data you're working with

play42:04

whether it's uh structured data whether

play42:07

it's time series data where whether it's

play42:09

textual data whether it's images right

play42:12

connecting all of that in a semantic

play42:14

knowledge base that is huge it's huge so

play42:20

so so sort of sort of a an evolution of

play42:23

of search or organization of information

play42:29

right just like a super quick addition I

play42:32

think what I'm also really looking

play42:33

forward to seeing is more use cases

play42:35

where you you have a true interaction

play42:38

which is kind of both sided so it's not

play42:39

just us learning how to better prompt to

play42:41

get a better result but it's also the

play42:43

model kind of gradually learning as we

play42:45

go along so this is something we are

play42:46

seeing as a trend not only in language

play42:48

but basically overall in machine

play42:50

learning and I think the coolest systems

play42:52

and it's something I'm super excited

play42:54

about in general is whenever you we're

play42:56

at a stage where you provide some input

play42:59

and the model maybe gives a suggestion

play43:01

and then you give some sort of feedback

play43:02

as a human and the model learns from the

play43:04

experts gradually as you advance and

play43:06

then step by step you can automate

play43:08

further essentially and I think that

play43:11

would be great to to see that also um

play43:13

more more with llms in general um but in

play43:16

general I also absolutely agree with

play43:18

what you've said already thanks we're

play43:20

going to take this

play43:21

question thanks for the discussion was

play43:24

really cool and I have

play43:26

question basic like actually it's two

play43:29

questions but I try to compile it in one

play43:32

um at which level of like let's say

play43:36

success rate do you think will llms

play43:38

actually be implemented is that like an

play43:41

do you think that's like an absolute

play43:43

like it shouldn't be like 99% it will

play43:46

give the correct answer and then 1% it's

play43:49

going to give you the wrong answer for

play43:51

like something where a customer asks you

play43:53

like I let's say at the um as a customer

play43:57

service right like and you ask a

play43:58

question how do I do exploit that at

play44:01

which point do you think companies will

play44:04

actually

play44:05

feel safe with llms and Which percentage

play44:09

and how do you ensure that they actually

play44:11

trust them because right now I think

play44:13

there's still like this ah yeah but

play44:15

maybe what about that 1% when they like

play44:18

go off and then they're like how do I I

play44:22

don't know update my iPhone and then it

play44:24

tells you just throw it into to the

play44:26

microwave or something I don't know um I

play44:29

know it's it will be more reasonable

play44:31

than that but like how do you see that

play44:33

um and how do you think or what do you

play44:37

think it's another question how do you

play44:39

think like the whole topic or like the

play44:41

whole argument of l&m's just being

play44:43

stochastic parrots um what do you think

play44:46

about that do you think they are

play44:48

stochastic parrots only or do you think

play44:50

they have like an inner World model that

play44:52

they actually can somewhat reason we

play44:56

already ran out of time but Hannah I

play44:57

would say um you will give give the

play44:59

answer and then um you're you're always

play45:02

welcome to then discuss afterwards but

play45:05

Hannah here you go okay I will I will

play45:07

hurry so that you also get a chance to

play45:08

comment if you like so stochastic parrot

play45:12

yes definitely I mean this is partially

play45:13

what they are built for that's great uh

play45:15

they they they give um the most

play45:17

probabilistic answer which is very very

play45:19

powerful but also comes with these

play45:20

limits of potentially giving wrong

play45:22

answers so I think in general you can

play45:24

think about it in two aess one is how

play45:26

good is the model compared to how good a

play45:28

human could do it and the other is um

play45:31

kind of how sensitive is the use case

play45:33

and whenever um the model is better than

play45:36

human and the case is not that sensitive

play45:39

or for full automation if the model is

play45:42

better than human or close to human but

play45:44

the case is somewhat sensitive you

play45:45

probably already reached the stage where

play45:47

you want to have a human in the loop and

play45:49

then we can have different stages of or

play45:51

different ways of having the human in

play45:53

the loop but you probably always want

play45:55

that and I think

play45:56

if you go for an approach of that it

play45:58

takes out a lot of risk while still

play46:00

utilizing the benefits and this is

play46:01

something I would always recommend to

play46:03

clients I think it's not reasonable to

play46:05

to automate fully if it's a very risky

play46:08

case or a human could do the job

play46:10

better I see someone and yob nodding so

play46:13

we're gonna take this thanks a lot Hanah

play46:15

um so thanks a lot for the uh L lovely

play46:20

discussion and like I said I I have the

play46:22

feeling that we could go on for some

play46:24

more minutes um one that I want to

play46:26

mention definitely is um the park on

play46:28

tomorrow um we we are at C and someone

play46:31

is giving a talk there as well on the

play46:34

interesting topic why causality is the

play46:36

ne next big step in AI uh it is in

play46:38

Cinema 4 at 2:30 so don't miss it I will

play46:41

be definitely there um and uh this

play46:45

leaves me with nothing less than uh

play46:46

expressing my gratitude to you three uh

play46:49

for sharing your knowledge to project a

play46:51

team and Google team and also for your

play46:54

questions and your

play46:56

thanks lot

play47:00

you

Rate This

5.0 / 5 (0 votes)

Related Tags
AI技術LMS進化ビジネス影響パネルディスカッション専門家分析自然言語処理機械学習技術進歩ビジネス価値人間の介入
Do you need a summary in English?