Rising Titan – Mistral's Roadmap in Generative AI | Slush 2023

Slush
5 Dec 202325:39

Summary

TLDRこのビデオスクリプトでは、ロンドンを拠点とするライトスピードのパートナーであるポール・マーフィーが、AI分野での投資経験と、新興企業Mistolの創設者アーサーとの出会いについて語ります。Mistolは、開発者がモデルをカスタマイズし、特定のタスクに適合させることができるようにすることを目指して、AIの基礎モデルを異なるアプローチで構築しています。彼らは既に7Bモデルをリリースし、コミュニティからの強い関心を集めています。アーサーは、オープンソースモデルの重要性、AIの安全性、規制へのアプローチ、そしてAIが社会にもたらすポジティブな変化についても議論しています。最終的に、ヨーロッパにおけるAI分野のリーダーシップの重要性を強調しています。

Takeaways

  • 🌐 Mistolは、AIの基礎モデルを他の会社とは異なる方法で開発し、よりオープンでアクセスしやすいモデルを目指している。
  • 🚀 Mistolは設立からわずか6ヶ月で、効率的なチームワークにより、最先端のモデルを迅速に開発している。
  • 💻 Mistolの戦略は、開発者向けに特化し、より小型で特定のタスクに適したモデルを提供することにある。
  • 🔍 Mistolの7Bモデルは、コミュニティからの高い関心を集め、多様な応用例が見られる。
  • 🤖 Mistolは、AIの安全性やバイアスの問題に対処するために、モデルの重みを公開し、カスタマイズを可能にしている。
  • 📈 Mistolは、将来的にモデルのホスティングサービスや新しい技術、プラットフォームの発表を計画している。
  • 🌍 ヨーロッパにおける人材獲得は競争が激しく、Mistolは最高のエンジニアと科学者の採用に注力している。
  • ⚖️ AIに関する規制や安全性について、Mistolはアプリケーションレイヤーに焦点を当てたアプローチを提唱している。
  • 🌱 AIは医療、教育、創造的思考などの分野で人々の生活を改善する可能性がある。
  • 🇪🇺 ヨーロッパにおけるAI分野のリーダーシップは、地域の価値観に合った技術の発展に不可欠である。

Q & A

  • ライトスピードはいつからヨーロッパで投資を始めましたか?

    -2007年から。

  • ライトスピードはAIにどのくらい投資していますか?

    -約10年間でAIに約50社に投資し、10億ドル以上を投資しています。

  • ミストールの創業者は誰ですか?

    -アーサー、グォム、ティモットです。

  • ミストールはどのようなビジョンを持っていますか?

    -基礎モデルを他の会社とは異なる方法で作成し、開発者がモデルを特化させて独自のものにするためのよりオープンなアクセスを提供することです。

  • ミストールの最初の目標は何でしたか?

    -7Bモデルの構築です。

  • ミストール7Bモデルはどのように迅速に開発されましたか?

    -優秀なチームを組み、データセットに80%のチームを割り当てることで、3か月以内に完成しました。

  • ミストール7Bモデルに対するコミュニティの反応はどうでしたか?

    -非常にポジティブで、多くの開発者がミストール7Bをカスタマイズして新しい機能やアプリケーションを作成しました。

  • ミストールは将来何を計画していますか?

    -新しいモデルと技術を発表し、モデルのホスティング能力を提供するプラットフォームの立ち上げです。

  • ミストールは採用においてどのような課題に直面していますか?

    -最高のエンジニアと科学者を採用することが非常に競争が激しいため、大きな課題です。

  • ミストールはオープンソースとAIモデルの関係をどのように定義していますか?

    -モデルの重みを公開しているため、完全にオープンソースとは言わず、'オープンウェイト'と表現しています。これにより、モデルの修正が可能になりますが、すべてを完全に理解するわけではありません。

Outlines

00:00

🚀 ライトスピードによるAIへの投資とミストールのビジョン

ポール・マーフィーは、ロンドンに拠点を置くライトスピードのパートナーとして、2007年からヨーロッパでの投資活動について紹介し、AI分野への約10年にわたる投資と50社以上への出資を強調しました。彼は、新興企業ミストールとその共同創設者アーサーに対する高い評価を表明し、AIの基盤モデルを異なるアプローチで開発しようとする彼らのビジョンを紹介しました。アーサーは、開発チームの迅速な行動とオープンソースモデルへの取り組みを通じて、開発者がモデルを特化させ、彼ら自身のものにすることを可能にするプラットフォームを構築していると述べました。

05:02

🌍 コミュニティとのエンゲージメントと次のステップ

アーサーは、ミストールの7Bモデルがコミュニティによってどのように受け入れられ、活用されているかを説明しました。特に、開発者が独自のタスクやデータセットにモデルを微調整し、新しい能力やトピックを探求している例を挙げました。さらに、将来のモデルや技術、プラットフォームの展開について触れ、ミストールがモデルホスティングと迅速な推論機能を提供する計画を示唆しました。

10:03

🤖 オープンソースとAIの課題

オープンソースの概念がAIとモデルにどのように適用されるかについて、アーサーは「オープンウェイト」という用語を使用して説明しました。これは、モデルの重みを公開することで、変更の可能性を提供するものですが、完全な透明性や理解を必ずしも意味するわけではありません。彼は、モデルの変更とカスタマイズが重要であり、オープンソースアプローチが科学と技術の進歩を加速させるために重要であると強調しました。また、バイアス制御と解釈可能性の向上についても言及しました。

15:05

⚖️ AI規制と責任の所在

アーサーはAI規制に関して、製品安全、国家安全保障、そして存在リスクという三つの異なる側面を識別しました。役割と責任に関しては、アプリケーション層に圧力をかけることで、間接的にモデルプロバイダーに対する市場圧力を生み出すべきだと主張しました。これにより、より制御可能で安全なモデルの開発が促されると述べています。また、規制に対するアプローチとして、アプリケーション層に焦点を当てることの重要性を強調しました。

20:07

🌟 AIによる未来の可能性

アーサーはAIの持つポジティブな可能性に焦点を当て、特に医療、教育、仕事の方法における革新的な変化を強調しました。AIがクリエイティブな思考を促進し、新しい社会を形成することへの期待を表明しました。さらに、気候変動などの世界的な課題に対してAIがどのように貢献できるかについても触れ、科学的な発見や技術的な進歩を加速させることができると語りました。

25:09

🇪🇺 ヨーロッパにおけるAIのリーダーシップとその重要性

AI技術の進展におけるヨーロッパの位置付けと、地域内での技術的リーダーシップの確立の重要性についてアーサーは言及しました。彼は、ヨーロッパが技術的な進歩を主導し、政策および技術の両面で提案を行うことで、民主主義と社会に対するヨーロッパの価値観を形成していくことの重要性を強調しました。

Mindmap

Keywords

The video is abnormal, and we are working hard to fix it.
Please replace the link and try again.

Highlights

Paul Murphy introduces Light Speed and its investment history in Europe, emphasizing their focus on AI with a portfolio of 50 companies and over a billion dollars invested.

Arthur shares the vision behind Mistol, aiming to make foundational AI models more accessible and customizable for developers, with a focus on open source models.

Mistol's rapid development of a 7B model within three months of funding, highlighting the efficiency and dedication of their team.

The community's enthusiastic reception of Mistol's 7B model, leading to thousands of derivative works and integrations in various open source projects.

Mistol's plans for releasing new models, techniques, and a platform for hosting models with fast inference capabilities.

Challenges in hiring top talent and building a community around Mistol's AI models, emphasizing the importance of community engagement.

Arthur's perspective on AI regulation, advocating for clear distinctions between product safety, national security, and existential risks.

The potential of AI in transforming healthcare and education through personalized and empathetic interactions.

AI's role in addressing climate change by enabling scientific breakthroughs in fields like chemistry and material science.

The significance of having a European champion in AI to shape the technology according to European values and democracy.

Mistol's commitment to open source as a core part of its philosophy, facilitating customization and addressing biases.

The distinction between open source software and open weights in AI, and how Mistol balances competition and collaboration.

The need for independent regulatory bodies to ensure AI safety and prevent regulatory capture by large tech companies.

Arthur's view on existential risks of AI as speculative and the importance of empirical evidence in regulatory discussions.

The future of AI in enabling more creative thinking by automating routine tasks and structuring knowledge.

Transcripts

play00:00

[Music]

play00:17

[Music]

play00:24

[Music]

play00:26

okay um welcome everyone really nice to

play00:29

see you uh very very happy to be back at

play00:31

slush especially this time with Arthur

play00:34

uh my name is Paul Murphy I'm a partner

play00:35

at light speed based in London uh just a

play00:39

real quick uh bit about light speed for

play00:41

those that don't know uh we have

play00:43

actually we're a Silicon Valley based

play00:44

fund but we've been investing Europe uh

play00:46

since 2007 we have over 30 companies now

play00:50

um in

play00:51

Europe uh and um yeah we're investing in

play00:54

pretty much every sector uh and every

play00:57

stage um we're talking about AI today

play00:59

and I think I think it's important to

play01:00

put some context around that from our

play01:02

perspective we actually have been

play01:03

investing in AI for nearly a decade we

play01:05

have about 50 companies um and have

play01:09

invested over a billion dollars into the

play01:11

category and that context is relevant

play01:14

because uh when we met Arthur and his

play01:17

co-founders we thank you we immediately

play01:20

fell in love with the vision of mistol

play01:23

um and so I thought the the best place

play01:25

to to start would be to ask you Arthur

play01:27

to tell us a little bit about what

play01:28

you're building at

play01:30

sure thank you very much slush for the

play01:33

invitation thank you Paul as well um so

play01:36

yeah we started mistal six months ago uh

play01:40

with guom and timot and our vision was

play01:43

that we wanted to make the foundational

play01:46

models a bit differently from the other

play01:49

companies uh we've been in the field for

play01:52

almost a decade now and we've seen it go

play01:54

from a cat and dog detector to something

play01:57

is very close to being humanlike

play01:59

intelligent or atast at least looks like

play02:00

it and we knew that with a very

play02:03

dedicated team we could develop uh

play02:06

state-of-the-art models very very

play02:08

quickly and we could actually take the

play02:11

field into something that is that would

play02:13

be more open where would give more

play02:15

access to developers so that they could

play02:17

specialize the models make them their

play02:19

own make them as small as possible to

play02:21

solve their task and for us the good way

play02:24

of doing it and the good way of starting

play02:25

that was to ship the best open source

play02:28

models create models that would be very

play02:31

easy to to use by individual developers

play02:35

and from then on build onto an

play02:37

Enterprise play to sell a platform that

play02:40

allows developers to take large language

play02:42

models and to make them their own to to

play02:45

create some differentiation on the

play02:47

application they're making and that's an

play02:49

differentiation which is currently hard

play02:51

to do when you only access apis of a

play02:53

couple of providers but if you have a

play02:55

deep access to the models you can create

play02:57

things that are much more interesting

play02:59

and this is what what we want to enable

play03:01

so when we we LED your seed round it

play03:03

wasn't that long ago you told us that

play03:06

you're first thing you're were going to

play03:07

do is to build your 7B model and then I

play03:11

think it was it was like 3 months from

play03:12

when we signed the docs on that round uh

play03:15

we got our message saying hey we're

play03:17

ready uh it's ready and it was faster

play03:19

than we had expected it was already

play03:21

incredibly ambitious I'm just I think

play03:23

everyone's probably wondering how you

play03:25

did how you did that so quickly well I

play03:28

think the secret is to have a a good

play03:29

team uh so we were joined by our first

play03:32

employees uh a dozens of them at the

play03:35

beginning of June and nobody took

play03:37

holidays uh we Rec created the wall what

play03:40

we call the machine learning Ops system

play03:43

so that's actually Fair simple you you

play03:45

need to create a very good Training code

play03:48

base you need to create a very good

play03:50

inference uh code base uh to to deploy

play03:54

the models you need to be able to

play03:55

evaluate the models and the one thing

play03:57

you do need the most and where we

play03:58

actually dedicated 80% of the team on uh

play04:02

for three months is to have some very

play04:04

good data sets so we we went to the open

play04:06

web took public domain knowledge created

play04:09

it so that we could just get the best of

play04:11

it filtered it did everything to get

play04:14

something very good did some work around

play04:17

how to better optimize the models and

play04:19

combine all of this uh and then train

play04:21

the model to get the 7B and we continue

play04:24

doing it uh with the new models we'll be

play04:26

soon announcing like when you say it's

play04:29

fairly easy I think maybe some people

play04:31

would disagree with you on that but you

play04:33

definitely made it look easy I think

play04:34

that's true um so I'm curious uh the

play04:37

community you know was has been very

play04:40

engaged with 7B model Since You released

play04:42

it I think it was you know trending on

play04:44

hugging face for multiple days top you

play04:46

know top top models um what kinds of

play04:49

things have you seen that have been

play04:50

interesting so far from the community so

play04:53

we've seen I think thousands of um D

play04:56

derivative work so uh Developers that

play04:59

took mistal 7B and fine tuned it on

play05:02

their task or on their data sets to make

play05:04

it special so we've seen new

play05:06

capabilities like longer context uh

play05:09

better instruction following capacities

play05:12

uh we've seen uh like new topics so

play05:15

we've seen like occult specialized

play05:17

models able to talk about uh post test

play05:20

experience and the like much better than

play05:22

what MB was able to do before so many

play05:25

kind of different applications uh some

play05:27

of them useful some of them

play05:30

just funny um we've seen integration in

play05:33

a lot of llm Open Source projects so the

play05:36

open source world around Genera T is

play05:39

pretty is is pretty involved already so

play05:41

you have retrieval augmentation systems

play05:44

you have projects that allow to deploy

play05:46

the models on your laptop you have all

play05:48

of these things and they adopted M 7B

play05:50

very quickly and I think it was the

play05:52

field was really missing an actor that

play05:54

would produce the best open source

play05:55

models and actively engage with the

play05:57

community and that's what we we uh we we

play06:00

are enabling okay and so now 7B is out

play06:02

there what comes next so we have um

play06:06

nothing announced yet but we we do have

play06:08

things in house that we'll be announcing

play06:09

before the end of the year uh new models

play06:12

uh new techniques uh and obviously the

play06:15

beginning of a platform so we're

play06:16

actively working on the product uh we'll

play06:19

be soon offering uh hosting capacities

play06:21

for our models uh with very fast uh

play06:24

influence uh capabilities and yeah

play06:27

that's for uh very soon okay okay I'll

play06:30

watch the space um so you're also while

play06:32

you're doing all this incredibly what I

play06:34

think most people would think of as

play06:36

quite challenging technical work you're

play06:38

also building a company and I know

play06:40

that's not easy haven't done it myself

play06:42

before um what's keeping you up at night

play06:44

right now what's your biggest headache

play06:46

um so hiring is obviously a very big

play06:49

challenge I think the only reason why we

play06:51

got there so fast is because we hired

play06:53

the the best engineers and the best

play06:54

scientist in the world it's a very

play06:56

competitive landscape uh Europe is full

play06:59

of talent especially the junior ones uh

play07:02

and so we we are this is some like a

play07:05

very big preoccupation for us like I'm

play07:07

constantly working on it so that's one

play07:09

thing um the other thing is like

play07:11

creating the community engaging with it

play07:14

uh so we started with the with mral 7B

play07:16

but we really need to uh yeah well

play07:20

facilitate the life of our users uh have

play07:23

them engage facilitate Upstream

play07:25

contribution facilitate the emergence of

play07:27

IDs that we could help enable

play07:30

so that's another thing we have a lot of

play07:32

um I guess policy matters uh that we did

play07:36

not expect but obviously this is an

play07:39

agenda that you don't select um there's

play07:43

we we so there's there's different

play07:45

tracks you have in the US you have in EU

play07:48

um we've been uh vocal about the fact

play07:51

that we wanted to have hard regulation

play07:53

on the product side because it's very

play07:54

important and we see ourselves as the

play07:57

provider of tools and a big enabler of

play07:59

compliance for the application makers so

play08:01

we've been saying that uh constantly and

play08:04

and and we've seen like the debate uh

play08:07

progress on these topics and so this is

play08:09

something that yeah we're very keen on

play08:11

trying to enable from a technical

play08:13

perspective because it's important that

play08:15

you have technical Founders that

play08:17

participate in that discussion uh and so

play08:19

that that has kept me up at night uh for

play08:22

for a while and I think you know the

play08:24

ambition was certainly to be able to

play08:25

build something that could rival other

play08:28

large companies like open Ai and I'm

play08:32

just curious what do you view as a

play08:34

differentiating philosophy or approach

play08:36

to companies like open AI I think a

play08:39

differentiating philosophy is that we

play08:41

really Target the developer space and we

play08:44

really think that when you're making an

play08:46

application that you want to put into

play08:47

production you do want to have several

play08:51

specialized models that are as many

play08:54

chips you you should see them as chips

play08:56

that you assemble in an application and

play08:59

it's actually not easy to make a very

play09:01

good chip for the use case you want so

play09:03

you can start with a very big model with

play09:05

thousands of billion well with hundreds

play09:08

of billions of parameters it's going to

play09:09

solve your task maybe but you could

play09:11

actually have something which is 100

play09:13

times smaller and when you make an a a

play09:15

production application that goes at

play09:16

scale and Target a lot of users you want

play09:19

to make the choices that lower the

play09:21

latency lower the costs uh and leverage

play09:23

the actual proprietary data that you may

play09:25

have and this is something that I think

play09:28

that that's not the the topic of our

play09:30

competitors they're really targeting

play09:32

like multi-usage very large models AGI

play09:36

we takeing very much much more pragmatic

play09:38

approach in enabling super useful

play09:40

application today uh that would be cost

play09:42

efficient that would be very low latency

play09:45

and that would enable strong

play09:46

differentiation through uh proprietary

play09:48

data okay and you've talked I think

play09:50

another key difference you've talked a

play09:52

lot about open source as being a core

play09:53

part of your DNA um and I think question

play09:57

I sort of wanted to ask uh Arthur by the

play10:00

way wouldn't look at these questions

play10:01

beforehand so he wasn't expecting this

play10:03

one but I understand the concept of Open

play10:06

Source software I think we all do we see

play10:08

the code you kind of can take it modify

play10:10

it um and use it but in the world of AI

play10:13

and and models the concept of Open

play10:16

Source just feels like it's maybe a bit

play10:17

different because actually some things

play10:19

you do keep for yourself or you have to

play10:21

what does open source mean in the

play10:23

context of llms and AI so we don't

play10:26

really call them open source so the the

play10:28

models we provide are open weight I

play10:30

think it's important to like keep a good

play10:31

distinction between the like the

play10:33

terminology we were using for software

play10:35

and the terminology we are using for

play10:36

models if you provide the weights of a

play10:38

models you're enabling modification

play10:40

you're not necessarily enabling like

play10:43

full understanding of what's going on

play10:45

but even if you do provide full

play10:46

transparency on the data sets and

play10:47

training you don't know what's going on

play10:49

cuz it's it's a bit opaque by Design so

play10:52

it's an empirical science when you

play10:53

create a model the only way to verify

play10:56

that the model is doing what you expect

play10:57

is to measure it with with with

play10:59

evaluation this something will be

play11:00

enabling and then it's to modify it with

play11:03

some signal coming from either humans or

play11:05

maybe machines to to modify the model so

play11:08

really the modification part is super

play11:10

important for differentiation and we're

play11:12

taking this approach there's a full open

play11:14

source approach which I think is very

play11:16

valid as well for science in which you

play11:18

disclose your data set you disclose

play11:20

everything that I think that's that's

play11:23

something that we would strive toward at

play11:24

some point but obviously it's super

play11:26

competitive and the data set part is

play11:28

very hard to to obtain it's also very

play11:31

Capital intensive you need a lot of gpus

play11:33

so right now we're taking a balanced

play11:35

approach in between what we uh opens

play11:38

what the open ways we provide the things

play11:40

we keep for ourselves to to get a

play11:42

competitive uh Edge and this is going to

play11:44

be a dynamic play and we expect it to to

play11:46

evolve with time and with technology

play11:48

okay and then does the does the open

play11:50

weight approach help with other

play11:52

challenges like biases and control yeah

play11:55

so it helps with basically two things

play11:56

the first thing is that you can modify

play11:58

the the biases you can have like a

play12:00

strong and fine uh modification

play12:02

capabilities on the editorial tone on

play12:05

the orientation and alignment of the

play12:08

model so we allow alignment of your own

play12:10

models to your own values and those can

play12:13

slightly differs um so like fine control

play12:16

of biases goes through fine deep access

play12:20

to to models that's the first thing the

play12:22

second thing it allows and we've seen it

play12:24

with active engagement of the AI safety

play12:27

community in particular around open

play12:28

Source models it allows to have better

play12:31

interpretability because you can see the

play12:33

inner activations of the of the models

play12:36

and and that tells you things about

play12:38

what's happening uh about why the model

play12:40

is taking a decision and not another so

play12:42

why is it outputting award and not

play12:44

another and so in the interpretability

play12:46

world it's also super useful it's also

play12:48

and I guess the last thing is that it's

play12:50

very useful to do red teaming because

play12:51

you have a deep access to to the model

play12:53

and so you can try to verify the the

play12:55

part of it which are a bit failing or

play12:57

behaving unexpectedly and these are

play12:59

things that you can then correct very

play13:01

similarly to uh what we've been doing in

play13:03

the open source software for security

play13:05

cyber security and the like okay and

play13:06

then what I mean what is sort of what do

play13:08

you view as at stake here you know why

play13:10

is this is this in other words is this a

play13:12

business Advantage for Mel or is it

play13:14

something more fundamental that you see

play13:16

as almost a

play13:17

responsibility so it's both a business

play13:20

Advantage because we allow further

play13:21

customization and differentiations and

play13:23

it's a very mature market and we expect

play13:26

that on the application space the one

play13:29

actors the actors that are going to

play13:30

survive and create some value are the

play13:32

one that will be able to strongly

play13:33

differentiate themselves and so they

play13:35

would need deep access to models so

play13:37

that's a business differentiator then

play13:39

there's a bit of an ideological

play13:41

differentiators in the sense that I've

play13:43

been contributing to open source for 10

play13:46

years G as well we really think that AI

play13:49

has been accelerated by open science by

play13:51

the circulation of knowledge and that's

play13:53

how we went in 10 years from something

play13:55

very very uh well interesting but that

play13:58

would just detect Tech boats and

play13:59

something that actually uh will speak

play14:02

the human language so this has been

play14:04

allowed because you had big tech labs

play14:07

you had the Academia as well that was

play14:09

all of them were communicating at

play14:11

conferences every every year and and

play14:14

information would circulate and that

play14:15

accelerated things and suddenly in 2020

play14:19

open I decided to stop publishing and it

play14:22

was followed by its competitors uh very

play14:24

closely after and so ever since 2022 we

play14:28

haven't seen like major advances in llm

play14:31

publicly announced and so we've seen

play14:34

currently there's like new architectures

play14:36

that are used internally by our

play14:37

competitors and that are not available

play14:39

out there this is something we will

play14:41

correct very soon okay great um so I

play14:44

want to shift Focus now talk about

play14:46

something you mentioned earlier which is

play14:47

regulation and it's a topic you kind of

play14:49

can't avoid I think you've thinking

play14:51

about AI um a lot of focus within Europe

play14:55

and and in the UK um and I think you at

play14:57

the safety Summit in the the AI safety

play14:59

Summit um last month there's a lot of

play15:02

ideas out there and I think um you know

play15:05

curious to hear your view is to what

play15:07

should be the priority how should

play15:09

regulation be prioritized and

play15:11

instrumented yeah so I think it's quite

play15:14

interest it's a very interesting topic

play15:16

for me and and we've been uh yeah we've

play15:19

been contributing IDs the one thing that

play15:21

I would start with is that we've been

play15:24

talking about regulation and safety and

play15:28

mixing Concepts very heavily so there's

play15:31

a matter of product safety which is

play15:33

answering the question of you deploy a

play15:36

diagnosis assistant in the hospital you

play15:38

want it to be safe you want to be able

play15:40

to measure whether the decision it's

play15:42

making is actually sound is actually

play15:44

correct so that's that's what we call

play15:45

product safety that's something you have

play15:47

when you buy a car you have product

play15:48

safety of your car and it should very

play15:50

much be similar for applications that's

play15:52

one thing and AI to some extent creates

play15:55

new problems because you have models

play15:58

that are not deterministic and so they

play16:00

behave in a potentially an expected way

play16:02

so it's useful to refine the hard lws

play16:04

that we have around uh product safety

play16:06

regulation now there's another topic

play16:09

that came up which is National Security

play16:11

so the question of whether the llms that

play16:13

we're training the LM that everyone is

play16:15

training is spreading too much knowledge

play16:17

so when you have access to llm you're

play16:19

effectively able to educate yourself on

play16:21

many topics and this is something that

play16:23

is a concern for different actors

play16:26

because you could have like small groups

play16:28

that are deemed bad that could use this

play16:30

knowledge to do bad things so this is

play16:33

this has been at the a central topic

play16:35

especially in the US um we're still

play16:37

lacking a lot of there's absolutely no

play16:40

public evidence that llms are

play16:41

facilitating anything so we're really we

play16:44

we've been advocating for for some

play16:46

empirical grounding of the discussion

play16:48

and this is something that's currently

play16:49

very much lacking and then there's a

play16:51

third thing which is kind of mixed with

play16:53

all this with with the two first which

play16:55

is existential risk so knowing whether

play16:58

the technology we're making is

play17:00

effectively on an unbounded exponential

play17:02

that will end up destroying us because

play17:04

as every exponential it kind of breaks

play17:07

the limits at some point and and that's

play17:09

well it becomes IL defined as we say in

play17:12

mathematics so this is something that

play17:14

for us is very much science fiction

play17:16

that's empirical evidences so what we've

play17:19

been saying is that we should really

play17:21

focus on the first topic which is

play17:23

imminent is something that is we do need

play17:25

to have product safety on AI because

play17:27

it's it's going to to otherwise it's

play17:29

going to break trust in the technology

play17:30

we're making and so we want to enable

play17:32

that on the second part we are lacking

play17:34

empirical evidence but I think this is

play17:36

something that we should monitor closely

play17:39

knowledge historically knowledge the

play17:41

spreading of knowledge has always had

play17:42

more benefits than uh than than

play17:45

drawbacks and we AI is not different in

play17:48

that respect but still it's something

play17:50

that that could do with monitoring

play17:51

because it's really new technology on

play17:53

the third aspect of AGI and and and the

play17:57

like and and the fact that that you

play17:58

could have an autonomous system that

play17:59

would go out of control this is

play18:01

something that we are not at heas

play18:03

discussing because we really think that

play18:06

as scientists we are lacking evidence of

play18:08

any existential risk and we think that

play18:09

it pollutes the discussion on the first

play18:11

aspect which is super important yeah and

play18:13

so if I just kind of make sure I

play18:15

understand this right the view is that

play18:18

the application layer is probably the

play18:19

one that has the most responsibility in

play18:22

terms of safety at least to consumers or

play18:24

end users whoever that is businesses but

play18:27

that perhaps the models could provide

play18:28

that as a feature or functionality but

play18:30

it's not the responsibility of the model

play18:32

to ensure that the ultimate data

play18:34

transmitted is itself safe exactly so we

play18:37

think that the correct way of putting

play18:39

some pressure on the model providers uh

play18:41

like us is to effectively say that any

play18:44

application which is deployed and that

play18:46

includes the application that we deploy

play18:48

uh should be should meet a certain

play18:50

number of safety standards so they

play18:53

should do what they're expected to do

play18:55

and if you do that then that means that

play18:57

the application providers will be

play18:59

looking at model providers that are

play19:02

controllable enough that can give some

play19:04

form of guarantees that can give some

play19:06

evaluation Tools around the fact that

play19:08

they're controllable and that they do

play19:10

what they're expected to do so you have

play19:12

some form of second order pressure that

play19:14

is put you put pressure on the

play19:16

application layer and that puts a market

play19:18

pressure on the foundational model

play19:21

developers and that's the correct way of

play19:24

making a healthy competition in making

play19:26

the most controllable models in making

play19:28

the best evaluation tools and making the

play19:30

best Guard railing tools and we think

play19:33

that it's a much better way of doing it

play19:35

than applying directly a pressure on the

play19:38

foundational model layer because if you

play19:39

do that well you're you're in a IL

play19:41

defined territory because you're trying

play19:44

to control something which is by Design

play19:47

super multi-purpose very akin to a

play19:49

programming language so you can't really

play19:51

regulate the programming language

play19:52

because you can do anything with it and

play19:54

so really there's a problem of

play19:55

definition and then there's an

play19:56

operational problem of the fact that if

play19:59

you put some heavy pressure on that

play20:00

layer you're effectively um favoring the

play20:04

big actors that have a lot of compliance

play20:06

capabilities and you're you're making it

play20:09

harder for startups with Innovative IDs

play20:12

to to come up and compete and

play20:14

so this like foundational models is a

play20:17

bad proxy for a market capture and so we

play20:20

believe that applying the regulation

play20:23

pressure on the application layer is the

play20:24

one thing to do because that's going to

play20:26

Foster competition and provide a safer

play20:29

world do you think that there's a role

play20:32

for an or you know an iaea kind of like

play20:35

organization to exist that helps to

play20:37

enforce or provide this guidance

play20:39

regulation so yes I think um this kind

play20:44

of Regulation if we need

play20:46

to monitor I think we we do need to have

play20:49

empirical evidence of what's happening

play20:51

in the space and we need to monitor the

play20:53

product side safety and one way of doing

play20:56

it is to enforce that we have very very

play20:59

independent uh organisms that actually

play21:02

monitor these things and when I say

play21:04

independent I mean that we should be

play21:07

very cautious of of preventing pressure

play21:11

and Regulatory capture of this things so

play21:13

setting standards but ensuring that no

play21:16

big actor is basically writing the

play21:18

standard themselves so what that means

play21:20

that if we are if we if we need to have

play21:24

this this form of organisms they need to

play21:25

be very well funded probably state

play21:27

funded

play21:28

and being completely screen from

play21:32

pressure from the industry okay so now I

play21:35

want to shift you know I think the

play21:36

regulation debate is largely many of the

play21:38

debates in AI are tend to sort of skew

play21:41

somewhat negative so let's dream for a

play21:43

second like how can AI make our lives

play21:46

better what do you see as the utopian

play21:50

future with

play21:51

AI so I think the there's so there's

play21:54

many vertical in which Ai and like

play21:57

interact ING with machines with natural

play22:00

language carry a lot of value uh so

play22:04

Healthcare is going to be completely

play22:05

changed by AI because you you will be

play22:08

able to

play22:09

interact uh with empathic beings uh that

play22:13

are actually super well grounded on

play22:15

statistics and that's really what you

play22:17

were expecting from medicine so we

play22:19

expect that AI is going to empower uh

play22:21

Physicians to be much better at what

play22:23

they're doing and to make better

play22:25

decisions um education is also a super

play22:28

interesting topic uh personalization of

play22:31

Education we know that it's super

play22:33

important to uh take the most most

play22:35

potential of of human beings and having

play22:38

some like your individual teacher being

play22:40

an assistant this is going to change a

play22:42

lot of things especially in the global

play22:44

South um so that's two things generally

play22:47

speaking this is going to change the way

play22:48

we work so it's a way the fact that it

play22:51

can interact with v structure knowledge

play22:53

and that it can do well act as if well a

play22:57

bit imita the boring task of of your

play22:59

daily life this is going to enable more

play23:03

space for creative thinking so we will

play23:06

be able to think more creatively and

play23:08

that's going to unleash I think a new

play23:09

Society very soon and if you think about

play23:11

some of the more existential risks we

play23:13

face in the world like climate change do

play23:16

you think that something like that can

play23:18

be addressed or at least improved yes so

play23:20

I think this is a frontier which which

play23:21

hasn't been completely addressed yet but

play23:23

this is really a promise of having

play23:25

better models the fact that if you if

play23:28

you have some ways of reasoning around a

play23:31

pool of science well you can enable

play23:33

scientists to come up with new ideas you

play23:35

can potentially unlock very precise

play23:37

things like create like in chemistry in

play23:40

accelerating uh chemical reaction so

play23:43

that you emit less CO2 for instance

play23:45

these things like Material Science

play23:47

chemistry Fus nuclear fusion as well all

play23:50

of these locks that we have and that are

play23:53

that we basically need to break in order

play23:55

to address climate change well I mean

play23:57

that's one of the way you can can

play23:57

address climate change obviously the one

play23:59

way is also to reduce consumption but

play24:02

the the these things we we think that AI

play24:05

is going to be an enabler of of of

play24:07

breaking these lcks it's not going to be

play24:09

an easy task uh there's still many

play24:12

things to invent and we think that going

play24:14

through the open science part uh

play24:17

fostering the the the keeping fostering

play24:19

the AI community that drove the field

play24:21

forward for 10 years is is super

play24:23

important to break these logs okay

play24:24

that's great so I think I want to come

play24:26

back to to Europe sort of for our last

play24:28

questions we're we're out of time um how

play24:31

I mean I think the fact that the company

play24:34

is being built in Europe is very

play24:35

important to you it was obvious to you

play24:37

and your co-founders when we invested um

play24:39

how important do you think it is for the

play24:41

industry that we have a European

play24:44

champion emerg in in the field of

play24:46

AI so the Technologies is AI generative

play24:51

AI is is really a wave you can it's

play24:54

going to change society quite

play24:56

significantly and in Europe we have a

play24:59

choice of either being on top of the

play25:00

wave and driving the technology forward

play25:02

or just looking at it happening in the

play25:05

US uh and in China and we think that in

play25:09

order to shape the technology to our

play25:10

values and to the way we think about

play25:12

democracy about Society we need to have

play25:15

very strong technological actors that

play25:17

are able to drive the field forward make

play25:20

proposals um both in term of policy and

play25:23

in term of technology and so that's why

play25:25

we believe it's super important that

play25:27

actors

play25:28

that we have strong actors in Europe

play25:30

great thank you so much Arthur really

play25:31

appreciate it

play25:38

amazing