Rising Titan – Mistral's Roadmap in Generative AI | Slush 2023

Slush
5 Dec 202325:39

Summary

TLDRIn a recent discussion at Slush, Paul Murphy, a partner at Light Speed, interviews Arthur, the co-founder of Mistral, a company focused on developing foundational AI models. Arthur explains that Mistral's vision is to create state-of-the-art models that are accessible to developers, allowing them to specialize and customize models for their specific tasks. They have already released a 7B model and have seen significant community engagement, with developers creating derivative works and integrating the model into various open-source projects. The conversation also touches on the importance of regulation in AI, with a focus on product safety and the need for empirical evidence in discussions about national security and existential risks. Arthur emphasizes the potential of AI to revolutionize sectors like healthcare and education and believes that fostering an open AI community is crucial for addressing global challenges like climate change. He stresses the significance of having strong European actors in AI to drive technological advancements and policy proposals that align with European values.

Takeaways

  • 🤖 **Investment in AI**: Light Speed, a Silicon Valley based fund, has been investing in Europe since 2007 with a focus on AI, investing over a billion dollars into the category.
  • 🚀 **Mistral's Vision**: Mistral, a startup co-founded by Arthur, aims to develop state-of-the-art AI models quickly and provide open access to developers for specialization.
  • 🧠 **Open Source Models**: Mistral's approach is to create easy-to-use open source models that enable developers to tailor large language models for their specific applications.
  • 💾 **Data and Training**: The success of Mistral's 7B model was attributed to a good team, a machine learning Ops system, and a focus on creating high-quality datasets.
  • 🔍 **Community Engagement**: The release of the 7B model led to thousands of derivative works, showcasing new capabilities and integrations within the open source community.
  • 🌐 **Upcoming Developments**: Mistral has plans for new models, techniques, and a platform offering hosting capacities with fast inference capabilities to be announced by year-end.
  • 🧑‍💼 **Building a Company**: Arthur highlights hiring the best talent and community engagement as key challenges while building Mistral.
  • 📈 **Regulation and Safety**: There's a call for hard regulation on the product side for safety and compliance, with a focus on empirical evidence over speculation.
  • 🌱 **Open Science**: Open source principles have accelerated AI advancements, and Mistral aims to continue this tradition by fostering knowledge circulation.
  • 🛡️ **Product Safety**: Arthur suggests that regulation should focus on the application layer, ensuring that deployed applications meet safety standards.
  • ⚖️ **Independence in Regulation**: For effective regulation, there's a need for independent oversight, possibly state-funded, to prevent industry bias and pressure.
  • 🌟 **Positive Impact**: AI has the potential to revolutionize sectors like healthcare and education, enabling more efficient and personalized services.
  • 🌍 **European Leadership**: It's important for Europe to have strong AI actors to drive technological advancements and influence policy to reflect European values.

Q & A

  • What is the name of the company that Paul Murphy is a partner at?

    -Paul Murphy is a partner at Light Speed, a Silicon Valley based fund with investments in Europe since 2007.

  • How long has Light Speed been investing in Europe?

    -Light Speed has been investing in Europe since 2007.

  • What is the focus of the company Mistral, as described by Arthur?

    -Mistral focuses on developing state-of-the-art models quickly and aims to provide open access to these models for developers to specialize and make them their own, thus creating more human-like intelligent applications.

  • What was the first major milestone for Mistral after securing their seed round of funding?

    -The first major milestone for Mistral was to build their 7B model, which they achieved in less than three months.

  • How did Mistral manage to develop their 7B model so quickly?

    -Mistral managed to develop their 7B model quickly by having a good team, creating a machine learning Ops system, focusing on good training and inference code bases, and dedicating a large part of the team to curate and optimize datasets.

  • What has been the community's response to Mistral's 7B model?

    -The community has been very engaged with the 7B model, with thousands of derivative works where developers fine-tuned the model for their specific tasks or datasets, resulting in new capabilities and applications.

  • What are the next steps for Mistral?

    -Mistral is working on new models, techniques, and the beginning of a platform. They plan to offer hosting capacities for their models with fast inference capabilities and are expected to announce these developments before the end of the year.

  • What are the two main challenges that Arthur identifies for building the company?

    -The two main challenges Arthur identifies are hiring the best engineers and scientists in a competitive landscape and creating an engaged community around their open-source models.

  • How does Arthur view the concept of open source in the context of AI models?

    -Arthur differentiates between open source software and open weight models in AI. While providing the weights allows for modification, it doesn't necessarily enable full understanding due to the opacity of the models. He also mentions a balanced approach between openness and maintaining a competitive edge.

  • What is Arthur's perspective on the role of regulation in AI?

    -Arthur believes that regulation should focus on product safety, ensuring that AI applications meet certain safety standards. He also emphasizes the importance of empirical evidence in discussions around national security and existential risks related to AI.

  • How does Arthur envision AI improving society in the future?

    -Arthur sees AI as a tool that can revolutionize healthcare, education, and enable more creative thinking in society. He also believes AI can contribute to addressing existential risks like climate change by unlocking new scientific discoveries.

  • Why is it important for Europe to have a strong presence in the field of AI, according to Arthur?

    -Arthur believes it's crucial for Europe to have strong technological actors in AI to drive the field forward, shape the technology according to European values, and ensure Europe is not just a spectator as the technology transforms society.

Outlines

00:00

🎉 Introduction and Company Vision

Paul Murphy, a partner at Light Speed, introduces himself and the company's investment history in Europe since 2007, highlighting their broad sector and stage investment approach. He emphasizes Light Speed's significant investment in AI, mentioning a decade of experience and over a billion dollars invested in the category. The conversation shifts to Arthur, who shares the story of Mistal's founding six months prior, with a vision to innovate foundational AI models. Arthur outlines their strategy to create open-source models that are user-friendly for developers, allowing them to specialize and optimize the models for their tasks. The rapid development of their 7B model is attributed to a dedicated team and efficient machine learning operations system.

05:02

🚀 Community Engagement and Future Models

The discussion moves to the community's engagement with Mistal's 7B model, which has seen thousands of derivative works and integration into numerous open-source projects. Arthur shares the new capabilities enabled by the model, such as longer context understanding and better instruction following. He also teases upcoming announcements of new models and techniques, hinting at a platform that will offer hosting capacities with fast inference capabilities. The focus then shifts to the challenges of building a company, with hiring and community engagement being top priorities for Arthur.

10:03

🤔 Open Source Philosophy and Differentiation

Arthur clarifies the concept of 'open weight' in the context of AI models, distinguishing it from traditional open-source software. He explains that while the weights of the models are made accessible for modification, full transparency is not always possible due to the models' complexity. The open weight approach is shown to be beneficial for bias control, interpretability, and security through red teaming. Arthur also discusses the business and ideological advantages of open-source, emphasizing the importance of knowledge circulation in accelerating AI advancements. He differentiates Mistal's approach from competitors by targeting developers and focusing on specialized, efficient models that can be customized using proprietary data.

15:05

🌐 Regulation and Safety in AI

The conversation delves into the topic of AI regulation, with a focus on product safety, national security, and existential risk. Arthur advocates for a balanced approach, emphasizing the need for empirical evidence to guide discussions and regulations. He suggests that the application layer should bear the responsibility for safety, with model providers offering controllable models and evaluation tools. Arthur also calls for independent regulatory bodies, possibly state-funded, to monitor AI safety without being influenced by industry pressures.

20:07

🌟 Positive Impacts and Utopian Vision of AI

Arthur envisions a utopian future where AI significantly improves various sectors, such as healthcare and education, by providing personalized assistance and enabling more creative thinking. He also sees AI as a potential tool to address climate change by accelerating scientific research and innovation in fields like chemistry and material science. Arthur stresses the importance of fostering an open AI community to drive advancements and overcome global challenges.

25:09

🇪🇺 The Importance of a European AI Champion

In the final segment, Arthur highlights the importance of establishing a European presence in the AI field. He argues that having a European champion is crucial for shaping technology according to European values and ensuring that the continent is not just a spectator as the AI wave progresses. Arthur sees the development of strong European technological actors as essential for driving policy and technological proposals, influencing the direction of AI globally.

Mindmap

Keywords

💡AI Investment

AI Investment refers to the financial backing provided to companies or projects that are focused on developing or utilizing artificial intelligence technologies. In the video, Light Speed, a Silicon Valley-based fund, has been investing in AI for nearly a decade with over a billion dollars invested in the category, showcasing the significance of AI in modern technology investments.

💡Mistral

Mistral is the name of the company being discussed in the video, which is focused on creating foundational models for AI differently from other companies. The company aims to develop state-of-the-art models quickly and provide more open access to developers, which is a key theme in the discussion about the future of AI development and accessibility.

💡Open Source Models

Open Source Models in the context of AI refer to the models whose underlying code or structure is made publicly available, allowing others to view, modify, and distribute the models. Arthur from Mistral emphasizes the importance of open source in AI, stating that it allows for better customization, differentiation, and community engagement, which is crucial for the evolution and adoption of AI technologies.

💡7B Model

The 7B Model mentioned in the video is a large-scale AI model developed by Mistral. The '7B' likely refers to the number of parameters the model has, indicating its complexity and capacity for learning. The development and release of this model faster than expected is a testament to the company's commitment to rapid innovation in the AI field.

💡Developer Access

Developer Access denotes the level of interaction and modification rights given to developers over AI models. Mistral's approach aims to provide developers with deep access to models, allowing them to specialize and tailor models to their specific needs. This is a core part of Mistral's philosophy to democratize AI and let developers create more efficient and differentiated applications.

💡AI Safety

AI Safety involves the practices and regulations aimed at ensuring that AI systems function correctly, ethically, and without causing harm. In the video, the discussion touches on the importance of AI safety and the need for regulation, reflecting the broader conversation about the ethical considerations and potential risks associated with advanced AI systems.

💡Product Safety

Product Safety in the context of AI refers to the reliability and safety of AI applications when they are deployed in real-world scenarios. The video emphasizes the need for AI applications to meet certain safety standards to build trust and ensure they perform as expected without causing harm or errors.

💡National Security

National Security, as discussed in the video, is a concern related to AI that involves the potential for AI systems to disseminate knowledge that could be misused by malicious actors. It's a part of the broader regulatory debate on AI, where the speakers advocate for empirical evidence and careful monitoring without stifling innovation.

💡Existential Risk

Existential Risk in the context of AI pertains to the theoretical possibility that advanced AI could become uncontrollable and pose a threat to human existence. The video speakers view this as a currently remote and speculative concern, advocating for a focus on more immediate and tangible issues like product safety and ethical AI development.

💡European AI Champion

A European AI Champion refers to a leading AI company or entity based in Europe that can influence and drive the direction of AI technology development according to European values and interests. The video highlights the importance of having such a champion to ensure Europe's active role in shaping the future of AI globally.

💡Regulation Pressure

Regulation Pressure is the influence that regulatory policies and standards have on the development and deployment of AI technologies. The discussion in the video suggests that applying regulation pressure at the application layer rather than the foundational model layer can foster innovation, competition, and safety in the AI industry.

Highlights

Paul Murphy, a partner at Light Speed, discusses the firm's extensive investment in AI, totaling over a billion dollars.

Mistal, a company founded by Arthur and his team, aims to create open-source models that are accessible to developers for specialization.

Mistal's vision is to democratize access to large language models, allowing developers to create more personalized and efficient applications.

The company has rapidly developed a 7B model, showcasing their ability to innovate quickly in the AI space.

Mistal's 7B model has been well-received by the community, leading to thousands of derivative works and specialized applications.

Arthur emphasizes the importance of a dedicated team and good data sets in achieving their rapid development of AI models.

The company is focused on building a platform that offers hosting capacities with fast inference capabilities for their models.

Hiring the best talent is a significant challenge for Mistal, as they aim to stay competitive in the European tech landscape.

Mistal is proactively engaging with policy matters, advocating for hard regulation on the product side to ensure compliance for application makers.

Arthur differentiates Mistal's approach from competitors like OpenAI by targeting developers and enabling them to build specialized, efficient applications.

Open weight models are crucial for Mistal's strategy, allowing for modification and customization to align with developers' values and use cases.

Mistal sees open science as a key driver in the rapid advancements of AI and aims to contribute to this tradition.

The company is committed to addressing biases and ensuring model interpretability through their open weight approach.

Arthur discusses the need for regulation in AI, particularly focusing on product safety and the importance of empirical evidence in policy discussions.

Mistal believes that regulation should primarily target the application layer, fostering competition and promoting safer AI technologies.

The potential for AI to improve sectors like healthcare, education, and address global challenges like climate change is highlighted as a utopian future.

Having a strong European presence in AI is seen as critical for shaping technology according to European values and ensuring a leading role in AI innovation.

Transcripts

play00:00

[Music]

play00:17

[Music]

play00:24

[Music]

play00:26

okay um welcome everyone really nice to

play00:29

see you uh very very happy to be back at

play00:31

slush especially this time with Arthur

play00:34

uh my name is Paul Murphy I'm a partner

play00:35

at light speed based in London uh just a

play00:39

real quick uh bit about light speed for

play00:41

those that don't know uh we have

play00:43

actually we're a Silicon Valley based

play00:44

fund but we've been investing Europe uh

play00:46

since 2007 we have over 30 companies now

play00:50

um in

play00:51

Europe uh and um yeah we're investing in

play00:54

pretty much every sector uh and every

play00:57

stage um we're talking about AI today

play00:59

and I think I think it's important to

play01:00

put some context around that from our

play01:02

perspective we actually have been

play01:03

investing in AI for nearly a decade we

play01:05

have about 50 companies um and have

play01:09

invested over a billion dollars into the

play01:11

category and that context is relevant

play01:14

because uh when we met Arthur and his

play01:17

co-founders we thank you we immediately

play01:20

fell in love with the vision of mistol

play01:23

um and so I thought the the best place

play01:25

to to start would be to ask you Arthur

play01:27

to tell us a little bit about what

play01:28

you're building at

play01:30

sure thank you very much slush for the

play01:33

invitation thank you Paul as well um so

play01:36

yeah we started mistal six months ago uh

play01:40

with guom and timot and our vision was

play01:43

that we wanted to make the foundational

play01:46

models a bit differently from the other

play01:49

companies uh we've been in the field for

play01:52

almost a decade now and we've seen it go

play01:54

from a cat and dog detector to something

play01:57

is very close to being humanlike

play01:59

intelligent or atast at least looks like

play02:00

it and we knew that with a very

play02:03

dedicated team we could develop uh

play02:06

state-of-the-art models very very

play02:08

quickly and we could actually take the

play02:11

field into something that is that would

play02:13

be more open where would give more

play02:15

access to developers so that they could

play02:17

specialize the models make them their

play02:19

own make them as small as possible to

play02:21

solve their task and for us the good way

play02:24

of doing it and the good way of starting

play02:25

that was to ship the best open source

play02:28

models create models that would be very

play02:31

easy to to use by individual developers

play02:35

and from then on build onto an

play02:37

Enterprise play to sell a platform that

play02:40

allows developers to take large language

play02:42

models and to make them their own to to

play02:45

create some differentiation on the

play02:47

application they're making and that's an

play02:49

differentiation which is currently hard

play02:51

to do when you only access apis of a

play02:53

couple of providers but if you have a

play02:55

deep access to the models you can create

play02:57

things that are much more interesting

play02:59

and this is what what we want to enable

play03:01

so when we we LED your seed round it

play03:03

wasn't that long ago you told us that

play03:06

you're first thing you're were going to

play03:07

do is to build your 7B model and then I

play03:11

think it was it was like 3 months from

play03:12

when we signed the docs on that round uh

play03:15

we got our message saying hey we're

play03:17

ready uh it's ready and it was faster

play03:19

than we had expected it was already

play03:21

incredibly ambitious I'm just I think

play03:23

everyone's probably wondering how you

play03:25

did how you did that so quickly well I

play03:28

think the secret is to have a a good

play03:29

team uh so we were joined by our first

play03:32

employees uh a dozens of them at the

play03:35

beginning of June and nobody took

play03:37

holidays uh we Rec created the wall what

play03:40

we call the machine learning Ops system

play03:43

so that's actually Fair simple you you

play03:45

need to create a very good Training code

play03:48

base you need to create a very good

play03:50

inference uh code base uh to to deploy

play03:54

the models you need to be able to

play03:55

evaluate the models and the one thing

play03:57

you do need the most and where we

play03:58

actually dedicated 80% of the team on uh

play04:02

for three months is to have some very

play04:04

good data sets so we we went to the open

play04:06

web took public domain knowledge created

play04:09

it so that we could just get the best of

play04:11

it filtered it did everything to get

play04:14

something very good did some work around

play04:17

how to better optimize the models and

play04:19

combine all of this uh and then train

play04:21

the model to get the 7B and we continue

play04:24

doing it uh with the new models we'll be

play04:26

soon announcing like when you say it's

play04:29

fairly easy I think maybe some people

play04:31

would disagree with you on that but you

play04:33

definitely made it look easy I think

play04:34

that's true um so I'm curious uh the

play04:37

community you know was has been very

play04:40

engaged with 7B model Since You released

play04:42

it I think it was you know trending on

play04:44

hugging face for multiple days top you

play04:46

know top top models um what kinds of

play04:49

things have you seen that have been

play04:50

interesting so far from the community so

play04:53

we've seen I think thousands of um D

play04:56

derivative work so uh Developers that

play04:59

took mistal 7B and fine tuned it on

play05:02

their task or on their data sets to make

play05:04

it special so we've seen new

play05:06

capabilities like longer context uh

play05:09

better instruction following capacities

play05:12

uh we've seen uh like new topics so

play05:15

we've seen like occult specialized

play05:17

models able to talk about uh post test

play05:20

experience and the like much better than

play05:22

what MB was able to do before so many

play05:25

kind of different applications uh some

play05:27

of them useful some of them

play05:30

just funny um we've seen integration in

play05:33

a lot of llm Open Source projects so the

play05:36

open source world around Genera T is

play05:39

pretty is is pretty involved already so

play05:41

you have retrieval augmentation systems

play05:44

you have projects that allow to deploy

play05:46

the models on your laptop you have all

play05:48

of these things and they adopted M 7B

play05:50

very quickly and I think it was the

play05:52

field was really missing an actor that

play05:54

would produce the best open source

play05:55

models and actively engage with the

play05:57

community and that's what we we uh we we

play06:00

are enabling okay and so now 7B is out

play06:02

there what comes next so we have um

play06:06

nothing announced yet but we we do have

play06:08

things in house that we'll be announcing

play06:09

before the end of the year uh new models

play06:12

uh new techniques uh and obviously the

play06:15

beginning of a platform so we're

play06:16

actively working on the product uh we'll

play06:19

be soon offering uh hosting capacities

play06:21

for our models uh with very fast uh

play06:24

influence uh capabilities and yeah

play06:27

that's for uh very soon okay okay I'll

play06:30

watch the space um so you're also while

play06:32

you're doing all this incredibly what I

play06:34

think most people would think of as

play06:36

quite challenging technical work you're

play06:38

also building a company and I know

play06:40

that's not easy haven't done it myself

play06:42

before um what's keeping you up at night

play06:44

right now what's your biggest headache

play06:46

um so hiring is obviously a very big

play06:49

challenge I think the only reason why we

play06:51

got there so fast is because we hired

play06:53

the the best engineers and the best

play06:54

scientist in the world it's a very

play06:56

competitive landscape uh Europe is full

play06:59

of talent especially the junior ones uh

play07:02

and so we we are this is some like a

play07:05

very big preoccupation for us like I'm

play07:07

constantly working on it so that's one

play07:09

thing um the other thing is like

play07:11

creating the community engaging with it

play07:14

uh so we started with the with mral 7B

play07:16

but we really need to uh yeah well

play07:20

facilitate the life of our users uh have

play07:23

them engage facilitate Upstream

play07:25

contribution facilitate the emergence of

play07:27

IDs that we could help enable

play07:30

so that's another thing we have a lot of

play07:32

um I guess policy matters uh that we did

play07:36

not expect but obviously this is an

play07:39

agenda that you don't select um there's

play07:43

we we so there's there's different

play07:45

tracks you have in the US you have in EU

play07:48

um we've been uh vocal about the fact

play07:51

that we wanted to have hard regulation

play07:53

on the product side because it's very

play07:54

important and we see ourselves as the

play07:57

provider of tools and a big enabler of

play07:59

compliance for the application makers so

play08:01

we've been saying that uh constantly and

play08:04

and and we've seen like the debate uh

play08:07

progress on these topics and so this is

play08:09

something that yeah we're very keen on

play08:11

trying to enable from a technical

play08:13

perspective because it's important that

play08:15

you have technical Founders that

play08:17

participate in that discussion uh and so

play08:19

that that has kept me up at night uh for

play08:22

for a while and I think you know the

play08:24

ambition was certainly to be able to

play08:25

build something that could rival other

play08:28

large companies like open Ai and I'm

play08:32

just curious what do you view as a

play08:34

differentiating philosophy or approach

play08:36

to companies like open AI I think a

play08:39

differentiating philosophy is that we

play08:41

really Target the developer space and we

play08:44

really think that when you're making an

play08:46

application that you want to put into

play08:47

production you do want to have several

play08:51

specialized models that are as many

play08:54

chips you you should see them as chips

play08:56

that you assemble in an application and

play08:59

it's actually not easy to make a very

play09:01

good chip for the use case you want so

play09:03

you can start with a very big model with

play09:05

thousands of billion well with hundreds

play09:08

of billions of parameters it's going to

play09:09

solve your task maybe but you could

play09:11

actually have something which is 100

play09:13

times smaller and when you make an a a

play09:15

production application that goes at

play09:16

scale and Target a lot of users you want

play09:19

to make the choices that lower the

play09:21

latency lower the costs uh and leverage

play09:23

the actual proprietary data that you may

play09:25

have and this is something that I think

play09:28

that that's not the the topic of our

play09:30

competitors they're really targeting

play09:32

like multi-usage very large models AGI

play09:36

we takeing very much much more pragmatic

play09:38

approach in enabling super useful

play09:40

application today uh that would be cost

play09:42

efficient that would be very low latency

play09:45

and that would enable strong

play09:46

differentiation through uh proprietary

play09:48

data okay and you've talked I think

play09:50

another key difference you've talked a

play09:52

lot about open source as being a core

play09:53

part of your DNA um and I think question

play09:57

I sort of wanted to ask uh Arthur by the

play10:00

way wouldn't look at these questions

play10:01

beforehand so he wasn't expecting this

play10:03

one but I understand the concept of Open

play10:06

Source software I think we all do we see

play10:08

the code you kind of can take it modify

play10:10

it um and use it but in the world of AI

play10:13

and and models the concept of Open

play10:16

Source just feels like it's maybe a bit

play10:17

different because actually some things

play10:19

you do keep for yourself or you have to

play10:21

what does open source mean in the

play10:23

context of llms and AI so we don't

play10:26

really call them open source so the the

play10:28

models we provide are open weight I

play10:30

think it's important to like keep a good

play10:31

distinction between the like the

play10:33

terminology we were using for software

play10:35

and the terminology we are using for

play10:36

models if you provide the weights of a

play10:38

models you're enabling modification

play10:40

you're not necessarily enabling like

play10:43

full understanding of what's going on

play10:45

but even if you do provide full

play10:46

transparency on the data sets and

play10:47

training you don't know what's going on

play10:49

cuz it's it's a bit opaque by Design so

play10:52

it's an empirical science when you

play10:53

create a model the only way to verify

play10:56

that the model is doing what you expect

play10:57

is to measure it with with with

play10:59

evaluation this something will be

play11:00

enabling and then it's to modify it with

play11:03

some signal coming from either humans or

play11:05

maybe machines to to modify the model so

play11:08

really the modification part is super

play11:10

important for differentiation and we're

play11:12

taking this approach there's a full open

play11:14

source approach which I think is very

play11:16

valid as well for science in which you

play11:18

disclose your data set you disclose

play11:20

everything that I think that's that's

play11:23

something that we would strive toward at

play11:24

some point but obviously it's super

play11:26

competitive and the data set part is

play11:28

very hard to to obtain it's also very

play11:31

Capital intensive you need a lot of gpus

play11:33

so right now we're taking a balanced

play11:35

approach in between what we uh opens

play11:38

what the open ways we provide the things

play11:40

we keep for ourselves to to get a

play11:42

competitive uh Edge and this is going to

play11:44

be a dynamic play and we expect it to to

play11:46

evolve with time and with technology

play11:48

okay and then does the does the open

play11:50

weight approach help with other

play11:52

challenges like biases and control yeah

play11:55

so it helps with basically two things

play11:56

the first thing is that you can modify

play11:58

the the biases you can have like a

play12:00

strong and fine uh modification

play12:02

capabilities on the editorial tone on

play12:05

the orientation and alignment of the

play12:08

model so we allow alignment of your own

play12:10

models to your own values and those can

play12:13

slightly differs um so like fine control

play12:16

of biases goes through fine deep access

play12:20

to to models that's the first thing the

play12:22

second thing it allows and we've seen it

play12:24

with active engagement of the AI safety

play12:27

community in particular around open

play12:28

Source models it allows to have better

play12:31

interpretability because you can see the

play12:33

inner activations of the of the models

play12:36

and and that tells you things about

play12:38

what's happening uh about why the model

play12:40

is taking a decision and not another so

play12:42

why is it outputting award and not

play12:44

another and so in the interpretability

play12:46

world it's also super useful it's also

play12:48

and I guess the last thing is that it's

play12:50

very useful to do red teaming because

play12:51

you have a deep access to to the model

play12:53

and so you can try to verify the the

play12:55

part of it which are a bit failing or

play12:57

behaving unexpectedly and these are

play12:59

things that you can then correct very

play13:01

similarly to uh what we've been doing in

play13:03

the open source software for security

play13:05

cyber security and the like okay and

play13:06

then what I mean what is sort of what do

play13:08

you view as at stake here you know why

play13:10

is this is this in other words is this a

play13:12

business Advantage for Mel or is it

play13:14

something more fundamental that you see

play13:16

as almost a

play13:17

responsibility so it's both a business

play13:20

Advantage because we allow further

play13:21

customization and differentiations and

play13:23

it's a very mature market and we expect

play13:26

that on the application space the one

play13:29

actors the actors that are going to

play13:30

survive and create some value are the

play13:32

one that will be able to strongly

play13:33

differentiate themselves and so they

play13:35

would need deep access to models so

play13:37

that's a business differentiator then

play13:39

there's a bit of an ideological

play13:41

differentiators in the sense that I've

play13:43

been contributing to open source for 10

play13:46

years G as well we really think that AI

play13:49

has been accelerated by open science by

play13:51

the circulation of knowledge and that's

play13:53

how we went in 10 years from something

play13:55

very very uh well interesting but that

play13:58

would just detect Tech boats and

play13:59

something that actually uh will speak

play14:02

the human language so this has been

play14:04

allowed because you had big tech labs

play14:07

you had the Academia as well that was

play14:09

all of them were communicating at

play14:11

conferences every every year and and

play14:14

information would circulate and that

play14:15

accelerated things and suddenly in 2020

play14:19

open I decided to stop publishing and it

play14:22

was followed by its competitors uh very

play14:24

closely after and so ever since 2022 we

play14:28

haven't seen like major advances in llm

play14:31

publicly announced and so we've seen

play14:34

currently there's like new architectures

play14:36

that are used internally by our

play14:37

competitors and that are not available

play14:39

out there this is something we will

play14:41

correct very soon okay great um so I

play14:44

want to shift Focus now talk about

play14:46

something you mentioned earlier which is

play14:47

regulation and it's a topic you kind of

play14:49

can't avoid I think you've thinking

play14:51

about AI um a lot of focus within Europe

play14:55

and and in the UK um and I think you at

play14:57

the safety Summit in the the AI safety

play14:59

Summit um last month there's a lot of

play15:02

ideas out there and I think um you know

play15:05

curious to hear your view is to what

play15:07

should be the priority how should

play15:09

regulation be prioritized and

play15:11

instrumented yeah so I think it's quite

play15:14

interest it's a very interesting topic

play15:16

for me and and we've been uh yeah we've

play15:19

been contributing IDs the one thing that

play15:21

I would start with is that we've been

play15:24

talking about regulation and safety and

play15:28

mixing Concepts very heavily so there's

play15:31

a matter of product safety which is

play15:33

answering the question of you deploy a

play15:36

diagnosis assistant in the hospital you

play15:38

want it to be safe you want to be able

play15:40

to measure whether the decision it's

play15:42

making is actually sound is actually

play15:44

correct so that's that's what we call

play15:45

product safety that's something you have

play15:47

when you buy a car you have product

play15:48

safety of your car and it should very

play15:50

much be similar for applications that's

play15:52

one thing and AI to some extent creates

play15:55

new problems because you have models

play15:58

that are not deterministic and so they

play16:00

behave in a potentially an expected way

play16:02

so it's useful to refine the hard lws

play16:04

that we have around uh product safety

play16:06

regulation now there's another topic

play16:09

that came up which is National Security

play16:11

so the question of whether the llms that

play16:13

we're training the LM that everyone is

play16:15

training is spreading too much knowledge

play16:17

so when you have access to llm you're

play16:19

effectively able to educate yourself on

play16:21

many topics and this is something that

play16:23

is a concern for different actors

play16:26

because you could have like small groups

play16:28

that are deemed bad that could use this

play16:30

knowledge to do bad things so this is

play16:33

this has been at the a central topic

play16:35

especially in the US um we're still

play16:37

lacking a lot of there's absolutely no

play16:40

public evidence that llms are

play16:41

facilitating anything so we're really we

play16:44

we've been advocating for for some

play16:46

empirical grounding of the discussion

play16:48

and this is something that's currently

play16:49

very much lacking and then there's a

play16:51

third thing which is kind of mixed with

play16:53

all this with with the two first which

play16:55

is existential risk so knowing whether

play16:58

the technology we're making is

play17:00

effectively on an unbounded exponential

play17:02

that will end up destroying us because

play17:04

as every exponential it kind of breaks

play17:07

the limits at some point and and that's

play17:09

well it becomes IL defined as we say in

play17:12

mathematics so this is something that

play17:14

for us is very much science fiction

play17:16

that's empirical evidences so what we've

play17:19

been saying is that we should really

play17:21

focus on the first topic which is

play17:23

imminent is something that is we do need

play17:25

to have product safety on AI because

play17:27

it's it's going to to otherwise it's

play17:29

going to break trust in the technology

play17:30

we're making and so we want to enable

play17:32

that on the second part we are lacking

play17:34

empirical evidence but I think this is

play17:36

something that we should monitor closely

play17:39

knowledge historically knowledge the

play17:41

spreading of knowledge has always had

play17:42

more benefits than uh than than

play17:45

drawbacks and we AI is not different in

play17:48

that respect but still it's something

play17:50

that that could do with monitoring

play17:51

because it's really new technology on

play17:53

the third aspect of AGI and and and the

play17:57

like and and the fact that that you

play17:58

could have an autonomous system that

play17:59

would go out of control this is

play18:01

something that we are not at heas

play18:03

discussing because we really think that

play18:06

as scientists we are lacking evidence of

play18:08

any existential risk and we think that

play18:09

it pollutes the discussion on the first

play18:11

aspect which is super important yeah and

play18:13

so if I just kind of make sure I

play18:15

understand this right the view is that

play18:18

the application layer is probably the

play18:19

one that has the most responsibility in

play18:22

terms of safety at least to consumers or

play18:24

end users whoever that is businesses but

play18:27

that perhaps the models could provide

play18:28

that as a feature or functionality but

play18:30

it's not the responsibility of the model

play18:32

to ensure that the ultimate data

play18:34

transmitted is itself safe exactly so we

play18:37

think that the correct way of putting

play18:39

some pressure on the model providers uh

play18:41

like us is to effectively say that any

play18:44

application which is deployed and that

play18:46

includes the application that we deploy

play18:48

uh should be should meet a certain

play18:50

number of safety standards so they

play18:53

should do what they're expected to do

play18:55

and if you do that then that means that

play18:57

the application providers will be

play18:59

looking at model providers that are

play19:02

controllable enough that can give some

play19:04

form of guarantees that can give some

play19:06

evaluation Tools around the fact that

play19:08

they're controllable and that they do

play19:10

what they're expected to do so you have

play19:12

some form of second order pressure that

play19:14

is put you put pressure on the

play19:16

application layer and that puts a market

play19:18

pressure on the foundational model

play19:21

developers and that's the correct way of

play19:24

making a healthy competition in making

play19:26

the most controllable models in making

play19:28

the best evaluation tools and making the

play19:30

best Guard railing tools and we think

play19:33

that it's a much better way of doing it

play19:35

than applying directly a pressure on the

play19:38

foundational model layer because if you

play19:39

do that well you're you're in a IL

play19:41

defined territory because you're trying

play19:44

to control something which is by Design

play19:47

super multi-purpose very akin to a

play19:49

programming language so you can't really

play19:51

regulate the programming language

play19:52

because you can do anything with it and

play19:54

so really there's a problem of

play19:55

definition and then there's an

play19:56

operational problem of the fact that if

play19:59

you put some heavy pressure on that

play20:00

layer you're effectively um favoring the

play20:04

big actors that have a lot of compliance

play20:06

capabilities and you're you're making it

play20:09

harder for startups with Innovative IDs

play20:12

to to come up and compete and

play20:14

so this like foundational models is a

play20:17

bad proxy for a market capture and so we

play20:20

believe that applying the regulation

play20:23

pressure on the application layer is the

play20:24

one thing to do because that's going to

play20:26

Foster competition and provide a safer

play20:29

world do you think that there's a role

play20:32

for an or you know an iaea kind of like

play20:35

organization to exist that helps to

play20:37

enforce or provide this guidance

play20:39

regulation so yes I think um this kind

play20:44

of Regulation if we need

play20:46

to monitor I think we we do need to have

play20:49

empirical evidence of what's happening

play20:51

in the space and we need to monitor the

play20:53

product side safety and one way of doing

play20:56

it is to enforce that we have very very

play20:59

independent uh organisms that actually

play21:02

monitor these things and when I say

play21:04

independent I mean that we should be

play21:07

very cautious of of preventing pressure

play21:11

and Regulatory capture of this things so

play21:13

setting standards but ensuring that no

play21:16

big actor is basically writing the

play21:18

standard themselves so what that means

play21:20

that if we are if we if we need to have

play21:24

this this form of organisms they need to

play21:25

be very well funded probably state

play21:27

funded

play21:28

and being completely screen from

play21:32

pressure from the industry okay so now I

play21:35

want to shift you know I think the

play21:36

regulation debate is largely many of the

play21:38

debates in AI are tend to sort of skew

play21:41

somewhat negative so let's dream for a

play21:43

second like how can AI make our lives

play21:46

better what do you see as the utopian

play21:50

future with

play21:51

AI so I think the there's so there's

play21:54

many vertical in which Ai and like

play21:57

interact ING with machines with natural

play22:00

language carry a lot of value uh so

play22:04

Healthcare is going to be completely

play22:05

changed by AI because you you will be

play22:08

able to

play22:09

interact uh with empathic beings uh that

play22:13

are actually super well grounded on

play22:15

statistics and that's really what you

play22:17

were expecting from medicine so we

play22:19

expect that AI is going to empower uh

play22:21

Physicians to be much better at what

play22:23

they're doing and to make better

play22:25

decisions um education is also a super

play22:28

interesting topic uh personalization of

play22:31

Education we know that it's super

play22:33

important to uh take the most most

play22:35

potential of of human beings and having

play22:38

some like your individual teacher being

play22:40

an assistant this is going to change a

play22:42

lot of things especially in the global

play22:44

South um so that's two things generally

play22:47

speaking this is going to change the way

play22:48

we work so it's a way the fact that it

play22:51

can interact with v structure knowledge

play22:53

and that it can do well act as if well a

play22:57

bit imita the boring task of of your

play22:59

daily life this is going to enable more

play23:03

space for creative thinking so we will

play23:06

be able to think more creatively and

play23:08

that's going to unleash I think a new

play23:09

Society very soon and if you think about

play23:11

some of the more existential risks we

play23:13

face in the world like climate change do

play23:16

you think that something like that can

play23:18

be addressed or at least improved yes so

play23:20

I think this is a frontier which which

play23:21

hasn't been completely addressed yet but

play23:23

this is really a promise of having

play23:25

better models the fact that if you if

play23:28

you have some ways of reasoning around a

play23:31

pool of science well you can enable

play23:33

scientists to come up with new ideas you

play23:35

can potentially unlock very precise

play23:37

things like create like in chemistry in

play23:40

accelerating uh chemical reaction so

play23:43

that you emit less CO2 for instance

play23:45

these things like Material Science

play23:47

chemistry Fus nuclear fusion as well all

play23:50

of these locks that we have and that are

play23:53

that we basically need to break in order

play23:55

to address climate change well I mean

play23:57

that's one of the way you can can

play23:57

address climate change obviously the one

play23:59

way is also to reduce consumption but

play24:02

the the these things we we think that AI

play24:05

is going to be an enabler of of of

play24:07

breaking these lcks it's not going to be

play24:09

an easy task uh there's still many

play24:12

things to invent and we think that going

play24:14

through the open science part uh

play24:17

fostering the the the keeping fostering

play24:19

the AI community that drove the field

play24:21

forward for 10 years is is super

play24:23

important to break these logs okay

play24:24

that's great so I think I want to come

play24:26

back to to Europe sort of for our last

play24:28

questions we're we're out of time um how

play24:31

I mean I think the fact that the company

play24:34

is being built in Europe is very

play24:35

important to you it was obvious to you

play24:37

and your co-founders when we invested um

play24:39

how important do you think it is for the

play24:41

industry that we have a European

play24:44

champion emerg in in the field of

play24:46

AI so the Technologies is AI generative

play24:51

AI is is really a wave you can it's

play24:54

going to change society quite

play24:56

significantly and in Europe we have a

play24:59

choice of either being on top of the

play25:00

wave and driving the technology forward

play25:02

or just looking at it happening in the

play25:05

US uh and in China and we think that in

play25:09

order to shape the technology to our

play25:10

values and to the way we think about

play25:12

democracy about Society we need to have

play25:15

very strong technological actors that

play25:17

are able to drive the field forward make

play25:20

proposals um both in term of policy and

play25:23

in term of technology and so that's why

play25:25

we believe it's super important that

play25:27

actors

play25:28

that we have strong actors in Europe

play25:30

great thank you so much Arthur really

play25:31

appreciate it

play25:38

amazing

Rate This

5.0 / 5 (0 votes)

Related Tags
AI InnovationDeveloper ToolsOpen SourceProduct SafetyRegulationEuropean TechAI EthicsHealthcare AIEducation AIClimate ChangeTech Policy