Migrate and modernize your database applications (L300) | AWS Events

AWS Events
5 Jun 202433:40

Summary

TLDRビデオスクリプトは、AWSのデータベースサービスの現代化について解説しています。Barry Morrisが主導するこのトピックでは、データがAIの基礎であり、クラウドへの移行がビジネスの機敏性、スケール、および可用性を高めると語られています。Amazon.comがリレーショナルデータベースからノンリレーショナルデータベースへの移行を通じて、検索、顧客フィードバック、推奨エンジンなどの機能を強化した経験が紹介されています。また、クラウドネイティブのデータベースサービスを使用することで、アプリケーションの柔軟性とスケーラビリティを高め、ビジネスのイノベーションサイクルに再投資することができると強調されています。

Takeaways

  • 🌟 AWSのデータベースサービスは、非関係型データベースを含む多岐に渡るデータベースサービスを提供しており、クラウドへのデータベースの移行とモダン化の過程で重要な役割を果たしている。
  • 🚀 Amazon.comは、従来のリレーショナルデータベース技術から、検索、キーバリュー、グラフ技術などの多様なデータベースに移行し、クラウド上で高いスケーラビリティを実現している。
  • 🔄 クラウドデータベースの移行とモダン化は、企業のビジネスの敏捷性、スケール、および24/7の可用性を高めるために重要なプロセスである。
  • 🛠 従来の3層アプリケーションアーキテクチャから、マイクロサービスやコンテナ化されたアプリケーションへの移行が、クラウドネイティブなアーキテクチャを実現している。
  • 🔑 データベースの選択は、アプリケーションの要件に応じて最適なものを選ぶことが重要で、これによりパフォーマンスと柔軟性が得られる。
  • 🔍 クラウドへの移行において、マルチテナントアーキテクチャを検討することは、コスト効率とリソースの最適化を可能にする。
  • 📈 Amazon Auroraは、従来のMySQLやPostgreSQLに比べて大幅なパフォーマンス向上を提供し、ストレージの自動スケーラビリティを備えている。
  • 🛑 Babble Fishは、SQL ServerからPostgreSQLへの移行を容易にし、ダウンタイムを最小限に抑えることができるサービスである。
  • 📊 Amazon TimestreamやNeptuneなどの特別なデータベースは、特定のユースケースに対応し、ハイパースケーラビリティやリアルタイムのデータ処理を可能にしている。
  • 🔌 AWSでは、データベース移行サービス(DMS)やスキーマ変換ツールなどのツールが提供され、データの移行と分析のためのETLプロセスを簡素化している。
  • 📚 現代のアプリケーションでは、データの分析と機械学習への活用が重要で、AWSはこれらのプロセスをサポートする包括的なサービスを提供している。

Q & A

  • AWSのデータベースサービスのどの部分をBarry Morrisが担当していますか?

    -Barry MorrisはAWSで非関係型データベースを含む多数のデータベースサービスを担当しています。

  • データベースの現代化とはどのような意味を持っていますか?

    -データベースの現代化とはクラウド上でデータベースを最新の技術に合わせて更新し、パフォーマンスや可用性、スケーラビリティを高めることを指します。

  • Amazon.comはどのようにしてデータベースの限界に達しましたか?

    -Amazon.comはリレーショナルデータベース技術をベースにいても、その限界に達し、その後にDynamo DBなどの新しいデータベース技術を導入してスケーラビリティを高めることを決めました。

  • AWSはどのようにしてデータベースの選択と使用を最適化していますか?

    -AWSは適切なデータベースを適切な用途に使用することで、パフォーマンスとアジャイル性を確保しています。これにより、データベースの選択がビジネスのニーズに合わせて最適化されます。

  • マイクロサービスとは何で、なぜクラウド上でスケーラブルですか?

    -マイクロサービスはアプリケーションを小さく、独立したサービスに分割することで、それぞれのサービスが独立してスケールできるアーキテクチャです。クラウド上で実行することで、リソースの柔軟な割り当てが可能でスケーラビリティが高まります。

  • AWSのデータベース移行サービスのデータコレクターとは何ですか?

    -データコレクターはAWSのデータベース移行サービスの一環で、自動的にデータベースのインベントリを収集し、分析することで、移行の計画を容易にします。

  • Amazon Auroraは何で、どのような利点がありますか?

    -Amazon Auroraは互換性のあるリレーショナルデータベースサービスで、MySQLやPostgreSQLと同様のSQLをサポートし、高性能とスケーラビリティを提供します。また、ストレージが自動的にスケーラブルであることが大きな利点です。

  • AWSではどのようにしてデータベースの多テナンシーを管理していますか?

    -AWSでは、単一のアプリケーションイメージとデータベースインスタンスを共有する完全な多テナンシーから、各顧客に専用のインスタンスを提供するまでのスペクトルで管理しています。顧客のニーズに応じて適切なアプローチを選択します。

  • Amazon Dynamo DBはどのような種類のデータベースですか?

    -Amazon Dynamo DBはキーバリューストア型のデータベースで、非常に高いスケーラビリティとパフォーマンスを持ち、大きなデータセットに対する高速アクセスに適しています。

  • AWSのデータベースサービスはどのようにしてビジネスの迅速なイノベーションサイクルに貢献していますか?

    -AWSのデータベースサービスは、データベースの管理や保守の重労働を自動化し、開発者がビジネスロジックやイノベーションに集中できるようにすることで、迅速なイノベーションサイクルに貢献しています。

Outlines

00:00

😀 クラウドデータベースの移行とモダン化

Barry Morris氏はAWSのデータベースサービスのリーダーとして、データがAIの基礎であり、オペレーションデータベースがデータの基礎だと語りました。顧客の多くがクラウドに移行し、クラウドデータベースへの移行を経験しています。Amazon自身もリレーショナルデータベースの限界に達し、多くのデータベースサービスに移行しました。また、現代のデータベースアプリケーションには、 Agility、Scale、および24/7の可用性が求められます。AWSはモノリスではなく、マイクロサービスアーキテクチャに移行し、スケールとコスト管理の両面で利点を享受しています。

05:01

😉 クラウドへの移行とマルチテナンシーの課題

クラウドへの移行は、データセンターの終了やコスト削減、デジタルトランスフォーメーションなどが理由として挙げられます。クラウドへの移行を通じて、顧客はコストを削減し、イノベーションサイクルへの再投資を望んでいます。ISVにとっては、クラウドへのサービス提供に伴いマルチテナンシーの問題があります。単純なマルチテナンシーでは、小規模な顧客に対して過剰なプロビジョニングが行われ、コストがかさんでしまう一方で、完全なマルチテナンシーでは、顧客間でアプリケーションとデータベースを共有するため、複雑さが増します。

10:01

🎯 データベースの分析と移行戦略

データベースの移行は、分析から始まるプロセスです。顧客は新しいアプリケーションを構築するか、既存のアプリケーションを分析し、削減、統合、リフトアンドシフト、またはモダン化を選んで移行します。AWSは、データベース移行サービスのデータコレクターツールを提供し、手動での監査を自動化して、データソースのインベントリを作成します。これにより、顧客はどのデータベースから始めるかのインフォーメーションを得ることができます。

15:03

🛠 データベース移行ツールの活用

AWSのデータベース移行サービス(DMS)とスキーマ変換ツールは、移行プロセスを支援します。スキーマ変換ツールは、顧客が最適なターゲットデータベースを選択し、スキーマ変換やコードオブジェクト変換を支援します。DMSは、データの移動と最小限の停止時間またはほぼゼロの停止時間のためのデータレプリケーションを支援します。また、AWSでは、Aurora PostgresやAurora for MySQLなどのマネージドサービスを提供しており、顧客はこれにより、インフラの管理から解放され、イノベーションに注力できます。

20:05

📈 パフォーマンスとスケーラビリティを兼ね備えるデータベースサービス

AWSは、さまざまなスケーラビリティとパフォーマンスを提供するデータベースサービスを提供しています。Amazon Auroraは、従来のMySQLやPostgreSQLに比べて大幅な性能向上を提供し、ストレージの自動拡張性を備えています。また、DynamoDBはキーバリューストアとして、膨大なスケールでの要求に対応できる能力を有しています。DocumentDBは、ドキュメントデータベースとして、MongoDBと互換性があり、スキーマの柔軟性を提供します。

25:08

🔄 データのリアルタイム分析と移行

リアルタイムデータ分析と移行は、ビジネスにとって重要です。AWSでは、ストリーミングデータのためのKinesis、MSK、バッチ処理のためのEMR、Glueなどのサービスを提供しています。これにより、顧客はデータパイプラインを構築し、データの移動と分析を効率化できます。また、Amazon QuickSightなどのビジネスインテリジェンスツールや、Amazon SageMakerのような機械学習サービスを活用して、データから洞察を得ることができます。

30:09

🚀 ベクターデータストアと生成的AIアプリケーション

生成的AIアプリケーションは、ベクターデータストアを活用して、新しいデータと既存のデータの間のギャップを埋めています。AWSでは、様々なデータベースサービスにベクターデータストアのサポートを提供しています。これにより、開発者は新しい機能を構築する際に、データストアとベクターデータストアを簡単に統合できます。

🛠 移行とモダン化のためのツールとサービス

AWSは、移行とモダン化のプロセスを支援するツールとサービスを提供しています。DMS、Fleet Advisor、Schema Conversion Toolなどがその例です。これらのツールは、データの分析、インベントリの作成、スキーマの変換を支援し、顧客が移行プロセスを円滑に進めることができます。また、AWSは、データ戦略チームやデータドリブンイノベーションセンターなどのプログラムを通じて、顧客のビジネスケースをサポートしています。

Mindmap

Keywords

💡データベース

データベースとは、大量のデータを効率的に管理するためのシステムであり、このビデオでは特にクラウド上で動作するデータベースサービスに焦点が当てられています。データはAIの基礎であり、データベースはデータの基盤を提供します。ビデオでは、AWSのデータベースサービスを率いるBarry Morrisが、顧客がクラウドに移行し、データベースを最新化する過程について話しています。

💡クラウド移行

クラウド移行とは、企業が自社のデータやアプリケーションをインターネット経由でアクセス可能なリモートサーバーに移すプロセスを指します。ビデオでは、クラウドへの移行がデータベースの最新化と密接に関係しており、顧客がクラウドデータベースに移行することで、柔軟性、スケーラビリティ、および可用性の向上が実現可能であると説明されています。

💡NoSQLデータベース

NoSQLデータベースは、非関係型のデータベースであり、ビデオでは特にDynamoDBというAWSのNoSQLデータベースサービスが紹介されています。NoSQLはスケーラビリティや柔軟性に優れており、ビデオではAmazonがリレーショナルデータベース技術からNoSQLへの移行を行い、その利点を強調しています。

💡マイクロサービス

マイクロサービスは、大規模なアプリケーションをより小さな、独立したサービスに分割するアーキテクチャスタイルです。ビデオでは、マイクロサービスがアプリケーションのスケーラビリティや開発者の迅速なイノベーションを可能にする重要な要素として触れられています。

💡スケーラビリティ

スケーラビリティとは、システムが負荷の増加に伴い、より多くのリソースを効果的に活用してパフォーマンスを維持または向上させる能力を指します。ビデオでは、スケーラビリティがクラウドデータベースの主要な利点として強調されており、DynamoDBがこの特性を持ち、膨大なアクセス負荷にも対応できると紹介されています。

💡可用性

可用性は、システムが必要な時に利用可能であることを保証するシステムの性質です。ビデオでは、現代のアプリケーションでは24/7の可用性が求められており、クラウドデータベースはシステム障害や極端な負荷下でも高い可用性を維持する能力を持っていると説明されています。

💡データマイグレーション

データマイグレーションとは、データをあるシステムから別のシステムに移動させるプロセスです。ビデオでは、AWSのDatabase Migration Serviceがデータマイグレーションのプロセスを自動化し、顧客がクラウドに移行するのを助けるツールとして紹介されています。

💡サーバーレスアーキテクチャ

サーバーレスアーキテクチャは、開発者がバックエンドインフラストラクチャの管理に関わる詳細を心配する必要なく、アプリケーションをビルドできるようにするアーキテクチャです。ビデオでは、サーバーレス技術が新しいアプリケーションの構築において技術的な負担を軽減し、開発者のイノベーションを促進する上で役立つと説明されています。

💡キャッシュ

キャッシュは、データのコピーを高速アクセス可能な場所に保存しておくことで、データへのアクセス時間を短縮する技術です。ビデオでは、Elastic CacheやMemoryDBがマイクロ秒レベルのレスポンスタイムを提供し、アプリケーションのパフォーマンスを向上させるために使われると紹介されています。

💡マルチテナンシー

マルチテナンシーは、ソフトウェアアプリケーションが複数の顧客(テナント)のデータを同じインスタンス上で独立して管理できるアーキテクチャスタイルです。ビデオでは、クラウドサービスにおいてマルチテナンシーがコスト効率とリソースの共有を提供し、クラウドへの移行において重要な概念であると説明されています。

Highlights

Barry Morris介绍了自己在AWS数据库服务中的职责,特别是非关系型数据库。

数据是AI的基础,操作数据库是数据的基础。

Amazon自身也经历了从关系型数据库到多种数据库技术的迁移。

现代数据库应用需求包括敏捷性、可扩展性和24/7可用性。

传统应用堆栈正在向微服务和解耦应用转变。

多租户架构是云服务提供商需要考虑的重要问题。

Amazon.com的数据库现代化之旅,包括从单体架构到微服务架构的转变。

客户迁移到云数据库的典型旅程,包括从EC2开始的迁移。

Venki介绍了数据库现代化的不同路径和模式。

使用数据库迁移服务(DMS)和模式转换工具简化迁移过程。

Amazon Aurora提供了与传统数据库相比更高的性能和可扩展性。

Babble Fish工具帮助减少从SQL Server迁移到PostgreSQL的复杂性。

Amazon Dynamo DB能够处理极高规模的请求,如Prime Day。

Amazon Document DB为文档模型提供了高性能的数据库解决方案。

Amazon Neptune和Timestream针对特定数据集提供了专业的数据库解决方案。

数据现代化不仅是数据库的迁移,还包括从交易型数据库到分析型数据库的转变。

AWS提供了多种工具和服务来支持数据库的现代化和迁移。

向量数据库在支持生成性AI应用方面变得越来越重要。

AWS和合作伙伴提供的工具和服务可以帮助规划迁移和现代化之旅。

AWS提供不同的计划和团队来帮助客户在数据现代化方面取得成功。

Transcripts

play00:02

um so um just want to welcome you uh my

play00:06

name is Barry Morris uh I I head up a

play00:08

number of the database services in na

play00:11

AWS uh notably non-relational databases

play00:14

work very closely with a number of the

play00:16

other uh database or all the other

play00:18

database leaders in the

play00:19

organization um so I just wanted what

play00:22

we're going to be talking about today is

play00:24

about databases of course data is the

play00:25

foundation of AI and operational

play00:27

databases are the foundation of data um

play00:30

uh and most of our customers are going

play00:32

through a a sort of a journey of

play00:34

migration onto on on onto the cloud and

play00:37

onto Cloud databases um and um and this

play00:41

is quite a fun thing for us to talk

play00:42

about because uh because Amazon's been

play00:45

through this ourselves um amazon.com not

play00:48

too many years ago was running

play00:49

essentially on relational database

play00:52

technology um and it got to its limits

play00:55

and uh and so what we've done

play00:57

subsequently if you look at the at the

play00:59

at the Amazon

play01:00

uh web page today the amazon.com page

play01:03

behind that application is many

play01:05

databases there's search technology

play01:08

that's driving the search there's

play01:09

there's there's key value technology

play01:11

that's that's driving the the customer

play01:13

feedback stuff there's graph technology

play01:16

that's driving the recommendations and

play01:17

so on and we'll talk a little bit more

play01:19

about that in a minute so this is a

play01:20

journey that we've been on in ourselves

play01:23

um we want to talk through really sort

play01:25

of what it means to modernize databases

play01:28

on on the cloud I'm going to give you a

play01:30

little bit of an overview and Van is

play01:32

going to give us a bit of a a deep dive

play01:35

so um excuse me the first thing is about

play01:39

the these modern modern database

play01:41

applications uh the requirements are are

play01:44

these uh first of all agility because

play01:45

your business is agile you've got a need

play01:48

to move your business forward you need

play01:50

your unlike old style applications that

play01:53

would be built rather like a bridge

play01:54

they'd be built and you move on and

play01:55

build the next Bridge today applications

play01:57

are live you keep working on them you

play01:59

move as far as as you can um and so

play02:02

agility is absolutely key to

play02:04

it um the second piece scale is actually

play02:07

the primary thing that drove Amazon to

play02:09

move to notably adding uh document this

play02:12

Dynamo DB into into amazon.com because

play02:16

uh every time that we every time that we

play02:18

came to a big uh retail day um we had to

play02:21

do enormous planning to make sure that

play02:23

the relational databases could keep up

play02:25

with the likely Peak load um since we

play02:28

put Dynamo DB in there we've never had

play02:30

to even plan for it um because it just

play02:32

scales and so that's really the the

play02:35

essence of this whole kind of

play02:36

modernization thing if you use the right

play02:38

database for the right thing um you get

play02:40

the performance you get the agility and

play02:42

so forth and lastly of course um getting

play02:45

to to to 247 availability even under

play02:49

situations of of system failure or of

play02:51

extreme load

play02:52

Etc so these are the sorts of

play02:55

requirements for mod modern applications

play02:57

um and traditional applications don't

play02:59

typ tyly do

play03:01

that uh on the left you see the sort of

play03:03

typical application stack that's been

play03:05

around for for a couple of decades uh

play03:07

and it's and it's you've got a you've

play03:09

got a presentation layer and you've got

play03:10

a business logic layer and you've got a

play03:11

storage layer uh kind of three tier type

play03:14

applications these are exactly the

play03:16

applications that people are running on

play03:18

premises and uh and which don't tend to

play03:20

scale very well on the cloud of course

play03:23

where we're going to is what's on the

play03:25

right you've still got those same basic

play03:27

layers they're all very different

play03:29

presentation layer is probably much more

play03:30

mobile oriented it's much more it's it's

play03:33

it's it's voice oriented it's natural

play03:34

language oriented perhaps it's targeting

play03:37

special devices set top boxes in in car

play03:39

and so forth um but it's the middle

play03:42

layer that's that's had the big change

play03:44

which is that we've moved towards

play03:45

microservices uh and to so disaggregated

play03:48

applications uh running on kubernetes or

play03:50

whatever uh and the goal there is that

play03:53

every piece of the application can not

play03:55

only scale independently but that it can

play03:57

have separate teams that are working on

play03:59

it and they can be going through their

play04:01

cicd deployment pipelines independently

play04:04

and so you start getting that that kind

play04:05

of application velocity it's absolutely

play04:08

at the Key of it and of course at the

play04:10

bottom each of those microservices will

play04:12

tend to use not only their language of

play04:15

choice and their Frameworks of choice

play04:16

and so forth but they'll tend to use

play04:18

that the database of choice and for some

play04:20

of them it's going to be a mongodb style

play04:23

document database and for some of them

play04:24

it's going to be relational and some of

play04:26

them it's going to be key value for many

play04:27

of them there's going to be a cach on

play04:29

top of it um and the and and what you've

play04:31

got is an architecture that doesn't look

play04:33

anything like that monolithic

play04:34

architecture that you had uh on

play04:38

premises additionally um there are if if

play04:42

you're if you if if you've historically

play04:44

been selling software to on premises

play04:47

customers uh so you're an isv in other

play04:49

words uh and you start wanting to Ser to

play04:52

serve th th those that application as a

play04:54

service on the cloud you've got this

play04:56

issue of multi-tenancy and uh and really

play04:59

this sort of multi tendency thing is

play05:01

that the naive way is that you're going

play05:03

to simply fire up an instance of that

play05:04

application stack uh for every customer

play05:07

and that's what you've got depicted here

play05:09

lots of people do this it's got various

play05:11

strengths and weaknesses single biggest

play05:13

weakness is that you you probably going

play05:16

to over provision for your smallest

play05:17

customers and it's going to cost you a

play05:18

lot of money um the alternative at the

play05:21

other end of the spectrum um is to is to

play05:24

have multi-tenancy full multi tendency

play05:26

where all your customers are sharing the

play05:27

same application image and database

play05:29

instances and so forth um that's hard to

play05:32

do as you're going through this whole

play05:34

process that's something if you're going

play05:35

to be doing that that's a big piece of

play05:37

work that's going to be part of your

play05:40

modernization in fact a lot of customers

play05:42

end up with uh a lot of a lot of of our

play05:45

customers will end up having um a mix of

play05:48

those two things because their largest

play05:50

customers have got so much more in terms

play05:53

of their needs um that it's quite it

play05:55

makes sense for them to have dedicated

play05:56

infrastructure and then the smaller ones

play05:58

to share and so there are advantages and

play06:01

disadvantages to to these two sides of

play06:03

it but the main message from the point

play06:05

of view of modernizing applications and

play06:07

certainly moving on premises to the

play06:09

cloud is think about uh multi- tendency

play06:12

up front it is a big part of of the

play06:15

need we mentioned Amazon uh amazon.com

play06:19

um this as I say is something which uh

play06:22

um is is is enormous scale um the you

play06:26

know this the the numbers of

play06:27

transactions that we have to deal with

play06:29

on a sort of on a busy day or measured

play06:32

in in trillions per day um so very very

play06:35

large uh data loads um and uh and we

play06:38

moved from this monolith we moved to

play06:40

microservices style architecture

play06:42

organizationally as well of course what

play06:45

we did is the typical Amazon thing of

play06:46

having small teams what we call two

play06:48

Pizza teams a team that can be fed by

play06:50

two pizzas um driving each of these

play06:53

microservices and so now you've got

play06:55

something where as I mentioned earlier

play06:57

we don't really have concerns about

play06:59

scale where like we had before uh we

play07:01

feel that we can cope with whatever the

play07:03

Peaks are on on a particular day um

play07:06

because of the of the modernization that

play07:08

we've

play07:09

done and then lastly just before I hand

play07:11

over to Vani let me just sort of picture

play07:14

for you what a typical Journey looks

play07:16

like for a

play07:18

customer um you know typically people

play07:20

are still from the bottom left and

play07:21

they're and they're moving from on

play07:23

premises to probably to ec2 um that'll

play07:26

be a kind of a lift and shift movement

play07:28

they're running essentially identical

play07:30

technology but now on the cloud they got

play07:32

a lot of benefit from that because they

play07:34

can do things like provisioning quickly

play07:36

or move to a bigger machine or add

play07:38

another location or whatever um and they

play07:40

they get onto both of these arrows the

play07:42

one of which is the arrow of innovation

play07:44

velocity and the other of which of

play07:46

course is cost management um and they're

play07:48

onto that but and and there's some

play07:50

customers that are happy to stop there

play07:52

but really what we've been talking about

play07:53

in terms of Amazon and and and and and

play07:55

and some of the more advanced customers

play07:57

they're getting to the far end they've

play07:59

moved through going through moving to

play08:01

manage databases where they no longer

play08:03

have to do the the the sort of what we

play08:05

call undifferentiated heavy lifting of

play08:07

running those database services and

play08:08

doing the backups and the patching and

play08:10

everything else we'll take care of that

play08:12

for you uh once they've done that quite

play08:15

often people that came onto the cloud

play08:16

running on a commercial database will

play08:18

feel well I don't need to do that I can

play08:20

actually run on postgress or my SQL or

play08:22

something so let's me move let me move

play08:25

away from that and ultimately moving to

play08:27

modernization which is to adding elastic

play08:30

Ash to make to because you can do that

play08:32

on the cloud and it's much harder to do

play08:34

on premises or adding Dynamo or adding

play08:36

whatever it is um moving document DB in

play08:39

because you've got content management to

play08:41

do and so forth so really that's what I

play08:43

wanted to do by way of introduction I'm

play08:45

going to hand over to venki um and he

play08:47

can take us into a deep

play08:53

dive thank you thank you

play08:56

Barry so perfect let's let's take a look

play08:59

into the pathways for modernization so

play09:02

what you have learned is customers would

play09:04

like to modernize as they migrate into

play09:06

the cloud right and as you all know

play09:09

modernization is not a process it's a

play09:12

journey and you may have different

play09:14

comping reasons to migrate and most of

play09:16

the common reasons we hear from the

play09:18

customers are data center exit or end of

play09:21

Hardware software licensing or simply

play09:23

for cost

play09:25

savings and the board of directors CTO

play09:28

CFOs

play09:30

are no longer expect uh they no longer

play09:32

expect cost savings although you do

play09:34

expect to save at least 25 to 50% when

play09:37

they move into on uh when they move to

play09:39

AWS when you compare it with on

play09:43

Prem they would also want to know

play09:45

transform the customer experience they

play09:47

would want to digitally transform the

play09:49

business and most importantly they would

play09:52

want to reinvest that cost savings back

play09:54

into the Innovation cycle and that is

play09:57

very important and for that you cannot

play09:59

modernize just the database alone right

play10:01

you know you have to start with the

play10:02

application no matter what so when you

play10:04

start building right you know you have

play10:06

to analyze your existing portfolio of

play10:09

applications now a lot of customers you

play10:12

know we have learned that they start

play10:13

from the scratch by building you know

play10:16

brand new applications mostly they use

play10:18

serverless as their building blocks and

play10:22

basically they are technical depth free

play10:23

right you know and then they look into

play10:26

the existing portfolio of applications

play10:28

they will try to to either Reduce by

play10:30

returing their applications that they no

play10:32

longer want or they try to consolidate

play10:35

by moving into multi tency they try to

play10:37

safy their

play10:39

applications then most of uh the

play10:41

customers would like to do a lift and

play10:43

shift or they migrate to AWS and the

play10:47

last pattern is the modernization right

play10:49

now this is where you would see

play10:51

refactoring and re-platforming and this

play10:53

is where we going to spend some time on

play10:58

it so here is a slide which is probably

play11:02

quite busy but then let's start from the

play11:05

left hand side or right hand side to me

play11:08

uh so that's the application you have

play11:10

databases and windows right you know the

play11:12

first step that most of the customers

play11:14

would like to do is rehost right you

play11:17

know you can rehost the application into

play11:19

the cloud on ec2 for example but then

play11:21

there are requirements from the customer

play11:23

where they do not want to move into

play11:25

Cloud probably due to some stringent

play11:27

requirements uh they can go for public

play11:29

CL if public cloud is not an option then

play11:31

you can go for AWS Outpost you know that

play11:33

is an option for customers to run AWS

play11:36

infrastructure on Prim to get a full

play11:39

hybrid

play11:40

experience and then when you want to

play11:42

completely refactor and rep platform

play11:45

your applications you know you can move

play11:47

into more of a cloud native architecture

play11:49

as you see on the

play11:50

screen so let's say you have to build an

play11:55

event driven architecture or or

play11:57

applications that are already being even

play12:00

driven you can go for Lambda right and

play12:03

then you can go for containers

play12:04

especially if you have any VMS you want

play12:06

to move into Cloud you can go for

play12:08

container especially they're great for

play12:10

monolithic applications or let's say

play12:13

that you want to run and deploy long

play12:15

running

play12:17

applications then let's say you want to

play12:19

have uh more of a controlled environment

play12:22

where you do not want to spend time on

play12:24

managing you could go for ECS which is

play12:27

again a managed container service where

play12:30

you can run deploy your cized

play12:33

applications you also have foget which

play12:36

is a computer engine for managing your

play12:38

uh conization then you have eks where

play12:42

you can again uh manage your conization

play12:46

applications again that is based on open

play12:48

Service open source kubernetes so that's

play12:52

there so from a developer perspective

play12:55

you know they would like to have more

play12:57

agility right you know we do have

play12:59

applications like codar where developers

play13:02

can build deployment workflows on AWS

play13:05

and for continuous delivery you have

play13:08

code builds uh code Pipeline and so on

play13:13

then moving into the data layer you know

play13:15

this is where you have mve to managed

play13:17

and mve to open source where either

play13:20

you're going to move from let's say SQL

play13:22

Server to audio SQL or let's say you

play13:24

want to move away from SQL Server to

play13:26

Aurora postgress so we will spend some

play13:29

time into the next few slides

play13:32

there so what what are the common

play13:34

patterns that we normally see from the

play13:37

customers right so let's say you have uh

play13:41

like thousands of enrollments at your on

play13:43

Prem and you want to make an inventory

play13:45

the most challenging part is like you

play13:47

know you need to do a manual auditing to

play13:48

find out what your data sources are how

play13:51

many databases you have and so on and so

play13:53

forth that could be a task which could

play13:56

take months right in order to make that

play13:58

job easier we have a tool from the

play14:00

database migration service called data

play14:02

collector so with that tool you know

play14:05

it's automated it's going to collect all

play14:07

your inventories probably in Hearts

play14:09

you're going to get a list and that list

play14:11

will have all your environment with the

play14:13

versions the database size and so on so

play14:16

you could make an informed decision of

play14:18

like you know okay which database I need

play14:20

to start with and so on and so forth

play14:22

then let's say you have commercial

play14:24

engines like SQL server and Oracle

play14:25

server and you want to make a decision

play14:28

okay I want to do

play14:29

migrate to an ec2 or I want to move into

play14:32

RDS the first option most of the

play14:34

customers see is ec2 obviously they use

play14:36

native tools to migrate but let's say

play14:39

they have some restrictions and they're

play14:40

unable to use uh Native tools then you

play14:43

have DMS and S what it stands for is

play14:45

database migration service and schema

play14:48

conversion tool so the schema conversion

play14:50

tool is probably the first step where

play14:53

you know it's going to create an

play14:54

assessment report and if you're unsure

play14:57

let's say I don't know where to migrate

play14:58

to from s you know should I pick my SQL

play15:00

or should I pick postris then schema

play15:03

conversion tool would give you an

play15:04

assessment report and that tool is going

play15:07

to tell you okay this is the best Target

play15:09

because the level of effort for you to

play15:10

migrate is probably the least so with

play15:14

that you're going to make a decision and

play15:17

not only that it it's going to also help

play15:19

you do schema conversion it's going to

play15:21

do code objects conversion and the most

play15:24

important part is the data migration

play15:25

which is taken care of by TMS and for

play15:29

minimal downtime or near zero downtime

play15:31

it also helps with data replication as a

play15:34

part of DMS what we call it as

play15:37

CDC the first pattern is pretty obvious

play15:40

that is M to manage so let's say you

play15:42

have an nonpr SQL you then to uh RDS SQL

play15:47

servers which is pretty straightforward

play15:48

you can use DMS in or native tools of

play15:51

your choice to migrate which is

play15:54

great and uh just before going there

play15:57

like you might wonder you know what is

play15:58

the benefit of moving to a manage

play16:00

service you know like Barry said manage

play16:01

services are great if you don't want to

play16:04

spend your time on undifferentiated

play16:05

heavy lift like you know provisioning

play16:07

patching upgrades security and so on so

play16:10

R is as a managed service we give you

play16:13

that experience so that you can focus on

play16:16

Innovation the next pattern which we

play16:18

commonly see from the customer is breakf

play16:20

free that is you know I want to move

play16:23

away from SQL to something else and this

play16:25

is where sat is going to help you you

play16:27

know it's going to create your this

play16:29

assessment report and you're going to

play16:30

move from there so in this case you know

play16:33

I picked Amazon Aurora so for those who

play16:36

is not aware of arur uh so this is

play16:39

compatible with both my SQL and postris

play16:42

which gives you five times more

play16:44

performance than uh vanilla my SQL and

play16:47

three times more performance than

play16:50

postris and the best part is you know

play16:52

the storage is distributed um you know

play16:55

it's automatically uh scalable and so on

play16:58

and so forth so I have two more uh

play17:00

elements there which I will cover after

play17:02

this so that's the babble fish again for

play17:05

those who is not aware of Babble fish

play17:07

you know moving away from SQL Server to

play17:11

let's say aor postris is a task on its

play17:13

own right you know it's timeconsuming is

play17:15

resource

play17:16

intensive but with Babble fish we have

play17:19

reduced the downtime or reduced the

play17:21

effort so what it means is uh let's say

play17:24

you can keep your applications that are

play17:26

written for the SQL Server post can able

play17:30

to understand so basically it can able

play17:31

to understand both tsql and psql so for

play17:35

the new codes that you're developing you

play17:37

can continue writing it on psql and for

play17:39

the existing codes you can take your

play17:40

time to convert and until then you can

play17:43

continue to use your tsql codes so it's

play17:45

basically a machine that can understand

play17:47

both

play17:49

languages and the last most popular

play17:52

option is the purpose B database because

play17:53

what we understood after working several

play17:56

with the customers is that there is not

play17:58

one database that suits for all your use

play18:01

cases right you that isn't a database so

play18:03

although Amazon Dynam DB is what I've

play18:05

used in the screen uh there are multiple

play18:07

databases which I will be talking about

play18:08

in the next few screens and before I go

play18:11

there I want to talk about Aurora

play18:12

Limitless for a bit so you know

play18:15

relational databases can one scale to an

play18:17

extent right whereas aora Limitless

play18:19

gives you an option to scale almost

play18:22

infinitely right you know you with the

play18:23

writer in node for example you can

play18:26

almost able to achieve million

play18:27

transactions per second you know that is

play18:30

unbelievable for a relational database

play18:32

yes you can able to achieve that with

play18:34

purpose build databases but from a

play18:36

relational perspective normally you are

play18:37

bound to hit some of the uh capacity or

play18:41

the limitations which we have taken a

play18:43

feedback and worked on a new tool that

play18:45

is Aurora

play18:49

Limitless So within this modernization

play18:53

uh pattern right you know what we' have

play18:55

seen is adding this caching yes you can

play19:00

work with the databases uh normally what

play19:02

happens is the data is going to be

play19:04

fetched from the disk which is an

play19:05

expensive process we going to put it

play19:07

into the memory and then you're going to

play19:08

read it from the memory which is all

play19:10

right but then as you keep doing you

play19:12

know that is going to add performance

play19:14

penalty to your applications but let's

play19:16

say your developers need microsc latency

play19:19

as a basic requirement then the

play19:22

relational databases can't do

play19:23

justification right and this is where

play19:26

AWS has given you an option that's

play19:28

elastic cache in memory DB now Amazon

play19:31

elastic cache is R is and M cache uh

play19:36

compatible so persistence is an option

play19:39

so you can enable or disable and it

play19:41

gives you microsc uh response times and

play19:45

on the other hand uh Amazon memory DB is

play19:47

fully compatible with redis and it's a

play19:49

persistent inmemory data store right so

play19:52

you get uh you know microsc latency for

play19:55

reads and millisecond latency for rights

play19:57

and keep in mind you can use this

play19:59

database as a persistent inmemory data

play20:02

store especially where the application

play20:05

sensitive application is Sensi to

play20:06

latency this is great and as you see you

play20:10

know you can achieve scalability and

play20:13

Agility with the help of a purpose build

play20:15

database like elastic cache and memory

play20:19

TV then the second subpattern again for

play20:22

non- relational uh I mean for the

play20:24

purpose buil is a non-relational query

play20:25

patterns yes you know you can scale to

play20:28

an ex with the relational databases but

play20:30

what if you have uh a requirement where

play20:33

you have to have an high performance

play20:35

application that you know that has to

play20:38

scale infinitely right such as

play20:40

amazon.com you know when when you have

play20:42

this

play20:42

amazon.com uh sales like Prime day for

play20:45

example uh just to give you a stats

play20:48

Amazon Dynamo DB was able to serve 45

play20:51

trillion requests a day so that is 45

play20:54

million request per second sorry 7

play20:58

trillion million requests uh a day and

play21:00

then 45 million requests per second so

play21:02

that's a lot imagine you know we don't

play21:04

expect customers to be scaling of that

play21:06

size but what I'm trying to say is you

play21:07

know you have a service that can able to

play21:09

scale that much uh so Dynamo DB is built

play21:13

for key value database like I said it

play21:14

can able to perform at any scale and uh

play21:20

so if you have an use case where your

play21:22

developers needed uh you needed you can

play21:26

go for 10B

play21:28

and the second popular engine is the

play21:31

document TB in this diagram uh so again

play21:33

this is a purpose buil database for the

play21:37

data document model so let's say you as

play21:40

a company that's working on a lot of

play21:42

documents be it nested documents that

play21:44

you'd want to query the document or

play21:45

index the document uh then doing it on a

play21:49

relational uh database is probably timec

play21:52

consuming right you know you have to

play21:53

normalize that you have to fit it on a

play21:55

relational table and that's not going to

play21:57

scale well and this is why we have come

play21:59

up with uh Amazon document DB which is

play22:02

fully compatible with mongodb so which

play22:05

means any apis drivers application codes

play22:08

that you're using already with mongodb

play22:10

you can able to use it with Amazon

play22:13

document DB and some of the benefits

play22:16

again same as what we seen before

play22:18

scalability agility and

play22:20

performance and uh the last uh is the

play22:25

specialized data sets what we call so

play22:27

these are like unique cases right so

play22:29

let's say you as a company that uh that

play22:32

is working with highly connected data

play22:34

sets uh let's say you are a company

play22:36

which has millions of followers and you

play22:38

want to establish a relationship between

play22:40

the followers and let's say an athlete

play22:42

for a sports store then uh you know

play22:45

social or sorry the graph database such

play22:48

as Amazon Neptune is a great fit for

play22:50

your use case and for uh time stream

play22:55

imagine you I have a data set uh based

play22:58

on time that's constantly changing and

play23:00

you have like uh GBS of data that's

play23:02

getting ingested per minute uh obviously

play23:05

in relational database can't do

play23:07

justification for that especially when

play23:10

you want to create index on a relational

play23:12

database for that size the indexes are

play23:14

going to be bit clunky right so you

play23:16

don't uh you know it's not going to

play23:18

scale well so this is where Amazon time

play23:20

stream is going to come in really handy

play23:23

and this is a database where you know

play23:26

you can ingest tens of GBS of data

play23:28

permanent and you can able to create

play23:30

that in fraction of

play23:32

second so that's the whole uh uh in a

play23:36

purpose built diagram with different use

play23:38

cases and now let's see what customers

play23:41

do let's say after migrating right so

play23:44

you know you have your applications now

play23:45

as a customer you have made your case

play23:48

you have migrated you have probably

play23:50

modernized some of your

play23:52

databases uh the next step

play23:55

is what to do with this data right now

play23:57

data is a foundation right so you would

play23:59

want to move some of the data from the

play24:01

transactional engine to the analytical

play24:03

engine and probably beyond for AML

play24:06

Innovation and for that what you seen is

play24:09

you know customers normally uh have

play24:11

these developers they would want to use

play24:13

new services but most of the times when

play24:15

they want to use this new service or a

play24:17

plugin they would have to

play24:18

rearchitecturing

play24:28

a service anywhere you can build a

play24:30

pipeline and you can move the data

play24:31

around so for streaming we have Amazon

play24:34

kindnesses and uh

play24:36

msk and for batch Eng you have uh

play24:39

services like TMS you have Amazon EMR

play24:41

and you can use open source Big Data

play24:43

Frameworks like Hive spark Presto and

play24:47

you also have glue where you want to do

play24:49

an ETL kind of you know extract

play24:51

transform and

play24:53

load and then you move your data build

play24:56

your analytics or your data L of that

play24:58

sort and then from there you know you

play25:01

can do a visualization using Amazon

play25:03

quick site which is a business

play25:04

intelligence tool or you can like you

play25:07

know stream it back to other backend

play25:09

applications or you could basically use

play25:13

this data to build train and model your

play25:16

uh machine learning llms right and

play25:18

that's where Amazon sagemaker comes in

play25:20

uh handy and you do also have Bedrock

play25:23

know we have a separate track for that

play25:24

so I'm not going to dive deep into that

play25:27

and one other area which I would want to

play25:28

highlight is the ZL uh here so ZL is

play25:32

another platform that's going to help

play25:34

you move your data with a click of a

play25:36

button from your analytical database

play25:39

sorry from your relational database to

play25:40

the analytical database so from Amazon

play25:42

Aurora to Red shift uh you have the zero

play25:46

zero ETL integration that works uh

play25:49

smoothly uh our next session is on zero

play25:51

ETL so I'm not going to jump too much

play25:53

into it so you can wait and listen to

play25:55

the next session to uh get more about it

play25:58

so these are some of the zero ETL

play25:59

features that's available as of

play26:02

today and again now it's easy you don't

play26:05

your developers don't need to spend time

play26:06

building this whole Pipeline and

play26:08

something breaks you know you don't want

play26:10

to spend time on investigating and

play26:11

fixing it uh we made your life easier

play26:14

and what's even interesting is quite

play26:16

recently we have made uh a new feature

play26:20

where you can take control over your

play26:22

Source database on like which tables

play26:24

that you want to migrate it from olp to

play26:27

oap so that's more on the next

play26:31

session and if you want to go seress

play26:33

obviously you have a lot of options on

play26:34

the application and here is a snapshot

play26:36

of the applications from the database

play26:38

side you have all this that's

play26:40

highlighted in white uh or the database

play26:42

layer of course so you can use uh Sous

play26:46

offerings

play26:48

here so what's next yes you have

play26:51

migrated you have modernized you have

play26:53

innovated like you know uh what's the

play26:55

next popular pattern that we have seen

play26:57

right you know with this whole whole

play26:58

generative era we have seen Vector

play27:01

databases so before I come in here I

play27:03

want to talk about this slide I'll go to

play27:06

that one later so probably have seen

play27:08

this uh slide before but I would want to

play27:11

focus just on the vector data store

play27:13

right you know so Vector data store is

play27:15

an emerging thing uh especially that

play27:18

supports this generative AI

play27:20

applications so again we have a separate

play27:23

topic running on this Vector but just to

play27:25

give you an overview let's say you as a

play27:27

customer or present with a prompt you

play27:29

enter uh a question and you're looking

play27:32

for an answer so what it's going to do

play27:34

is it's going to look into the llm and

play27:36

it's going to fetch and it's going to

play27:37

provide a context as an output but what

play27:41

if the data that you have trained uh was

play27:43

trained on let's say 3 months ago right

play27:47

know and there's a lot of data that's

play27:48

probably has been populated right after

play27:50

that and if you asked any question of up

play27:52

toate data probably you're not going to

play27:53

get any up toate information so how are

play27:56

you going to tackle it and what we call

play27:57

it as h ination in in a world so in

play28:00

order to you know tackle this

play28:02

hallucination problem you have this rag

play28:05

based solution or retrival augumented

play28:07

generation and to support that you have

play28:10

Vector data store so basically you're

play28:11

going to create embeddings and you're

play28:12

going to store into a vector database

play28:15

Ting is nothing but a format that

play28:17

machine can understand to train uh so

play28:21

with this what it's going to happen is

play28:23

uh with all the new data that you're

play28:25

getting from the right hand side we're

play28:27

going to create embeddings we're going

play28:28

to put it into the vector data store and

play28:31

for the questions that's being asked

play28:33

from the user we're going to use the uh

play28:36

context and we going to do a similarity

play28:39

search we're going to get the context

play28:42

and with the given input we're going to

play28:43

add this additional context to the

play28:45

ground Alum so that it gives you an

play28:47

accurate information and this is where

play28:50

Vector databases are popular so I'm

play28:51

going to go back to the previous slide

play28:54

and then I'm going to jump onto that so

play28:56

you might wonder like you know why

play28:58

having this vectors and data store uh

play29:01

are important like why do you want to

play29:02

keep it together yes you do have a

play29:04

separate uh Vector data store but then

play29:07

as a developer you don't want to build

play29:09

another pipeline to move your data from

play29:11

your data store to another Vector data

play29:13

store right you want to keep everything

play29:15

in sync so that you don't you avoid this

play29:17

data moment your developers probably

play29:21

will have to learn a new language to or

play29:23

create a new apas and so on and so forth

play29:25

so you're going to avoid all of this if

play29:27

you are keeping your vector and your

play29:30

databases together so to make that

play29:33

easier I'm going to move now on to the

play29:35

next one sorry for

play29:38

that so we have all of these services

play29:41

that you on the screen that supports

play29:43

Vector database now many of you have

play29:44

already used or some of you have already

play29:46

used the service so we have Vector data

play29:49

store support in all of it right so you

play29:52

can continue using your database as you

play29:55

do and then if you have a new

play29:57

functionality of you are building a new

play29:58

generated U application then you can

play30:01

continue using the vector data store uh

play30:03

in this

play30:06

service so

play30:08

now what are the tools that's available

play30:11

you know we have spoken about everything

play30:12

now what are the tools that's available

play30:15

so you have DMS Fleet adviser which you

play30:17

already seen that helps you basically

play30:19

doing an inventory analysis and you can

play30:22

create a business use case basically to

play30:24

perform this migration and you have

play30:26

schema conversion tool that helps you

play30:28

assessment report schema conversion code

play30:31

conversion and so on and the database

play30:33

migration service itself which would

play30:36

help you to remove the data and helps

play30:39

you replicate then you know we have

play30:40

folks like us essay Partners in the room

play30:44

and outside and uh again internally we

play30:47

have dmas then you have a lot of other

play30:49

programs which you'll be hearing

play30:51

throughout the session here in other uh

play30:53

areas so I'm not going to spend much

play30:54

time on

play30:55

it uh and we spoke about schema

play30:58

conversion tool so I still add a slide

play31:00

here so why is it used like I said you

play31:03

know create this assessment convert

play31:05

schema code objects and not only for

play31:07

relational database as you see on the

play31:09

far right it also helps you extract data

play31:12

warehouse uh data to Amazon red

play31:16

shift so what SS and Target to DMS

play31:20

support as of today so here is a

play31:22

snapshot of that uh and this list is

play31:24

growing so you can take a picture and

play31:27

when we show this probably you'll have

play31:28

new sources next

play31:31

time so when we speak with the customers

play31:34

right now we start with the data

play31:35

flywheel I mean you may have a different

play31:37

data flywheel but then we work with the

play31:39

customers we create uh you know their

play31:41

own custom data flywheel and if you Noti

play31:44

is you know it's made up of few key

play31:45

initiatives and one key initiative

play31:48

doesn't work on its own it has to be a

play31:50

combination for the business to grow in

play31:54

long term uh so the businesses that

play31:57

actually Focus is on each part of this

play31:59

flywheel has a long momentum right you

play32:02

know this is where AWS will work with

play32:04

you Partners will work with you in

play32:06

creating this flywheel and we help you

play32:10

succeed so what are the takeaways you

play32:12

know like I initially said modernization

play32:14

is not a process but rather a journey

play32:17

and you can use the pathways and decide

play32:19

like you know where do I start and where

play32:21

I am and then from there you can move

play32:23

your next step because you know it's

play32:24

it's it's a journey like we said and

play32:27

with tools that we have shown from AWS

play32:30

as well as from Partners you can start

play32:32

planning your migration and

play32:34

modernization

play32:36

journey so we have a lot of programs uh

play32:39

again you would be seeing this in almost

play32:41

all the sessions but just to give you a

play32:43

high level overview if you need P level

play32:46

executive guidance we have data data

play32:47

strategy team so they are basically a

play32:50

former cxos from large Enterprises we

play32:53

will team you up with those uh members

play32:56

and we can help you create a business

play32:57

use case and we have D2 or data driven

play33:00

everything which is again a program

play33:02

where we'll work with your C Level uh

play33:04

and we'll create a business use case or

play33:06

ideation day and we will help you uh

play33:09

achieve that uh for a long-term

play33:12

progress then you have reimagine data

play33:14

and you have AWS generative AI

play33:16

Innovation Center where we will again

play33:18

team you up with uh you know folks with

play33:21

the skill set on that particular domain

play33:23

and we will help you achieve what you

play33:25

want to build in the uh next in ation

play33:28

cycle on the generative

play33:30

Ai and with that I'm I'm going to

play33:33

conclude the presentation thank you very

play33:34

much for listening to our special thing

play33:36

I appreciate that

Rate This

5.0 / 5 (0 votes)

Related Tags
AWSデータベースクラウド移行スケールアジャイルマイグレーションNoSQLDynamoDBAI分析
Do you need a summary in English?