Using AI in Software Design: How ChatGPT Can Help With Creating a Solution Architecture | R. Müller

iSAQB
6 Feb 202447:49

Summary

TLDRRal Miller, an expert in software and architecture, discusses the utilization of AI in software design, specifically focusing on the capabilities of GPT models. He highlights the importance of context in AI interactions and shares practical tips for effectively using chatbots like GPT for complex tasks such as preparing for the iSAQB Advanced exam. Miller emphasizes the iterative process of refining prompts and the potential of AI as a collaborative tool in architectural decision-making, while also addressing considerations around data protection and the ethical use of AI.

Takeaways

  • 🚀 Ral Miller is a renowned software and architecture expert, author, and accredited trainer known for initiating the open-source project doc toolchain.
  • 📈 Ral discusses the effective use of AI, specifically chatbots like CPT, in software design and creating solution architectures.
  • 💡 The importance of data protection and copyright is highlighted when using AI, emphasizing the need to handle personal and company information with care.
  • 📚 Ral shares his experience and tips on using AI for preparing for the iSAQB Advanced exam, noting the differences between GPT-3 and GPT-4 models.
  • 🌟 GPT-4 is recognized for its superior performance in a medical exam study, showcasing its ability to provide higher quality responses compared to GPT-3.
  • 🧠 The core functionality of AI models like GPT-4 is likened to an all-knowing monkey rather than a statistical parrot, indicating a more complex and knowledgeable system.
  • 🔍 Ral introduces the concept of 'embeddings' and their role in enhancing AI models, allowing them to access and utilize additional information beyond the core neural network.
  • 🔗 The significance of context in AI interactions is discussed, with strategies provided for maintaining context and ensuring effective communication with AI.
  • 📝 Tips for priming AI models are shared, including the use of custom instructions to establish long-term memory and refine the AI's responses.
  • 🛠️ Ral demonstrates the practical application of AI in solving complex tasks, such as the iSAQB Advanced exam, by iteratively working with the AI to generate and refine solutions.
  • 📈 The potential of AI as an architectural co-pilot is emphasized, with Ral noting that while AI can provide a good starting point, human guidance and review are still necessary.

Q & A

  • What is Ral Miller known for in the field of software and architecture?

    -Ral Miller is a well-known software and architecture expert, author, and accredited trainer. He started the open-source project doc toolchain, which focuses on the effective documentation of software architecture. He is also responsible for creating various formats of the arc 42 template.

  • What is the main topic of Ral Miller's session?

    -The main topic of Ral Miller's session is 'Using AI in software design', specifically discussing how chat can assist in creating a solution architecture.

  • What are the potential dangers of using CCK GPT mentioned in the transcript?

    -The potential dangers of using CCK GPT include issues with data protection, both for personal data and company information, and copyright concerns when working with non-public files.

  • How does Ral Miller describe the core functionality of large language models like GPT?

    -Ral Miller describes the core functionality of large language models like GPT as being based on a neural network that performs auto-completion, similar to how a mobile phone predicts text based on probabilities.

  • What is the difference in performance between GPT-3 and GPT-4 as discussed in the transcript?

    -The transcript mentions a study where GPT-4 scored an 82% accuracy rate on a medical exam with roughly 1,000 questions, compared to GPT-3.5 which scored 65% on average and GPT-3 with 75%. This indicates that GPT-4 has a higher quality and more capabilities than GPT-3.

  • How does Ral Miller suggest maintaining context when interacting with chatbots like GPT?

    -Ral Miller suggests starting every session with a prompt that includes a special character like a greater sign (>) to help identify whether the system is still within the context. As long as the context is active, the first paragraph of each output will include the dash. Once the context is lost, the system will no longer display the quote character.

  • What is the role of embeddings in enhancing the capabilities of a neural network like GPT?

    -Embeddings are used to extend the neural network with additional data. They are created from text fragments and stored in a vector database. When interacting with the system, it may query this database for text fragments that fit the prompt, pulling them into the context and enhancing the system's understanding and responses.

  • How does Ral Miller prepare the chatbot for a specific task?

    -Ral Miller prepares the chatbot by providing it with a detailed context, or 'priming', which includes information about his background, the goal of the session, and any specific requirements or constraints. He also uses custom instructions to guide the chatbot's responses and to create a 'long-term memory' for the chatbot.

  • What is the significance of the context size for GPT models?

    -The context size determines how much of the chat history the model takes into account when generating responses. GPT-3 has a context size of 2,000 tokens, while GPT-4 has increased this to 128,000 tokens. A larger context size allows the model to consider more information, potentially leading to more accurate and relevant responses.

  • How does Ral Miller use chat GPT to assist with an advanced certification exam?

    -Ral Miller uses chat GPT to generate solutions for the tasks in the exam by providing detailed prompts and context. He also uses it to understand the task better by asking the chatbot questions and refining the prompts based on the responses. The chatbot helps in creating documents, diagrams, and strategies as part of the exam.

  • What are Ral Miller's views on the future capabilities of chatbots like GPT?

    -Ral Miller believes that chatbots like GPT can serve as an architectural body co-pilot, helping to generate ideas and make decisions. He expects that advancements in the model will lead to an even larger context size, which will help to overcome current limitations. He also sees potential in using GPT for passing oral exams by integrating it with virtual avatars and video technology.

Outlines

00:00

🤖 Introduction to AI in Software Design

Ral Miller, a renowned software and architecture expert, discusses the use of AI in software design. He highlights the importance of considering data protection and copyright when using AI tools like GPT. Ral introduces the concept of chatbots in creating solution architectures and shares his experience with GPT, emphasizing that while GPT is often seen as a stochastic parrot, he views it as an all-knowing monkey due to its extensive data training and ability to provide detailed answers.

05:02

🧠 Understanding the AI Model's Architecture

The speaker delves into the architecture of AI models, comparing GPT-3 and GPT-4, and noting the differences in their training data and parameters. He explains that GPT-4 has a more complex structure and can process multimodal inputs, including images and voice. The discussion includes the concept of 'embeddings' and how they extend the AI's knowledge base beyond its training data, suggesting that the AI model is more than just a neural network.

10:02

📝 The Role of Context in AI Interaction

The importance of context when interacting with AI is emphasized, with the speaker explaining how the AI model uses context to generate accurate responses. He provides tips on maintaining context during a chat session and how the model's context window can affect the quality of its answers. The speaker also discusses the concept of priming the AI with information to ensure it understands the discussion's goals and background.

15:06

🎯 Preparing for the iSAQB Advanced Certification

The speaker shares his approach to preparing for the iSAQB Advanced Certification using AI. He outlines the process of priming the AI with personal and professional context, discussing his preferred programming languages and architectural styles. The speaker demonstrates how to use the AI to understand and work with exam materials, including rules and glossaries, to build a solid foundation for tackling the certification exam.

20:06

📚 AI's Ability to Interpret and Generate Content

The speaker explores AI's capability to interpret and generate content, such as diagrams and tables, from provided documents. He discusses the limitations of AI in reading images within PDFs and shares his experience in guiding the AI to produce the desired outputs. The speaker also highlights the AI's ability to ask relevant questions about a task, showing its active engagement in understanding and solving problems.

25:10

🛠️ Iterative Approach to AI-assisted Task Solving

The speaker describes an iterative approach to solving tasks with AI assistance, where the AI generates parts of the response to refine the solution. He discusses the process of working with limited output space and the need to edit prompts to keep the context concise. The speaker provides examples of how he refined his prompts to achieve better results, such as creating a utility tree and assessing quality scenarios.

30:11

🏗️ Building a Solution Architecture with AI

The speaker details the process of building a solution architecture with AI, discussing the creation of a context diagram, business models, and a technology stack. He shares his experiences in refining prompts to improve the AI's outputs and emphasizes the importance of checking and iterating on the AI's results. The speaker also discusses the potential of AI to serve as an architectural co-pilot, providing valuable insights and options for problem-solving.

35:11

🚀 The Future of AI in Architectural Decision-Making

In the concluding remarks, the speaker reflects on the potential of AI in architectural decision-making. He discusses the advancements in AI models and their ability to transform input into solution outputs. The speaker acknowledges that while AI has limitations and requires guidance, it serves as a valuable tool for generating ideas, making decisions, and identifying errors and inconsistencies in architectural designs.

40:13

💬 Q&A Session with Ral Miller

The Q&A session with Ral Miller covers topics such as the time and effort required to solve problems with AI, the creation of embeddings, and the AI's ability to surprise with new ideas. Ral discusses the ethical considerations of using AI in exams and his experiences with using AI for open-source software inquiries. He also addresses concerns about the use of IP in training sets and the importance of using AI responsibly.

Mindmap

Keywords

💡AI in software design

AI, or Artificial Intelligence, plays a pivotal role in the field of software design by assisting in the creation of solution architectures. In the context of the video, AI is used to enhance the design process through chatbots like GPT, which can generate ideas, make decisions, and provide architectural recommendations. The speaker discusses how AI tools like GPT can be utilized to improve efficiency and innovation in software design, as demonstrated by their application in tackling a complex exam task.

💡Chatbot

A chatbot is an AI-powered conversational agent designed to interact with humans through text or voice interfaces. In the video, the chatbot GPT is used as a tool to assist in the software design process. It is capable of understanding context, generating responses, and even asking questions to clarify or expand on the given task. The chatbot's ability to process natural language and provide meaningful outputs makes it a valuable asset in the design and decision-making process.

💡Solution architecture

Solution architecture refers to the high-level design of a system that outlines the structure, components, and their relationships. It is a blueprint for the software's construction and serves as a guide for developers. In the video, the speaker discusses how AI, specifically through the use of chatbots, can aid in creating effective solution architectures by generating ideas, analyzing requirements, and proposing designs that meet specific objectives.

💡GPT-3 and GPT-4

GPT-3 and GPT-4 are versions of Generative Pre-trained Transformer models developed by OpenAI. These models are advanced language models capable of understanding and generating human-like text based on the input they receive. GPT-4, being the newer version, has been trained on more data and has more parameters, resulting in improved performance and understanding compared to GPT-3. In the video, the speaker contrasts the capabilities of these models and discusses how GPT-4's enhanced abilities make it more effective for tasks like software design and exam problem-solving.

💡Context

In the context of AI and chatbots, context refers to the background information or previous interactions that the system uses to generate relevant and coherent responses. It is crucial for understanding the continuity of a conversation and providing accurate information. The video emphasizes the importance of context in chatbot interactions, as it allows the AI to maintain a consistent and informed dialogue, which is essential for complex tasks like solving an exam problem.

💡Embeddings

Embeddings are a critical concept in natural language processing and AI, where words or phrases are represented as numerical vectors in a high-dimensional space. These vectors capture the semantic meaning of the text, allowing the AI to understand relationships between words and generate contextually relevant responses. In the video, the speaker mentions embeddings as a way to extend the AI's knowledge beyond its training data by incorporating additional text fragments into the chatbot's responses.

💡Prompt engineering

Prompt engineering is the process of crafting effective prompts or questions for AI models to elicit desired responses. It involves carefully designing the input to guide the AI towards specific outcomes, making it a crucial skill when using chatbots for complex tasks. In the video, the speaker emphasizes the importance of prompt engineering in working with GPT, as it allows for better control over the AI's output and helps in achieving the objectives of the task at hand.

💡Data protection

Data protection refers to the measures and policies implemented to ensure the privacy and security of personal and sensitive data. It is a critical concern when using AI tools, as these systems can handle large amounts of information, some of which may be confidential. In the video, the speaker issues a disclaimer about the importance of data protection, cautioning against the use of non-public files and personal data with AI tools like GPT to prevent potential breaches.

💡Advanced software architecture certification

An advanced software architecture certification is a professional qualification that validates an individual's expertise and knowledge in the field of software architecture. The video focuses on the speaker's attempt to use AI to assist in achieving such a certification, specifically through the iSAQB (International Software Architecture Qualification Board) exam. The process involves solving complex architectural tasks and demonstrating a deep understanding of software design principles.

💡Multimodal capabilities

Multimodal capabilities refer to the ability of an AI system to process and understand multiple types of inputs, such as text, images, and voice. This enhances the system's versatility and interaction capabilities. In the video, the speaker mentions that GPT-4 has multimodal capabilities, allowing it to not only work with text but also process images and generate speech, which indicates a more comprehensive and interactive AI experience.

Highlights

Ral Miller is a well-known expert in software and architecture, and the creator of the open-source project doc toolchain.

Ral Miller emphasizes the importance of considering data protection and copyright when using AI in software design.

GPT-4 has been trained on terabytes of text data, resulting in more parameters for providing better answers compared to GPT-3.

GPT-4 scores higher on tests, with an 82% accuracy compared to GPT-3's 65% and human average of 75%.

Ral Miller refers to GPT-4 as an 'all-knowing monkey' rather than a 'statistical parrot' due to its extensive knowledge base and improved capabilities.

GPT-4 has a context size of 128,000 tokens, a significant increase from GPT-3's 32,000 tokens, allowing for a larger window of context.

Ral Miller shares his experience using GPT-4 to prepare for the iSAQB Advanced exam, demonstrating its potential as a study tool.

The speaker discusses the use of embeddings to extend the neural network with additional data, which is stored in a vector database.

Ral Miller provides tips on maintaining context during chat sessions with GPT-4 by starting each session with a specific prompt.

The importance of priming the system with relevant context, such as profession and preferred technologies, is emphasized for effective communication with GPT-4.

Ral Miller demonstrates the ability of GPT-4 to understand and work with complex diagrams and images from PDFs.

The speaker highlights the iterative process of refining prompts and checking the model's understanding to achieve high-quality outputs.

Ral Miller shares his success in using GPT-4 to generate a solution strategy for a software architecture task, showcasing its problem-solving capabilities.

The speaker discusses the potential of GPT-4 as an architectural body co-pilot, aiding in decision-making and providing different perspectives on architecture.

Ral Miller addresses the issue of bias in AI systems and shares his experience with how the system's generated image reflected a bias.

The speaker emphasizes the importance of checking and refining the results produced by GPT-4, as it serves as a starting point rather than a final solution.

Ral Miller concludes that GPT-4, with its advancements, is a valuable tool for transforming input into solution outputs, though it requires guidance and analysis.

Transcripts

play00:00

Ral Miller is a very well-known software

play00:02

architecture expert author and iqb

play00:06

accredited trainer Ral started the

play00:09

open-source project doc toolchain which

play00:12

deals with the effective

play00:14

documentation of software

play00:16

architecture uh he's also responsible

play00:19

for creating the various formats of the

play00:21

arc 42

play00:23

template yeah and enjoy now his session

play00:26

on using AI in software design

play00:31

thank you MCO for the uh introduction so

play00:34

I can skip my first slide hello and

play00:37

welcome to my talk using Ai and software

play00:40

design how chat can help with creating a

play00:43

solution architecture let's skip the

play00:46

slide because you know already know

play00:49

everything about me and um time is

play00:53

precious today because I have lots of

play00:55

slides but first up a

play00:58

disclaimer be aware of the dangers when

play01:01

you use cck GPT keep data protection in

play01:04

mind both in regards of personal data

play01:08

and company

play01:10

information copyright might be an issue

play01:13

when you work with um with CPT if we

play01:16

drop files which are not public

play01:19

available and we will I will show you

play01:23

how to solve the um isq aqb Advanced

play01:28

exam with chpt but don't do this with a

play01:33

real exam you not allowed to do

play01:36

so

play01:38

now another

play01:40

disclaimer I am not an AI expert I am

play01:45

only a prompt engineer I'm a user of

play01:48

chat GPT and um I think I I gathered

play01:52

quite some experience by now so I will

play01:56

show you today my tips and tricks

play02:01

most of the time when um people talk

play02:04

about CPT and other large language

play02:07

medals models they say it's a stochastic

play02:12

parrot because there's a neuron Network

play02:15

at the core of the system and it just

play02:19

yeah does some kind of autoc completion

play02:23

like when you type in text on your

play02:25

mobile phone based on

play02:28

probabilities that's the core of the

play02:31

large language

play02:32

model but I believe it is more the all

play02:36

knowing monkey and why I think so I will

play02:41

now first try to tell you and introduce

play02:45

some Basics about

play02:48

chpt so if you compare the freely aaable

play02:54

CET

play02:55

gpt3 with the not free aaable CG 4 you

play03:00

will notice that there's a huge

play03:03

difference um cpt4 has been trained on

play03:08

lots of data terabytes of Text data by

play03:12

the way most of the information I

play03:14

present here I Googled in the inter on

play03:17

the internet but um there's not much

play03:20

reliable data out there but it gives you

play03:24

some yeah rough estimates

play03:28

so much more data in the training and

play03:32

many more

play03:35

parameters which are needed to give good

play03:39

answers so gpt3 is available for free so

play03:44

many people who talk about CET GPT and

play03:48

who have checked it out are talking

play03:51

about the free

play03:53

model because you have to pay 20 bucks

play03:56

per month to use the GPT 4 model at

play03:59

least at least if you use a um chat GPT

play04:02

4 by open

play04:05

Ai and if

play04:07

you search deeper for uh comparisons

play04:12

here's for instance a study about a

play04:15

medical exam with roughly 1,000

play04:19

questions and here you see the

play04:21

difference between those two models GPT

play04:25

3.5 scores uh 65% of the test humans on

play04:31

average

play04:32

75% GPT 4

play04:36

82% so it has really a different

play04:40

quality the GPT 4 model than the gpt3

play04:46

model and it is more than just the

play04:50

neural

play04:51

network but it's hard to find details

play04:55

about it um it's based on the

play04:58

Transformer model architecture and you

play05:01

find this picture on the internet and

play05:05

yes we are Architects and we should be

play05:07

able to read this uh this diagram but um

play05:12

for me it's not so yeah doesn't give me

play05:17

so much meaning so much background as I

play05:21

expected so for this talk I came up with

play05:25

my own diagram and it's um yeah

play05:30

partly guesses it's quite abstract but

play05:34

you will notice what I want to tell you

play05:37

at the core there's a neuron

play05:40

Network and you input text to the model

play05:45

and before this text hits the neural

play05:48

network we already have a natural

play05:52

language processor so it's not only the

play05:55

neural network at the core but there's

play05:57

something in front of it

play06:00

and then when we output data there's an

play06:04

output

play06:05

processor and then we get our text

play06:09

output and um there's also the system

play06:13

also feeds context to the NLP

play06:18

processor we will talk in a minute about

play06:21

the

play06:23

context and it seems to be not only one

play06:27

neural network in at core but a mixture

play06:32

of

play06:33

experts so in the past when um you asked

play06:37

a question how to mathematical question

play06:42

what's 1 +

play06:44

one it went through the normal neuronal

play06:47

Network and it had some problems with

play06:51

bigger numbers or more complex

play06:54

calculations now at the core seems to be

play06:57

this mixture of experts so something

play07:02

decides uh to which expert model the

play07:06

question should be

play07:08

rooted and so it gained much more

play07:11

knowledge and can um lead to better

play07:15

answers but it also now is a

play07:20

multimodal um yeah tool because it not

play07:25

only works on text but you can now also

play07:28

drop images which are then processed and

play07:33

fed to the NLP

play07:36

processor and even um you can ask it to

play07:39

Output images through Deli

play07:43

3 and voice you can speak to

play07:49

it and it can generate speech so you can

play07:54

um not only type in your questions but

play07:57

you can just um work with it with your

play08:00

voice there's a Code

play08:03

interpretor uh which is capable of um

play08:06

yeah doing using Python scripts to work

play08:10

on your

play08:12

prompt and there are many more plugins

play08:16

aailable which makes the model quite

play08:19

powerful so you see it's not only just a

play08:23

neon Network at a core it's many more of

play08:28

it and there was one article where they

play08:31

stated that jet gptt 4 has around

play08:36

110 layers whatever this means I think

play08:40

it's not the layer of the neural network

play08:44

it's the

play08:45

layers before and after

play08:48

it but as I said I'm not the AI expert

play08:53

I'm the prompt

play08:55

engineer so let's talk about context

play09:00

when you chat with the

play09:03

system you type in your text and first

play09:07

should be some kind of priming you

play09:11

should give the system some context who

play09:14

you are what you know and what you're

play09:17

are going to do in this example hey

play09:21

let's work on a web- based application

play09:23

with spring Boot and

play09:26

MySQL and then you chat with a system

play09:30

and

play09:32

everything which is here in the chat in

play09:36

the history will be part of the context

play09:39

of the system will be fed to the

play09:41

neuronal network this is quite important

play09:46

because if you continue your

play09:48

chat there's a window of

play09:51

context and

play09:53

now some messages has left this context

play09:58

will not be fed anymore to the

play10:02

model and this means that your first

play10:05

statements the priming leaves the

play10:08

context and the system doesn't know

play10:10

anymore what we are talking about and it

play10:14

will start to give wrong

play10:17

answers so how important is this

play10:22

fact chck gpt3 has a context of 2,000

play10:27

tokens GPT 4 or has a context yeah in

play10:31

the past of uh 32,000

play10:36

tokens since roughly four weeks um they

play10:41

increased managed to increase the

play10:44

context to

play10:47

128,000 tokens a token is not a word

play10:51

it's a part of a word so those um 32,000

play10:56

tokens are roughly 20,000 worth

play11:00

so it takes some time to get out of this

play11:04

context but it depends on the model you

play11:07

work with and for each model there are

play11:10

different models with different context

play11:13

sizes so the context size of

play11:16

128 uh K is only aailable through the

play11:20

API at the

play11:22

moment so here's my first tip to see

play11:27

whether you are still in cont text or

play11:31

not just start every session with this

play11:35

prompt which tells the system to start

play11:38

every response with a greater

play11:41

sign because it

play11:43

outputs um it's text with markdown

play11:46

formatted the greater sign will be

play11:50

outputed as a

play11:53

quote which will look like this so the

play11:56

first paragraph of each output

play11:59

will be this

play12:03

Dash as long as we are in

play12:07

context the first time we leave the

play12:10

context it will forget about this prompt

play12:13

and will also not display this uh quote

play12:19

anymore but I wanted to tell you

play12:23

that the system is not just a

play12:27

probability machine not Sy stastical

play12:29

parot and for this embeddings are quite

play12:35

important um you want to extend the

play12:39

network the neural network with your own

play12:41

data but that's not possible because

play12:44

it's already fully trained and you can't

play12:48

easily train it with additional data the

play12:51

solution for this is to use

play12:57

text

play13:02

and create vectors from these um text

play13:06

fragments um there are some algorithms

play13:10

which um create those vectors um which

play13:15

just check out the words and see how

play13:20

yeah how near they are to each

play13:23

other quite complex I don't want to get

play13:26

get into detail here but you have to

play13:28

know about those embeddings that they

play13:31

are different than what the model knows

play13:35

um within the neuron Network and it's

play13:39

also not yeah how it how it belongs to

play13:43

the context because those text fragments

play13:46

those vectors are stored in a vector

play13:50

database and when you talk to the system

play13:53

it always for each prompt might query

play13:56

the vector database check out out

play13:59

whether there is a text fragment which

play14:02

fits your prompt and pull it into your

play14:08

context so if it does this your context

play14:12

will be smaller than what you chat with

play14:16

a system so that's why it's important to

play14:20

know about it and even if you have a big

play14:24

context it's always a question how much

play14:27

context you you really

play14:30

have again GPT 4 32,000

play14:35

tokens and we will see when we are in

play14:40

the chat model the the chat

play14:43

GPT we will have less we will have

play14:47

something about 4,000 or 8,000

play14:52

tokens so going back to the all knowing

play14:55

monkey or the stastical parrot I I

play14:58

believe with this approach those

play15:01

multiple layers multimodal approach and

play15:05

those

play15:07

embeddings um the system behaves like

play15:12

Yeah The all- Knowing monkey which has

play15:15

access to all the knowledge of the world

play15:17

at least the knowledge with which is

play15:20

aailable to

play15:23

it but let's get started with the real

play15:27

details with the

play15:30

yeah with trying to get a solution for

play15:33

the advanced certification of the is

play15:39

aqb so as I already said we have to

play15:42

prepare the model we have to do some

play15:44

priming every session every chat session

play15:47

starts from scratch the model doesn't

play15:50

know what you chatted with it before in

play15:54

another session so you have to tell it

play15:58

the context of your discussion it

play16:01

doesn't know you you have to tell it

play16:04

what it should know about you you have

play16:08

to tell it what's the goal of the

play16:10

session and maybe also what the solution

play16:13

should look

play16:16

like so for instance my context is that

play16:20

I'm an experienced software architect in

play16:22

the field of web development my favorite

play16:24

programming language is Java or groovy

play16:27

JavaScript and Cor responding Frameworks

play16:29

are not my thing I prepare to get by

play16:31

with minimal JavaScript okay that's me

play16:35

interesting I um want to create an image

play16:38

for this and here you see some kind of

play16:42

bias and it's so funny that not even I

play16:46

am depicted as as an old man it's also

play16:50

that there really old um technology

play16:53

calculator on the table so I gave it a

play16:56

try and um see what kind of bias it uses

play17:00

when I tell it U my favorite programming

play17:03

language is Python and JavaScript that's

play17:05

cool and I love modern JavaScript

play17:08

Frameworks hey now I'm young and cool

play17:11

and

play17:12

tip so bias is a problem even where you

play17:17

don't expect

play17:18

it but let's move on with the

play17:22

priming

play17:24

um I shouldn't not only tell the system

play17:28

yeah this this small paragraph I also

play17:32

extended it to tell it more about um how

play17:36

I work which Frameworks I I like to use

play17:39

by the way there's lots of text on those

play17:41

slides you don't have to read it um you

play17:45

will get the slides

play17:46

later so this is really enough context

play17:53

uh for the model to understand how I

play17:55

think how I work um what's my profession

play17:59

is

play18:00

um it isn't long ago that you had to

play18:04

copy and paste such a priming every time

play18:07

into your chat session which was not fun

play18:11

at all but um now catp uh has custom

play18:15

instructions and you have two text

play18:18

Fields where you can put in um what

play18:21

would you like chpt to know about you to

play18:24

provide better responses so that is the

play18:27

text from the last SL side and I call

play18:29

this long-term memory because I can put

play18:33

in there something which CHP should

play18:37

memorize and the second part of the

play18:40

custom instruction is how would you like

play18:42

chpt to respond and there I also have

play18:46

lots of prompts uh for instance start

play18:49

every response with a greater sign um we

play18:52

already had this but also many things

play18:56

for instance um

play18:59

be excellent at reasoning don't mention

play19:01

your knowledge cut off um I already know

play19:04

this and um that really

play19:08

helps and um I also instructed that I

play19:13

want to create ASD do output for my

play19:18

documents and even how I want to have my

play19:21

ask do output

play19:24

formatted okay even more priming

play19:29

I want to have the best basis for my

play19:34

chat session with cck

play19:36

GPT so I tell it today I want to talk

play19:40

with you about a software architecture

play19:42

certification do you know the is

play19:45

aqb every time I start a session and um

play19:49

talk with chat chpt I ask the model

play19:54

whether it knows about the

play19:56

Technologies um or other abstract things

play20:00

I want to talk about it um so it first

play20:06

answers yes it knows it then it

play20:08

describes the is aqb and this tells me

play20:13

that it yeah knows what it's talking

play20:17

about and it also creates some

play20:20

context for instance here I ask about

play20:24

the advanced level

play20:26

certification and it repeat the key

play20:29

aspects of the certification adding this

play20:32

to the context and helping both chpt and

play20:37

me um for this

play20:40

session by the way I think the answer

play20:42

was much longer but um that isn't too

play20:47

important for this

play20:49

session now we are still at the priming

play20:52

mode um so there are some um public

play20:57

aaable PDFs for the exam and I um just

play21:03

dropped in here the official

play21:05

rules the model should know about the

play21:09

rules I also dropped the glossery

play21:14

because I'm not a native speaker of the

play21:17

English language and um when we talk

play21:21

chat GPT and me about um the the exam we

play21:26

should use the same wording

play21:28

and so I think it's important that it

play21:31

knows about this

play21:32

document and as you can see um those

play21:35

documents are now put into the system as

play21:42

embeddings and that should be enough for

play21:45

the priming at a moment now we can start

play21:49

with the real task we have a third PDF

play21:53

document the example task big spender

play21:57

for a um Advanced

play22:02

certification I asked chpt to read the

play22:05

document carefully but not start to

play22:08

solve the task yet why every time you

play22:13

drop something into chat GPT it tries to

play22:16

come up with a good answer and here it

play22:20

would try to yeah already um solve the

play22:25

certification task but this would

play22:29

clutter up the the context

play22:33

so even with this it just tells me

play22:37

something about uh the document that it

play22:41

understood it and

play22:44

um ask me if I have any

play22:47

questions feel free to ask

play22:50

great the PDF contains two images and I

play22:55

already experienced that somehow images

play22:59

in PDFs are hard to read for C

play23:02

GPT but if I just drop them into the

play23:06

chat and tell him hey here are my images

play23:11

from the

play23:12

PDF it really manages to read and

play23:17

understand those

play23:19

diagrams it can read the text in the

play23:22

boxes it uh will notice the connections

play23:26

and uh output what it sees there which

play23:29

is quite

play23:31

helpful 6 months ago when I submitted

play23:34

the idea for the talk um jgpt wasn't

play23:38

capable of doing all of this and I

play23:42

wasn't sure whether I can create a

play23:44

solution uh with chpt for it um it

play23:48

couldn't read those images but what was

play23:50

quite interesting there was a

play23:52

description underneath um this class

play23:54

diagram in in form of a table and I did

play23:58

it the other way around I put in the

play24:00

description and asked CHP to create a

play24:03

diagram and the diagram didn't look like

play24:06

this one and I was looking for the

play24:09

reason and found out that the

play24:12

description didn't fit the diagram in

play24:15

all parts which is also quite

play24:18

interesting one last

play24:21

instruction I learned that CH gbt is

play24:25

capable of asking me questions

play24:29

so before you solve a task check if you

play24:32

have any questions regarding the task

play24:35

and ask me those

play24:37

questions I pressed return and oops Yeah

play24:42

it already had some questions it came up

play24:45

with a list of 10 questions about the

play24:49

task and I think quite good questions

play24:53

like regulatory

play24:56

requirements it does doesn't know that

play24:58

I'm from Germany and U maybe uh one

play25:02

regulatory requirement is a um data

play25:05

Protection Law and other German laws so

play25:09

this was quite

play25:11

interesting and I answered all those

play25:13

questions and you can answer them one by

play25:16

one and every time you submit one it

play25:18

will tell you okay now I understand but

play25:22

please answer also the the other

play25:24

questions so you can turn side

play25:30

but the First subtask Quality

play25:35

attributes I always let CH gbt repeat

play25:40

the objective for the

play25:43

subtask and um it repeats it adds to the

play25:48

context and I see that it really knows

play25:51

about the subtask

play25:53

great so this worked this is the subtask

play25:57

the objective now we can go on derve the

play26:02

quality scenarios trade me an asky do

play26:06

document okay and wow I got a table

play26:13

quite nice I scanned it and looks

play26:18

reasonable the importance we could work

play26:21

on this it shouldn't be everything high

play26:25

but quite a good

play26:28

sample now we also want to have a

play26:30

utility tree I asked CH to create one in

play26:36

plant oml and as you can see here this

play26:39

says three of

play26:41

three so when I work with those prompts

play26:46

I refine them and it doesn't make sense

play26:49

because of the limited context to refine

play26:52

The Prompt in the next prompt it makes

play26:56

sense to edit your prompt

play26:58

to keep the context short and here you

play27:01

can see um I had three tries and uh

play27:06

first I I just had the idea to Output it

play27:10

as a plant uml mind map but then I

play27:13

noticed it didn't came up with the um

play27:16

IDs of the quality scenarios so I

play27:18

refined this

play27:21

and wow it tells me it created a file

play27:24

which can be downloaded I wasn't aware

play27:27

well that it can do so but it worked it

play27:31

was not a hallucination so I downloaded

play27:34

this file displayed it in plant uml and

play27:38

here you go we have a utility

play27:41

tree we have the quality attributes

play27:45

everything's

play27:48

fine now you can

play27:50

see um yeah that's that's a third step

play27:54

for this

play27:56

subtask um again a table and um it's a

play28:00

special table because I again refined my

play28:05

prompt because first it had some

play28:08

problems um with the motivation and I

play28:12

then asked uh the system to to put the

play28:16

motivation below those lines with a call

play28:20

span of three and it knows how to work

play28:23

with asid do tables and uh came up with

play28:26

a good result

play28:28

the result is a little bit questionable

play28:31

I would have uh post other weights on it

play28:35

so system reliable and availability only

play28:40

10% yeah okay I think it should be more

play28:44

but these are all things you can discuss

play28:46

with a model and um so change its

play28:50

mind so as you have seen I I let the

play28:56

model create parts of the uh

play28:59

response um just to to work on the

play29:03

details because it has a limited output

play29:06

space and uh it will not output the the

play29:10

whole um solution for this one task if

play29:13

you do not um work iteration by

play29:17

iteration on it and so I then ask it to

play29:21

uh give a solution for the whole subtask

play29:25

and here we are it repeats the other

play29:28

things it already created and looks good

play29:32

to

play29:32

me solution strategy that was an easy

play29:36

one um

play29:38

so but I had a problem with repeating

play29:42

the objective um you can see I refined

play29:45

The Prompt as stated in the document was

play29:49

important because if I do not say as

play29:52

stated in the document it just uh tries

play29:55

to get it from the cont context if I say

play29:58

as stated in the document it starts to

play30:01

read the document again and outputs what

play30:06

it finds in a document and this seems to

play30:10

be quite okay it's it's a word nearly

play30:14

word by word um repetion of the the

play30:18

objective from the document so with this

play30:22

I just um told it hey give me a solution

play30:26

strategy

play30:29

and um ah yeah and again I I told it if

play30:35

you have any questions about this task

play30:37

ask them

play30:39

now and again I was surprised it came up

play30:43

with some

play30:44

questions and um it doesn't work as

play30:48

expected that when I at the start tell

play30:51

it whenever you have question ask a

play30:54

questions you always have to repeat this

play30:57

um the thing but it came up with those

play31:00

questions I answered those questions

play31:03

told it to um give me a solution

play31:06

strategy and it came up as expected uh

play31:10

with a key constraints assumptions and

play31:12

basic

play31:14

principles talking about the chat

play31:17

context I can ask it how much of the

play31:20

context is left and it tells me 2,478

play31:25

tokens of 4,000 aable tokens so we are

play31:29

still in context

play31:31

great I repeat it and it tells me 2,300

play31:37

tokens okay that's an

play31:41

estimate of 8,000 tokens so it doesn't

play31:44

know whether it has 4,000 or 8,000

play31:48

tokens but we are still in context I

play31:51

believe because what happened here we

play31:54

don't have the quote anymore here in

play31:58

front of it but that is just uh because

play32:03

it sometimes forgets to Output it here

play32:06

it is

play32:08

again again uh objective for the third

play32:12

statement for third subtask this time it

play32:16

forgot something overview of purchased

play32:19

and open source

play32:22

Parts okay so in my

play32:26

prompts

play32:28

I yeah in the next prompt I I tell it to

play32:32

that it forgot it but first I wanted to

play32:36

to have the context diagram I thought

play32:39

this is an easy task I told it to create

play32:42

it as plant TL in the C4 notation but

play32:46

what I got was too simple far too simple

play32:50

it didn't remember what it needed for

play32:53

this context diagram so this is

play32:56

worthless

play32:57

give the model time to think is an a

play33:00

prompting approach where you yeah let

play33:03

the model first generate the information

play33:06

it needs so first name all actors who

play33:09

will interact with the system name all

play33:11

external systems and create a table with

play33:15

uh the external

play33:18

interfaces and now we get an answer and

play33:23

with this answer we just repeat our time

play33:27

ask our prompt from

play33:30

um just two prompts ago and the result

play33:35

is quite better and now I'm I'm happy

play33:38

with um the actors and the external

play33:43

systems let's go on with business

play33:47

models we need a business

play33:50

structure and um chat GPT can generate

play33:54

it again we have a plant ml diagram uh

play33:58

as

play33:59

download and again I'm not happy with

play34:03

this

play34:05

diagram and again I do did some prompt

play34:09

engineering and toed the system to

play34:12

display all external components from the

play34:14

subtask three in the diagram actors and

play34:18

external systems and that changed the

play34:21

display the the generated

play34:25

diagram

play34:28

the fifth subtask wants to have a

play34:30

technology

play34:32

stack again let's repeat the

play34:37

objective and hey that is not the whole

play34:41

OB objective because if you take a look

play34:45

at um at the document subtask five is a

play34:50

long

play34:51

subtask and the problem seems to be that

play34:55

it goes over a page break

play34:57

this is a problem for

play34:59

chip and in this case we I fix it by

play35:04

putting all the text in the

play35:08

context when working with uh those

play35:11

embeddings it just fetches for each

play35:13

answer those embeddings it needs now I

play35:17

put it in the context and grow my

play35:21

context and with this full

play35:25

context um

play35:28

I can

play35:29

get a technology stack I'm happy with

play35:33

this and uh it was also requested to um

play35:38

explain the technology stack with the

play35:41

workflow through the system and yeah it

play35:44

tells me the system parts involved for

play35:46

each step and the contribution to the

play35:49

Quality goals looks fine for

play35:52

me let's check the context again

play35:58

tells me we used up 4,600 tokens out of

play36:03

8,000 tokens

play36:06

okay let's go

play36:08

forward the final subtask

play36:13

evaluation and in this case I extended

play36:17

my prompt to repeat the full

play36:21

objective so you have to be quite

play36:24

detailed and concise with your prompt

play36:28

and then I got um the the full objective

play36:33

fine and this time I asked whether it

play36:37

understands the objective and whether it

play36:40

has questions regarding it and I was

play36:43

surprised it says yes I understand the

play36:47

objective and I don't have any specific

play36:50

questions okay so go ahead identify the

play36:56

top five riskiest and most important

play36:58

quality scenarios here they are

play37:03

fine and then I asked to create the

play37:06

whole

play37:07

document and as you can see data in inte

play37:12

Integrity it has a rational essential

play37:15

for legal compliance great architectural

play37:18

decision implementing strict Access

play37:20

Control good tradeoffs this might may

play37:24

slightly increase system complexity but

play37:27

it's

play37:28

necessary I think those results are

play37:31

quite good and now we finished the

play37:36

exam and the question is how much

play37:38

context is now left and now it tells me

play37:42

we used up 5,000 tokens out of 4,000

play37:45

tokens and the earliest parts of the

play37:48

conversation are no longer in the

play37:52

immediate

play37:54

context we just used up all the the

play37:58

context so I guess when you do it in a

play38:02

real

play38:03

scenario you will start the context from

play38:08

fresh for each

play38:11

subtask and

play38:13

um I I guess it's also just a matter of

play38:17

time that we get a chat GPT with a

play38:20

really huge context and then these

play38:23

problems will be

play38:25

gone

play38:30

conclusion CH GPT is a generative

play38:33

pre-tin Transformer that's what GPT is

play38:36

about and it is capable of transforming

play38:40

the advanced exam input to a solution

play38:44

output I think that's quite great as I

play38:47

said six months ago I think that

play38:50

wouldn't be such a good solution but now

play38:54

with the advancements in the model I

play38:58

think the solution is quite okay um it

play39:01

has some problems but it is a good

play39:04

starting point it's it's a architectural

play39:09

body co-pilot for me and I now can work

play39:14

from this Solution on and make it

play39:18

better it needs Guidance the results

play39:22

have to be checked

play39:24

yes it's quite helpful can walk alone

play39:27

that's what I mean it's a buddy um to

play39:29

help me to get good ideas to um make

play39:34

good decisions and to um for for

play39:38

instance um get different views on uh on

play39:42

the

play39:45

architecture interly interesting if you

play39:49

ask CHT to analyze the

play39:52

results and ask it to find errors and

play39:57

inconsistencies it will find the

play39:59

shortcomings by its own so this is also

play40:03

quite helpful to just ask the

play40:05

system whether the things it produced

play40:09

are good or

play40:13

not now the advanced exam um consists of

play40:19

two parts the homework which we just did

play40:23

with

play40:24

chup and an oral exam

play40:27

where you have to show up and um answer

play40:30

some questions and will jgpt be able to

play40:35

pass the oral exam the technology for

play40:38

this is already there you can create

play40:41

virtual avatars you can speak to the

play40:44

system and you can even uh fake video so

play40:49

the technology is there but I think that

play40:55

the it will be yeah you will notice that

play41:01

the the solution was not created by a h

play41:05

a human

play41:07

architect and so you still have to

play41:10

analyze the

play41:12

solution understand it and be prepared

play41:15

for the oral

play41:17

exam to say it with the words of William

play41:20

Gibson the future is already here it's

play41:22

just not very evenly distributed so when

play41:26

you think of gpt3 the free model and GPT

play41:29

4 where you have to pay 20 bucks that's

play41:33

not evenly distributed not everybody has

play41:36

access to the better

play41:38

models but it's already a very

play41:41

interesting um

play41:44

technology so thank you for your

play41:47

attention the slides are aailable behind

play41:51

this QR code there's also a discussion

play41:54

board there and um the full transcript

play41:57

and the solution as

play42:01

PDF thank you for your attention I think

play42:04

we still have three minutes left for

play42:08

questions thank you Ral this was an

play42:11

amazing talk and about a very important

play42:14

topic um I have two questions for you um

play42:19

the first would

play42:20

be how long did it take you to solve the

play42:24

problem how many attempts were

play42:27

needed H it was quite fast it um I would

play42:32

say two days to work with the system and

play42:37

to um rework some of the

play42:40

prompts but I also have to say that I I

play42:44

only did a quick view over the results

play42:48

they

play42:49

look from a first view quite good but um

play42:54

you will find uh problem s with

play42:58

it okay thank you um the another

play43:01

question um are there other means of

play43:05

creating embeddings other than priming

play43:10

interactivity uh by dropping documents

play43:12

in the CET

play43:13

GPT so when you just work with CET GPT

play43:17

um I think this is a um main way to do

play43:21

embeddings you can also by now um State

play43:25

some URL which it then goes and fetches

play43:28

them and um if you create your own

play43:31

systems through the API uh you can use

play43:35

the API to create embeddings to um for

play43:39

instance use a larger um set of PDFs

play43:43

create your own Vector database and thus

play43:46

create those embeddings you then have

play43:49

have the full control of it and um

play43:52

because you always um you yeah you

play43:56

create the chat you put in the the

play44:00

context to the API and you decide how

play44:03

much of the context you will fill with

play44:06

the

play44:09

embeddings um thank you uh there here's

play44:13

another question um did Jet Chet GPT

play44:17

really help you or did you try to make

play44:19

Chet GPT understand insights you already

play44:23

had

play44:24

beforehand did it Sur surpris you with

play44:27

good ideas you did not have

play44:30

before yes it it's always surprises me

play44:34

with good ideas and uh always helps me

play44:37

um for instance it was closer to the

play44:41

objectives of the task and I was and um

play44:45

that was something where I thought yes

play44:48

it's it's right these these were the

play44:50

expected results I would have written

play44:55

much more not just to to um tell how I

play45:00

came to those uh

play45:02

results but that's one thing I was

play45:05

surprised of another thing um that it

play45:09

asked me questions about the task and

play45:13

the if you take a look at those

play45:15

questions they make

play45:18

sense and this also helps me to get a

play45:21

better understanding of the task and the

play45:24

next thing is when you work on

play45:27

architectural decision records with cat

play45:30

GPT um you can ask it for several

play45:34

options how to solve the problem and it

play45:38

might find options you didn't think of

play45:42

so yes it helps it helps to extend your

play45:47

your

play45:49

mind okay and uh now a last question

play45:53

because uh the time is running um for

play45:56

the Break um what is your view on the IP

play45:59

used in the training sets and in whether

play46:03

your input or questions were then added

play46:07

as training for the

play46:10

model so most of the time when I work

play46:13

with CET GPT um I uh asked very general

play46:17

questions about um open source software

play46:20

so that that's something I do not

play46:23

consider as a problem when you work um

play46:26

with a exam I could only work with a um

play46:30

public aaable um example you're not

play46:33

allowed to do this with a real exam and

play46:37

also um I wouldn't put in um some

play46:40

internal documents or something like

play46:42

this or personal data um but the

play46:46

interesting point is that um all those

play46:48

companies who provide those large

play46:50

language models um if you buy the

play46:54

Enterprise Edition they promise you that

play46:57

they don't use the data you put into the

play47:00

chat for trainings or something like

play47:03

this so I think there will be

play47:06

yeah a switch in how you think about

play47:10

this problem and how it will be

play47:13

solved okay thank you uh there is one

play47:17

thing uh um someone is asking for the QR

play47:21

code uh maybe you can bring it up again

play47:23

the QR code but um

play47:26

I think it was in in your um

play47:28

presentation you you can download the

play47:32

presentations to the attendees you can

play47:34

download all presentations uh uh later

play47:37

uh we will provide it for you yeah so uh

play47:41

I think or maybe I think I will also put

play47:43

the link in in the chat okay this is

play47:47

great okay thank you

Rate This

5.0 / 5 (0 votes)

Related Tags
AI in DesignSoftware ArchitectureGPT-4Certification ExamsChatbot TutorialAI ExpertiseData ProtectionPrompt EngineeringMultimodal ToolsVector Databases