makerday ft chris 09/07

Dineshman Bajracharya
7 Sept 202417:25

Summary

TLDRThe speaker emphasizes the importance of maintaining clean and compliant code repositories, particularly on platforms like GitHub, to ensure security and avoid legal issues. They discuss the use of tools like GitHub Copilot for generating compliant code and the necessity of understanding how repositories work. The speaker also touches on the role of data cleansing and the use of vector and graph databases in enhancing model understanding. They advocate for a holistic approach to problem-solving, combining scientific and engineering thinking, and stress the value of good documentation and output analysis in machine learning projects.

Takeaways

  • 💻 Ensure repositories are clean and follow rules, especially when using platforms like GitHub.
  • 🔍 Understand how repositories work and perform checks to ensure code is compliant and secure.
  • 🏢 For large institutions like Assurance, code must be in compliance with government regulations to avoid rework.
  • 🔒 Use tools like GitHub Copilot for code recommendations, which automatically censor sensitive information like SSH keys.
  • 🛠️ Developers should consider compliance when coding, even if they initially prioritize freedom and creativity.
  • 📝 Maintain good documentation in repositories, such as README files, to aid language models in understanding and using the code.
  • 🔑 Use 'key vault' or secure databases to store sensitive information that should not be exposed in repositories.
  • 📈 Consider using vector and graph databases to enhance how models understand and interact with your data.
  • 🧠 Think holistically about problem-solving, starting with a scientific approach and then applying engineering to implement solutions.
  • 🔍 Analyze model outputs to understand performance and determine if adjustments in methodology or post-processing are needed.

Q & A

  • Why is it important to keep repositories clean?

    -Repositories need to be clean because they are often subject to compliance checks and rules, especially when they are public on platforms like GitHub. Clean code ensures that there are no security risks or violations of privacy, which is crucial for both the developers and the users of the code.

  • What does the speaker mean by 'repositories' in the context of GitHub?

    -In the context of GitHub, 'repositories' refers to the projects or collections of files that developers use to store their code. These repositories can be public or private and are a central part of version control and collaboration in software development.

  • What is the significance of compliance in the context of the script?

    -Compliance is significant because it ensures that the code and practices followed by the developers adhere to legal and regulatory standards, especially important for large institutions and when dealing with sensitive data or government-related projects.

  • What is the role of 'Assurance' mentioned in the script?

    -Assurance is likely a B2B firm mentioned in the script, which may be involved in providing compliance checks and ensuring that the models and code developed by the team are in line with government regulations.

  • Why is it necessary to keep production code intact and how does scanning contribute to this?

    -Production code needs to be kept intact to ensure reliability and security. Scanning the code helps identify any vulnerabilities or compliance issues, thus preventing potential risks before the code is deployed or handed off to other entities like the government.

  • What is the purpose of 'key vault' as mentioned in the script?

    -A 'key vault' is a secure storage mechanism used to safeguard sensitive information like API keys and passwords. It ensures that only authorized personnel can access these critical pieces of information, enhancing security within the development environment.

  • How does GitHub Co-pilot provide recommendations while ensuring compliance?

    -GitHub Co-pilot provides recommendations by understanding the context of the code and the developer's intent. It ensures compliance by not exposing sensitive information like SSH keys and by hashing out any potentially sensitive data before presenting suggestions to the user.

  • What is the significance of using proper documentation like README files in repositories?

    -Proper documentation, such as README files, is significant because it provides clear instructions and information about the project, which is essential for understanding the project's purpose, dependencies, and how to run the code. This information is crucial for both humans and machine learning models that may use the repository.

  • Why is it important to think like a scientist when approaching a problem in the context of the script?

    -Thinking like a scientist is important because it encourages a holistic and innovative approach to problem-solving. It involves considering the process flow and desired outcomes without being limited by current technological constraints, which can lead to more effective and creative solutions.

  • What does the speaker suggest about the role of output analysis in understanding model performance?

    -The speaker suggests that output analysis is crucial for understanding why a model is performing in a certain way. By analyzing the output, developers can determine if the issue lies with the underlying methodology of the large language model or if additional post-processing is required.

  • Why is it recommended to collect repositories that are useful and relevant to the specific query?

    -Collecting repositories that are useful and relevant ensures that the data and code used are directly applicable to the problem at hand. This targeted approach can lead to more efficient and effective solutions, as opposed to using a broad and potentially irrelevant dataset.

Outlines

00:00

🛠️ Repository Compliance and Code Quality

The speaker emphasizes the importance of maintaining clean and compliant repositories, particularly on platforms like GitHub. They discuss the necessity of following rules and ensuring code is cleansed before it's handed off to authorities, such as the government. The speaker also touches on the use of tools like GitHub Copilot for generating compliant code and the importance of guarding against data theft and ensuring privacy. They provide examples of how to use GitHub Copilot effectively and securely, including the use of SDKs and handling of secrets.

05:00

🔐 Data Privacy and Internal Repositories

This paragraph delves into the concept of data privacy within an internal ecosystem, such as the Assurance ecosystem mentioned. The speaker discusses the use of private repositories and the importance of having similar data for effective machine learning models. They explain the use of key vaults for secure data storage and the significance of proper documentation in repositories, such as README files, for aiding language models in understanding project requirements. The speaker also highlights the importance of thinking like a scientist when approaching problems and the role of documentation in machine learning projects.

10:01

🧠 Thinking Holistically in Problem Solving

The speaker encourages a holistic approach to problem-solving, suggesting that one should first think like a scientist to conceptualize a solution and then like an engineer to implement it. They advocate for not limiting oneself with preconceived notions of what technology can or cannot do, and instead, to think creatively and outside the box. The speaker also discusses the importance of algorithm design and the potential pitfalls of relying too heavily on patches rather than creating robust solutions. They suggest analyzing output to understand model performance and to identify whether additional post-processing or changes to the underlying methodology are needed.

15:01

📚 Effective Repository Utilization and Continuous Learning

In the final paragraph, the speaker advises on how to effectively utilize repositories by selecting those that are useful and relevant to the problem at hand. They recommend unit testing and output analysis to understand the results produced by language models. The speaker also stresses the importance of continuous learning, as the field of language models is relatively new and constantly evolving. They offer to share resources on measuring the effectiveness of language model programs and encourage reaching out for help, emphasizing the importance of a comprehensive and open-minded approach to learning and problem-solving.

Mindmap

Keywords

💡Repositories

Repositories refer to storage locations where code and other project files are kept, typically on platforms like GitHub. In the context of the video, repositories are emphasized as being 'clean' to ensure compliance with rules and regulations. This is crucial as the video discusses the importance of maintaining high standards for code quality and security, especially when dealing with sensitive data or in regulated industries.

💡Compliance

Compliance in the video script refers to adhering to laws, regulations, and standards that govern how data and code are handled. It is highlighted as a critical aspect of software development, particularly when the code is used by large institutions or submitted to government entities. The speaker mentions that ensuring code compliance before submission can prevent the need for rework due to errors or non-compliance issues.

💡Code Cleansing

Code cleansing is the process of reviewing and refining code to remove any issues or vulnerabilities that could compromise its integrity or security. The video emphasizes the importance of cleansing code before it is shared or used in production environments. This is to ensure that the code does not contain any harmful elements, such as unauthorized access keys or security vulnerabilities.

💡GitHub

GitHub is a web-based platform for version control and collaboration, where developers can store and manage their code repositories. The speaker in the video uses GitHub as an example of a platform where repositories are managed and checked for cleanliness and compliance. GitHub's role in the script illustrates the practical application of the concepts being discussed.

💡Co-pilot

Co-pilot, as mentioned in the script, likely refers to AI-assisted coding tools that provide developers with code suggestions and recommendations. The speaker discusses how these tools can help ensure that code is in compliance and does not contain sensitive information. It also touches on the idea that these tools can learn from the context and provide relevant suggestions, which is a key aspect of their utility in development.

💡Key Vaults

Key Vaults are secure storage systems for sensitive information, such as API keys or passwords. In the video, the speaker uses the term to describe a method of securely storing information that should not be exposed, such as SSH keys. The concept is tied to the broader theme of data security and the importance of protecting sensitive data within development environments.

💡Vector Database

A vector database is a type of database that stores and retrieves data based on vector representations of information. In the context of the video, the speaker suggests using vector databases to help AI models understand and process information more effectively. This concept is part of the broader discussion on how to enhance AI capabilities and improve the quality of code recommendations.

💡Graph Database

A graph database is a type of database that uses graph structures with nodes, edges, and properties to represent and store data. The speaker in the video mentions graph databases as a potential tool for organizing and querying data in a way that can enhance the capabilities of AI models. This is related to the overall theme of leveraging advanced data structures to improve development processes.

💡Documentation

Documentation in the video refers to the written materials that accompany code, such as README files or project descriptions. The speaker stresses the importance of good documentation for helping AI models understand the context and requirements of a project. Documentation is also crucial for human developers to quickly grasp the purpose and setup of a project, as illustrated by the speaker's reference to a well-documented Capstone project.

💡Output Analysis

Output analysis is the process of examining the results produced by a model or system to understand its performance and identify areas for improvement. The speaker in the video encourages developers to analyze the output of their models to determine if the results are consistent and accurate. This analysis can reveal whether the underlying methodology or additional post-processing is needed to achieve the desired outcomes.

Highlights

Emphasis on maintaining clean repositories due to compliance with rules.

The importance of understanding how repositories work, especially on platforms like GitHub.

The necessity of code cleansing to ensure it is compliant before being handed off to the government.

Use of scans to ensure production code compliance with large institutions like Assurance.

The role of GitHub co-pilot in providing recommendations while adhering to compliance standards.

Explanation of how GitHub co-pilot avoids exposing sensitive information like SSH keys.

The concept of key vaults as a secure way to store sensitive information.

The significance of having similar data in private repositories for model training.

The use of vector and graph databases to enhance language model understanding.

The value of good documentation in helping language models learn and reducing data cleansing time.

The shift from data imputation to allowing models to handle it themselves.

Encouragement to think like a scientist first, then an engineer when approaching problems.

The importance of algorithm design and thinking holistically about solutions.

Advice on not limiting oneself with preconceived notions of what technology can or cannot do.

The significance of analyzing output to understand model performance and identify necessary adjustments.

The challenge of ensuring consistent results from language models and the role of randomness.

Recommendation to collect repositories that are useful and relevant to specific queries.

The importance of unit testing and output analysis in the development process.

Offer to share articles on measuring the effectiveness of language model programs.

Transcripts

play00:00

um yes you do want to make sure you're

play00:04

you're understand that our repositories

play00:06

are def definitely very clean because we

play00:08

do have to follow certain rules right we

play00:11

do have when you go to like GitHub for

play00:13

example github.com

play00:16

um you look at the respositories

play00:24

right oh my God let me sign let me sign

play00:26

my account first

play00:28

sorry uh

play00:37

okay so

play00:38

like definitely we go to like different

play00:41

responsories for

play00:42

example um people already started doing

play00:46

different checks on respositories and

play00:48

they already make sure their code is

play00:50

cleansed and good to go because first

play00:53

understand how respository works okay I

play00:54

know that we all very smart we know how

play00:57

respositories works right but even for

play00:59

me I had I actually learned something

play01:01

new yesterday I learned about

play01:02

environments yesterday this new um thing

play01:05

of how we keep our our production code

play01:08

intact these days most of our production

play01:11

code for example has certain

play01:15

um scans from it they scan our code make

play01:19

sure it's in compliance because um think

play01:21

about large institution like Assurance

play01:24

um I know some you guys still don't

play01:26

really know assur is it's quite um it's

play01:29

a B2B firm but so we do report both of

play01:32

our models and and everything through

play01:34

your government at the end of the day so

play01:36

when you go to like a Banking Company

play01:38

you go to any companies out there

play01:41

definitely you have to make sure you're

play01:42

in compliance right there and how do we

play01:45

make sure we're in compliance of course

play01:47

we want to make sure our cod in

play01:48

compliance before we hand it off to the

play01:51

government and then if they tell us that

play01:53

oh all these errors are wrong then we

play01:56

have to go back and do all the rework

play01:57

right just just just in general right

play02:00

like um that's why you could hear people

play02:02

say oh my God we need to make sure that

play02:04

our code is fit to criteria it's like

play02:07

open source it it's not going to be

play02:09

stealing people's informations all the

play02:11

stuff you hear about all these different

play02:12

types of guard rails that you hear from

play02:14

people these days right realistically

play02:16

that makes sense you know as a developer

play02:18

you sometimes like who cares right

play02:20

that's probably the first you ask that's

play02:23

true um you should definitely have the

play02:25

freedom to develop as much as you want

play02:27

but you Al also consider that in their

play02:30

mindsets of these developers they always

play02:32

make sure that their code is in

play02:34

compliance so when you want those

play02:36

recommendations from C- pilot you want

play02:38

recommendations that are in compliance

play02:41

that actually helps your customer at the

play02:43

end of the day you don't want just some

play02:44

random code out there right so for

play02:47

example let me give you a couple

play02:48

examples when you use GitHub co-pilot

play02:50

for example and let's just do one

play02:53

interaction real quick you get compilot

play02:56

um when they offer you

play02:58

recommendations um

play03:01

for example let me see have

play03:05

here like they will they will always

play03:07

hash out those keys for you they will

play03:09

never

play03:11

um they will never put those um SSH keys

play03:14

for you because they don't want you to

play03:16

see other people's codes right however

play03:18

they understand the structure because

play03:20

they designed the algorithm for you

play03:21

already right so you might want to ask

play03:23

like how do I connect to Azure um

play03:31

machine Learning Studio for example

play03:33

right you might ask something like this

play03:35

because you want to understand how can I

play03:36

connect to this type of service so that

play03:39

I can run my code for example

play03:42

right and

play03:44

then you will wait for co-pilot to tell

play03:48

you what to do right so let's make this

play03:50

little

play03:51

bigger co-pilot will tell you okay Chris

play03:54

what you need to do

play03:56

install the

play03:58

SDK um the sofware development kit and

play04:01

then run your code okay Second Step

play04:03

import this third step they'll tell you

play04:05

is you need to include these different

play04:08

types of Secrets right back right here

play04:11

right back in the day what they did

play04:13

before disc copil got updated they put

play04:15

xxxxx so they put these guard rails

play04:18

already for your program so that people

play04:21

because sometimes it's very dangerous

play04:23

because as human human beings we don't

play04:26

take it for granted that um putting

play04:29

these secrets sometimes is very

play04:30

dangerous but most of the times in your

play04:33

programs in these um gbt programs

play04:36

post-processing work before the before

play04:38

the agent actually sends you the result

play04:40

and gives it to you as his output they

play04:42

always do a little clensing to it they

play04:44

cens the program and they make sure that

play04:46

these are already hashed out already so

play04:48

from here you might not see it but

play04:50

definitely um it could be there for

play04:54

example so this is just how for example

play04:57

how they provide you the best

play04:58

recommendation you know but for your

play05:00

case right this is specifically

play05:02

generated for the specific generality of

play05:05

your customer right but your customer is

play05:06

only me it's only internal so if your

play05:09

data is already already private and we

play05:11

live in the same ecosystem the same

play05:13

Assurance ecosystem then you shouldn't

play05:15

have to worry about that okay so that is

play05:18

one thing about how there's difference

play05:19

between co-pilot versus using private

play05:21

responsories because you need to have

play05:23

the mindset that like this is going to

play05:26

be sharable with me and most of the

play05:27

times these things are already in the

play05:30

what what what what it calls them is

play05:33

called key vault key vault is just a

play05:35

fancy word to call database a way for me

play05:38

to store this information so people

play05:40

cannot view it unless you're admins for

play05:42

example that's just a fancy way to say

play05:44

that right there um so for going back to

play05:47

your problem right here is after you

play05:49

create that um private respository make

play05:52

sure you have some type of data that's

play05:54

um that's very similar to each other too

play05:57

you know because you have to understand

play05:58

a model right when a model Auto actually

play06:01

um learns from all this information and

play06:03

we expose the API right they're

play06:06

basically you know using this as like a

play06:10

vector database right like how you guys

play06:12

were describing where I think it was you

play06:14

guys right like Tre this like a vector

play06:17

database for example and quering that

play06:18

information right and then using that

play06:21

chat gbt model to help incorporate that

play06:22

information right so this is how I think

play06:25

it was a great idea when your team

play06:26

members talked about can use rag this is

play06:28

a great way too

play06:30

another great way too is also use graph

play06:31

databases for example when you go to

play06:34

certain responsory this is just a demo

play06:36

responsory there's certain key wordss

play06:38

that you would put inside of here for

play06:40

example and those key wordss is

play06:42

typically what you would want your um

play06:45

Vector database to understand and come

play06:48

up with right for example so actually

play06:50

let me give you a more realistic example

play06:53

this is from one of my Capstone projects

play06:55

I work for our team

play07:01

oh please do not hack me okay that's

play07:04

what I say please do not

play07:07

hack me

play07:09

see GI I don't know my password so I

play07:12

need to look it up h

play07:31

sorry I'm just siging in you guys it's

play07:33

quite slow apologize yeah no worries

play07:35

yeah um I was sort of like doing a bit

play07:38

of research on my end like beforehand um

play07:40

something I was looking into was sort of

play07:42

like getting that data from the repos so

play07:44

I was looking at like documentation for

play07:46

um the GitHub rest API right so you can

play07:49

get like all the different information

play07:50

like you can get information from like

play07:52

the readme.md or if you're doing like a

play07:54

python like project get Thea from appp

play07:58

yeah yeah so yep definitely and that's

play08:01

where we would get that Source

play08:02

information because if you create a

play08:04

really good re me file like this is one

play08:06

of my Capstone projects I can share with

play08:08

you guys um that we work with Dr Tech on

play08:12

um this is one of their products that

play08:13

they help us develop um this is you

play08:17

know they could incorporate little

play08:19

information that's fine but at least you

play08:20

see how like re me files should look

play08:22

like they tell you to project name for

play08:24

example they tell you like specific

play08:27

keywords like about um what type of

play08:30

packages you need to be installed right

play08:31

these are all important information your

play08:32

language model needs to understand right

play08:36

dependencies um how to run the pipeline

play08:38

for example these are all information

play08:41

you're feeding to your model for example

play08:43

right and parameters explanation is just

play08:46

another way for us to identify those

play08:48

certain parameters that needed in order

play08:50

to run your program for example right so

play08:53

you have to understand like um a human

play08:56

how we think you know and how does a

play08:58

computer think there actually very

play08:59

similar because like when we think yes

play09:02

we have that contextual awareness

play09:04

because we're able to look at one

play09:06

problem but we're also able to um cross

play09:09

reference other materials that from our

play09:10

past experience right now you have think

play09:13

of a computer a computer Only Knows yes

play09:14

or no so how can you help your computer

play09:17

understand what's happening right here

play09:18

that's why people are like oh let's

play09:19

create a vector database let's create a

play09:21

graph database so we can incorporate

play09:23

that information that our computer

play09:24

doesn't know because at the end of the

play09:26

day it's true that model that language

play09:28

large language model can only understand

play09:30

a certain amount of information but how

play09:32

can you help your model understand more

play09:35

this is how you can if yes definitely

play09:37

having good documentation is very good

play09:40

because it can help your model to learn

play09:41

it and another thing that's very

play09:43

important too is um you spend less time

play09:46

on

play09:48

um on trying to cleans your data because

play09:51

I know lots of us spend lots of times

play09:53

during our machine learning courses to

play09:55

cleanse our data right clean our data

play09:57

make sure our data is clean make sure

play09:58

that our data impute it and stuff like

play10:00

that right I will tell you today that

play10:02

nobody I don't I don't even do

play10:03

imputation anymore okay I let the model

play10:05

do it itself okay that's what I do okay

play10:08

because I'm lazy I'm a very lazy person

play10:10

I if I can find something that can help

play10:12

me solve the problem faster I would do

play10:14

it these days this is just more for you

play10:16

to have think like a first think like a

play10:19

scientist okay so first put your

play10:21

scientist hat on don't think about

play10:24

engineering don't think about the

play10:25

limitations of what your program can do

play10:27

that's the that's the first thing don't

play10:28

think about that first have an idea how

play10:32

the process flow and what you want now

play10:35

after you have that put down your

play10:36

scientist hat put back on your

play10:38

engineering hat on in your engineering

play10:41

hat how can you make it happen that's

play10:43

why I would say best for you because

play10:45

sometimes as human beings we do have a

play10:47

lot of bias because we're like oh we

play10:49

were taught that we cannot do it this

play10:51

way because the technology doesn't exist

play10:54

but to be honest if you want something

play10:55

to happen you can make it happen it's

play10:58

just whether or not you don't want to

play10:59

make it happen so that's my best tip to

play11:01

you guys is don't think like that it's

play11:03

very if you want to be a scientist a

play11:06

scientist we have to think outside the

play11:08

box like right right now for example you

play11:11

could tell me that um to gather data is

play11:13

using a gury right and you're saying

play11:15

Chris how can I do this what I would

play11:18

tell you is Anything Can Happen what you

play11:21

can do is simulate the gur yourself

play11:24

write the write the code yourself and

play11:25

simulate it you know so that's where I

play11:27

would say know don't

play11:30

because I know people always put these

play11:31

limitations on how you think but I was

play11:34

just say probably don't think like that

play11:36

because that will give you more more

play11:38

trouble for yourself because then you'll

play11:39

be trying to find different types of

play11:41

tools to put them together and you won't

play11:43

be able to create a holistic solution so

play11:46

that's some things I've noticed from

play11:48

some people my past Capon projects that

play11:50

they're all fixated on these fancy tools

play11:53

but they're not thinking like a

play11:55

scientist they're thinking more like an

play11:56

engineer if something breaks How can I

play11:59

how can I put a patch on it every single

play12:01

time but I guarantee you putting patches

play12:04

is a good thing for short-term purposes

play12:06

but long-term perspective something is

play12:08

going to break so you have to understand

play12:11

algorithm design is very important so

play12:13

it's very important for you guys to

play12:14

think holistically how the solution

play12:16

should look like so this is the reason

play12:18

why people have now started investing

play12:21

investing companies have started

play12:22

investing so much money on cleansing

play12:24

this data making a format very pretty so

play12:28

you can use it and that's the reason why

play12:30

your professors would teach you okay

play12:33

guys in order to do professional

play12:34

documentation you need to write

play12:36

something like this right here because

play12:38

not only I mean of course your professor

play12:40

didn't back in the day they didn't do it

play12:42

because they want they wanted for you to

play12:44

use your large English model but the

play12:46

idea is now flowing that you do need it

play12:48

okay so that's what I'm trying to make

play12:50

you guys think like outside a box and

play12:52

see how everything fits together and

play12:54

don't think just like a like um in one

play12:56

side think about from all perspectives

play12:58

how this really helps you right here um

play13:02

and then you know some of your keyword

play13:04

from your program when you when you send

play13:06

that query to um co-pilot you want to

play13:10

have you want to understand certain

play13:12

scenarios right so in your codebase

play13:14

there could be different types of um

play13:17

known bugs for example so these are

play13:19

released notes that they put under here

play13:20

right there are special keywords that

play13:22

they tell you what your program can and

play13:25

cannot do so yes there could be certain

play13:27

programs in your um co-pilot program

play13:29

there certain key wordss that can help

play13:31

your co-pilot pick up these words and

play13:34

send a quick query to your to your

play13:36

customer for example that's where you

play13:38

have to think about as a human being

play13:40

this is how I think but now how can I

play13:42

help my computer my machine to think

play13:45

like how I want to think like right

play13:46

there so this is one example I can give

play13:49

you guys for example of how to look at

play13:52

the problem differently if that makes

play13:53

sense um because another way too also um

play13:59

I can tell you guys that this problem

play14:00

might be a little diff difficult is um

play14:03

after you look at the result you're like

play14:05

the result is wrong I have to change

play14:08

something but I want you to spend a

play14:09

little bit of time to analyze your

play14:12

output analyze your output if you do

play14:14

some output

play14:15

analysis understand why your model is

play14:17

performing this way after you understand

play14:19

why your model is performing this way it

play14:22

could either be you need to add an extra

play14:24

postprocessing to your program or is it

play14:27

because the underlying l l m methodology

play14:30

is wrong so spend a little time to

play14:32

analyze that part because that part is

play14:34

the hard part right there of how to

play14:36

analyze your results because they could

play14:38

have hallucinations and stuff like that

play14:40

right so this is where I do see some

play14:44

difficulty in this problem right there

play14:45

is um how do you know your results would

play14:48

be consistent you don't know because you

play14:50

add a temperature to your problem you

play14:52

add some Randomness and you want that

play14:53

Randomness because you want something

play14:55

like that so just make sure that um

play15:01

just make sure about that those are also

play15:02

some some tips I can tell you guys um

play15:05

about how you can approach this problem

play15:07

a little bit okay like um just go

play15:11

through some very good and also just

play15:14

because you're collecting a lot of

play15:15

respositories make sure to collect some

play15:17

respositories that actually useful okay

play15:19

that can help you solve the problem okay

play15:21

pick some like how I would post a

play15:23

problem is I would probably have like

play15:27

um just a scratch not if I have

play15:30

suppositories 1 a 1 b 1 C for example I

play15:35

and we have up to like what one Z for

play15:37

example at least pick the ones that's

play15:38

correlate to your specific query that

play15:40

you want to and test it do those unit

play15:42

testing okay so how can you test it do

play15:44

some unit

play15:45

testing and

play15:47

see what results you get you

play15:51

know you get and then analyze it and

play15:54

then do then do output

play15:57

analysis and see what's happening and

play15:59

then you can understand what this will

play16:01

help you do is understand why your

play16:04

output is like

play16:06

that is it my llm

play16:10

methodology or

play16:13

um um an extra an interpret an extra

play16:18

post processing postprocessing I get

play16:22

it so these are the things I would tell

play16:24

you to think about you know when you

play16:25

start when you start you know

play16:27

implementation on your next step let's

play16:29

think about these things right here and

play16:30

then you then it will help you to

play16:32

understand what type of responsories you

play16:34

actually need for your program right

play16:36

here and I'll send I'll send you guys

play16:38

this don't don't take a picture I'll

play16:39

send it to you guys okay don't worry

play16:41

okay I'll send you guys this okay um but

play16:45

yeah and like I said you need any help

play16:47

feel free to reach out okay um but these

play16:50

are the high Lev things I could think of

play16:52

from my top of my head about what what

play16:54

could be possible and I'll send you a

play16:55

little bit of articles about um how

play16:58

people measure the effectiveness of the

play17:01

of their Pro of their LM program right

play17:03

here because this this space is actually

play17:06

quite new too even to me it's quite new

play17:08

okay so yeah I I'm still learning this

play17:10

part too okay um so hopefully this will

play17:13

help you guys I know you guys have

play17:14

another session but um like I said you

play17:17

need anything feel free to reach out to

play17:18

me

play17:20

okay well hopefully you guys have rest

play17:23

rest nice of your weekend okay yeah

Rate This

5.0 / 5 (0 votes)

Связанные теги
Clean CodeComplianceGitHubCoding PracticesData SecurityMachine LearningDocumentationModel ComplianceCode RepositoriesSoftware Development
Вам нужно краткое изложение на английском?