ADDRESSING OVERFITTING ISSUES IN THE SPARSE IDENTIFICATION OF NONLINEAR DYNAMICAL SYSTEMS

Leonardo Santos de Brito Alves
15 Nov 202013:27

Summary

TLDRIn this video, Leo Alves from UFF discusses his research on mitigating overfitting in Symbolic Regression, a machine learning technique. Sponsored by CNPQ and the US Air Force, the study collaborates with UCLA's Mechanical Aerospace Engineering department. Alves explores the use of model development from first principles and data-based approaches, focusing on the challenges of convergence and error propagation when increasing system nonlinearity or state vector size. He examines the impact of regularization, sampling rates, and the condition number on model accuracy, suggesting that alternative polynomial bases may improve Symbolic Regression's performance.

Takeaways

  • 📚 The speaker, Leo Alves from UFF, discusses his work on addressing overfitting issues in 'CINDY', a project sponsored by CNPQ and the US Air Force, in collaboration with UCLA's Mechanical Aerospace Engineering department.
  • 🌟 The script contrasts two approaches to model development: traditional first principles and data-based approaches, with historical examples including Galileo, Newton, and Kepler.
  • 🤖 The focus is on machine learning, specifically symbolic regression, which uses regression analysis to find models that fit available data, citing significant papers in the field.
  • 🔍 The script explains the process of using CINDY, starting from data compression, building a system of equations, and making assumptions about the state vector size and sparsity of dependencies.
  • 📈 The importance of defining a sampling rate and period is highlighted, which are crucial for building matrices and evaluating the state vector at different times.
  • 🔧 The process involves creating a library of candidate functions using a polynomial representation with a monomial basis, which is a power series.
  • 🧬 The script delves into the use of genetic programming and compressed sensing for identifying nonlinear differential equations that model data.
  • 📉 The impact of increasing nonlinearity order on error propagation and coefficient accuracy is discussed, showing how regularization techniques like Lasso can help.
  • 📊 The Lawrence equations are used as test cases to illustrate different regimes: chaotic, double periodic, and periodic, each with distinct frequency spectra and time series behavior.
  • 📌 The script shows that the condition number of the candidate function matrix is a good proxy for relative error, and how increasing sampling rate or period affects this.
  • 🔍 The final takeaway is the recognition of the Van der Monde structure in the library of candidate functions, which is known to be ill-conditioned, and the need to explore different bases to overcome this issue and minimize error propagation.

Q & A

  • What is the main topic of discussion in this video by Leo Alves?

    -The main topic is addressing overfitting issues in the context of symbolic regression, particularly focusing on a method called Cindy, which is used for model development from data.

  • What is the role of CNPQ and the US Air Force in Leo Alves' work?

    -CNPQ and the US Air Force have sponsored Leo Alves' work, indicating financial or strategic support for the research on addressing overfitting in symbolic regression.

  • What is symbolic regression and why is it significant in machine learning?

    -Symbolic regression is a form of machine learning that uses regression analysis to find models that best fit available data. It is significant because it allows for the discovery of underlying equations from data, which can be crucial for understanding complex systems.

  • Who are some key researchers mentioned in the script that have contributed to symbolic regression?

    -Key researchers mentioned include Lipson and co-workers, who used genetic programming for symbolic regression, and Brenton W. B. Procter and co-authors, who applied ideas from compressed sensing and sparse regression to the field.

  • What are some common data compression methods mentioned in the script?

    -Some common data compression methods mentioned are projection methods like collocation, principal component analysis, and proper orthogonal decomposition.

  • What assumptions does the Cindy method make about the state vector and its relationship with the system?

    -The Cindy method assumes that the state vector size is arbitrary but small, and that the dependence of the function on the state vector is sparse, meaning that each function may depend on only a few elements of the state vector.

  • How does the script describe the process of building a library of candidate functions for Cindy?

    -The process involves defining a sampling rate and period, building matrices based on the state vector and its time derivatives, and creating a polynomial representation using a monomial basis, which includes all possible combinations of terms up to a certain order.

  • What is the role of regularization in the context of the Cindy method?

    -Regularization, specifically Lasso in the script, is used to minimize the objective function and prevent overfitting by automatically removing terms that are deemed unnecessary, thus improving the model's generalizability.

  • What are the Lawrence equations mentioned in the script, and what is unique about their parameter involvement?

    -The Lawrence equations are a set of differential equations used as test cases in the script. They are unique because the control parameters appear only in the linear terms of each equation, and the maximum linearity order is quadratic.

  • How does the script discuss the impact of increasing nonlinearity order on the performance of the Cindy method?

    -The script discusses that increasing the nonlinearity order leads to error propagation and incorrect coefficients, highlighting the need for regularization techniques to mitigate these issues.

  • What insights does the script provide regarding the relationship between the condition number of the candidate function matrix and the relative error in the model?

    -The script suggests that the condition number of the candidate function matrix is a good proxy for the relative error in the model. It increases with the nonlinearity order and the size of the state vector, indicating more error propagation and the limitations of using the Cindy method.

  • What is the proposed next step to overcome the limitations discussed in the script?

    -The proposed next step is to use a different basis for representing the unknown system to overcome the issue of error propagation associated with the van der Monde structure of the current candidate function matrix.

Outlines

00:00

📚 Introduction to Overfitting in Symbolic Regression

The speaker, Leo Alves from UFF, introduces the topic of addressing overfitting issues in Symbolic Regression, a project sponsored by CNPQ and the US Air Force and conducted in collaboration with UCLA's Mechanical Aerospace Engineering department. The talk focuses on the use of machine learning, specifically symbolic regression, to develop models from data. Symbolic regression is compared to traditional model development from first principles as well as historical data-based approaches by scientists like Galileo and Johann Kepler. The speaker also mentions significant papers in the field, including work by Lipson and Brenton Proctor, and discusses the challenges of convergence problems in increasing system nonlinearity or state vector size.

05:03

🔍 Analyzing Symbolic Regression and Regularization Techniques

This paragraph delves into the technical aspects of symbolic regression, explaining the process of transforming nonlinear ordinary differential equations into an algebraic system using matrices and linear regression. The use of an objective function with regularization, specifically Lasso, is highlighted for its ability to automatically remove unnecessary terms. The speaker uses the Lawrence equations as test cases to demonstrate different regimes of behavior, including chaotic, double periodic, and periodic. The effects of increasing nonlinearity order on error propagation and coefficient accuracy are discussed, emphasizing the importance of regularization in managing these issues.

10:03

📉 Exploring the Impact of Non-linearity Order and Condition Number

The final paragraph presents an analysis of the impact of non-linearity order on the condition number of the candidate function matrix and the relative error in model fitting. It is shown that increasing the sampling rate or period can improve the condition number and reduce error up to a certain point, after which there is no further improvement. The chaotic condition is found to produce the smallest condition numbers and errors, which is counterintuitive but explained by the random distribution of matrix elements in chaotic systems. The paragraph concludes with insights on the van der Monde structure of the candidate function library and its implications for error propagation, suggesting that using a different basis for polynomial representation could potentially overcome these issues.

Mindmap

Keywords

💡Overfitting

Overfitting refers to a model's tendency to perform well on training data but poorly on new, unseen data due to its complexity and excessive adaptation to the training set. In the video, overfitting is an issue that the speaker aims to address in the context of using symbolic regression within the framework of the 'Cindy' project, which is a method for model development from data.

💡Symbolic Regression

Symbolic regression is a subset of machine learning that involves using regression analysis to find mathematical expressions that best fit a set of data. The speaker discusses this concept as a key aspect of their work, where they apply it to model development, contrasting it with traditional model development from first principles.

💡First Principles

First principles refer to the fundamental theories or laws that form the basis for understanding and explaining phenomena. In the script, the speaker mentions that traditional model development in science is done from first principles, as exemplified by Galileo and Newton's approach to developing the laws of motion.

💡Data-Based Approaches

Data-based approaches involve using empirical data to develop models or theories. The video contrasts these approaches with those based on first principles, highlighting Johann Kepler's use of data from planetary orbits to formulate his laws of planetary motion as an historical example.

💡Genetic Programming

Genetic programming is a technique inspired by the process of natural selection that evolves computer programs to perform a user-defined task. In the script, it is mentioned as a method used by Lipson and coworkers to identify nonlinear differential equations that model data.

💡Compressed Sensing

Compressed sensing is a signal processing technique that allows the reconstruction of sparse signals from a small number of linear projections. The speaker refers to the work of Brenton W. B. Procter and his use of compressed sensing in the context of symbolic regression to improve model fitting.

💡Van der Monde Matrix

A Van der Monde matrix is a type of Vandermonde matrix that is used in the context of symbolic regression to represent a polynomial basis. The speaker discusses the structure of the candidate function library in 'Cindy' as having a Van der Monde matrix structure, which is known to be ill-conditioned and leads to error propagation.

💡Regularization

Regularization is a technique used to prevent overfitting by adding a penalty term to the loss function, which discourages overly complex models. The script mentions Lasso regularization, which is used to automatically remove unnecessary terms from the model, as a method to combat overfitting in symbolic regression.

💡Condition Number

The condition number of a matrix is a measure of its sensitivity to changes or errors in the data it represents. In the video, the speaker discusses how the condition number can be used as a proxy for the relative error in the model and how it increases with the nonlinearity order, indicating more error propagation.

💡Asymptotic Behavior

Asymptotic behavior refers to the long-term behavior of a system as time approaches infinity. The speaker mentions analyzing the asymptotic behavior of the Lawrence equations to understand the system's behavior over large times, ignoring transient behaviors.

💡Eigenvalues

Eigenvalues are scalar values that characterize a linear transformation, represented by a matrix. The script discusses how the eigenvalues of the candidate function library matrix move away from the unit circle as the system transitions from chaotic to periodic and double periodic conditions, affecting the condition number.

Highlights

Leo Alves from UFF discusses overfitting issues in Cindy, a project sponsored by CNPQ and the US Air Force.

Collaboration with the Mechanical Aerospace Engineering Department at UCLA on using machine learning for model development.

Traditional model development from first principles as proposed by Galileo and Newton, contrasted with data-based approaches.

Introduction of symbolic regression in machine learning to find models that fit available data.

Citation of key papers by Lipson and coworkers, and Brenton Proctor and Cuts, that brought symbolic regression to the forefront.

Addressing convergence problems in Cindy, particularly when increasing system nonlinearity or state vector size.

Assumption in Cindy that the state vector size is arbitrary but small, and the time history in data is known.

Explanation of how to build a system of equations in Cindy with state vector size and state function.

Importance of having the derivative time derivative of data, either measured directly or approximated.

Building a library of candidate functions using a polynomial representation with a monomial basis.

Transformation of non-linear ordinary differential equations into an algebraic system for solving.

Use of an objective function with regularization, specifically Lasso, to minimize error and remove unnecessary terms.

Test cases using the Lawrence equations to analyze different regimes: chaotic, double periodic, and periodic.

Observation of error propagation and coefficient inaccuracies when increasing nonlinearity order without regularization.

Demonstration of regularization's effectiveness in eliminating unnecessary terms but challenges remaining with physical terms.

Analysis of relative error behavior and condition number of the candidate function matrix with different sampling rates and periods.

Condition number as a proxy for relative error, useful when the exact solution is unknown.

Van der Monde structure of the library of candidate functions and its impact on error propagation and conditioning.

Proposal to use a different basis to represent unknown systems to overcome issues with error propagation.

Invitation for questions and closing remarks, emphasizing the importance of addressing overfitting in machine learning models.

Transcripts

play00:00

good morning afternoon or night

play00:02

depending on when you're watching this

play00:03

my name is leo alves i'm from uff

play00:07

i'm here to talk about my work on

play00:08

addressing overfitting issues in cindy

play00:11

this has been sponsored by cnpq also the

play00:14

us air force

play00:16

and it's been doing in collaboration

play00:17

this work has been done in collaboration

play00:19

with

play00:20

some folks at the mechanical aerospace

play00:22

engineering department at ucla

play00:24

so in general we're going to talk about

play00:28

the use of cindy from model development

play00:30

right model development has been

play00:32

traditionally done as it's known in

play00:34

science

play00:34

modern science from first principles

play00:37

this is what's been proposed by galileo

play00:40

from the beginning this is what for

play00:42

instance what as everybody knows

play00:44

that zach newton has done when he

play00:46

developed his uh

play00:47

laws of motion from first principles

play00:49

using calculus and his experiments

play00:52

but this can also be done using

play00:56

through uh data-based approaches and

play00:59

this is nothing new johann kepler did

play01:02

that

play01:03

which is a contemporary of the previous

play01:05

two people i mentioned

play01:06

and when he developed his laws of

play01:08

planetary motion

play01:10

using data from the plant orbits

play01:13

that he collected using his telescopes

play01:17

and we so what we're going to focus on

play01:19

is on the use of machine learning to do

play01:21

that

play01:22

and we're going to focus on a specific

play01:23

aspect of machine learning that is known

play01:25

as symbolic regression

play01:27

which is nothing but the use of

play01:28

regression analysis

play01:30

to search for models that best fit

play01:33

the available data and

play01:36

so to mention a few

play01:40

work a few papers on symbolic regression

play01:44

that really has

play01:44

brought it to the forefront of machine

play01:46

learning uh

play01:47

is the work of lipson and coworkers in

play01:50

which they use genetic programming to

play01:52

identify these

play01:54

nonlinear differential equations that

play01:56

model data and

play01:57

more recently there's been the work of

play01:59

brenton proctor and cuts

play02:02

using ideas of compressed sensing and

play02:05

sparse regression

play02:08

to be able to do the same but based on

play02:10

linear regression

play02:11

there's a lot of more work based on the

play02:14

stuff that these guys have done

play02:16

but i'm just citing here the main papers

play02:19

the original papers in which they

play02:20

developed those

play02:21

but so we're going to focus on this work

play02:24

on cindy

play02:25

and specific issues associated with each

play02:29

which is

play02:29

convergence problems that you have with

play02:31

it and

play02:32

they happen usually when you try to

play02:34

increase the nonlinearity order of your

play02:36

system

play02:36

they also happen when you increase the

play02:39

vector state size in your system

play02:42

and so i'm going to try to address those

play02:43

issues here

play02:45

so basically we assume that you have

play02:47

some

play02:48

data there is like a compressed data set

play02:51

that represents

play02:52

your problem right that compressed data

play02:54

can be compressing the data can be done

play02:56

in different ways

play02:58

um for instance with projection methods

play03:00

like colurking

play03:01

you can use principal component analysis

play03:04

or for instance like

play03:05

proper tonal decomposition dynamic

play03:07

multicoop is a number of ways in which

play03:09

you can do that

play03:10

and then you can build a system of

play03:12

equations like

play03:14

that in which you have your state vector

play03:16

size and you have your state function

play03:18

and so cindy there's a few assumptions

play03:21

behind it

play03:22

and the main one is that your state

play03:25

vector

play03:27

size is arbitrary but it's a small right

play03:30

so

play03:30

n is arbitrary but small and you know

play03:33

the time history in your data so do you

play03:35

know

play03:36

how these guys vary with time and you

play03:39

assume that

play03:40

whatever dependence f has on

play03:43

your state vector x is this

play03:46

is although it's not known is very

play03:48

sparse

play03:49

for instance f1 depends only on x1 and

play03:51

x2 but not on the others and so on

play03:55

so cindy the way the first thing you do

play03:57

is you define a sampling rate

play03:58

m and a sampling period tau and

play04:02

um which is you know depends on the

play04:05

initial

play04:05

and final time that you use to where to

play04:08

extract your data from

play04:09

then you build your matrices and you can

play04:12

you have your state size state vector

play04:14

and then you evaluate it at the

play04:15

different times for which you have that

play04:17

data

play04:17

and you build that matrix but what you

play04:19

really need is the derivative time

play04:21

derivative of that data

play04:22

and you can either measure that directly

play04:24

or approximate it numerically from

play04:27

the original data that you have then you

play04:30

build your library of candidate

play04:31

functions

play04:32

and the way you do that is is usually is

play04:35

a polynomial representation

play04:37

using a monomial basis which is nothing

play04:40

but a power series

play04:41

you can see here you have 1 x x

play04:44

squared and so on right and then all

play04:46

possible combinations for instance

play04:47

quadratic terms not only have to be

play04:49

x1 squared but also x1 times x2 and so

play04:52

on and you do that up to

play04:53

whatever order that you want to use that

play04:56

we call the highest value of which we

play04:58

call p

play04:59

and that will give you a possibility of

play05:02

terms that you can combine to fit your

play05:04

data

play05:05

at having q terms and then you of course

play05:09

you're going to have

play05:10

coefficients that are going to be in

play05:11

front of each one of these terms

play05:13

that you use to create your coefficient

play05:15

matrix right then you can propose

play05:17

transform that non-linear

play05:21

ordinary differential equations that we

play05:23

talked about before

play05:25

which is here you can transform that

play05:27

based on these matrixes on like an

play05:29

algebraic system

play05:31

and you can do that for each lines in

play05:33

your coefficient

play05:34

matrix or your uh the state

play05:38

vector derivative matrix and you solve

play05:41

it in stages

play05:42

and of course and you do linear

play05:44

regression to find out which

play05:45

coefficients you need to put in front of

play05:46

which

play05:47

these terms in order to fit the data for

play05:49

x dot

play05:50

and you do that also but to do that of

play05:53

course you need an objective function to

play05:55

minimize

play05:56

and in this case is essentially the

play05:57

left-hand side minus the right-hand side

play06:00

and we included here some regularization

play06:02

in this case the specific one is lasso

play06:04

which is a well-known one that is nice

play06:07

because it

play06:08

takes it removes terms that it deems

play06:11

unnecessary automatically for you

play06:14

so the test cases that we're going to

play06:15

use are going to be done the lawrence

play06:17

equations

play06:19

interesting thing about it is that the

play06:20

control parameters appear only on the

play06:22

linear terms in each equation

play06:24

sigma rho and beta and these equations

play06:26

that the maximum

play06:27

linearity order in them is quadratic

play06:29

right you have x times z and x times y

play06:31

here

play06:32

there are three typical scenarios that

play06:34

we can analyze

play06:35

one is a chaotic regime the phase of the

play06:38

parameters to get those the ones that we

play06:40

used here at least

play06:41

given here so you can see like a broad

play06:44

band of frequency spectra here

play06:46

and that explains that the time series

play06:48

behavior that you see on the left

play06:51

and there's a double periodic regime in

play06:53

which you get

play06:54

two dominant uh frequencies in your

play06:57

specular the corresponding time series

play06:58

is on the left

play06:59

and there's a periodic regime when you

play07:02

have a single dominant frequency that

play07:04

controls the behavior

play07:05

um so of course i'm talking about here

play07:08

asymptotic

play07:09

behavior so for very large times so

play07:12

you're ignoring the data for the early

play07:14

times

play07:15

in which you can have linear growth of

play07:17

disturbances and whatever transient

play07:18

behavior

play07:19

before you reach this asymptotic trends

play07:21

that i just showed

play07:23

um so the the first results i'm going to

play07:26

look into is without regularization

play07:28

we're going to look at the values of the

play07:29

three main parameters that's been

play07:31

increased in non-linearity order

play07:33

and i need to note here that one doesn't

play07:35

mean linear approximation

play07:37

it just means like that and we use an

play07:40

inverse problem type of approach

play07:42

in which i feed to my library of

play07:44

candidate functions exactly

play07:47

the same terms that appear in the

play07:50

lawrence equations although the

play07:51

coefficients in the parameters in front

play07:53

of each one are not yet known

play07:54

and so as an increase in nonlinearity

play07:57

order

play07:58

we can see that we have a lot of error

play07:59

propagation these results in the first

play08:01

line are bang on

play08:03

like across machine precision accuracy

play08:06

uh with the ones used to generate the

play08:07

data but that's the increase in only

play08:09

entire the order

play08:10

the there's a lot of error propagation

play08:13

and the coefficients become

play08:14

very very wrong and also not only that

play08:19

i'm illustrating here the case of the

play08:21

fifth order ones we have this has been

play08:23

normalized by the highest coefficient

play08:25

which you can see

play08:26

is when multiplying x in the equation

play08:28

for i which is exactly

play08:29

rho so that's why the maximum one is one

play08:32

here

play08:33

and you can see there's a lot of terms

play08:34

that appear in your equation that should

play08:36

be

play08:36

zero but are there although even for

play08:39

instance the high order ones are small

play08:42

nothing steady in you really that it

play08:44

doesn't exist and it's supposed to be

play08:45

small or if it shouldn't exist at all

play08:48

so this is the reason why people use

play08:49

lots of regularization because it knocks

play08:51

off those terms

play08:52

it does a pretty good job of eliminating

play08:54

most of those terms but still

play08:56

you still have on physical terms in in

play08:59

whatever model that you get back

play09:02

independent of whatever value you use

play09:04

for the regularization parameter

play09:06

this has been done in the previous case

play09:08

for the doubly periodic case that's why

play09:10

rho is 165 and here for rho equals 28

play09:13

is the chaotic case but the similar

play09:15

trend happens for all three cases

play09:18

so then we move on to looking into the

play09:22

how the relative error behaves and also

play09:24

associate with the condition number of

play09:26

the library of the matrix of the library

play09:28

of candidate functions

play09:29

and we can see that if we increase the

play09:33

the sampling rate

play09:34

after some point it does nothing to

play09:36

change the condition number of our

play09:38

matrix

play09:39

and it doesn't do anything to reduce the

play09:41

relative error either

play09:43

also if you increase the period uh the

play09:46

sampling period

play09:47

after it does decrease the condition

play09:49

number and the error

play09:51

but after a certain point it doesn't

play09:53

improve the results any further

play09:55

so you can tell that the condition

play09:57

number is a pretty good proxy for the

play09:58

behavior of the relative error

play10:00

which is a very good thing because in

play10:02

this particular case we generated our

play10:04

data from the largest equation so we

play10:06

know the exact solution

play10:08

but in general we do not so it we can't

play10:10

you really calculate this guy

play10:12

so knowing that the conditional number

play10:13

which you can calculate for any problem

play10:15

is a good proxy for this behavior is a

play10:17

good thing just noting that for very

play10:19

small sampling

play10:21

sampling rates the conditional number

play10:23

decreases because the matrix size

play10:25

decreases but

play10:26

we don't have enough data to create a

play10:28

proper model for our data that's why the

play10:30

error is still

play10:31

large so if you understand that then you

play10:33

can use the conditional number as a

play10:35

proxy for your relative error

play10:36

and in this last plot you can summarize

play10:40

our results and we show that as we

play10:43

increase the non-linearity order

play10:45

the conditional number of the matrix of

play10:48

associated with the library of candidate

play10:50

functions

play10:50

increases which means there's more error

play10:52

propagation which is why the relative

play10:54

error also

play10:54

increases a less intuitive thing

play10:58

is that the chaotic condition is the one

play11:01

that produces the smallest

play11:03

condition numbers although it still

play11:05

increases the non-integrity order

play11:07

and also largest errors smallest errors

play11:11

which is sort of counterintuitive but if

play11:13

you understand that

play11:15

um the fact that the solution is chaotic

play11:19

probably moves that matrix towards more

play11:22

of our like

play11:23

its elements having a random

play11:25

distribution of values and we know

play11:27

that matrix that matrices that are

play11:29

generated with whose elements are

play11:31

regenerated randomly

play11:32

they have very small conditionals you

play11:34

can prove that so maybe

play11:36

by by being a chaotic behavior or moving

play11:38

towards that scenario that's why the

play11:39

condition number decreases

play11:41

and also a way to approximately

play11:44

calculate the condition number is the

play11:45

ratio between largest and smaller

play11:47

smallest eigenvalues and as you go to

play11:50

periodic and double periodic conditions

play11:52

the eigenvalues move away from the unit

play11:54

circle so that ratio becomes larger

play11:56

which is probably why the condition

play11:58

number also increases in those cases

play12:01

so to summarize um

play12:04

it turns out the library of candidate

play12:06

functions has a vundermond

play12:09

structure right and van der monde

play12:11

matrices are known to be u-conditioned

play12:13

and the condition number of those

play12:15

matrices actually increases as the size

play12:18

of the matrix increases

play12:20

and by increasing the non-linearity

play12:22

order

play12:24

and we haven't done that but by proxy

play12:27

by increasing the state vector size we

play12:29

increase the size of that matrix

play12:31

and then as a result we're going to have

play12:34

larger matrix larger matrix means more

play12:36

eu conditioning

play12:38

because it's a van dermont type matrix

play12:40

so we are going to have more

play12:41

error propagation and so which limits

play12:44

our ability to use cindy

play12:47

the reason why these things happen is

play12:49

because

play12:50

we have chosen a monomial basis for a

play12:52

polynomial representation

play12:54

and so

play12:57

as a next that's why it has a van der

play12:59

man type of structure it's it's a

play13:01

well-known

play13:02

fact from interpolation theory and

play13:04

numerical analysis

play13:05

so the next step moving forward is to

play13:07

try to use a different basis and we're

play13:09

gonna we are currently trying

play13:11

or talking all bases to represent our

play13:13

unknown system to try to overcome this

play13:15

issue

play13:16

and minimize error propagation so thank

play13:18

you for your time if you have any

play13:20

questions please place it in the chat

play13:22

and i'll get back to you thank you so

play13:24

much and take care

Rate This

5.0 / 5 (0 votes)

相关标签
Machine LearningOverfittingSymbolic RegressionModel DevelopmentCindy AlgorithmFirst PrinciplesData-Based ApproachesGenetic ProgrammingCompressed SensingSparse RegressionError Propagation
您是否需要英文摘要?