Regularization and Generalization in Deep Learning

About Regularization and Generalization

Regularization is one of the most important concepts in machine learning. In mathematics and statistics, finance, computer science, machine learning and inverse problems regularization is the process of adding information to solve an ill-posed problem. In the context of machine learning optimization problems, it applies a modification to the objective functions to reduce generalization error of the learning model even at the cost of increased training error. Generalization refers to the capability of a trained model to make the right predictions when faced with unknown input data during its operational life.

In the rest of this article, we try to gain an intuitive understanding of the mathematical basis of regularization theory for inverse problems and its application to improve the generalization performance of learning algorithms. I have included some mathematical content to emphasize that, there is sufficiently strong mathematical foundations that support machine learning algorithms. Very often it is easy to subscribe to these algorithms with available APIs without having any understanding of their mathematical basis. However, this knowledge is shallow. Hope readers will appreciate that comment.

Forward and Inverse problems in modelling

Analysis in science and engineering involves developing models of physical systems. These models can be used to predict how the systems react to its environment with an excitation at the input (which is the cause) to produce the response at the output (which is the effect).



Figure-1

This method of analysing and predicting a system behaviour is called a direct or forward problem.

Forward or Direct problems

A direct or a forward problem starts with a cause, as an example a pattern xn Rfrom the input data space X which is I - dimensional when transformed through a system results in a desired output yn Rd  which is an observable effect in the output data space Y with d - dimensions.

Consider the dataset X = {x1, x2, ....., xN}   and correspondingly Y = {y1, y2, ....., yN}. It is assumed that X and Y are in linear vector spaces. The mapping


is a forward transformation represented by a function f(xn), where n = 1, 2, …., N. This means


 

Forward transformation function f(.) may either be linear or non-linear. The goal is to predict the desired unique output given the input data using an appropriate physical or mathematical model that represents the transformation. Once the model is determined it is used to predict the effect given the cause.

Figure-2

In general forward transformations are represented by an operator A which may be either a matrix, a differential equation or even a transformation of a cause into an effect that can be measured as in an instrument (e.g., thermal expansion of material resulting in an indication for temperature, voltage across a piezo-electric cell for pressure, displacement of a needle for speed etc. In our context we continue to use f that operates on x to denote such transformations to y.


Figure-3


Forward problems are not always well-posed, but in most cases they are. Well-posed problems have unique solutions exist that are insensitive to small perturbations in input data (can be due to noise) or in the initial values. 

Figure-4


Insensitiveness to small changes in data and other conditions indicates to the stability of the solution.


Figure-5

The figure above illustrates an unstable situation where small changes in the data causes different solutions which are the blue and pink coloured responses.

Forward-Inverse problems

A forward problem indicates the existence of a “inverse problem”. For forward inverse or in short called inverse problems, the task is to recover the cause given the effect. In that sense they are antonyms to forward problems.

Figure-6 

Very often all characteristics and parameters of the physical system are not known. It is therefore required to infer these characteristics from known responses (e.g., the RLC circuit output voltage e0 ) of the system.

Consider the problem of inferring the animal class that caused the observed footprints in the figure below. These types of problems are known as inverse problems.

Figure-7 

Inverse problems are concerned with determining the causes xn of a desired effect or an observed effect yn even though the physical properties of the model are unknown. In other words the inverse problem tries to infer the inputs by observing the outputs. This is similar to the situation that we observe the symptoms (effects) and try to answer “What is disease (cause) that resulted in the symptom?” or “What is question to which the answer is Thiruvanthapuram?” 

Solving these problems amount to identifying inverse mapping which can be expressed as 


 

These problems are very active in the field of research in applied sciences such as signal processing, machine learning, computer vision, astronomy, solutions of differential equations, various areas of engineering, etc.

Figure-8 

In mathematics the existence of inverse of function is an important property. The inverse mapping from space Y to space X exists if-and-only-if the forward mapping from X to Y is “one-to-one” and “onto”. Such mappings are said to be both injective and surjective, hence bijective.

Well posed and ill posed problems

Hadamard’s Definition of well posed problems.

The concept of a well-posed problem is due to the French mathematician Jacques. Hadamard (1923), who took the point of view that every mathematical modelling problem corresponding to some physical or technological phenomenon must be well-posed.



 Figure-9 

Hadamard postulated that inverse problems are well posed if their solutions satisfy three conditions.

  • The solutions exist. (i.e., for every f (xn) there exists a desired output yn.). 
  • The solutions are unique such that (f (xn) ≠ f (xj) for all n j). 
  • The solution is continuously dependent on the input. (Small changes in the input will only lead to small changes in output.)

This also means that a well-posed problem is always well defined, unambiguous (easily identifiable), the solution is a single correct answer, and free from internal contradictions. 

Figure-10

Another definition according to Nashed (1987) a problem is well-posed if the set of data/observations is a closed set (i.e., the range of forward mapping is closed). In the following discussion we consider Hadamard’s definition.

The following figure illustrates the notion of ill posed problems.

Figure-11

It can be seen when the output space is mapped to input space small changes in output space Y can  result in large fluctuations and undesired oscillations in the input space X causing instability.

Ill-posed problems

If any of the conditions according to Hadamard’s definition is not satisfied, the problem is ill posed. This statement holds true in general for inverse problems (and could be used as a definition). 

A broad class of the so-called inverse problems arise in physics, technology and other branches of science. In particular problems of data processing of physical experiments belongs to the class of ill-posed problems.

Ill-posed problems in purchase decisions

Figure-12 

The ill-posed inverse problem is related to solving for x given y, when f -1may not even exist or is not continuous. 

If the stability condition is violated, the numerical solution of the inverse problem by standard methods is difficult and often yields instability even if the data are exact (since any numerical method has internal errors acting like noise). 

Therefore, special techniques, the so-called regularization methods must be used to obtain a stable approximation of the solution. The appropriate construction and analysis of regularization methods and subsequently (or simultaneously) of numerical schemes is the major issue in the solution of inverse problems. 

Without regularization and without further information, the error between the exact and noisy solutions can be arbitrarily large, even if the noise is arbitrarily small. 

Inverse problems and regularization 

Inverse problem can be abstractly stated as follows: 

Given “f(.)” and “y” determine “x”, when the true solution denoted by 


 

may not exist. 

We denote the unknown solution with superscript x which is a minimum norm solution. The difficulty of the solution is such that even if  f -1(y) exists it might not be computable. 

Usually, we do not have the exact data but only the noisy data 


 

where the magnitude of



Regularization as a method to solve ill-posed inverse problems 

Regularization is applied to approximate the inverse of forward function f -1 by a family of stable regularization operators Rα where α being the regularization parameter. Stability is achieved by reducing the effect of noise amplification. 

The problem can be restated as 



and 

 

Rα is a continuous approximation of f -1.

Since the observed output data y contains random noise (δ), then the estimate of input data by observing the output data is computed as 


  
and the total error 

i.e.,

total error = data error + approximation error.

All terms both right and left hand sides are error norms. The parameter α controls the effect of regularization. When α is small, Rα is a good approximation of f -1, but not stable. When α is large, Rα is a bad approximation, but stable. The trade-off between stability and approximation as function of the positive parameter α is shown in following figure.

Figure-13 

The Tikhonov functional 

In the next few paragraphs, we only discuss the regularization theory proposed by Tikhonov (1963) and his colleagues for linear mappings between input space X to output space Y. Tikhonov’s method for nonlinear mapping and other regularization methods are beyond the scope of this article. 

Tikhonov’s method has been used effectively for machine learning problems and is closely related to support vector machines. The basic idea by Tikhonov et.al., is to stabilize the solution by means of an auxiliary nonnegative functional that embeds prior information about the solution. This prior information assumes that the input-output mapping is smooth. Tikhonov’s regularization theory replaces the standard squared error minimization methods with minimization of a regularization risk functional that comprises of two terms. 

This functional (a functional is function of another function that performs a linear mapping from a vector space to the real line) is defined as



where α is a positive real number called the regularization (penalty) parameter and Jα(x) is called the Tikhonov functional. The notation EN stands for the expected value over the dataset comprising of N samples of the squared terms. 

The space of this mapping is a linear vector space of functions on which the norm is defined and is typically a Hilbert space. (For those who are not familiar vector space and functional analysis do not bother about this new term, it enough to know that Hilbert space is a generalization of Euclidean space). 

Estimating the approximation to find


is performed by minimising, optimizing the above functional in the least squared sense, i.e.,


The second term α||x||2 is called the smoothness term or a stabilizer because it stabilizes the solution of the inverse problem.   If α is correctly chosen then solution converges



 and

Addition of regularization term can encode prior knowledge of x which helps to convert the ill-posed inverse problem to a well-posed one. Regularization coefficient α controls the importance of data dependent error and regularization term. Regularization parameter thus represents a trade-off between closeness to data and smoothness. The limiting values of α are 0 and . With α tending to 0, the problem is unconstrained with the solution being completely determined from the training samples and is same as the standard least square error solution. With α tending to , the solutions are unreliable. 

Regularization and Generalization for Machine learning and modelling 

The classical approach of modelling a linear or nonlinear transformation by a matrix or a differential equation or a measurement process etc, is not suitable for all applications and has certain drawbacks. Such analytical models are not always complete. Solving a partial differential equation with big data is computationally complex can take very long amount of time to achieve a solution. 

Machine learning models are data driven algorithms which are not explicitly programmed. They perform learning from large amount of data in most cases which are typically high dimensional to make inference, predictions and take decisions. They learn to infer an underlying mapping from an input data space X to an output data space Y. Machine learning models learn the input output relationships from a given finite sized dataset that is expected to represent a much larger collection of available data. That means the learned model must be capable to generalize or make the right estimates or prediction with data samples not seen during training. 

Deep Learning and Ill-posed Inverse problems

 

Machine learning models offer an alternative to analytical approaches by learning to infer from the given dataset {(x1,y1), (x2,y2), ……,(xNyN)}. The problem of learning from examples is an example of inverse problem. These models learn to infer from a given finite sized available dataset which is partitioned for training and testing. The main goal is to generalize well on unseen data which is much larger in size than the dataset. 


Figure-14

A machine learning model uses the set of training samples to approximate a function (predictor) called a hypothesis that maps input data variable to their corresponding targeted output data variables (responses). The learning algorithm governs the learning process with a set of well-defined rules.  

These algorithms must choose a hypothesis h from a set of predictors called the hypothesis space H such that its error over the available dataset is minimized.  Once trained the model is expected to predict correctly even with unseen input data. Learning to produce the right response for any random input data from a finite set of samples is an inductive process which is an ill-posed problem. 

This problem is not solvable without making additional assumptions to make it well defined. By restricting the learner to choose a predictor from space H, we bias it towards a particular set of predictors. The choice of the restriction of predictor hypothesis space H is done based on some prior knowledge about the problem to be learned. This method of selecting the optimal h from restricted set of predictors by observing a finite set training data results in the phenomenon called inductive bias (a.k.a. learning bias).  

Regularization theory proposed for ill-posed inverse problems can be easily adapted to learning models. The work of Vapnik and his colleagues in statistical learning theory to control the model complexity to arrive at optimal solutions is the basis of applying regularization methods to ill-posed problems. The solutions correspond to a target function that performs the mapping between input data and output data. Regularized solutions in learning problems provide stable approximate solutions and gives continuous estimates of the ill-posed problems.  

A machine learning model uses a set of training samples to approximate a function (predictor) called a hypothesis that maps input data variable to their corresponding targeted output data variables (responses). The prescribed set of well-defined rules which governs the learning process is called a learning algorithm. 

Advantage of Deep Learning for modelling physical systems

Deep artificial neural network models with large number of hidden layers are universal function approximators. An ANN is an abstract machine which creates a non-linear mapping between an I-dimensional input data space and a d-dimension output space. 

Figure-15 

This non-linear mapping is captured in the weight and bias parameters of the network during the learning process of a neural network. Learning methods are essentially iterative gradient descent based methods. The "art" of training a neural network is to control the learning such that the resulting mapping is robust to noise or errors in the data. 

The optimization algorithm is typically based on back-propagation which find weights and bias parameters that minimize the error metric between computed output values and the correct output values in an iterative manner.

During training, the back-propagation algorithm iteratively adds a delta value (which can be positive or negative) to each weight and bias. The weight/bias delta is a fraction controlled by the learning rate (usually represented by η) of the weight gradient. The weight gradient is the calculus derivative of the error function. 

Due to availability powerful parallel computing facility deep learning algorithms can be efficiently implemented. Their performance improves with bigger amounts of data and can capture multiscale information. The optimization of the functional Jα(x) is achieved by iterative optimization methods. These factors have enabled solutions involving practical data that are traditionally difficult with analytical methods, and lead to faster and more effective algorithms. 


Figure-16 

These methods attempt to achieve a stable approximate solution to the exact solution of

as shown in the following figure for an image restoration problem

Figure-17 

Why Regularization is required in deep  learning 

An overly complex model can overfit any given dataset. Minimization of the cost function in the least squared sense can result in unstable solutions.  Regularization methods in deep learning helps to achieve the following objectives. 

  • Minimize model complexity by punishing the weight parameters of the model.
  • Eliminate overfitting. 
  • Improves generalization

Consider a cost function comprising of the standard mean squared error

  


(For brevity of notatons we consider the targeted output values of all instances are scalars therefore we replace the vector notation of output with scalar representation and the weight matrix with a vector.)

The degree of the polynomial increases as the model gets complex and can fit to all the data points in the dataset. In deep learning number of learnable parameters is often considered a measure of model complexity. Model complexity can be minimized during training by punishing the higher order weight values to move close to zero . 

Regularization thus provides a fundamental framework to solve learning problems and design learning algorithms. 

Generalization by Regularization in Deep Learning 

Generalization capability of learning models refers to their ability to make accurate predictions on unknown test data input not observed during the training process. For classical machine learning algorithms generalization performance is influenced the bias-variance dilemma. This means models that are over trained or with more than a certain complexity level tend to overfit on the training data perform poorly on test data. Similarly  a model which is not trained enough or without sufficient complexity will underfit.  To improve generalization by minimizing overfitting we apply an explicit regularization term to impose an additional cost for model complexity which effectively reduce complexity level. 

In any machine learning inductive bias induces some sort of capacity control that restricts the predictors to be “simple”, which in turn allows for generalization. The success of simple model that learned to fit on the training data depends on how well the model generalizes on real data. 

An interesting characteristic of deep neural networks is its implicit regularization capability i.e., their ability to generalize well on test data even with an over capacitated architecture without explicit regularization which is contrary to the usual understanding of the bias-variance trade-off. In deep networks learning biases induced by training procedures and optimization algorithms can cause implicit regularization. 

For deep networks with implicit regularization we add explicit methods. Such regularization techniques include Ridge regression (also known as Tikhonov regularization), Lasso and Elastic net algorithms. In particular Lasso method can be used for feature selection since it forces a model to use fewer parameter coefficients. 

Methods of Weight Regularization 

An extra cost associated for larger valued weights is added to the loss function. This method of penalizing the network when weight values become more irregular is called weight regularization. Examples are L2 regularization, L1 regularization and L1 - Lregularization. By punishing the network weights values these methods achieve model complexity reduction and improve generalization. 

L2 or Ridge Regularization 

Ridge regularization was introduced by Hoerl and Kennard. This method is the most common one and is a.k.a. weight decay regularization uses the L2 norm for the parameter coefficients. Since the standard mean squared error function is sensitive to random errors and outliers in the data, if the weight values are not constrained, they will tend to become large valued and explode. Therefore, a ridge constraint is imposed and the new optimization problem is defined as 



We assume that the input dataset X is a standardized so that it is zero centered having unit variance and the output values from Y are also zero centered. Then the L2 cost function or the L2 Penalized Residual Sum of Squared errors (PRSS) can be written in terms of modified cost function that includes the additional regularization term and is expressed as 


 

Or


  

where I is the dimensionality of the input data feature vector x, y a scalar target and N is number of data sample available. 

The first term measures the discrepancy between the predicted output and the true label values. The α value in the second term controls the strength of the regularization. 

The weight update equation is

 

L2 regularization has an advantage, since the cost function includes quadratic term, minimizing the function w.r.t. the weight values is a convex optimization problem has therefore a unique solution. It thus has a closed form solution. 

The selection of α value controls the shrinkage of the weight values. As α → 0, the cost function reduces to the original residual sum of squared errors. As α, parameter values → 0. The optimal value of α is chosen such that it minimizes the expected prediction error.  The L2 method does not force any parameter values and therefore feature data variables to be zero, however it selectively assigns more importance to those features which has more variance (more information) useful to minimize the prediction error performance. It shrinks the weight coefficients of low variance feature variables. This method is good for high dimensional dataset if all features are considered important.


Figure-18 

L2 regularization is highly sensitive to multi-colinearity in data, i.e., when multiple two or more predictor variables in data exhibit linear dependence and lack independence. Then least squared estimates of the estimates of the weight coefficients become extremely sensitive to random errors in the data. 

Figure shows the geometric interpretation of L2 method. The objective is to minimize the cost function under the constraint that is to stay within the gray-shaded ball. The elliptical contours represent equal valued unregularized cost function values. The gray shaded ball is the region of equal valued L2 regularized functions represented by circles. The optimal set of weight values are obtained by solving the constrained optimization problem. The solution is shown at the intersection of the region with minimal cost function. The penalty term is proportionate to the squared L2 norm of model parameters. 

L1 Regularization or Lasso regularization

This method was introduced Tibshirani in 1996. The optimization problem is defined as 


Hence the cost function can be written as



The weight update equation is




The gradient is defined as



When wi is negative adding α will force it to more positive and closer to zero and vice-versa. This change in weight values can result in the lesser significant feature values being removed from the weight update equation.  

Unlike the Ridge method the Lasso method can penalize weight coefficients for features and force them to zero. Hence it can be used for feature selection by selecting one variable when there are a set of highly correlated feature variables and ignore the other correlated ones. It thus enables feature size reduction and offers a sparse solution when the feature dimensionality is high. 

A drawback for Lasso is used when feature dimensionality I is large and number of training samples “ N ” is relatively less. In such cases where N > I, the method selects only “ N ” feature variables. The Lasso method selection of feature is highly dependent on dataset. 


Figure-19 

The L1-regularization method is similar to L2 regularization. The model parameters are penalized by its own absolute weight coefficients within the constraints formed by the straight edges. Figure also illustrates how L1 method induces sparsity. The gray shaded square is region of equal valued L1 regularized functions represented by the edges. 

Elastic net 

Each of the above regularization technique offers advantages and disadvantages for certain use cases.  The Lasso method helps to reduce the feature variables, the Ridge has the advantage that it has unique optimal (minimal) solution. Elastic net combines these two methods to include both the advantages.  

The method is to minimize the following cost function which is defined as



Or



The second order (quadratic) penalty term makes the cost function strongly convex. This results in a unique minimum solution. Both Ridge and Lasso methods can be considered special cases of Elastic net.

 

The parameter 𝜆 is called the mixing coefficient. For Lasso 𝜆 = 1 and for Ridge 𝜆 = 0. For 𝜆 > 0, minimization of the cost function is always a convex optimization problem.

 

In the naive implementation of elastic net method finds an optimal set of weight values in a two-stage method. First the ridge coefficients are determined and then the {\displaystyle \lambda _{2}}a lasso type shrinkage performed. This two-step method causes double shrinkage of weight coefficients. The prediction capability of the model decreases due to increased bias. To compensate for this the estimated coefficients can be multiplied by (1+ α2).

Figure-20 


The above figure shows a comparison between the above methods. Two-dimensional contour plots of the ridge penalty; lasso penalty and the elastic net penalty with α = 0.5. Vertices are point of singularities. For lasso the edges are straight lines. For both ridge and elastic net, the edges are strictly convex; for elastic net, the strength of convexity varies with α

 

Other methods to tackle overfit in learning models

 

Dropout: This method is used for deep artificial neural network models. While training during each update cycle, a neuron output is active only with a certain probability “p”. Each dropout layer chooses a set of random units with probability “1-p” and set their outputs to zero and the synaptic weights are not updated.   The random dropout of nodes is performed only during training and not done during testing.


Figure-21

 

Batch Normalization: It is general practice to initialize the network parameters with zero mean and unit variance. As training progresses the set of weight values loses this property.   Using batch normalization of layer weights re-establishes this property. It also helps to reduce the need for dropout.

 

Combining Multiple Learners (Ensemble method)

According to the “No Free Lunch Theorem” there is no single learning algorithm that is always the most accurate in any problem domain. The usual approach is to try many and choose the one that performs the best on a separate validation. The simplest way to combine multiple learners corresponds to taking a linear combination of the L base learners to reduce the problem of overfit and reduce variance. Important characteristics of base learners are 

a) Diversity- independence and lack of correlation

b) Accuracy and

c) Computational speed. 

There are two different ways the multiple base-learners that complement each other are combined to generate the final output.

Multi-expert Combination 

Multi-expert combination methods have base-learners that work in parallel. Examples are voting and stacking.

For class predictions a majority vote is considered and for regression averaged output is used. These learners use a bagging scheme whereby the L different and independent base learners are trained over slightly L different training sets which are randomly chosen from the set with replacement. Bagging is a short form for bootstrap aggregation. Random Forest Classifiers are examples of ensemble learning that use bagging.

Figure-22 

Model stacking is an efficient ensemble method in which the predictions, generated by using various machine learning algorithms, are used as inputs in a second-layer learning algorithm. This second-layer algorithm is trained to optimally combine the model predictions to form a new set of predictions. For example, when linear regression is used as second-level/layer modelling, it estimates these weights by minimizing the least square errors. However, the second-layer modeling is not restricted to only linear models; the relationship between the predictors can be more complex, opening the door to employing other machine learning algorithms.


Figure-23

Multistage Combination 

Multistage combination methods use a serial approach where the next base-learner is trained with or tested on only the instances where the previous base-learners are not accurate enough. The idea is that the leaners are sorted in increasing complexity so that a strong and complex learner is not used (or its complex representation is not extracted) unless the preceding simpler weak learners are not confident. Boosting uses simple base models and tries to “boost” their aggregate complexity. Unlike bagging methods where individual learners are independent, boosting processes are sequential and iterative. Adaboost (AdaBoost is short for Adaptive Boosting and is a very popular boosting technique) and Gradient Boosting machines are examples of this type (XGBoost is one of the fastest implementations of gradient boosted trees.).

Figure-24

 

Early Stopping: According to G. Hinton “early stopping is free-lunch”. During validation the model performance is monitored and number training epochs when no further improvement is observed. In the figure below the number of training epochs for the deep learning model can be limited to 7 so that the validation error is minimizes even though the training loss seems continues to decrease when the epochs are continued.

Figure-25

 

 

Data Augmentation

 

Data augmentation for training is a widely used technique to improve generalization performance of machine learning models particularly in image and natural language processing related datasets.

 

Unlike traditional models the performance of deep learning architectures consistently improves with increased dataset sizes. However datasets with large sizes are not easily obtainable.  Hence techniques to synthesize additional data sample by manipulating the original one is an easy and cheaper alternative. Figure below is block diagram that involves both human and deterministic sequence of transformations on the original dataset to augment. Data augmentation may be done such that it does not exceed an upper bound that can result in considerable difference between the original set and the enhanced one, causing adversarial effect in model performance. Generative adversarial networks are being utilized for automatic data enhancement.

Figure-26

 

Adding random noise in feature data during training is an augmentation strategy. Another method is to add noise in network weights to make the model insensitive to small weight changes.

 

 

References

  1. Heinz W. Engl, Martin Hanke, Andreas Neubauer, “Regularization of Inverse Problems”, Springer 
  2. Ajay Verma, Michael W. OppenheimerDavid B. DomanOn-Line Adaptive Estimation and Trajectory Reshaping, September 2005, DOI: 10.2514/6.2005-6436 
  3. Kshitij TayalChieh-Hsin LaiVipin KumarJu Sun, Inverse Problems, Deep Learning, and Symmetry Breaking,  arXiv:2003.09077v1 [cs.LG]  
  4. Lucas A., Michael Iliadis, R. Molina, A. Katsaggelos, Using Deep Neural Networks for Inverse Problems in Imaging: Beyond Analytical Methods,Published in IEEE Signal Processing Magazine 2018 
  5. Connor Shorten &  Taghi M. Khoshgoftaar , “A survey on Image Data Augmentation for Deep Learning”, Journal of Big Data volume 6, Article number: 60 (2019) 
  6. Aristide Baratin, Thomas George, César Laurent, R Devon Hjelm,
  7. Guillaume Lajoie, Pascal Vincent, Simon Lacoste-Julien, “Implicit Regularization via Neural Feature Alignment”, arXiv:2008.00938v2 [cs.LG] 28 Oct 2020
  8. Gitta Kutyniok, Solving Mathematical “Problems by Deep Learning:
  9. Inverse Problems”, Woudschoten Conference Zeist, The Netherlands, October 9{11, 2019
  10. Sargur Srihari, “Regularization in Neural Networks”, buffalo.edu
  11. Ernesto De Vito, Umberto De Giovannini, Lorenzo Rosasco, Francesca Odone. “Learning from Examples as an Inverse Problem.”, Article in Journal of Machine Learning Research · May 2005
  12. Hyeontae Jo, Hwijae Son,  and Hyung Ju Hwang, Eun Heui Kim,  “Deep Neural Network Approach to, Forward-Inverse problems”,  Networks and Heterogeneous media, American Institute of Mathematical Sciences, Volume 15, Number 2, June 2020 pp. 247
  13. Karl-Heinz Ilk, “On the Regularization of Ill-Posed Problems” , researchgate.net/publication/234423030
  14. Devis Tuia, Remi Flamary, Michel Barlaud, To Be or Not To Be Convex? A Study on Regularization in Hyperspectral Image Classification, Published in2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS),  DOI: 10.1109/IGARSS.2015.7326942
  15. https://statweb.stanford.edu/~owen/courses/305a/Rudyregularization.pdf
  16. https://en.wikipedia.org/wiki/Machine_learning#Training_models
  17. http://www.statistics4u.com/fundstat_eng/cc_ann_recurrentnet.html
  18. http://www.kjdaun.uwaterloo.ca/research/inverse.html
  19. https://in.mathworks.com/discovery/regularization.html
  20. http://d2l.ai/chapter_multilayer-perceptrons/weight-decay.html
  21. https://blog.datadive.net/selecting-good-features-part-ii-linear-models-and-regularization/ 
  22. https://www.coursera.org/lecture/ml-regression/can-we-use-regularization-for-feature-selection-0FyEi
  23. https://www.cs.ubc.ca/~murphyk/Teaching/CS540-Fall08/L13.pdf
  24. https://ml-cheatsheet.readthedocs.io/en/latest/regularization.html
  25. https://www.youtube.com/watch?v=MiFQt5CYM4Y
  26. https://www.youtube.com/watch?v=dYMCwxgl3vk


Image Credits

Figure-1: slideshare.net

Figure-4: Reference 2

Figure-5: Reference 2

Figure-7: shutterstock.com/508591663.jpg

Figure-9: mathsisfun.com

Figure-10: siltanen-research.net

Figure-11: static-01.hindawi.com

Figure-12: kjdaun.uwaterloo.ca

Figure-13:

Figure-14: guru99.com

Figure-15:

Figure-16: groundai.com

Figure-17: Reference 4

Figure-18: rasbt.github.io

Figure-19: rasbt.github.io

Figure-20: drek4537l1klr.cloudfront.net

Figure-21: cs.toronto.edu

Figure-22: researchgate.net

Figure-23: blogs.sas.com

Figure-24: miro.medium.com

Figure-25: Reference 6

Figure-26: ai.stanford.edu



Comments

  1. This comment has been removed by the author.

    ReplyDelete
  2. Excellent Blog! I would Thanks for sharing this wonderful content. Its very useful to us.I gained many unknown information, the way you have clearly explained is really fantastic.keep posting such useful information.

    Top Voice recognition softwares provides multi-lingual facilities, voice navigation, and customer analytics. The rapid digital transformation enabled the boom of the Voice recognition market.

    Voice Recognition Software, Analytics Insight, Voice Recognition, Speech Recognition, Digital Transformation, Customer Analytics, Voice Navigation

    ReplyDelete
  3. This comment has been removed by the author.

    ReplyDelete
  4. Thanks for Sharing This Article.It is very so much valuable content. I hope these Commenting lists will help to my website
    best data science course in delhi

    ReplyDelete
  5. Thanks for Sharing This Article.It is very so much valuable content. I hope these Commenting lists will help to my website
    data science course delhi

    ReplyDelete
  6. This comment has been removed by the author.

    ReplyDelete
  7. Wonderful article, Thank you for sharing amazing blog write-ups.

    You can also check out another blog on Cryptography and Network Security

    ReplyDelete
  8. Thanks for sharing this article. I was much inspired by this topic.I love to learn more about AI.

    If anybody having a questions regarding Mobile App Development reach Way2Smile Solutions. Mobile App Development Chennai

    ReplyDelete
  9. Nice post.The blog has information related Artificial intelligence. The content is effective and knowledgeable.
    Artificial Intelligence services

    ReplyDelete
  10. This post is filled with unique good ideas! Thanks for sharing your experience. As a leading web and mobile application development company in New York, we are committed to providing customers with quality services. Please keep updating your site as I am regular visitor to your site. Here I am talking about RisingMax which is top rated IT Consulting Company NYC provides real estate auction software that defines and creates innovative and robust mobile application experiences, no matter how complex or diverse your needs are.

    ReplyDelete
  11. Great Post, Really very happy to say,your post is very interesting to read.I never stop myself to say something about it.You’re doing a great job.Keep it up
    artificial intelligence-course

    ReplyDelete
  12. The information you provided in the blog that is really unique. Thanks for sharing this blog. SRB Technology is the Best Artificial Intelligence Course in Muscat, anybody need our service feel free to contact us. For more details visit our official website. Robotic And Coding For Kids in Muscat | Digital Marketing Training in Muscat | SEO Services in Muscat

    ReplyDelete

  13. Quite Informative!
    Recently consulted a Web Design Company in India, they were suggesting the same points.
    Thanks for posting.

    ReplyDelete
  14. Nice ,

    If any one wants to develop own app Please click here

    ReplyDelete
  15. Once again you provide several doses of reality which explore the complete explanation of packing and moving companies in Bangalore . This article don't have to be that long. I simply couldn't leave your web site before suggesting that I actually loved the usual info on packing and movers services in Bangalore. I just want to know what is the best way to get real service.

    ReplyDelete
  16. Excellent idea. Thank you for sharing the useful information. Share more updates.
    Deep Learning with Tensorflow Online Course
    Pytest Online Training

    ReplyDelete
  17. valuable blog,Informative content...thanks for sharing, Waiting for the next update...
    Loadrunner Training Online
    Loadrunner Online Training

    ReplyDelete
  18. Such a nice blog with the attractive reference links which give the basic ideas on the topic.

    AI Courses in Chennai
    Learn Artificial Intelligence Online
    AI Courses in Bangalore

    ReplyDelete
  19. Thanks for this great and useful information. RisingMax is one of the leading IT consulting companies in NYC offering 7 phases of the System Development Life Cycle, and information about software development cost breakdown all over the world. Reduce your cost by up to 55-65% by outsourcing your software development with us.

    ReplyDelete
  20. Companies have started leaning towards artificial intelligence (AI) solutions for a variety of reasons. Today's businesses need all the help they can get to keep up with competitors who are investing heavily in artificial intelligence research and development to stay ahead of the competition.

    ai development company

    ReplyDelete
  21. When your website or blog goes live for the first time, it is exciting. That is until you realize no one but you and your. lhd machine in mining

    ReplyDelete
  22. Thanks for Share the Details of Machine Learning with Python Training, Machine Learning with Python Courses, Machine Learning with Python Certifications Process and Understand the Clear Concept.

    Machine Learning with Python Training in Bangalore
    Machine Learning Python Training in Bangalore
    Machine Learning Training in Bangalore
    Machine Learning course in Bangalore

    ReplyDelete
  23. Good blog!!! It is more impressive... thanks for sharing with us...
    iOS Vs Android
    Is iOS Better Than Android?

    ReplyDelete
  24. AI & ML in Dubai
    https://www.nsreem.com/ourservices/ai-ml/
    Artificial intelligence is very widespread today. In at least certainly considered one among its various forms has had an impact on all major industries in the world today, NSREEM is #1 AI & ML Service Provider in Dubai
    1633063143436-10

    ReplyDelete
  25. AI & ML in Dubai
    https://www.nsreem.com/ourservices/ai-ml/
    Artificial intelligence is very widespread today. In at least certainly considered one among its various forms has had an impact on all major industries in the world today, NSREEM is #1 AI & ML Service Provider in Dubai
    1633063517606-10

    ReplyDelete
  26. AI & ML in Dubai
    https://www.nsreem.com/ourservices/ai-ml/
    Artificial intelligence is very widespread today. In at least certainly considered one among its various forms has had an impact on all major industries in the world today, NSREEM is #1 AI & ML Service Provider in Dubai
    1633072575733-10

    ReplyDelete
  27. Really thanks for the informative blog. Regards Best Deep learning company in london (https://datalabs.optisolbusiness.com/?lang=gb)

    ReplyDelete
  28. this blog is really interesting, if you want you can search for more information on data science course in bangalore

    ReplyDelete
  29. Know every aspect of Artificial Intelligence with Tecdecod. Here, I will share with you some unknown and secret hacks and tips of AI. There are many things you need to know about AI, and this guide will help you know everything. So, what are you waiting for? Subscribe my blog and understand the whole AI concept.

    ReplyDelete
  30. Thank you for a very interesting article on lead generation. I greatly appreciate the time you take to do all the research to put together your posts. I especially enjoyed this one!!

    ReplyDelete
  31. Techokids offers Online AI courses for kids from Highly Selective With Strong Technical and Industry Background, Experienced Mentors.

    ReplyDelete

  32. Good work! Your post is an excellent example of why I keep coming back to read your excellent quality content.

    facial recognition software
    e-signature software

    ReplyDelete
  33. I Like to add one more important thing here, The swarm intelligence market is expected to be valued at US$ 447.2 Million by 2030 at a CAGR 40%.

    ReplyDelete
  34. Really Fantastic blog, awesome information and knowledgeable content. Keep sharing more blogs. Thanks for sharing this blog with us.
    Data Science Training

    ReplyDelete
  35. I Like to add one more important thing here, The Artificial Intelligence Market is expected to be around US$ 190 Billion by 2025 at a CAGR of 37% in the given forecast period.

    ReplyDelete
  36. I Like to add one more important thing here, The Deep Learning Market is expected to be around US$ 25.50 Billion by 2025 at a CAGR of 42% in the given forecast period.

    ReplyDelete
  37. This comment has been removed by the author.

    ReplyDelete
  38. Thanks for sharing the good post, it is very informative and detailed article, to know more you can join our course Machine Learning With Data Science Course

    ReplyDelete
  39. Truly I love this blog. Directly I am found which I truly need. please visit our website for more information about facial gesture analysis


    ReplyDelete
  40. Hi dear,

    Thank you for this wonderful post. It is very informative and useful. I would like to share something here too.Se sei disoccupato e ricevi sussidi, puoi imparare GRATUITAMENTE. I corsi si svolgono durante il giorno e la sera. La priorità sui corsi diurni è data ai disoccupati. Un corso giornaliero è gratuito per tutti i disoccupati; Ottieni i corsi gratuiti per disoccupati online.


    Ripetizioni scolastiche




    ReplyDelete
  41. Hi dear,

    Thank you for this wonderful post. It is very informative and useful. I would like to share something here too.Abbiamo molte possibilità per il tuo successo professionale. Partecipa ai nostri corsi di formazione online, Partecipa ai nostri corsi di recupero online. Iniziare a imparare gratuitamente con un'ampia gamma di corsi online gratuiti che coprono diverse materie. Corsi online gratuiti per raggiungere i tuoi obiettivi, forniamo anche lezioni di recupero per gli student.


    Corsi gratuiti per disoccupati



    ReplyDelete
  42. I am really very happy to visit yourblog. Directly I am found which I truly need. please visit our website for more information about Web Scraping Service in USA

    ReplyDelete
  43. Hi dear,

    Thank you for this wonderful post. It is very informative and useful. I would like to share something here too.Our highly professional team provide complete IT solutions that specializes in custom mobile and web application development. Call us at (+91) 9001721837.


    open source cms development

    ReplyDelete
  44. Awesome blog and great content, AI objectives are really clear and well mentioned. I love reading such blogs, specially with point to point clear explanations.

    AI Solutions Development

    ReplyDelete

  45. Hi dear,

    Thank you for this wonderful post. It is very informative and useful. I would like to share something here too.Loop of Words is an innovative digital marketing agency dedicated to enhancing your brand’s image and customer base. The latest tools, powerful strategies, and data-driven results are our power pillars to deliver the best results.



    https://loopofwords.in/website-development/

    ReplyDelete

  46. Hi dear,

    Thank you for this wonderful post. It is very informative and useful. I would like to share something here too.Loop of Words is an innovative digital marketing agency dedicated to enhancing your brand’s image and customer base. The latest tools, powerful strategies, and data-driven results are our power pillars to deliver the best results.



    https://loopofwords.in/website-development/>website development services

    ReplyDelete
  47. Thanks for such a great post and the review, I am totally impressed! Keep stuff like this coming.
    cyber security course

    ReplyDelete
  48. I have read this blog. This is really amazing and helpful. thanks for sharing the information.
    Artificial intelligence for kids

    ReplyDelete
  49. Thank you for this wonderful post.

    One of the augmented analytics platforms used for consumers' purposes. Attend a trained session on life sciences data and other patient analytics as a sole! Hurry up to take a weekly demo on life science by the best expertise on this field.
    Pharmaceutical Business Intelligence Pharmaceutical Business Intelligence


    Life Sciences Data Analytics Life Sciences Data Analytics
    Enterprise Analytics Solutions Enterprise Analytics Solutions

    WhizAi provides excellent Security Services in Chennai,

    Business Intelligence For Life Sciences
    Enterprise Analytics Solutions

    ReplyDelete
  50. Nice blog..! I really loved reading through this article... Thanks for sharing such an amazing post with us and keep blogging. we also have a website that provides best Artificial Intelligence services India

    ReplyDelete
  51. Web&システム開発のエンジニアリソースにお悩みの方へ。 gjnetwork型オフショア開発サービス ベトナムでは様々なスキルを持ったエンジニアをアサインできます。

    a, href="https://gjnetwork.jp/%E3%82%AA%E3%83%95%E3%82%B7%E3%83%A7%E3%82%A2%E9%96%8B%E7%99%BA/">オフショア開発 ベトナム

    ReplyDelete
  52. This comment has been removed by the author.

    ReplyDelete

  53. Your article may better understand the problem presented https://mateuszlomber.pl/

    ReplyDelete
  54. Fantastic Blog ! Machine Learning is an evolving field, this blog shows the importance of ML and Best Machine Learning training in Noida helps to grow skills in this field.

    ReplyDelete
  55. This comment has been removed by the author.

    ReplyDelete
  56. Machine learning is revolutionizing FP&A (Financial Planning & Analysis) by enhancing data-driven decision-making. It utilizes algorithms to analyze historical data and uncover patterns, improving forecasting accuracy. ML models can process extensive datasets quickly, providing real-time insights that enable agile responses to market shifts.
    Machine learning in fp&a

    ReplyDelete
  57. Nice Blog! Easy to understand and very informative. If you are looking for machine learning training in noida

    ReplyDelete

Post a Comment

Popular posts from this blog

Modeling Threshold Logic Neural Networks: McCulloch-Pitts model and Rosenblatt’s Perceptrons

Gradient Descent rule and Widrow Hoff rule for Deep Learning