!!! Overview
[{$pagename}] is the average of the [Loss function] of the entire [Training dataset].
The desire is to find the parameters 𝑤 𝑎𝑛𝑑 𝑏 that minimize the overall [{$pagename}].
[Cost function/Screen Shot 2017-12-26 at 06.17.00.png]
\\
Where:
* w is the [weight] a [Vector]
* b is the [bias] a [Scalar]
* L is the [Loss function]
* m is number of examples in the [Training dataset]
* 𝑦̂ is the [Predictor variable] output vector. It can also be denoted a^(number of layers)
* y is the truth from [Training dataset]
[{$pagename}] sum of [Loss functions] over your [training dataset] plus some model complexity penalty.
A [loss function] is a part of a [{$pagename}] which is a type of an [objective function].
[{$pagename}] is generally represented by "J" and
The entire concept of "Training a [Artificial Neural network] is minimizing the [{$pagename}]
Normally, you must optimise the [Training dataset] and the [weights] on the [synapses] as you will __NOT__ be able to control over the input [data]
!! Common [Examples]:
* [Mean Squared Error] ([MSE]): MSE(θ)=1N∑Ni=1(f(xi|θ)−yi)2MSE(θ)=1N∑i=1N(f(xi|θ)−yi)2
* SVM cost function: SVM(θ)=‖θ‖2+C∑Ni=1ξiSVM(θ)=‖θ‖2+C∑i=1Nξi \\(there are additional constraints connecting ξiξi with CC and with [Training dataset])
* J = ∑1/2(y-yHat)exp(2)
!! Category
%%category [Artificial Intelligence]%%
!! More Information
There might be more information for this subject on one of the following:
[{ReferringPagesPlugin before='*' after='\n' }]