!!! Overview [{$pagename}] is the average of the [Loss function] of the entire [Training dataset]. The desire is to find the parameters 𝑤 𝑎𝑛𝑑 𝑏 that minimize the overall [{$pagename}]. [Cost function/Screen Shot 2017-12-26 at 06.17.00.png] \\ Where: * w is the [weight] a [Vector] * b is the [bias] a [Scalar] * L is the [Loss function] * m is number of examples in the [Training dataset] * 𝑦̂ is the [Predictor variable] output vector. It can also be denoted a^(number of layers) * y is the truth from [Training dataset] [{$pagename}] sum of [Loss functions] over your [training dataset] plus some model complexity penalty. A [loss function] is a part of a [{$pagename}] which is a type of an [objective function]. [{$pagename}] is generally represented by "J" and The entire concept of "Training a [Artificial Neural network] is minimizing the [{$pagename}] Normally, you must optimise the [Training dataset] and the [weights] on the [synapses] as you will __NOT__ be able to control over the input [data] !! Common [Examples]: * [Mean Squared Error] ([MSE]): MSE(θ)=1N∑Ni=1(f(xi|θ)−yi)2MSE(θ)=1N∑i=1N(f(xi|θ)−yi)2 * SVM cost function: SVM(θ)=‖θ‖2+C∑Ni=1ξiSVM(θ)=‖θ‖2+C∑i=1Nξi \\(there are additional constraints connecting ξiξi with CC and with [Training dataset]) * J = ∑1/2(y-yHat)exp(2) !! Category %%category [Artificial Intelligence]%% !! More Information There might be more information for this subject on one of the following: [{ReferringPagesPlugin before='*' after='\n' }]