Because it gives more weight to one big error then to multiple small ones with the same sum.
We want the errors to be noise and not systematic. Noise usually has a gaussian distribution. And in a gaussian distribution multiple small values are more likely than one big one.
So Predic1 was better then Predic2? No. Because correctly predicting the one outlier shows more predictive power then staying close to the average. Therefore we use SumOfSquerrors:
SumOfSquerrors(Predic1) is 58
SumOfSquerrors(Predic2) is 45
This shows that Predic2 is "better" and we are happy :)
But what I've never understood is that if your objective is to magnify errors, why not cube it? Why not to a greater power still? If the other benefit is that all negative values to an even power become positive, then why not take the absolute value of the cube? No matter what, the degree to which we magnify errors strikes me as arbitrary.
Because it gives more weight to one big error then to multiple small ones with the same sum.
We want the errors to be noise and not systematic. Noise usually has a gaussian distribution. And in a gaussian distribution multiple small values are more likely than one big one.