cost and loss functions are synonymous (some people also call it error function). The more general scenario is to define an objective function first, which you want to optimize. This objective function could be to - maximize the posterior probabilities (e.g., naive Bayes) - maximize a fitness function (genetic programming) - maximize the total reward/value function (reinforcement learning) - maximize information gain/minimize child node impurities (CART decision tree classification) - minimize a mean squared error cost (or loss) function (CART, decision tree regression, linear regression, adaptive linear neurons, ... - maximize log-likelihood or minimize cross-entropy loss (or cost) function - minimize hinge loss (support vector machine) A loss function is a part of a cost function which is a type of an objective function. The purpose of Cost Function is to be either: Minimized — then returned value is usually called cost, loss or error. The goal is to find...