• Boltzmann Machine

    A Boltzmann machine is a type of unsupervised machine learning algorithm that uses a graphical representation, much like a neural network. This architecture expresses the probability distribution in terms of visible and hidden units.
  • Stochastic Gradient Descent

    Stochastic gradient descent is an algorithm for online-optimization. Its purpose is to estimate the optimal parameters of the learning hypotheses. Unlike gradient descent, the iterative process requires only a small amount of data at a time, which is very useful for large datasets.
  • Neural Network

    A neural network is a graphical representation of a set of linear and non-linear operations acting on an input data-point. In a feed-forward neural network, we stack several layers sequentially. The input data cross multiple layers, changing its feature's representation along the way. This process allows the creation of very complex predictors.
  • Support Vector Machine (SVM)

    An SVM algorithm learns the data by engineering the optimal separating line/curve between the classes. Compare with the other algorithms which try to do this by first determining the Bayes predictor. The SVM has robust generalization properties. However, they are usually tough to train.
  • Gradient Boosting

    Gradient boosting is another boosting algorithm. It uses gradient descent to minimize the loss function and hence the name. However, unlike in other algorithms, the learning rate is adjusted at every step.