□ Anomaly Detection using Gaussian distribution - example: detect overloaded server. Usage examples: intrusion detection, fraud detection, system health monitoring, removing anomalous data from the dataset etc. Anomaly DetectionĪnomaly detection (also outlier detection) is the identification of rare items, events or observations which raise suspicions by differing significantly from the majority of the data. □ K-means algorithm - example: split data into three clusters. Usage examples: market segmentation, social networks analysis, organize computing clusters, astronomical data analysis, image compression, etc. The algorithm itself decides what characteristic to use for splitting. In clustering problems we split the training examples by unknown characteristics. Instead of responding to feedback, unsupervised learning identifies commonalities in the data and reacts based on the presence or absence of such commonalities in each new piece of data. Unsupervised learning is a branch of machine learning that learns from test data that has not been labeled, classified or categorized. □ Logistic Regression - examples: microchip fitness detection, handwritten digits recognitions using one-vs-all approach. Usage examples: spam-filters, language detection, finding similar documents, handwritten letters recognition, etc. In classification problems we split input examples by certain characteristic. □ Linear Regression - example: house prices prediction. Usage examples: stock price forecast, sales analysis, dependency of any number, etc. Basically we try to draw a line/plane/n-dimensional plane along the training examples. In regression problems we do real value predictions. The ultimate purpose is to find such model parameters that will successfully continue correct input→output mapping (predictions) even for new input examples. Then we're training our model (machine learning algorithm parameters) to map the input to the output correctly (to do correct prediction). In supervised learning we have a set of training data as an input and a set of labels or "correct answers" for each training set as an output. In most cases the explanations are based on this great machine learning course. The purpose of this repository was not to implement machine learning algorithms using 3 rd party libraries or Octave/MatLab "one-liners" but rather to practice and to better understand the mathematics behind each algorithm. This repository contains MatLab/Octave examples of popular machine learning algorithms with code examples and mathematics behind them being explained. For Python/Jupyter version of this repository please check homemade-machine-learning project.
0 Comments
Leave a Reply. |