Data Science combines training disciplines in statistics, mathematics, and computer science. Acquiring a degree in Data Science can help a student make informed decisions using data.

With each passing day, data generated by companies and organizations is increasing, and with that, the need for data science experts who can utilize this to make a positive impact is also increasing. Universities in India are constantly developing courses and graduate programmes to cater to this need.

Check out the list of universities offering these programmes :  

  1. IIT Madras - B.Sc Programming and Data Science - Click here
  2. NMIMS - B.Sc. Data Science- Click here
  3. IIT Bombay - M.Sc. Applied Statistics and Informatics - Click here
  4. Symbiosis University - B.Sc. Data Science - Click here

Apart from the full-time degrees from these reputed universities, you can also take a look at some short duration certificate courses :

  1. IIT Delhi - 6-month programme in Data Science and Machine Learning - Click here
  2. BITS Pilani - 11-month programme in Artificial Intelligence and Machine Learning - Click here
  3. IIM, Ahmedabad - 1-month programme on Artificial Intelligence and Machine Learning for Business - Click here

All these courses have certain  ML Algorithms in common, such as Linear Regression, Logistic Regression, Naïve Bayes Classifier and more. In this post, I have curated 'Data Science Notes' by compiling a list of all such algorithms and their respective summaries.

Let's begin!

Linear Regression

• Regression is a statistical way to establish a relationship between a dependent variable and a set of the independent variable(s).
• This technique is applicable for Supervised Learning Regression problems where we try to predict a continuous variable.
• Linear Regression can be further classified into two types – Simple and Multiple Linear Regression.

Linear regression - Wikipedia

How does it work?
• Linear regression “fits a line” that best describes a linear relationship between the two variables (say height and weight) by the method of Ordinary Least Square (OLS).
• Ordinary Least Squares minimize the sum of squared residuals, residuals being the red and blue lines represented in the below graph.
• The final regression equation is in the form :

Y = β0   + β1*X

Check out a project on Linear Regression using the Sales and Advertising data:

Simple-Linear-Regression-Project/ at master · pb111/Simple-Linear-Regression-Project
Contribute to pb111/Simple-Linear-Regression-Project development by creating an account on GitHub.

Industry use of Linear Regression :

Potential Trend for Online Shopping Data Based on the Linear Regression and Sentiment Analysis
How to reduce the cost of competition in the industry, identify effective customers, and understand the emotional needs and consumer preferences of customers, so as to carry out fast and accurate commercial marketing, is an important research topic. In this paper, we discussed the method for the ana…

K-Nearest Neighbour

• It is a classification algorithm that takes a bunch of labelled points and uses them to learn how to label other points. It classifies cases based on their similarity to other cases.
• Cases that are near each other are known as neighbours.

k-nearest neighbors algorithm - Wikipedia

How does it work?
• Pick a value for 'k'.
• Calculate the distance (Euclidean distance) of the unknown case from all the cases.
• Select the k observations in the training data that are "nearest" to the unknown data point.
• Predict the response of the unknown data point using the most popular response value from k-nearest neighbours.


Example of k-NN classification: The test sample (green dot) should be classified either into blue squares or red triangles. If k = 3 (solid line circle) it is assigned to the red triangles because there are 2 triangles and only 1 square inside the inner circle. If k = 5 (dashed line circle) it is assigned to the blue squares (3 squares vs. 2 triangles inside the outer circle).

How to pick 'k'?
• A small value of k causes a highly complex model that results in overfitting of the data. Whereas, if we choose a high value of k, the model becomes overly generalized.
• Data scientists usually choose an odd number if the number of classes is 2 and another simple approach to select k is to set k=sqrt(n).

Industry use of KNN :

Novartis AG uses Amazon OpenSearch Service K-Nearest Neighbor (KNN) and Amazon SageMaker to power search and recommendation (Part 3/4) | Amazon Web Services
September 8, 2021: Amazon Elasticsearch Service has been renamed to Amazon OpenSearch Service. See details. This is the third post of a four-part series on the strategic collaboration between AWS and Novartis AG, where the AWS Professional Services team built the Buying Engine platform. In this seri…

Logistic Regression

• Logistic regression is a classification method.
• Its main goal is learning a function that returns a yes or no answer when presented as input a so-called feature vector.
•  From this, we can infer answers to classification problems, such as 'will an employee stay with an organisation or not ?'

Logistic regression - Wikipedia

How does it work?

• The data structure of logistic regression is pretty much similar to that of linear regression - you have a set of independent variables (height in the previous example) and a dependent variable (weight ).
• Function behind it :

• Logistic regression will estimate the bo parameter, which in turn can be used to calculate the probability using a new set of independent variables. If the probability turns out to be less than 0.5 then it's classified as a 'no'.

Industry use of Logistic Regression:

Credit Scoring using Logistic Regression by Deloitte

Naïve Bayes Classifier

• Naïve Bayes Classifier is a probabilistic machine learning model that is used for classification problems.
•They are known by a variety of names such as simple Bayes and independence Bayes
• For example, a fruit may be considered to be an apple if it is red, round, and about 10 cm in diameter. A naive Bayes classifier considers each of these features to contribute independently to the probability that this fruit is an apple, regardless of any possible correlations between the colour, roundness, and diameter features.

Naive Bayes classifier - Wikipedia

How does it work?
• The core principle of the classifier is based on Bayes Theorem

• By using this principle, Naïve Bayes Classifier calculates the probability and separates it into a bivariate target variable using a threshold value of 0.5.

Industry use of Naïve Bayes Classifier:

Biological Leukemia Classification - TIBCO

Decision Tree

• Decision Tree uses a tree-like structure and their possible combinations to solve a particular problem.
• It belongs to the class of supervised learning algorithms where it can be used for both classification and regression purposes.

1.10. Decision Trees
Decision Trees (DTs) are a non-parametric supervised learning method used for classification and regression. The goal is to create a model that predicts the value of a target variable by learning s...

How does it work?
• For each attribute in the dataset, the Decision-Tree algorithm forms a node. The most important attribute is placed at the root node.
• For evaluating the task in hand, we start at the root node and we work our way down the tree by following the corresponding node that meets our condition or decision.
• This process continues until a leaf node is reached. It contains the prediction or the outcome of the Decision Tree.

Industry use of Decision Tree:

User Guide

Random Forest

• A random forest is a supervised machine learning technique that uses decision tree algorithms to build it. This algorithm is used to anticipate behaviour.
• Its ease of use and flexibility have fueled its adoption, as it handles both classification and regression problems.

Random forest - Wikipedia

How does it work?
• Random forest builds several decision trees and combines them to make predictions more reliable and stable.
• The random forest has the same hyperparameters as the decision tree or the baggage classifier.
• The Random Forest adds additional randomness to the model as the trees expand.

Industry use of Random Forest

Predicting System Disruption in Industry using Random Forest

Support Vector Machines

• Support vector machines (SVMs) are a particularly powerful and flexible class of supervised algorithms for both classification and regression.
• Naïve Bayes is an example of a generative classifier whereas Support Vector Machine is an example of a discriminative classifier.
• SVM divides the classes by simply drawing a line or a curve.

Support-vector machine - Wikipedia

How does it work?
• Consider the below data, the separation of classes is rather easy and can be done manually.

• There are 3 different separators, which perfectly discriminate between the two classes.

• Rather than simply drawing a zero-width line between the classes, we can draw around each line a margin of some width, up to the nearest point. In support vector machines, the line that maximizes this margin is the one we will choose as the optimal model.

Industry use of Support Vector Machines:

Biological applications of support vector machines
Abstract. One of the major tasks in bioinformatics is the classification and prediction of biological data. With the rapid increase in size of the biological da


I hope this post guides you through all the basic ML Algorithms and their inner functionalities.

For more such content, - consider subscribing.

We are on Twitter! To stay up-to-date on the newest posts on the blog. 

Follow @DatascienceFM

See you next time!