Simple explanation regarding K-means Clustering in Unsupervised Learning and simple practice with sklearn in python
Previous article :
In the previous article, I was explained regarding Clustering Intuition.
Clustering : grouping data based on similarity patterns based on distance
Simple explanation regarding Clustering in Unsupervised Learning
According to Wikipedia :
Unsupervised learning is a type of machine learning that looks for previously undetected patterns in a data set with no pre-existing labels and with a minimum of human supervision. In contrast to supervised learning that usually makes use of human-labeled data, unsupervised learning, also known as self-organization allows for modeling of probability densities over inputs. …
What Is Machine Learning? What Is Supervised Learning? and What Is Unsupervised Learning? — Simple explanation regarding Machine Learning
An introduction and intuition, how evaluate regression and Classification model in general
Data scientists often use machine learning models to generate insight but wow does a Data scientist make a decision, whether the model will be implemented or not? When the model is implemented, there will be negative and positive impacts on the business. In order to prevent or minimize negative impacts, it is necessary to evaluate the model, so that it can estimate the positive impact and negative impact generated. Is this, model evaluation is one of the most important parts of machine learning.
In this moment, we’ll learn…
Intuition, Motivation, and How it works on the Random Forest Method
Random Forest lives up to its name : simply, made up of several trees. In more details, Random Forest is a set of decision trees built on random samples with different policies for splitting a node . The implementation, random forest uses the bootstrap method in building decision trees and there are two ways to interpret these results; the more common approach is based on a majority vote in classification case and an average in regression case.
The ideas behind Random Forest. in fact, are containing so many topics…
Intuition, motivation, and how it works on the bootstrap method
In statistics, the bootstrap is a widely applicable and extremely powerful statistical tool that can be used to quantify the uncertainty associated with a given estimator or statistical learning method. The bootstrap can derive a strong estimate of a population parameter such as standard deviation, mean, median, standard error, etc.
In wikipedia — the bootstrap is statistical method for estimating the sampling distribution of an estimator by sampling with replacement from the original sample, most often with the purpose of deriving robust estimates of standard errors and confidence intervals of…
Decision Tree Algorithms — Part 3
As has been explained, Decision Trees is the non-parametric supervised learning approach. In addition to classification with continuous data on the target, we also often find cases with discrete data on the target called regression. In the regression, the simple way can be to use Linear Regression to solve this case. …
Decision Tree Algorithms — Part 2
CART (Classification And Regression Tree) is a decision tree algorithm variation, in the previous article — The Basics of Decision Trees. Decision Trees is the non-parametric supervised learning approach. CART can be applied to both regression and classification problems.
As we know, data scientists often use decision trees to solve regression and classification problems and most of them use scikit-learn in decision tree implementation. Based on documentation, scikit-learn uses an optimised version of the CART algorithm
in the previous article it was explained that CART uses Gini Impurity in the process of splitting the…
Decision Tree Algorithms - Part 1
Decision Trees is the non-parametric supervised learning approach, and can be applied to both regression and classification problems. In keeping with the tree analogy, decision trees implement a sequential decision process. Starting from the root node, a feature is evaluated and one of the two nodes (branches) is selected, Each node in the tree is basically a decision rule. This procedure is repeated until a final leaf is reached, which normally represents the target. Decision trees are also attractive models if we care about interpretability.
There are algorithms for creating decision trees :
Linear regression is the supervised learning approach. In particular, linear regression is a useful method for predicting continuous values (target) and attempts to model the linear relationship between target and one or more predictor.
2.Simple Linear Regression (one predictor)
Simple linear regression lives up to its name: simply, finding a relationship between one predictor (𝑥) and target (y). Mathematically, we can write this linear relationship as
In Equation 1, ŷ is target, 𝑥 is Predictor, 𝜷₀ and 𝜷₁ (coefficients) are two unknown constants that represent the intercept and slope terms in the linear model. Simple linear regression attempts to…
Data Scientist and Artificial Intelligence Enthusiast