ๅฐŠๆ•ฌ็š„ ๅพฎไฟกๆฑ‡็Ž‡๏ผš1ๅ†† โ‰ˆ 0.046089 ๅ…ƒ ๆ”ฏไป˜ๅฎๆฑ‡็Ž‡๏ผš1ๅ†† โ‰ˆ 0.04618ๅ…ƒ [้€€ๅ‡บ็™ปๅฝ•]
SlideShare a Scribd company logo
LightGBM
Introduction
โ Boosting is an ensemble learning method that combines a set of weak learners into a
strong learner to minimize training errors.
โ Gradient Boosting is a powerful boosting algorithm that combines several weak
learners into strong learners, in which each new model is trained to minimize the loss
function such as mean squared error or cross-entropy of the previous model using
gradient descent.
โ LightGBM is a gradient boosting framework that uses tree based learning algorithms.
Advantages
โ— Faster training speed and higher efficiency.
โ— Lower memory usage.
โ— Better accuracy.
โ— Support of parallel, distributed, and GPU learning.
โ— Capable of handling large-scale data efficiently.
โ— Can handle categorical variable directly without the need for one-hot encoding.
What Makes LightGBM faster?
1. Histogram or bin way of splitting
For e.g. BU dataset has a column CSE-Students, in which weโ€™ve students from 6th,7th,
8th, 9th and 10th batch. Now, in other boosting methods all the batch will be tested that
wonโ€™t be minimal. So, now we can split the students into two bins, 6th-8th batch, and 9th-
10th batch. This will reduces the memory usage and speeds up the training process,.
What Makes LightGBM faster?(Cont.)
2. Exclusive Feature Building (EFB)
For e.g. weโ€™re considering gender of the respondents. If the respondents is a male, it will
enter 1 in the male column, 0 in female column, or if the respondents is a female, it will
enter 1 in the female column, 0 in male column. There is no chances to enter 1 in both
column at the same time. This type of features are called exclusive feature. LightGBM
will bundle this feature, by reducing two dimension into one dimension, through creating
a new feature, such as BF, that will contain 11 for male and 10 for female.
What Makes LightGBM faster?(Cont.)
3. GOSS (Gradient based One Side Sampling)
โ— It sees a error and decide how to create this sample
โ— For e.g. your baseline model is M0 on 500 records, i.e. you willโ€™ve 500 gradients or error. Let this is G1,G2,G3,โ€ฆ, G500.
Now LightGBM will sort it in descending order. Suppose, first gradient number 48 have have highest gradient record than 14,
and so on. So it will be now: G48, G14,..., G4.
Now certain percentage( usually 20%) from this record will be taken as one part (as top 20%) and from the remaining 80%
randomly selected certain percentage( usually 10%) will come out (as bottom subset 10%). Now these two are combined to
create new subsample.
Now If gradient is low, that means in this 80% the model performs good we donโ€™t need to train it again and again, but if in the
20% if the model is not performing well( gradients are high , errors are high), then it should train more. As a result top will take
high priority and sampling is done only from one side(right side ,80%).
LightGBM tree โ€“ growth strategies
โ— Light GBM grows tree vertically
while other algorithm grows trees
horizontally meaning that Light
GBM grows tree leaf- wise while
other algorithm grows level- wise.
โ— It will choose the leaf with max
delta loss to grow. When growing
the same leaf, Leaf-wise algorithm
can reduce more loss than a level-
wise algorithm
Where should we use LightGBM?
โ In our local machine, or anywhere where there is no gpu or no clustering
โ For performing faster machine learning tasks such as classification, regression and
ranking
LightGBM disadvantages
โ— Too many parameters
โ— Slow to tune parameters
โ— GPU configuration can be tough
โ— No GPU support on scikit โ€“learn API
Multilayer Perceptron (MLP)
Introduction
โ A multi-layer perceptron is a type of
Feed Forward Neural Network with
multiple neurons arranged in layers.
โ The network has at least three layers
with an input layer, one or more
hidden layers. and an output layer.
โ All the neurons in a layer are fully
connected to the neurons in the next
layer.
Working Process
โ The input layer is the visible layer.
โ It just passes the input to the next
layer.
โ The layers following the input layer
are the hidden layers.
โ The hidden layers neither directly
receive inputs nor send outputs to
the external environment.
โ The final layer is the output layer
which outputs a single value or a
vector of values.
Working Process(Cont.)
โ The activation functions used in the
layers can be linear or non-linear
depending on the type of the
problem modelled.
โ Typically, a sigmoid activation
function is used if the problem is a
binary classification problem and a
softmax activation function is used
in a multi-class classification
problem.
MLP Algorithms
Input: Input vector (x1, x2 ......, xn)
Output: Yn
Learning rate: ฮฑ
Assign random weights and biases for every connection in the network in the range [-0.5, +0.5].
Step 1: Forward Propagation
1. Calculate Input and Output in the Input Layer:
Input at Node j 'Ij' in the Input Layer is:
Where,
ฯฐj, is the input received at Node j
Output at Node j 'Oj' in the Input Layer is:
MLP Algorithms
Net Input at node j in the output layer is
๐ผ๐‘— = ๐›ด๐‘–=1
๐‘›
๐‘‚๐‘–๐‘ค๐‘–๐‘— + ๐‘ฅ0 * ๐œƒ๐‘—
where,
๐‘‚๐‘– is the output from Node i
๐‘ค๐‘–๐‘— is the weight in the link from Node i to Node j
๐‘ฅ0 is the input to the bias node โ€˜0โ€™ which is always assumed as 1
๐œƒ๐‘— is the weight in the link from the bias node โ€˜0โ€™ to Node j
Output at Node j:
๐‘‚๐‘— =
1
1 + โ…‡โˆ’๐ผ๐‘—
Where, ๐ผ๐‘— is the input received at Node j.
MLP Algorithms
โ— Estimated error at the node in the Output Layer:
Error = ๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ - ๐‘‚๐ธ๐‘ ๐‘ก๐‘–๐‘š๐‘Ž๐‘ก๐‘’๐‘‘
where,
๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ is the desired output value of the Node in the Output Layer
๐‘‚๐ธ๐‘ ๐‘ก๐‘–๐‘š๐‘Ž๐‘ก๐‘’๐‘‘ is the estimated output value of the Node in the Output Layer
MLP Algorithms
โ— Step 2: Backward Propagation
1. Calculated Error at each node:
For each Unit k in the Output Layer
๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘˜ = ๐‘‚๐‘˜(1-๐‘‚๐‘˜) (๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ -๐‘‚๐‘˜)
where,
๐‘‚๐‘˜ is the output value at Node k in the Output Layer
๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ is the desired output value at Node in the Output Layer
For each unit j in the Hidden Layer
๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ
๐‘— = ๐‘‚๐‘—(1-๐‘‚๐‘—)๐›ด๐‘˜๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘˜๐‘ค๐‘—๐‘˜
where,
๐‘‚๐‘— is the output value at Node j in the Hidden Layer
๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘˜ is the error at Node k in the Output Layer
๐‘ค๐‘—๐‘˜ is the weight in the link from Node j to Node k
MLP Algorithms
2. Update all weights and biases:
Update weights
where,
๐‘‚๐‘– is the output value at Node i
๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ
๐‘— is the error at Node j
๐›ผ is the learning rate
๐‘ค๐‘–๐‘— is the weight in the link from Node i to Node j
ฮ”๐‘ค๐‘–๐‘— is the difference in weight that has to be added to ๐‘ค๐‘–๐‘—
ฮ”๐‘ค๐‘–๐‘— = ๐›ผ * ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘— * ๐‘‚๐‘–
๐‘ค๐‘–๐‘— = ๐‘ค๐‘–๐‘— + ฮ”๐‘ค๐‘–๐‘—
MLPs Algorithms
Update Biases
where,
๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘— is the error at Node j
๐›ผ is the learning rate
๐œƒ๐‘— is the bias value from Bias Node 0 to Node j.
ฮ”๐œƒ๐‘— is the difference in bias that has to be added to ๐œƒ๐‘—.
ฮ”๐œƒ๐‘— = ๐›ผ * ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ
๐‘—
๐œƒ๐‘— =๐œƒ๐‘— + ฮ”๐œƒ๐‘—

More Related Content

Similar to LightGBM and Multilayer perceptron (MLP) slide

Lec 6-bp
Lec 6-bpLec 6-bp
Lec 6-bp
Taymoor Nazmy
ย 
deep CNN vs conventional ML
deep CNN vs conventional MLdeep CNN vs conventional ML
deep CNN vs conventional ML
Chao Han chaohan@vt.edu
ย 
Training Neural Networks.pptx
Training Neural Networks.pptxTraining Neural Networks.pptx
Training Neural Networks.pptx
ksghuge
ย 
Methods of Optimization in Machine Learning
Methods of Optimization in Machine LearningMethods of Optimization in Machine Learning
Methods of Optimization in Machine Learning
Knoldus Inc.
ย 
ICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via Randomization
ICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via RandomizationICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via Randomization
ICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via Randomization
Hidekazu Oiwa
ย 
Machine Learning With Neural Networks
Machine Learning  With Neural NetworksMachine Learning  With Neural Networks
Machine Learning With Neural Networks
Knoldus Inc.
ย 
Artificial Neural Network for machine learning
Artificial Neural Network for machine learningArtificial Neural Network for machine learning
Artificial Neural Network for machine learning
2303oyxxxjdeepak
ย 
Pregel: A System For Large Scale Graph Processing
Pregel: A System For Large Scale Graph ProcessingPregel: A System For Large Scale Graph Processing
Pregel: A System For Large Scale Graph Processing
Riyad Parvez
ย 
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
Thien Q. Tran
ย 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks
Tamer Ahmed Farrag, PhD
ย 
Transfer Learning _ Monk AI _ GCOEN.pptx
Transfer Learning _ Monk AI _ GCOEN.pptxTransfer Learning _ Monk AI _ GCOEN.pptx
Transfer Learning _ Monk AI _ GCOEN.pptx
GDSCGCOEN
ย 
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
Tahmid Abtahi
ย 
Machine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMachine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester Elective
MayuraD1
ย 
Feedforward neural network
Feedforward neural networkFeedforward neural network
Feedforward neural network
Sopheaktra YONG
ย 
Machine Learning - Supervised Learning
Machine Learning - Supervised LearningMachine Learning - Supervised Learning
Machine Learning - Supervised Learning
Giorgio Alfredo Spedicato
ย 
Multilayer & Back propagation algorithm
Multilayer & Back propagation algorithmMultilayer & Back propagation algorithm
Multilayer & Back propagation algorithm
swapnac12
ย 
eam2
eam2eam2
eam2
butest
ย 
Using CNTK's Python Interface for Deep LearningDave DeBarr -
Using CNTK's Python Interface for Deep LearningDave DeBarr - Using CNTK's Python Interface for Deep LearningDave DeBarr -
Using CNTK's Python Interface for Deep LearningDave DeBarr -
PyData
ย 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
aciijournal
ย 
Eye deep
Eye deepEye deep
Eye deep
sveitser
ย 

Similar to LightGBM and Multilayer perceptron (MLP) slide (20)

Lec 6-bp
Lec 6-bpLec 6-bp
Lec 6-bp
ย 
deep CNN vs conventional ML
deep CNN vs conventional MLdeep CNN vs conventional ML
deep CNN vs conventional ML
ย 
Training Neural Networks.pptx
Training Neural Networks.pptxTraining Neural Networks.pptx
Training Neural Networks.pptx
ย 
Methods of Optimization in Machine Learning
Methods of Optimization in Machine LearningMethods of Optimization in Machine Learning
Methods of Optimization in Machine Learning
ย 
ICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via Randomization
ICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via RandomizationICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via Randomization
ICML2013่ชญใฟไผš Large-Scale Learning with Less RAM via Randomization
ย 
Machine Learning With Neural Networks
Machine Learning  With Neural NetworksMachine Learning  With Neural Networks
Machine Learning With Neural Networks
ย 
Artificial Neural Network for machine learning
Artificial Neural Network for machine learningArtificial Neural Network for machine learning
Artificial Neural Network for machine learning
ย 
Pregel: A System For Large Scale Graph Processing
Pregel: A System For Large Scale Graph ProcessingPregel: A System For Large Scale Graph Processing
Pregel: A System For Large Scale Graph Processing
ย 
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
Set Transfomer: A Framework for Attention-based Permutaion-Invariant Neural N...
ย 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks
ย 
Transfer Learning _ Monk AI _ GCOEN.pptx
Transfer Learning _ Monk AI _ GCOEN.pptxTransfer Learning _ Monk AI _ GCOEN.pptx
Transfer Learning _ Monk AI _ GCOEN.pptx
ย 
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
A Framework for Scene Recognition Using Convolutional Neural Network as Featu...
ย 
Machine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMachine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester Elective
ย 
Feedforward neural network
Feedforward neural networkFeedforward neural network
Feedforward neural network
ย 
Machine Learning - Supervised Learning
Machine Learning - Supervised LearningMachine Learning - Supervised Learning
Machine Learning - Supervised Learning
ย 
Multilayer & Back propagation algorithm
Multilayer & Back propagation algorithmMultilayer & Back propagation algorithm
Multilayer & Back propagation algorithm
ย 
eam2
eam2eam2
eam2
ย 
Using CNTK's Python Interface for Deep LearningDave DeBarr -
Using CNTK's Python Interface for Deep LearningDave DeBarr - Using CNTK's Python Interface for Deep LearningDave DeBarr -
Using CNTK's Python Interface for Deep LearningDave DeBarr -
ย 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
ย 
Eye deep
Eye deepEye deep
Eye deep
ย 

Recently uploaded

Diversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT KanpurDiversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT Kanpur
Quiz Club IIT Kanpur
ย 
Library news letter Kitengesa Uganda June 2024
Library news letter Kitengesa Uganda June 2024Library news letter Kitengesa Uganda June 2024
Library news letter Kitengesa Uganda June 2024
Friends of African Village Libraries
ย 
Images as attribute values in the Odoo 17
Images as attribute values in the Odoo 17Images as attribute values in the Odoo 17
Images as attribute values in the Odoo 17
Celine George
ย 
Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...
Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...
Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...
Nguyen Thanh Tu Collection
ย 
Post init hook in the odoo 17 ERP Module
Post init hook in the  odoo 17 ERP ModulePost init hook in the  odoo 17 ERP Module
Post init hook in the odoo 17 ERP Module
Celine George
ย 
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
Kalna College
ย 
Diversity Quiz Prelims by Quiz Club, IIT Kanpur
Diversity Quiz Prelims by Quiz Club, IIT KanpurDiversity Quiz Prelims by Quiz Club, IIT Kanpur
Diversity Quiz Prelims by Quiz Club, IIT Kanpur
Quiz Club IIT Kanpur
ย 
Keynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse CityKeynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse City
PJ Caposey
ย 
How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17
Celine George
ย 
The Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teachingThe Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teaching
Derek Wenmoth
ย 
Slides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptxSlides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptx
shabeluno
ย 
Creativity for Innovation and Speechmaking
Creativity for Innovation and SpeechmakingCreativity for Innovation and Speechmaking
Creativity for Innovation and Speechmaking
MattVassar1
ย 
Cross-Cultural Leadership and Communication
Cross-Cultural Leadership and CommunicationCross-Cultural Leadership and Communication
Cross-Cultural Leadership and Communication
MattVassar1
ย 
How to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRMHow to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRM
Celine George
ย 
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
Kalna College
ย 
What are the new features in the Fleet Odoo 17
What are the new features in the Fleet Odoo 17What are the new features in the Fleet Odoo 17
What are the new features in the Fleet Odoo 17
Celine George
ย 
Hospital pharmacy and it's organization (1).pdf
Hospital pharmacy and it's organization (1).pdfHospital pharmacy and it's organization (1).pdf
Hospital pharmacy and it's organization (1).pdf
ShwetaGawande8
ย 
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
biruktesfaye27
ย 
Erasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES CroatiaErasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES Croatia
whatchangedhowreflec
ย 
Opportunity scholarships and the schools that receive them
Opportunity scholarships and the schools that receive themOpportunity scholarships and the schools that receive them
Opportunity scholarships and the schools that receive them
EducationNC
ย 

Recently uploaded (20)

Diversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT KanpurDiversity Quiz Finals by Quiz Club, IIT Kanpur
Diversity Quiz Finals by Quiz Club, IIT Kanpur
ย 
Library news letter Kitengesa Uganda June 2024
Library news letter Kitengesa Uganda June 2024Library news letter Kitengesa Uganda June 2024
Library news letter Kitengesa Uganda June 2024
ย 
Images as attribute values in the Odoo 17
Images as attribute values in the Odoo 17Images as attribute values in the Odoo 17
Images as attribute values in the Odoo 17
ย 
Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...
Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...
Bร€I TแบฌP Bแป” TRแปข 4 Kแปธ Nฤ‚NG TIแบพNG ANH LแปšP 9 - GLOBAL SUCCESS - FORM MแปšI 2025 - C...
ย 
Post init hook in the odoo 17 ERP Module
Post init hook in the  odoo 17 ERP ModulePost init hook in the  odoo 17 ERP Module
Post init hook in the odoo 17 ERP Module
ย 
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...220711130095 Tanu Pandey message currency, communication speed & control EPC ...
220711130095 Tanu Pandey message currency, communication speed & control EPC ...
ย 
Diversity Quiz Prelims by Quiz Club, IIT Kanpur
Diversity Quiz Prelims by Quiz Club, IIT KanpurDiversity Quiz Prelims by Quiz Club, IIT Kanpur
Diversity Quiz Prelims by Quiz Club, IIT Kanpur
ย 
Keynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse CityKeynote given on June 24 for MASSP at Grand Traverse City
Keynote given on June 24 for MASSP at Grand Traverse City
ย 
How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17How to Create User Notification in Odoo 17
How to Create User Notification in Odoo 17
ย 
The Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teachingThe Science of Learning: implications for modern teaching
The Science of Learning: implications for modern teaching
ย 
Slides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptxSlides Peluncuran Amalan Pemakanan Sihat.pptx
Slides Peluncuran Amalan Pemakanan Sihat.pptx
ย 
Creativity for Innovation and Speechmaking
Creativity for Innovation and SpeechmakingCreativity for Innovation and Speechmaking
Creativity for Innovation and Speechmaking
ย 
Cross-Cultural Leadership and Communication
Cross-Cultural Leadership and CommunicationCross-Cultural Leadership and Communication
Cross-Cultural Leadership and Communication
ย 
How to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRMHow to Create a Stage or a Pipeline in Odoo 17 CRM
How to Create a Stage or a Pipeline in Odoo 17 CRM
ย 
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
220711130086 Sukanta Singh E learning and mobile learning EPC 3 Internal Asse...
ย 
What are the new features in the Fleet Odoo 17
What are the new features in the Fleet Odoo 17What are the new features in the Fleet Odoo 17
What are the new features in the Fleet Odoo 17
ย 
Hospital pharmacy and it's organization (1).pdf
Hospital pharmacy and it's organization (1).pdfHospital pharmacy and it's organization (1).pdf
Hospital pharmacy and it's organization (1).pdf
ย 
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
Ethiopia and Eritrea Eritrea's journey has been marked by resilience and dete...
ย 
Erasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES CroatiaErasmus + DISSEMINATION ACTIVITIES Croatia
Erasmus + DISSEMINATION ACTIVITIES Croatia
ย 
Opportunity scholarships and the schools that receive them
Opportunity scholarships and the schools that receive themOpportunity scholarships and the schools that receive them
Opportunity scholarships and the schools that receive them
ย 

LightGBM and Multilayer perceptron (MLP) slide

  • 2. Introduction โ Boosting is an ensemble learning method that combines a set of weak learners into a strong learner to minimize training errors. โ Gradient Boosting is a powerful boosting algorithm that combines several weak learners into strong learners, in which each new model is trained to minimize the loss function such as mean squared error or cross-entropy of the previous model using gradient descent. โ LightGBM is a gradient boosting framework that uses tree based learning algorithms.
  • 3. Advantages โ— Faster training speed and higher efficiency. โ— Lower memory usage. โ— Better accuracy. โ— Support of parallel, distributed, and GPU learning. โ— Capable of handling large-scale data efficiently. โ— Can handle categorical variable directly without the need for one-hot encoding.
  • 4. What Makes LightGBM faster? 1. Histogram or bin way of splitting For e.g. BU dataset has a column CSE-Students, in which weโ€™ve students from 6th,7th, 8th, 9th and 10th batch. Now, in other boosting methods all the batch will be tested that wonโ€™t be minimal. So, now we can split the students into two bins, 6th-8th batch, and 9th- 10th batch. This will reduces the memory usage and speeds up the training process,.
  • 5. What Makes LightGBM faster?(Cont.) 2. Exclusive Feature Building (EFB) For e.g. weโ€™re considering gender of the respondents. If the respondents is a male, it will enter 1 in the male column, 0 in female column, or if the respondents is a female, it will enter 1 in the female column, 0 in male column. There is no chances to enter 1 in both column at the same time. This type of features are called exclusive feature. LightGBM will bundle this feature, by reducing two dimension into one dimension, through creating a new feature, such as BF, that will contain 11 for male and 10 for female.
  • 6. What Makes LightGBM faster?(Cont.) 3. GOSS (Gradient based One Side Sampling) โ— It sees a error and decide how to create this sample โ— For e.g. your baseline model is M0 on 500 records, i.e. you willโ€™ve 500 gradients or error. Let this is G1,G2,G3,โ€ฆ, G500. Now LightGBM will sort it in descending order. Suppose, first gradient number 48 have have highest gradient record than 14, and so on. So it will be now: G48, G14,..., G4. Now certain percentage( usually 20%) from this record will be taken as one part (as top 20%) and from the remaining 80% randomly selected certain percentage( usually 10%) will come out (as bottom subset 10%). Now these two are combined to create new subsample. Now If gradient is low, that means in this 80% the model performs good we donโ€™t need to train it again and again, but if in the 20% if the model is not performing well( gradients are high , errors are high), then it should train more. As a result top will take high priority and sampling is done only from one side(right side ,80%).
  • 7. LightGBM tree โ€“ growth strategies โ— Light GBM grows tree vertically while other algorithm grows trees horizontally meaning that Light GBM grows tree leaf- wise while other algorithm grows level- wise. โ— It will choose the leaf with max delta loss to grow. When growing the same leaf, Leaf-wise algorithm can reduce more loss than a level- wise algorithm
  • 8. Where should we use LightGBM? โ In our local machine, or anywhere where there is no gpu or no clustering โ For performing faster machine learning tasks such as classification, regression and ranking
  • 9. LightGBM disadvantages โ— Too many parameters โ— Slow to tune parameters โ— GPU configuration can be tough โ— No GPU support on scikit โ€“learn API
  • 11. Introduction โ A multi-layer perceptron is a type of Feed Forward Neural Network with multiple neurons arranged in layers. โ The network has at least three layers with an input layer, one or more hidden layers. and an output layer. โ All the neurons in a layer are fully connected to the neurons in the next layer.
  • 12. Working Process โ The input layer is the visible layer. โ It just passes the input to the next layer. โ The layers following the input layer are the hidden layers. โ The hidden layers neither directly receive inputs nor send outputs to the external environment. โ The final layer is the output layer which outputs a single value or a vector of values.
  • 13. Working Process(Cont.) โ The activation functions used in the layers can be linear or non-linear depending on the type of the problem modelled. โ Typically, a sigmoid activation function is used if the problem is a binary classification problem and a softmax activation function is used in a multi-class classification problem.
  • 14. MLP Algorithms Input: Input vector (x1, x2 ......, xn) Output: Yn Learning rate: ฮฑ Assign random weights and biases for every connection in the network in the range [-0.5, +0.5]. Step 1: Forward Propagation 1. Calculate Input and Output in the Input Layer: Input at Node j 'Ij' in the Input Layer is: Where, ฯฐj, is the input received at Node j Output at Node j 'Oj' in the Input Layer is:
  • 15. MLP Algorithms Net Input at node j in the output layer is ๐ผ๐‘— = ๐›ด๐‘–=1 ๐‘› ๐‘‚๐‘–๐‘ค๐‘–๐‘— + ๐‘ฅ0 * ๐œƒ๐‘— where, ๐‘‚๐‘– is the output from Node i ๐‘ค๐‘–๐‘— is the weight in the link from Node i to Node j ๐‘ฅ0 is the input to the bias node โ€˜0โ€™ which is always assumed as 1 ๐œƒ๐‘— is the weight in the link from the bias node โ€˜0โ€™ to Node j Output at Node j: ๐‘‚๐‘— = 1 1 + โ…‡โˆ’๐ผ๐‘— Where, ๐ผ๐‘— is the input received at Node j.
  • 16. MLP Algorithms โ— Estimated error at the node in the Output Layer: Error = ๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ - ๐‘‚๐ธ๐‘ ๐‘ก๐‘–๐‘š๐‘Ž๐‘ก๐‘’๐‘‘ where, ๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ is the desired output value of the Node in the Output Layer ๐‘‚๐ธ๐‘ ๐‘ก๐‘–๐‘š๐‘Ž๐‘ก๐‘’๐‘‘ is the estimated output value of the Node in the Output Layer
  • 17. MLP Algorithms โ— Step 2: Backward Propagation 1. Calculated Error at each node: For each Unit k in the Output Layer ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘˜ = ๐‘‚๐‘˜(1-๐‘‚๐‘˜) (๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ -๐‘‚๐‘˜) where, ๐‘‚๐‘˜ is the output value at Node k in the Output Layer ๐‘‚๐ท๐‘’๐‘ ๐‘–๐‘Ÿ๐‘’๐‘‘ is the desired output value at Node in the Output Layer For each unit j in the Hidden Layer ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ ๐‘— = ๐‘‚๐‘—(1-๐‘‚๐‘—)๐›ด๐‘˜๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘˜๐‘ค๐‘—๐‘˜ where, ๐‘‚๐‘— is the output value at Node j in the Hidden Layer ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘˜ is the error at Node k in the Output Layer ๐‘ค๐‘—๐‘˜ is the weight in the link from Node j to Node k
  • 18. MLP Algorithms 2. Update all weights and biases: Update weights where, ๐‘‚๐‘– is the output value at Node i ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ ๐‘— is the error at Node j ๐›ผ is the learning rate ๐‘ค๐‘–๐‘— is the weight in the link from Node i to Node j ฮ”๐‘ค๐‘–๐‘— is the difference in weight that has to be added to ๐‘ค๐‘–๐‘— ฮ”๐‘ค๐‘–๐‘— = ๐›ผ * ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘— * ๐‘‚๐‘– ๐‘ค๐‘–๐‘— = ๐‘ค๐‘–๐‘— + ฮ”๐‘ค๐‘–๐‘—
  • 19. MLPs Algorithms Update Biases where, ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ๐‘— is the error at Node j ๐›ผ is the learning rate ๐œƒ๐‘— is the bias value from Bias Node 0 to Node j. ฮ”๐œƒ๐‘— is the difference in bias that has to be added to ๐œƒ๐‘—. ฮ”๐œƒ๐‘— = ๐›ผ * ๐ธ๐‘Ÿ๐‘Ÿ๐‘œ๐‘Ÿ ๐‘— ๐œƒ๐‘— =๐œƒ๐‘— + ฮ”๐œƒ๐‘—

Editor's Notes

  1. Gradient descent is an optimization algorithm used to train machine learning models by minimizing errors between predicted and actual results.
  2. If in this 20% the model performs good we donโ€™t need to train it again and again, but if the results is bad i.e. error is high
  3. In your local machine, or anywhere where there is gpu or clustering, use XGBM
  ็ฟป่ฏ‘๏ผš