SlideShare a Scribd company logo
1 of 31
Download to read offline
Machine Learning for Language Technology 2015
http://stp.lingfil.uu.se/~santinim/ml/2015/ml4lt_2015.htm
Basic Concepts of Machine Learning
Induction & Evaluation
Marina Santini
santinim@stp.lingfil.uu.se
Department of Linguistics and Philology
Uppsala University, Uppsala, Sweden
Autumn 2015
Acknowledgments
• Daume’ (2015), Alpaydin (2010), NLTK
website, other web sites.
Lecture 3: Basic Concepts of ML 2
Outline
• Induction
– Induction pipeline
• Training set, test set and development set
• Parameters
• Hyperparameters
• Accuracy, precision, recall, f-measure
• Confusion matrix
• Crossvalidation
• Leave one out
• Stratification
Lecture 3: Basic Concepts of ML 3
Induction
• Induction is the process of reaching a general
conclusion from specific examples.
Lecture 3: Basic Concepts of ML 4
Inductive Machine Learning
• The goal of inductive machine learning is to take
some training data and use it to induce a function
(model, classifier, learning algorithm).
• This function will be evaluated on the test data.
• The machine learning algorithm has succeeded if
its performance on the test data is high.
Lecture 3: Basic Concepts of ML 5
Pipeline
• Induction pipeline
Lecture 3: Basic Concepts of ML 6
Task
• Predict the class for this ”unseen” example:
Sepal length – Sepal width – Petal length – Petal width - Type
5.2 3.7 1.7 0.3 ???
Lecture 1: What is Machine Learning? 7
Require us to
generalize from
the training data
Splitting data to measure performance
• Training data& Test Data
– Common splits: 80/20; 90/10
• NEVER TOUCH THE TEST DATA!
• TEST DATA MUST BELONG TO THE SAME
STATISTICAL DISTRIBUTION AS THE TRAINING DATA
Lecture 3: Basic Concepts of ML 8
Modelling
• ML uses formal models that might perform well
on our data.
• The choice of using one model rather than
another is our choice.
• A model tells us what sort of things we can learn.
• A model tells us what our inductive bias is.
Lecture 3: Basic Concepts of ML 9
Parameters
• Models can have many parameters and
finding the best combination of parameters is
not trivial.
Lecture 3: Basic Concepts of ML 10
Hyperparameters
• A hyperparameter is a parameter that controls
other parameters of the model.
Lecture 3: Basic Concepts of ML 11
Development Set
• Split your data into 70% training data, 10% development
data and 20% test data.
• For each possible setting of the hyperparameters:
– Train a model using that setting on the training data
– Compute the model error rate on the development
data
– From the above collection of medels, choos the one
that achieve the lowest error rate on development
data.
– Evaluate that model on the test data to estimate
future test performance.
Lecture 3: Basic Concepts of ML 12
Accuracy
• Accuracy measures the percentage of correct
results that a classifier has achieved.
Lecture 3: Basic Concepts of ML 13
True and False Positives and Negatives
• True positives are relevant items that we correctly identified as relevant.
• True negatives are irrelevant items that we correctly identified as
irrelevant.
• False positives (or Type I errors) are irrelevant items that we incorrectly
identified as relevant.
• False negatives (or Type II errors) are relevant items that we incorrectly
identified as irrelevant.
Lecture 3: Basic Concepts of ML 14
Precision, Recall, F-Measure
• Given these four numbers, we can define the
following metrics:
– Precision, which indicates how many of the items that
we identified were relevant, is TP/(TP+FP).
– Recall, which indicates how many of the relevant
items that we identified, is TP/(TP+FN).
– The F-Measure (or F-Score), which combines the
precision and recall to give a single score, is defined to
be the harmonic mean of the precision and recall: (2
× Precision × Recall) / (Precision + Recall).
Lecture 3: Basic Concepts of ML 15
Accuracy, Precision, Recall, F-measure
• Accuracy = (TP + TN)/(TP + TN + FP + FN)
• Precision = TP / TP + FP
• Recall = TP / TP + FN
• F-measure = 2*((precision*recall)/(precision+recall))
Lecture 3: Basic Concepts of ML 16
Confusion Matrix
• This is a useful table that presents both the class
distribution in the data and the classifiers
predicted class distribution with a breakdown of
error types.
• Usually, the rows are the observed/actual class
labels and the columns the predicted class labels.
• Each cell contains the number of predictions
made by the classifier that fall into that cell.
Lecture 3: Basic Concepts of ML 17
actual
predicted
Multi-Class Confusion Matrix
• If a classification system has been trained to
distinguish between cats, dogs and rabbits, a
confusion matrix will summarize the results:
Lecture 3: Basic Concepts of ML 18
Cross validation
• In 10-fold cross-validation you break you
training data up into 10 equally-sized
partitions.
• You train a learning algorithm on 9 of them
and tst it on the remaining 1.
• You do this 10 times, each holding out a
different partition as the test data.
• Typical choices for n-fold are 2, 5, 10.
• 10-fold cross validation is the most common.
Lecture 3: Basic Concepts of ML 19
Leave One Out
• Leave One Out (or LOO) is a simple cross-
validation. Each learning set is created by
taking all the samples except one, the test set
being the sample left out.
Lecture 3: Basic Concepts of ML 20
Stratification
• Proportion of each class in the traning set and
test sets is the same as the proportion in the
original sample.
Lecture 3: Basic Concepts of ML 21
Weka Cross validation
• 10-fold cross validation
Lecture 3: Basic Concepts of ML 22
Weka: Output
• Classifier output
Lecture 3: Basic Concepts of ML 23
Remember: Underfitting & Overfitting
Underfitting: the model has not learned enough
from the data and is unable to generalize
Overfitting: the model has learned too many
idiosyncrasies (noise) and is unable to generalize
Lecture 3: Basic Concepts of ML 24
Summary: Performance of a learning
model: Requirements
• Our goal when we choose a machine learning
model is that it does well on future, unseen data.
• The way in which we measure performance
should depend on the problem we are trying to
solve.
• There should be a strong relationship between
the data that our algorithm sees at training time
and the data it sees at test time.
Lecture 3: Basic Concepts of ML 25
Not everything is learnable
– Noise at feature level
– Noise at class label level
– Features are insufficient
– Labels are controversial
– Inductive bias not appropriate for the kind of
problem we try to learn
Lecture 3: Decision Trees (1) 26
Quiz 1: Stratification
• What does it mean ”stratified” cross validation?
1. The examples of a class are all in the training set, and the rest
of the classes are in the test set.
2. The proportion of each class in the sets ae the same as the
proportion in the original sample
3. None of the above.
Lecture 3: Basic Concepts of ML 27
Quiz 2: Accuracy
• Why is accuracy alone an unreliable measure?
1. Because it can be biassed towards the most frequent
class.
2. Because it always guesses wrong.
3. None of the above
Lecture 3: Basic Concepts of ML 28
Quiz 3: Data Splits
• Which are recommended splits between
training and test data?
1. 80/20
2. 50/50
3. 10/90
Lecture 3: Basic Concepts of ML 29
Quiz 4: Overfitting
• What does it mean overfitting?
1. the model has not learned enough from the data and
is unable to generalize
2. The proportion of each class in the sets is the same as
the proportion in the original sample
3. None of the above.
Lecture 3: Basic Concepts of ML 30
The End
Lecture 3: Basic Concepts of ML 31

More Related Content

What's hot

Basics of Machine Learning
Basics of Machine LearningBasics of Machine Learning
Basics of Machine Learning
butest
 

What's hot (20)

Machine learning
Machine learning Machine learning
Machine learning
 
Ensemble learning
Ensemble learningEnsemble learning
Ensemble learning
 
Breast Cancer Detection with Convolutional Neural Networks (CNN)
Breast Cancer Detection with Convolutional Neural Networks (CNN)Breast Cancer Detection with Convolutional Neural Networks (CNN)
Breast Cancer Detection with Convolutional Neural Networks (CNN)
 
Bagging.pptx
Bagging.pptxBagging.pptx
Bagging.pptx
 
Classification in data mining
Classification in data mining Classification in data mining
Classification in data mining
 
Classification Based Machine Learning Algorithms
Classification Based Machine Learning AlgorithmsClassification Based Machine Learning Algorithms
Classification Based Machine Learning Algorithms
 
Machine Learning - Splitting Datasets
Machine Learning - Splitting DatasetsMachine Learning - Splitting Datasets
Machine Learning - Splitting Datasets
 
Machine learning
Machine learningMachine learning
Machine learning
 
Ensemble learning
Ensemble learningEnsemble learning
Ensemble learning
 
Introduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-LearnIntroduction to Machine Learning with SciKit-Learn
Introduction to Machine Learning with SciKit-Learn
 
Machine learning
Machine learningMachine learning
Machine learning
 
Data Mining: Concepts and Techniques_ Chapter 6: Mining Frequent Patterns, ...
Data Mining:  Concepts and Techniques_ Chapter 6: Mining Frequent Patterns, ...Data Mining:  Concepts and Techniques_ Chapter 6: Mining Frequent Patterns, ...
Data Mining: Concepts and Techniques_ Chapter 6: Mining Frequent Patterns, ...
 
Unsupervised learning (clustering)
Unsupervised learning (clustering)Unsupervised learning (clustering)
Unsupervised learning (clustering)
 
Lecture 2 Basic Concepts in Machine Learning for Language Technology
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyLecture 2 Basic Concepts in Machine Learning for Language Technology
Lecture 2 Basic Concepts in Machine Learning for Language Technology
 
supervised learning
supervised learningsupervised learning
supervised learning
 
Confusion Matrix
Confusion MatrixConfusion Matrix
Confusion Matrix
 
Basics of Machine Learning
Basics of Machine LearningBasics of Machine Learning
Basics of Machine Learning
 
L2. Evaluating Machine Learning Algorithms I
L2. Evaluating Machine Learning Algorithms IL2. Evaluating Machine Learning Algorithms I
L2. Evaluating Machine Learning Algorithms I
 
DMTM Lecture 06 Classification evaluation
DMTM Lecture 06 Classification evaluationDMTM Lecture 06 Classification evaluation
DMTM Lecture 06 Classification evaluation
 
Id3 algorithm
Id3 algorithmId3 algorithm
Id3 algorithm
 

Viewers also liked

Information Gain
Information GainInformation Gain
Information Gain
guest32311f
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
butest
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learning
butest
 
Machine Learning and Inductive Inference
Machine Learning and Inductive InferenceMachine Learning and Inductive Inference
Machine Learning and Inductive Inference
butest
 
CS364 Artificial Intelligence Machine Learning
CS364 Artificial Intelligence Machine LearningCS364 Artificial Intelligence Machine Learning
CS364 Artificial Intelligence Machine Learning
butest
 
Towards Contextualized Information: How Automatic Genre Identification Can Help
Towards Contextualized Information: How Automatic Genre Identification Can HelpTowards Contextualized Information: How Automatic Genre Identification Can Help
Towards Contextualized Information: How Automatic Genre Identification Can Help
Marina Santini
 

Viewers also liked (20)

Lecture 3b: Decision Trees (1 part)
Lecture 3b: Decision Trees (1 part)Lecture 3b: Decision Trees (1 part)
Lecture 3b: Decision Trees (1 part)
 
Lecture 4 Decision Trees (2): Entropy, Information Gain, Gain Ratio
Lecture 4 Decision Trees (2): Entropy, Information Gain, Gain RatioLecture 4 Decision Trees (2): Entropy, Information Gain, Gain Ratio
Lecture 4 Decision Trees (2): Entropy, Information Gain, Gain Ratio
 
Lecture 2: Preliminaries (Understanding and Preprocessing data)
Lecture 2: Preliminaries (Understanding and Preprocessing data)Lecture 2: Preliminaries (Understanding and Preprocessing data)
Lecture 2: Preliminaries (Understanding and Preprocessing data)
 
Machine Learning with Applications in Categorization, Popularity and Sequence...
Machine Learning with Applications in Categorization, Popularity and Sequence...Machine Learning with Applications in Categorization, Popularity and Sequence...
Machine Learning with Applications in Categorization, Popularity and Sequence...
 
Lecture: Semantic Word Clouds
Lecture: Semantic Word CloudsLecture: Semantic Word Clouds
Lecture: Semantic Word Clouds
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
 
Information Gain
Information GainInformation Gain
Information Gain
 
Aisb cyberbullying
Aisb cyberbullyingAisb cyberbullying
Aisb cyberbullying
 
Practical machine learning - Part 1
Practical machine learning - Part 1Practical machine learning - Part 1
Practical machine learning - Part 1
 
Overfitting and-tbl
Overfitting and-tblOverfitting and-tbl
Overfitting and-tbl
 
How Emotional Are Users' Needs? Emotion in Query Logs
How Emotional Are Users' Needs? Emotion in Query LogsHow Emotional Are Users' Needs? Emotion in Query Logs
How Emotional Are Users' Needs? Emotion in Query Logs
 
Text Analytics for Semantic Computing
Text Analytics for Semantic ComputingText Analytics for Semantic Computing
Text Analytics for Semantic Computing
 
Bridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
Bridging the Gap: Machine Learning for Ubiquitous Computing -- EvaluationBridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
Bridging the Gap: Machine Learning for Ubiquitous Computing -- Evaluation
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
 
Comparing Machine Learning Algorithms in Text Mining
Comparing Machine Learning Algorithms in Text MiningComparing Machine Learning Algorithms in Text Mining
Comparing Machine Learning Algorithms in Text Mining
 
Introduction to machine learning
Introduction to machine learningIntroduction to machine learning
Introduction to machine learning
 
Ijcai ip-2015 cyberbullying-final
Ijcai ip-2015 cyberbullying-finalIjcai ip-2015 cyberbullying-final
Ijcai ip-2015 cyberbullying-final
 
Machine Learning and Inductive Inference
Machine Learning and Inductive InferenceMachine Learning and Inductive Inference
Machine Learning and Inductive Inference
 
CS364 Artificial Intelligence Machine Learning
CS364 Artificial Intelligence Machine LearningCS364 Artificial Intelligence Machine Learning
CS364 Artificial Intelligence Machine Learning
 
Towards Contextualized Information: How Automatic Genre Identification Can Help
Towards Contextualized Information: How Automatic Genre Identification Can HelpTowards Contextualized Information: How Automatic Genre Identification Can Help
Towards Contextualized Information: How Automatic Genre Identification Can Help
 

Similar to Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation

LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptxLETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
shamsul2010
 
Machine Learning
Machine LearningMachine Learning
Machine Learning
butest
 
Introduction to machine learning-2023-IT-AI and DS.pdf
Introduction to machine learning-2023-IT-AI and DS.pdfIntroduction to machine learning-2023-IT-AI and DS.pdf
Introduction to machine learning-2023-IT-AI and DS.pdf
SisayNegash4
 

Similar to Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation (20)

Lecture 9: Machine Learning in Practice (2)
Lecture 9: Machine Learning in Practice (2)Lecture 9: Machine Learning in Practice (2)
Lecture 9: Machine Learning in Practice (2)
 
1. Demystifying ML.pdf
1. Demystifying ML.pdf1. Demystifying ML.pdf
1. Demystifying ML.pdf
 
Top 10 Data Science Practioner Pitfalls - Mark Landry
Top 10 Data Science Practioner Pitfalls - Mark LandryTop 10 Data Science Practioner Pitfalls - Mark Landry
Top 10 Data Science Practioner Pitfalls - Mark Landry
 
Top 10 Data Science Practitioner Pitfalls
Top 10 Data Science Practitioner PitfallsTop 10 Data Science Practitioner Pitfalls
Top 10 Data Science Practitioner Pitfalls
 
Presentation on supervised learning
Presentation on supervised learningPresentation on supervised learning
Presentation on supervised learning
 
Intro to Machine Learning for non-Data Scientists
Intro to Machine Learning for non-Data ScientistsIntro to Machine Learning for non-Data Scientists
Intro to Machine Learning for non-Data Scientists
 
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptxLETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
LETS PUBLISH WITH MORE RELIABLE & PRESENTABLE MODELLING.pptx
 
Lecture 09(introduction to machine learning)
Lecture 09(introduction to machine learning)Lecture 09(introduction to machine learning)
Lecture 09(introduction to machine learning)
 
crossvalidation.pptx
crossvalidation.pptxcrossvalidation.pptx
crossvalidation.pptx
 
Machine Learning
Machine LearningMachine Learning
Machine Learning
 
Statistical Learning and Model Selection (1).pptx
Statistical Learning and Model Selection (1).pptxStatistical Learning and Model Selection (1).pptx
Statistical Learning and Model Selection (1).pptx
 
Data Science Chapter 4: Machine Learning 101
Data Science Chapter 4: Machine Learning 101Data Science Chapter 4: Machine Learning 101
Data Science Chapter 4: Machine Learning 101
 
Intro to machine learning
Intro to machine learningIntro to machine learning
Intro to machine learning
 
H2O World - Top 10 Data Science Pitfalls - Mark Landry
H2O World - Top 10 Data Science Pitfalls - Mark LandryH2O World - Top 10 Data Science Pitfalls - Mark Landry
H2O World - Top 10 Data Science Pitfalls - Mark Landry
 
4.1.pptx
4.1.pptx4.1.pptx
4.1.pptx
 
Application of Machine Learning in Agriculture
Application of Machine  Learning in AgricultureApplication of Machine  Learning in Agriculture
Application of Machine Learning in Agriculture
 
Model Selection Techniques
Model Selection TechniquesModel Selection Techniques
Model Selection Techniques
 
ai4.ppt
ai4.pptai4.ppt
ai4.ppt
 
Introduction to machine learning-2023-IT-AI and DS.pdf
Introduction to machine learning-2023-IT-AI and DS.pdfIntroduction to machine learning-2023-IT-AI and DS.pdf
Introduction to machine learning-2023-IT-AI and DS.pdf
 
Machine Learning Unit 1 Semester 3 MSc IT Part 2 Mumbai University
Machine Learning Unit 1 Semester 3  MSc IT Part 2 Mumbai UniversityMachine Learning Unit 1 Semester 3  MSc IT Part 2 Mumbai University
Machine Learning Unit 1 Semester 3 MSc IT Part 2 Mumbai University
 

More from Marina Santini

More from Marina Santini (20)

Can We Quantify Domainhood? Exploring Measures to Assess Domain-Specificity i...
Can We Quantify Domainhood? Exploring Measures to Assess Domain-Specificity i...Can We Quantify Domainhood? Exploring Measures to Assess Domain-Specificity i...
Can We Quantify Domainhood? Exploring Measures to Assess Domain-Specificity i...
 
Towards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology ApplicationsTowards a Quality Assessment of Web Corpora for Language Technology Applications
Towards a Quality Assessment of Web Corpora for Language Technology Applications
 
A Web Corpus for eCare: Collection, Lay Annotation and Learning -First Results-
A Web Corpus for eCare: Collection, Lay Annotation and Learning -First Results-A Web Corpus for eCare: Collection, Lay Annotation and Learning -First Results-
A Web Corpus for eCare: Collection, Lay Annotation and Learning -First Results-
 
An Exploratory Study on Genre Classification using Readability Features
An Exploratory Study on Genre Classification using Readability FeaturesAn Exploratory Study on Genre Classification using Readability Features
An Exploratory Study on Genre Classification using Readability Features
 
Lecture: Ontologies and the Semantic Web
Lecture: Ontologies and the Semantic WebLecture: Ontologies and the Semantic Web
Lecture: Ontologies and the Semantic Web
 
Lecture: Summarization
Lecture: SummarizationLecture: Summarization
Lecture: Summarization
 
Relation Extraction
Relation ExtractionRelation Extraction
Relation Extraction
 
Lecture: Question Answering
Lecture: Question AnsweringLecture: Question Answering
Lecture: Question Answering
 
IE: Named Entity Recognition (NER)
IE: Named Entity Recognition (NER)IE: Named Entity Recognition (NER)
IE: Named Entity Recognition (NER)
 
Lecture: Vector Semantics (aka Distributional Semantics)
Lecture: Vector Semantics (aka Distributional Semantics)Lecture: Vector Semantics (aka Distributional Semantics)
Lecture: Vector Semantics (aka Distributional Semantics)
 
Lecture: Word Sense Disambiguation
Lecture: Word Sense DisambiguationLecture: Word Sense Disambiguation
Lecture: Word Sense Disambiguation
 
Lecture: Word Senses
Lecture: Word SensesLecture: Word Senses
Lecture: Word Senses
 
Sentiment Analysis
Sentiment AnalysisSentiment Analysis
Sentiment Analysis
 
Semantic Role Labeling
Semantic Role LabelingSemantic Role Labeling
Semantic Role Labeling
 
Semantics and Computational Semantics
Semantics and Computational SemanticsSemantics and Computational Semantics
Semantics and Computational Semantics
 
Lecture 8: Machine Learning in Practice (1)
Lecture 8: Machine Learning in Practice (1) Lecture 8: Machine Learning in Practice (1)
Lecture 8: Machine Learning in Practice (1)
 
Lecture 5: Interval Estimation
Lecture 5: Interval Estimation Lecture 5: Interval Estimation
Lecture 5: Interval Estimation
 
Lecture 1: Introduction to the Course (Practical Information)
Lecture 1: Introduction to the Course (Practical Information)Lecture 1: Introduction to the Course (Practical Information)
Lecture 1: Introduction to the Course (Practical Information)
 
Lecture: Joint, Conditional and Marginal Probabilities
Lecture: Joint, Conditional and Marginal Probabilities Lecture: Joint, Conditional and Marginal Probabilities
Lecture: Joint, Conditional and Marginal Probabilities
 
Mathematics for Language Technology: Introduction to Probability Theory
Mathematics for Language Technology: Introduction to Probability TheoryMathematics for Language Technology: Introduction to Probability Theory
Mathematics for Language Technology: Introduction to Probability Theory
 

Recently uploaded

1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
QucHHunhnh
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
QucHHunhnh
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please Practise
AnaAcapella
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 

Recently uploaded (20)

SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
Spatium Project Simulation student brief
Spatium Project Simulation student briefSpatium Project Simulation student brief
Spatium Project Simulation student brief
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
Dyslexia AI Workshop for Slideshare.pptx
Dyslexia AI Workshop for Slideshare.pptxDyslexia AI Workshop for Slideshare.pptx
Dyslexia AI Workshop for Slideshare.pptx
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please Practise
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 

Lecture 3: Basic Concepts of Machine Learning - Induction & Evaluation

  • 1. Machine Learning for Language Technology 2015 http://stp.lingfil.uu.se/~santinim/ml/2015/ml4lt_2015.htm Basic Concepts of Machine Learning Induction & Evaluation Marina Santini santinim@stp.lingfil.uu.se Department of Linguistics and Philology Uppsala University, Uppsala, Sweden Autumn 2015
  • 2. Acknowledgments • Daume’ (2015), Alpaydin (2010), NLTK website, other web sites. Lecture 3: Basic Concepts of ML 2
  • 3. Outline • Induction – Induction pipeline • Training set, test set and development set • Parameters • Hyperparameters • Accuracy, precision, recall, f-measure • Confusion matrix • Crossvalidation • Leave one out • Stratification Lecture 3: Basic Concepts of ML 3
  • 4. Induction • Induction is the process of reaching a general conclusion from specific examples. Lecture 3: Basic Concepts of ML 4
  • 5. Inductive Machine Learning • The goal of inductive machine learning is to take some training data and use it to induce a function (model, classifier, learning algorithm). • This function will be evaluated on the test data. • The machine learning algorithm has succeeded if its performance on the test data is high. Lecture 3: Basic Concepts of ML 5
  • 6. Pipeline • Induction pipeline Lecture 3: Basic Concepts of ML 6
  • 7. Task • Predict the class for this ”unseen” example: Sepal length – Sepal width – Petal length – Petal width - Type 5.2 3.7 1.7 0.3 ??? Lecture 1: What is Machine Learning? 7 Require us to generalize from the training data
  • 8. Splitting data to measure performance • Training data& Test Data – Common splits: 80/20; 90/10 • NEVER TOUCH THE TEST DATA! • TEST DATA MUST BELONG TO THE SAME STATISTICAL DISTRIBUTION AS THE TRAINING DATA Lecture 3: Basic Concepts of ML 8
  • 9. Modelling • ML uses formal models that might perform well on our data. • The choice of using one model rather than another is our choice. • A model tells us what sort of things we can learn. • A model tells us what our inductive bias is. Lecture 3: Basic Concepts of ML 9
  • 10. Parameters • Models can have many parameters and finding the best combination of parameters is not trivial. Lecture 3: Basic Concepts of ML 10
  • 11. Hyperparameters • A hyperparameter is a parameter that controls other parameters of the model. Lecture 3: Basic Concepts of ML 11
  • 12. Development Set • Split your data into 70% training data, 10% development data and 20% test data. • For each possible setting of the hyperparameters: – Train a model using that setting on the training data – Compute the model error rate on the development data – From the above collection of medels, choos the one that achieve the lowest error rate on development data. – Evaluate that model on the test data to estimate future test performance. Lecture 3: Basic Concepts of ML 12
  • 13. Accuracy • Accuracy measures the percentage of correct results that a classifier has achieved. Lecture 3: Basic Concepts of ML 13
  • 14. True and False Positives and Negatives • True positives are relevant items that we correctly identified as relevant. • True negatives are irrelevant items that we correctly identified as irrelevant. • False positives (or Type I errors) are irrelevant items that we incorrectly identified as relevant. • False negatives (or Type II errors) are relevant items that we incorrectly identified as irrelevant. Lecture 3: Basic Concepts of ML 14
  • 15. Precision, Recall, F-Measure • Given these four numbers, we can define the following metrics: – Precision, which indicates how many of the items that we identified were relevant, is TP/(TP+FP). – Recall, which indicates how many of the relevant items that we identified, is TP/(TP+FN). – The F-Measure (or F-Score), which combines the precision and recall to give a single score, is defined to be the harmonic mean of the precision and recall: (2 × Precision × Recall) / (Precision + Recall). Lecture 3: Basic Concepts of ML 15
  • 16. Accuracy, Precision, Recall, F-measure • Accuracy = (TP + TN)/(TP + TN + FP + FN) • Precision = TP / TP + FP • Recall = TP / TP + FN • F-measure = 2*((precision*recall)/(precision+recall)) Lecture 3: Basic Concepts of ML 16
  • 17. Confusion Matrix • This is a useful table that presents both the class distribution in the data and the classifiers predicted class distribution with a breakdown of error types. • Usually, the rows are the observed/actual class labels and the columns the predicted class labels. • Each cell contains the number of predictions made by the classifier that fall into that cell. Lecture 3: Basic Concepts of ML 17 actual predicted
  • 18. Multi-Class Confusion Matrix • If a classification system has been trained to distinguish between cats, dogs and rabbits, a confusion matrix will summarize the results: Lecture 3: Basic Concepts of ML 18
  • 19. Cross validation • In 10-fold cross-validation you break you training data up into 10 equally-sized partitions. • You train a learning algorithm on 9 of them and tst it on the remaining 1. • You do this 10 times, each holding out a different partition as the test data. • Typical choices for n-fold are 2, 5, 10. • 10-fold cross validation is the most common. Lecture 3: Basic Concepts of ML 19
  • 20. Leave One Out • Leave One Out (or LOO) is a simple cross- validation. Each learning set is created by taking all the samples except one, the test set being the sample left out. Lecture 3: Basic Concepts of ML 20
  • 21. Stratification • Proportion of each class in the traning set and test sets is the same as the proportion in the original sample. Lecture 3: Basic Concepts of ML 21
  • 22. Weka Cross validation • 10-fold cross validation Lecture 3: Basic Concepts of ML 22
  • 23. Weka: Output • Classifier output Lecture 3: Basic Concepts of ML 23
  • 24. Remember: Underfitting & Overfitting Underfitting: the model has not learned enough from the data and is unable to generalize Overfitting: the model has learned too many idiosyncrasies (noise) and is unable to generalize Lecture 3: Basic Concepts of ML 24
  • 25. Summary: Performance of a learning model: Requirements • Our goal when we choose a machine learning model is that it does well on future, unseen data. • The way in which we measure performance should depend on the problem we are trying to solve. • There should be a strong relationship between the data that our algorithm sees at training time and the data it sees at test time. Lecture 3: Basic Concepts of ML 25
  • 26. Not everything is learnable – Noise at feature level – Noise at class label level – Features are insufficient – Labels are controversial – Inductive bias not appropriate for the kind of problem we try to learn Lecture 3: Decision Trees (1) 26
  • 27. Quiz 1: Stratification • What does it mean ”stratified” cross validation? 1. The examples of a class are all in the training set, and the rest of the classes are in the test set. 2. The proportion of each class in the sets ae the same as the proportion in the original sample 3. None of the above. Lecture 3: Basic Concepts of ML 27
  • 28. Quiz 2: Accuracy • Why is accuracy alone an unreliable measure? 1. Because it can be biassed towards the most frequent class. 2. Because it always guesses wrong. 3. None of the above Lecture 3: Basic Concepts of ML 28
  • 29. Quiz 3: Data Splits • Which are recommended splits between training and test data? 1. 80/20 2. 50/50 3. 10/90 Lecture 3: Basic Concepts of ML 29
  • 30. Quiz 4: Overfitting • What does it mean overfitting? 1. the model has not learned enough from the data and is unable to generalize 2. The proportion of each class in the sets is the same as the proportion in the original sample 3. None of the above. Lecture 3: Basic Concepts of ML 30
  • 31. The End Lecture 3: Basic Concepts of ML 31