SlideShare a Scribd company logo
1 of 16
On Selecting Feature-Value
Pairs on Smart Phones
for Activity Inferences
Gunarto Sindoro Njoo,
Yu-Hsiang Peng,
Kuo-Wei Hsu,
Wen-Chih Peng
Introduction
• Smart phone is getting smarter and smarter
2
Introduction
• Computer
• RAM: 2~8 GB on average
• Storage: >500 GB
• Power: Hundreds Watts
• Smartphone
• RAM: 512 MB~3 GB
• Storage: >4 GB
• Power: a few Watts
• Sensor Hub
• RAM: 16~64 KB
• Storage: 64 KB~256 KB
• Power: mW
3
Activity Inference Process
Raw data
Discretization
• MDLP
• LGD
Classifier
construction
• Decision Tree
• Naïve Bayesian
• K-Nearest Neighbor
• SVM
4
Activity Inference Process
Raw data
Discretization
•MDLP
•LGD
Feature-Value
selection
•ONEFVAS
•GIFVAS
•CBFVAS
Classifier
construction
•Decision Tree
•Naïve Bayesian
•K-Nearest Neighbor
•SVM
5
Feature-Value Selection
• What is feature-value?
• A range of sensor reading
• e.g. Accelerometer magnitude high, GPS at home, light bright
• Why using feature-value?
• Sensor reading relation with activity: relevant or not
• e.g. Accelerometer magnitude value reading
Accelerometer: LowAccelerometer: Low Accelerometer: high 6
FEATURE-VALUE METHODS
One-Cut
Iteration-based
Correlated-based
7
One-Cut (ONEFVAS)
• Entropy-based selection using threshold
Sensor Value Entropy
Accel Avg. 9.75 0.75
Accel Avg. 12 0.1
Accel Avg. 15 0
GPS Speed 0~3 0.95
GPS Speed 3~10 0.2
TimeDay Morning 0.61
TimeDay Afternoon 0.88
TimeDay Evening 0.37
TimeDay Night 0.23
Extracted Sensor Value Pair
Sensor Value Entropy
Accel Avg. 9.75 0.75
Accel Avg. 12 0.1
Accel Avg. 15 0
GPS Speed 0~3 0.95
GPS Speed 3~10 0.2
TimeDay Morning 0.61
TimeDay Afternoon 0.88
TimeDay Evening 0.37
TimeDay Night 0.23
Extracted Sensor Value Pair
Sensor Value Entropy
Accel Avg. 12 0.1
GPS Speed 3~10 0.2
TimeDay Evening 0.37
TimeDay Night 0.23
Extracted Sensor Value Pair
Entropy <= 0.5
8
Iteration-based (GIFVAS)
• Looping on the threshold, selecting feature-values iteratively.
• Evaluating accuracy for each iteration
• If accuracy reduction is big
• Then cancel the selection on this iteration, tag any feature-value
as special
• Any special feature-value will be remained until the last iteration
• Special feature-value
• Frequent but confusing
• Pure but infrequent
0.00%
50.00%
100.00%
1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0.0
Entropy Threshold
Accuracy
0.00%
50.00%
100.00%
1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0.0
Entropy Threshold
Feature-Value Pairs
9
Correlation-based
• Using Pearson correlation in the feature level
• Using entropy in the feature-value level
• For each feature-value pair
• Generate correlated feature-value
• Sort the correlated feature-value using entropy
• Pick only the best-N feature-value from it
• Discard other feature-value
80.00%
85.00%
90.00%
95.00%
100.00%
1 3 5 7 9 11 13 15 17
Best-N feature-value remained
Accuracy
Correlation
Original
350.00 KB
550.00 KB
750.00 KB
950.00 KB
1150.00 KB
1 3 5 7 9 11 13 15 17
Best-N feature-value remained
Model Size
Correlation
Original
10
Experiments
• Environments:
• Intel Quad Core 2.66GHz
• RAM 8 GB
• Java 7
• Weka 3.6.11 (all default parameter)
• Datasets:
• Collect from 11 participants
• At least 2 different activities, up to 6 activities
• Average 3 weeks, maximum 2 months
• Classifier Algorithm:
• Naïve Bayesian
• Decision Tree (J48)
• SVM (SMO)
• k-Nearest Neigbor (kNN) 11
Experiments (Model Size)
• Feature-value selection is not effective on Naïve Bayesian
• In general, feature-value selection works best on decision tree
0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Original ONEFVAS GIFVAS CBFVAS
Model Size (LGD)
0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Original ONEFVAS GIFVAS CBFVAS
Model Size (MDLP)
Naïve Bayes
Decision Tree
kNN
SVM
12
Experiments (LGD)
0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Original ONEFVAS GIFVAS CBFVAS
Model Size (LGD)
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Original ONEFVAS GIFVAS CBFVAS
Accuracy (LGD)
Decision Tree
kNN
SVM
• ONEFVAS gives the biggest saving on model size, but accuracy is low
• Most stable accuracy is on CBFVAS, while reducing model size well
13
Experiments (MDLP)
0.00%
10.00%
20.00%
30.00%
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Original ONEFVAS GIFVAS CBFVAS
Model Size (MDLP)
40.00%
50.00%
60.00%
70.00%
80.00%
90.00%
100.00%
Original ONEFVAS GIFVAS CBFVAS
Accuracy (MDLP)
Decision Tree
kNN
• Accuracy are more stable, while reductions on model size are well
• In most cases, decision tree gets the most benefit.
14
Conclusions
• Proposed feature-value selection for reducing model size
• ONEFVAS – Using entropy threshold
• GIFVAS – Using iteration on entropy threshold
• CBFVAS – Using correlation and entropy
• Proposed method is able to reduce model size while
maintaining accuracy performance
• Performance varies on discretization and classification algorithms
• Decision Tree gets the most benefit
15
Thank you
On Selecting Feature-Value Pairs on Smart Phones for Activity Inferences
Presented by: Gunarto Sindoro Njoo
GUNARTO.NCTU@GMAIL.COM
16

More Related Content

Viewers also liked

Myself presentation Machekhin Alexander
Myself presentation Machekhin AlexanderMyself presentation Machekhin Alexander
Myself presentation Machekhin AlexanderAlexander Machekhin
 
Razones para estudiar contaduria publica
Razones para estudiar contaduria publicaRazones para estudiar contaduria publica
Razones para estudiar contaduria publicacolombia59681072
 
Portfolio-anis
Portfolio-anisPortfolio-anis
Portfolio-anisartcitizen
 
Minha Apresentação AD1
Minha Apresentação AD1Minha Apresentação AD1
Minha Apresentação AD1cederj edu
 
Electronics NQF_3
Electronics NQF_3Electronics NQF_3
Electronics NQF_3Jan Barnard
 

Viewers also liked (8)

Myself presentation Machekhin Alexander
Myself presentation Machekhin AlexanderMyself presentation Machekhin Alexander
Myself presentation Machekhin Alexander
 
Razones para estudiar contaduria publica
Razones para estudiar contaduria publicaRazones para estudiar contaduria publica
Razones para estudiar contaduria publica
 
Mireneir taldea
Mireneir taldeaMireneir taldea
Mireneir taldea
 
Portfolio-anis
Portfolio-anisPortfolio-anis
Portfolio-anis
 
Minha Apresentação AD1
Minha Apresentação AD1Minha Apresentação AD1
Minha Apresentação AD1
 
Electronics NQF_3
Electronics NQF_3Electronics NQF_3
Electronics NQF_3
 
Poster.
Poster.Poster.
Poster.
 
Menendez Reecommendation
Menendez ReecommendationMenendez Reecommendation
Menendez Reecommendation
 

Similar to 20141030 Feature-Value DSAA 2014 V6_FINAL

CMU Trecvid sed11
CMU Trecvid sed11CMU Trecvid sed11
CMU Trecvid sed11Lu Jiang
 
(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems
(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems
(BDT207) Real-Time Analytics In Service Of Self-Healing EcosystemsAmazon Web Services
 
Contour-Constrained Superpixels for Image and Video Processing
Contour-Constrained Superpixels for Image and Video ProcessingContour-Constrained Superpixels for Image and Video Processing
Contour-Constrained Superpixels for Image and Video ProcessingNAVER Engineering
 
Cvpr 2018 papers review (efficient computing)
Cvpr 2018 papers review (efficient computing)Cvpr 2018 papers review (efficient computing)
Cvpr 2018 papers review (efficient computing)DonghyunKang12
 
Introduction to ML.NET
Introduction to ML.NETIntroduction to ML.NET
Introduction to ML.NETMarco Parenzan
 
CAMCOS_final Presentation_Group2.pptx
CAMCOS_final Presentation_Group2.pptxCAMCOS_final Presentation_Group2.pptx
CAMCOS_final Presentation_Group2.pptxhuzaifaazam3
 
LEXT OLS4100 3D Measuring Laser Microscope
LEXT OLS4100 3D Measuring Laser MicroscopeLEXT OLS4100 3D Measuring Laser Microscope
LEXT OLS4100 3D Measuring Laser MicroscopeOlympus IMS
 
Wits presentation 6_28072015
Wits presentation 6_28072015Wits presentation 6_28072015
Wits presentation 6_28072015Beatrice van Eden
 
Improving Hardware Efficiency for DNN Applications
Improving Hardware Efficiency for DNN ApplicationsImproving Hardware Efficiency for DNN Applications
Improving Hardware Efficiency for DNN ApplicationsChester Chen
 
[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...
[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...
[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...DataScienceConferenc1
 
Evolving Custom Communication Protocols
Evolving Custom Communication ProtocolsEvolving Custom Communication Protocols
Evolving Custom Communication ProtocolsWesley Faler
 
ML for detecting products on grocery shelves
ML for detecting products on grocery shelvesML for detecting products on grocery shelves
ML for detecting products on grocery shelvesKonstantin Bychenkov
 
Temporal Superpixels Based on Proximity-Weighted Patch Matching
Temporal Superpixels Based on Proximity-Weighted Patch MatchingTemporal Superpixels Based on Proximity-Weighted Patch Matching
Temporal Superpixels Based on Proximity-Weighted Patch MatchingNAVER Engineering
 
Superb Supervision of Short-lived Servers with Sensu
Superb Supervision of Short-lived Servers with SensuSuperb Supervision of Short-lived Servers with Sensu
Superb Supervision of Short-lived Servers with SensuPaul O'Connor
 
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...Sathishkumar Samiappan
 
Network selection techniques:SAW and TOPSIS
Network selection techniques:SAW and TOPSISNetwork selection techniques:SAW and TOPSIS
Network selection techniques:SAW and TOPSISYashwant Dagar
 
Donatas Mažionis, Building low latency web APIs
Donatas Mažionis, Building low latency web APIsDonatas Mažionis, Building low latency web APIs
Donatas Mažionis, Building low latency web APIsTanya Denisyuk
 

Similar to 20141030 Feature-Value DSAA 2014 V6_FINAL (20)

CMU Trecvid sed11
CMU Trecvid sed11CMU Trecvid sed11
CMU Trecvid sed11
 
(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems
(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems
(BDT207) Real-Time Analytics In Service Of Self-Healing Ecosystems
 
Contour-Constrained Superpixels for Image and Video Processing
Contour-Constrained Superpixels for Image and Video ProcessingContour-Constrained Superpixels for Image and Video Processing
Contour-Constrained Superpixels for Image and Video Processing
 
Cvpr 2018 papers review (efficient computing)
Cvpr 2018 papers review (efficient computing)Cvpr 2018 papers review (efficient computing)
Cvpr 2018 papers review (efficient computing)
 
Introduction to ML.NET
Introduction to ML.NETIntroduction to ML.NET
Introduction to ML.NET
 
CAMCOS_final Presentation_Group2.pptx
CAMCOS_final Presentation_Group2.pptxCAMCOS_final Presentation_Group2.pptx
CAMCOS_final Presentation_Group2.pptx
 
LEXT OLS4100 3D Measuring Laser Microscope
LEXT OLS4100 3D Measuring Laser MicroscopeLEXT OLS4100 3D Measuring Laser Microscope
LEXT OLS4100 3D Measuring Laser Microscope
 
WRAIR
WRAIRWRAIR
WRAIR
 
Aerospike & GCE (LSPE Talk)
Aerospike & GCE (LSPE Talk)Aerospike & GCE (LSPE Talk)
Aerospike & GCE (LSPE Talk)
 
Wits presentation 6_28072015
Wits presentation 6_28072015Wits presentation 6_28072015
Wits presentation 6_28072015
 
Improving Hardware Efficiency for DNN Applications
Improving Hardware Efficiency for DNN ApplicationsImproving Hardware Efficiency for DNN Applications
Improving Hardware Efficiency for DNN Applications
 
[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...
[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...
[Q-tangled 22] Deconstructing Quantum Machine Learning Algorithms - Sasha Laz...
 
Evolving Custom Communication Protocols
Evolving Custom Communication ProtocolsEvolving Custom Communication Protocols
Evolving Custom Communication Protocols
 
ML for detecting products on grocery shelves
ML for detecting products on grocery shelvesML for detecting products on grocery shelves
ML for detecting products on grocery shelves
 
Temporal Superpixels Based on Proximity-Weighted Patch Matching
Temporal Superpixels Based on Proximity-Weighted Patch MatchingTemporal Superpixels Based on Proximity-Weighted Patch Matching
Temporal Superpixels Based on Proximity-Weighted Patch Matching
 
Superb Supervision of Short-lived Servers with Sensu
Superb Supervision of Short-lived Servers with SensuSuperb Supervision of Short-lived Servers with Sensu
Superb Supervision of Short-lived Servers with Sensu
 
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
Non-Uniform Random Feature Selection and Kernel Density Scoring With SVM Base...
 
Network selection techniques:SAW and TOPSIS
Network selection techniques:SAW and TOPSISNetwork selection techniques:SAW and TOPSIS
Network selection techniques:SAW and TOPSIS
 
FallThesisFinal
FallThesisFinalFallThesisFinal
FallThesisFinal
 
Donatas Mažionis, Building low latency web APIs
Donatas Mažionis, Building low latency web APIsDonatas Mažionis, Building low latency web APIs
Donatas Mažionis, Building low latency web APIs
 

20141030 Feature-Value DSAA 2014 V6_FINAL

  • 1. On Selecting Feature-Value Pairs on Smart Phones for Activity Inferences Gunarto Sindoro Njoo, Yu-Hsiang Peng, Kuo-Wei Hsu, Wen-Chih Peng
  • 2. Introduction • Smart phone is getting smarter and smarter 2
  • 3. Introduction • Computer • RAM: 2~8 GB on average • Storage: >500 GB • Power: Hundreds Watts • Smartphone • RAM: 512 MB~3 GB • Storage: >4 GB • Power: a few Watts • Sensor Hub • RAM: 16~64 KB • Storage: 64 KB~256 KB • Power: mW 3
  • 4. Activity Inference Process Raw data Discretization • MDLP • LGD Classifier construction • Decision Tree • Naïve Bayesian • K-Nearest Neighbor • SVM 4
  • 5. Activity Inference Process Raw data Discretization •MDLP •LGD Feature-Value selection •ONEFVAS •GIFVAS •CBFVAS Classifier construction •Decision Tree •Naïve Bayesian •K-Nearest Neighbor •SVM 5
  • 6. Feature-Value Selection • What is feature-value? • A range of sensor reading • e.g. Accelerometer magnitude high, GPS at home, light bright • Why using feature-value? • Sensor reading relation with activity: relevant or not • e.g. Accelerometer magnitude value reading Accelerometer: LowAccelerometer: Low Accelerometer: high 6
  • 8. One-Cut (ONEFVAS) • Entropy-based selection using threshold Sensor Value Entropy Accel Avg. 9.75 0.75 Accel Avg. 12 0.1 Accel Avg. 15 0 GPS Speed 0~3 0.95 GPS Speed 3~10 0.2 TimeDay Morning 0.61 TimeDay Afternoon 0.88 TimeDay Evening 0.37 TimeDay Night 0.23 Extracted Sensor Value Pair Sensor Value Entropy Accel Avg. 9.75 0.75 Accel Avg. 12 0.1 Accel Avg. 15 0 GPS Speed 0~3 0.95 GPS Speed 3~10 0.2 TimeDay Morning 0.61 TimeDay Afternoon 0.88 TimeDay Evening 0.37 TimeDay Night 0.23 Extracted Sensor Value Pair Sensor Value Entropy Accel Avg. 12 0.1 GPS Speed 3~10 0.2 TimeDay Evening 0.37 TimeDay Night 0.23 Extracted Sensor Value Pair Entropy <= 0.5 8
  • 9. Iteration-based (GIFVAS) • Looping on the threshold, selecting feature-values iteratively. • Evaluating accuracy for each iteration • If accuracy reduction is big • Then cancel the selection on this iteration, tag any feature-value as special • Any special feature-value will be remained until the last iteration • Special feature-value • Frequent but confusing • Pure but infrequent 0.00% 50.00% 100.00% 1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0.0 Entropy Threshold Accuracy 0.00% 50.00% 100.00% 1.0 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0.0 Entropy Threshold Feature-Value Pairs 9
  • 10. Correlation-based • Using Pearson correlation in the feature level • Using entropy in the feature-value level • For each feature-value pair • Generate correlated feature-value • Sort the correlated feature-value using entropy • Pick only the best-N feature-value from it • Discard other feature-value 80.00% 85.00% 90.00% 95.00% 100.00% 1 3 5 7 9 11 13 15 17 Best-N feature-value remained Accuracy Correlation Original 350.00 KB 550.00 KB 750.00 KB 950.00 KB 1150.00 KB 1 3 5 7 9 11 13 15 17 Best-N feature-value remained Model Size Correlation Original 10
  • 11. Experiments • Environments: • Intel Quad Core 2.66GHz • RAM 8 GB • Java 7 • Weka 3.6.11 (all default parameter) • Datasets: • Collect from 11 participants • At least 2 different activities, up to 6 activities • Average 3 weeks, maximum 2 months • Classifier Algorithm: • Naïve Bayesian • Decision Tree (J48) • SVM (SMO) • k-Nearest Neigbor (kNN) 11
  • 12. Experiments (Model Size) • Feature-value selection is not effective on Naïve Bayesian • In general, feature-value selection works best on decision tree 0.00% 10.00% 20.00% 30.00% 40.00% 50.00% 60.00% 70.00% 80.00% 90.00% 100.00% Original ONEFVAS GIFVAS CBFVAS Model Size (LGD) 0.00% 10.00% 20.00% 30.00% 40.00% 50.00% 60.00% 70.00% 80.00% 90.00% 100.00% Original ONEFVAS GIFVAS CBFVAS Model Size (MDLP) Naïve Bayes Decision Tree kNN SVM 12
  • 13. Experiments (LGD) 0.00% 10.00% 20.00% 30.00% 40.00% 50.00% 60.00% 70.00% 80.00% 90.00% 100.00% Original ONEFVAS GIFVAS CBFVAS Model Size (LGD) 40.00% 50.00% 60.00% 70.00% 80.00% 90.00% 100.00% Original ONEFVAS GIFVAS CBFVAS Accuracy (LGD) Decision Tree kNN SVM • ONEFVAS gives the biggest saving on model size, but accuracy is low • Most stable accuracy is on CBFVAS, while reducing model size well 13
  • 14. Experiments (MDLP) 0.00% 10.00% 20.00% 30.00% 40.00% 50.00% 60.00% 70.00% 80.00% 90.00% 100.00% Original ONEFVAS GIFVAS CBFVAS Model Size (MDLP) 40.00% 50.00% 60.00% 70.00% 80.00% 90.00% 100.00% Original ONEFVAS GIFVAS CBFVAS Accuracy (MDLP) Decision Tree kNN • Accuracy are more stable, while reductions on model size are well • In most cases, decision tree gets the most benefit. 14
  • 15. Conclusions • Proposed feature-value selection for reducing model size • ONEFVAS – Using entropy threshold • GIFVAS – Using iteration on entropy threshold • CBFVAS – Using correlation and entropy • Proposed method is able to reduce model size while maintaining accuracy performance • Performance varies on discretization and classification algorithms • Decision Tree gets the most benefit 15
  • 16. Thank you On Selecting Feature-Value Pairs on Smart Phones for Activity Inferences Presented by: Gunarto Sindoro Njoo GUNARTO.NCTU@GMAIL.COM 16

Editor's Notes

  1. [Quick Explanation] Motivation on activity inference Helping on giving service by learning the context of the user [1] Silent, Phone/Message Filtering [2] Navigations, reading message out [3] Tips, Review, Recommendations
  2. [Quick Explanation] Motivation on reducing the storage size for activity classification Comparing between Computer, Smart phone, and Sensor Hub – Power Consumption Computer : Hundreds Watts Smart Phones: Several Watts Sensor Hubs : mW
  3. Activity inference process in general (Using Supervised learning) Why we need discretization? Because some of classification algorithms need interval so that it can work well, e.g. decision tree, rule based. In here we consider 2 supervised discretization methods, using information theory (GINI for LGD and Information Gain for MDLP)
  4. Feature-Value selection is inserted in the middle of discretization and classifier building We introduce 3 methods to do feature-value selection.
  5. Before going further, we need to explain what is feature value first. Feature-Value selection’s goal is: reduce model size of classifier because some of the intervals have less meaning. Why not feature selection ? Because number of sensors in smart phone are limited and removing one will reduce the accuracy greatly. We could also address that this approach could be a complement to feature-selection, and could do well in fewer features.
  6. We could select the feature-value using threshold. By doing so, we could reduce the number of feature-value well. The problem is how we set the entropy threshold.
  7. Doing iteration in searching for the best threshold. Even though there are some “special feature-value” added back, the number of feature-value pairs still reduced a lot. The problem here is that selection process is slow.
  8. By using a big value for N, then the model size is bigger too. Feature-value are grouped by their togetherness in the datasets: If feature are correlated If feature-value are together in the datasets.
  9. Activities: Working, Moving (Transportation), Dining, Shopping, Doing entertainment, Exercising (Doing sport)
  10. Original is “Without feature-value selection” ONEFVAS is using entropy threshold 0. LGD and MDLP are discretization methods. Naïve Bayesian uses matrix to represent the classification model, so the size is not reduced. Only the statistical attributes (mean, stdev, var) are changed. So on the following slides, we remove Naïve Bayesian and focus on the rest. On term of model size, ONEFVAS in Decision Tree is the best, but CBFVAS is more stable on most case. SVM on MDLP couldn’t be run because of some limitation in weka, due to huge number of intervals generated by MDLP discretization.
  11. Because of weka’s limitation on number of intervals, SVM can’t be run. Based on those graphs, we can see that Decision Tree gets the most benefit (Low model size and Small accuracy reduction)