*Interesting article posted by By John Langford. John Langford is a machine learning research scientist, a field which he says "is shifting from an academic discipline to an industrial tool". He is the author of the blog hunch.net. John works at Microsoft Research, and was previously affiliated with Yahoo Research,Toyota Technological Institute, and IBM's Watson Research Center. He studied Physics and Computer Science at the California Institute of Technology, earning a double bachelor's degree in 1997, and received his PhD in Computer Science from Carnegie Mellon University** in 2002.*

Here I have included excerpts of his article, with a link to the full version at the bottom.

**John's article**

Attempts to abstract and study machine learning are within some given framework or mathematical model. It turns out that all of these models are significantly flawed for the purpose of studying machine learning. I've created a table (below) outlining the major flaws in some common models of machine learning.

The point here is not simply "woe unto us". There are several implications which seem important.

- The multitude of models is a point of continuing confusion. It is common for people to learn about machine learning within one framework which often becomes there "home framework" through which they attempt to filter all machine learning. (Have you met people who can only think in terms of kernels? Only via Bayes Law? Only via PAC Learning?) Explicitly understanding the existence of these other frameworks can help resolve the confusion. This is particularly important when reviewing and particularly important for students.
- Algorithms which conform to multiple approaches can have substantial value. "I don't really understand it yet, because I only understand it one way". Reinterpretation alone is not the goal - we want algorithmic guidance.
- We need to remain constantly open to new mathematical models of machine learning. It's common to forget the flaws of the model that you are most familiar with in evaluating other models while the flaws of new models get exaggerated. The best way to avoid this is simply education.
- The value of theory alone is more limited than many theoreticians may be aware. Theories need to be tested to see if they correctly predict the underlying phenomena.

Here is a summary what is wrong with various frameworks for learning. To avoid being entirely negative, I added a column about what's right as well.

**Bayesian Learning**

**Methodology: **You specify a prior probability distribution over data-makers,*P(datamaker)* then use Bayes law to find a posterior *P(datamaker|x)*. True Bayesians integrate over the posterior to make predictions while many simply use the world with largest posterior directly.

**What is wrong:**

- Information theoretically problematic. Explicitly specifying a reasonable prior is often hard.
- Computationally difficult problems are commonly encountered.
- Human intensive. Partly due to the difficulties above and partly because "first specify a prior" is built into framework this approach is not very automatable.

**Graphical/generative Models**

**Methodology: **Sometimes Bayesian and sometimes not. Data-makers are typically assumed to be IID samples of fixed or varying length data. Data-makers are represented graphically with conditional independencies encoded in the graph. For some graphs, fast algorithms for making (or approximately making) predictions exist.

**What is wrong:**

- Often (still) fails to fix problems with the Bayesian approach.
- In real world applications, true conditional independence is rare, and results degrade rapidly with systematic misspecification of conditional independence.

**Convex Loss Optimization**

**Methodology: **Specify a loss function related to the world-imposed loss fucntion which is convex on some parametric predictive system. Optimize the parametric predictive system to find the global optima.

**What is wrong:**

- The temptation to forget that the world imposes nonconvex loss functions is sometimes overwhelming, and the mismatch is always dangerous.
- Limited models. Although switching to a convex loss means that some optimizations become convex, optimization on representations which aren't single layer linear combinations is often difficult.

**Gradient Descent**

**Methodology: **Specify an architecture with free parameters and use gradient descent with respect to data to tune the parameters.

**What is wrong:**

- Finicky. There are issues with paremeter initialization, step size, and representation. It helps a great deal to have accumulated experience using this sort of system and there is little theoretical guidance.
- Overfitting is a significant issue.

**Kernel-based learning**

**Methodology: **You chose a kernel *K(x,x')* between datapoints that satisfies certain conditions, and then use it as a measure of similarity when learning.

**What is wrong: **Specification of the kernel is not easy for some applications (this is another example of prior elicitation). *O(n2)* is not efficient enough when there is much data.

**Boosting**

**Methodology: **You create a learning algorithm that may be imperfect but which has some predictive edge, then apply it repeatedly in various ways to make a final predictor.

**What is wrong: **The boosting framework tells you nothing about how to build that initial algorithm. The weak learning assumption becomes violated at some point in the iterative process.

**Online Learning with Experts**

**Methodology: **You make many base predictors and then a master algorithm automatically switches between the use of these predictors so as to minimize regret.

**What is wrong: **Computational intractability can be a problem. This approach lives and dies on the effectiveness of the experts, but it provides little or no guidance in their construction.

**Learning Reductions**

**Methodology: **You solve complex machine learning problems by reducing them to well-studied base problems in a robust manner.

**What is wrong: **The existence of an algorithm satisfying reduction guarantees is not sufficient to guarantee success. Reductions tell you little or nothing about the design of the base learning algorithm.

**PAC Learning**

**Methodology: **You assume that samples are drawn IID from an unknown distribution D. You think of learning as finding a near-best hypothesis amongst a given set of hypotheses in a computationally tractable manner.

**What is right: **The focus on computation is pretty right-headed, because we are ultimately limited by what we can compute.

**What is wrong: **There are not many substantial positive results, particularly when D is noisy. Data isn’t IID in practice anyways.

**Statistical Learning Theory**

**Methodology: **You assume that samples are drawn IID from an unknown distribution D. You think of learning as figuring out the number of samples required to distinguish a near-best hypothesis from a set of hypotheses.

**What is wrong: **The data is not IID. Ignorance of computational difficulties often results in difficulty of application. More importantly, the bounds are often loose (sometimes to the point of vacuousness).

**Decision tree learning**

**Methodology: **Learning is a process of cutting up the input space and assigning predictions to pieces of the space.

**What is wrong: **There are learning problems which can not be solved by decision trees, but which are solvable. It’s common to find that other approaches give you a bit more performance. A theoretical grounding for many choices in these algorithms is lacking.

**Algorithmic complexity**

**Methodology: **Learning is about finding a program which correctly predicts the outputs given the inputs.

**What is wrong: **The theory literally suggests solving halting problems to solve machine learning.

**RL, MDP learning**

**Methodology: **Learning is about finding and acting according to a near optimal policy in an unknown Markov Decision Process.

**What is wrong: **Has anyone counted the number of states in real world problems? We can’t afford to wait that long. Discretizing the states creates a POMDP (see below). In the real world, we often have to deal with a POMDP anyways.

**RL, POMDP learning**

**Methodology: **Learning is about finding and acting according to a near optimaly policy in a Partially Observed Markov Decision Process

**What is wrong: **All known algorithms scale badly with the number of hidden states.

This set is incomplete of course, but it forms a starting point for understanding what’s out there. (Please fill in the what/pro/con of anything I missed.)

To read the full article on Hunch.net - including what is right about these methods, click here. You might also want to read the 8 worst predictive modeling techniques.

**DSC Resources**

- Career: Training | Books | Cheat Sheet | Apprenticeship | Certification | Salary Surveys | Jobs
- Knowledge: Research | Competitions | Webinars | Our Book | Members Only | Search DSC
- Buzz: Business News | Announcements | Events | RSS Feeds
- Misc: Top Links | Code Snippets | External Resources | Best Blogs | Subscribe | For Bloggers

**Additional Reading**

- Data Scientist Reveals his Growth Hacking Techniques
- 10 Modern Statistical Concepts Discovered by Data Scientists
- Top data science keywords on DSC
- 4 easy steps to becoming a data scientist
- 13 New Trends in Big Data and Data Science
- 22 tips for better data science
- Data Science Compared to 16 Analytic Disciplines
- How to detect spurious correlations, and how to find the real ones
- 17 short tutorials all data scientists should read (and practice)
- 10 types of data scientists
- 66 job interview questions for data scientists
- High versus low-level data science

Follow us on Twitter: @DataScienceCtrl | @AnalyticBridge

© 2020 Data Science Central ® Powered by

Badges | Report an Issue | Privacy Policy | Terms of Service

**Most Popular Content on DSC**

To not miss this type of content in the future, subscribe to our newsletter.

- Book: Statistics -- New Foundations, Toolbox, and Machine Learning Recipes
- Book: Classification and Regression In a Weekend - With Python
- Book: Applied Stochastic Processes
- Long-range Correlations in Time Series: Modeling, Testing, Case Study
- How to Automatically Determine the Number of Clusters in your Data
- New Machine Learning Cheat Sheet | Old one
- Confidence Intervals Without Pain - With Resampling
- Advanced Machine Learning with Basic Excel
- New Perspectives on Statistical Distributions and Deep Learning
- Fascinating New Results in the Theory of Randomness
- Fast Combinatorial Feature Selection

**Other popular resources**

- Comprehensive Repository of Data Science and ML Resources
- Statistical Concepts Explained in Simple English
- Machine Learning Concepts Explained in One Picture
- 100 Data Science Interview Questions and Answers
- Cheat Sheets | Curated Articles | Search | Jobs | Courses
- Post a Blog | Forum Questions | Books | Salaries | News

**Archives:** 2008-2014 |
2015-2016 |
2017-2019 |
Book 1 |
Book 2 |
More

**Most popular articles**

- Free Book and Resources for DSC Members
- New Perspectives on Statistical Distributions and Deep Learning
- Time series, Growth Modeling and Data Science Wizardy
- Statistical Concepts Explained in Simple English
- Machine Learning Concepts Explained in One Picture
- Comprehensive Repository of Data Science and ML Resources
- Advanced Machine Learning with Basic Excel
- Difference between ML, Data Science, AI, Deep Learning, and Statistics
- Selected Business Analytics, Data Science and ML articles
- How to Automatically Determine the Number of Clusters in your Data
- Fascinating New Results in the Theory of Randomness
- Hire a Data Scientist | Search DSC | Find a Job
- Post a Blog | Forum Questions

## You need to be a member of Data Science Central to add comments!

Join Data Science Central