*This article was written by Harshit Sharma.*

Successful detection and classification of traffic signs is one of the important problem to be solved if we want self driving cars. Idea is to make automobiles smart enough so as to achieve least human interaction for successful automation. Swift rise in dominance of deep learning over classical machine learning methods which is complemented by advancement in GPU (Graphics Processing Unit) has been astonishing in fields related to image recognition, NLP, self-driving cars etc.

The chart only goes to 2015 and 2016 NIPS had over 6,000 attendees. Chart displays surge in interest in deep learning and related techniques.

Deep learning is a class of machine learning technique where artificial neural network uses multiple hidden layers. Lot of credit goes to David Hubel and Torsten Wiesel, two famous neurophysiologists, who showed how neurons in the visual cortex work. Their work determined how neurons with similar functions are organized into columns, tiny computational machines that relay information to a higher region of the brain, where visual image is progressively formed.

*In layman’s term* brain combines low level features such as basic shapes, curves and builds more complex shapes out of it. A deep convolutional neural network is similar. It first identifies low level features and then learns to recognize and combines these features to learn more complicated patterns. These different levels of features come from different layers of the network.

Tensorflowis used to implement deep conv net for traffic sign classification. RELU was used for activation to introduce non linearity and drop out of varying percentages was used to avoid overfitting at different stages. It was hard to believe that simple deep net was able to achieve high accuracy on the training data. Following is the architecture used for traffic sign classification.

I decided to start with simple architecture after exploring LeNet. Idea is to start simple and add more complexity if required . Architecture is same as diagram posted above . Converting images to Grayscale really helped to achieve better accuracy. Model was trained for *40 epochs* for 256 *batch size *using* “adam” *optimizer, *learn rate* initially was set to 0.0001 which was further increased to *0.001 *for converging to target accuracy. This is my first model on top of tensorflow. Following is the small snippet of the model.

def deepnn_model(x,train=True):

# Arguments used for tf.truncated_normal, randomly defines variables for the weights and biases for each layer

x = tf.nn.conv2d(x, layer1_weight, strides=[1, 1, 1, 1], padding='VALID')

x = tf.nn.bias_add(x, layer1_bias)

x = tf.nn.relu(x)

x = tf.nn.max_pool(x,ksize=[1, 2, 2, 1],strides=[1, 2, 2, 1],padding='SAME')

if(train):

x = tf.nn.dropout(x, dropout1)

x = tf.nn.conv2d(x, layer2_weight, strides=[1, 1, 1, 1], padding='VALID')

x = tf.nn.bias_add(x, layer2_bias)

x = tf.nn.relu(x)

conv2 = tf.nn.max_pool(x,ksize=[1, 2, 2, 1],strides=[1, 2, 2, 1],padding='SAME')

if(train):

conv2 = tf.nn.dropout(conv2, dropout2)

fc0 = flatten(conv2)

fc1 = tf.add(tf.matmul(fc0, flat_weight),bias_flat)

fc1 = tf.nn.relu(fc1)

if(train):

fc1 = tf.nn.dropout(fc1, dropout3)

fc1 = tf.add(tf.matmul(fc1, flat_weight2),bias_flat2)

fc1 = tf.nn.relu(fc1)

if(train):

fc1 = tf.nn.dropout(fc1, dropout4)

fc1 = tf.add(tf.matmul(fc1, flat_weight3),bias_flat3)

logits = tf.nn.relu(fc1)

return logits

My aim was to achieve 90 percent *accuracy* to start with simple architecture and I was surprised to see it close to *94 percent* in the first run. Further adding more layers and complexity into the model complemented by data augmentation can achieve accuracy as high as 98 percent.

To read the full original article click here. For more deep learning related articles on DSC click here.

**DSC Resources**

- Services: Hire a Data Scientist | Search DSC | Classifieds | Find a Job
- Contributors: Post a Blog | Ask a Question
- Follow us: @DataScienceCtrl | @AnalyticBridge

Popular Articles

© 2021 TechTarget, Inc. Powered by

Badges | Report an Issue | Privacy Policy | Terms of Service

**Most Popular Content on DSC**

To not miss this type of content in the future, subscribe to our newsletter.

- Book: Applied Stochastic Processes
- Long-range Correlations in Time Series: Modeling, Testing, Case Study
- How to Automatically Determine the Number of Clusters in your Data
- New Machine Learning Cheat Sheet | Old one
- Confidence Intervals Without Pain - With Resampling
- Advanced Machine Learning with Basic Excel
- New Perspectives on Statistical Distributions and Deep Learning
- Fascinating New Results in the Theory of Randomness
- Fast Combinatorial Feature Selection

**Other popular resources**

- Comprehensive Repository of Data Science and ML Resources
- Statistical Concepts Explained in Simple English
- Machine Learning Concepts Explained in One Picture
- 100 Data Science Interview Questions and Answers
- Cheat Sheets | Curated Articles | Search | Jobs | Courses
- Post a Blog | Forum Questions | Books | Salaries | News

**Archives:** 2008-2014 |
2015-2016 |
2017-2019 |
Book 1 |
Book 2 |
More

**Most popular articles**

- Free Book and Resources for DSC Members
- New Perspectives on Statistical Distributions and Deep Learning
- Time series, Growth Modeling and Data Science Wizardy
- Statistical Concepts Explained in Simple English
- Machine Learning Concepts Explained in One Picture
- Comprehensive Repository of Data Science and ML Resources
- Advanced Machine Learning with Basic Excel
- Difference between ML, Data Science, AI, Deep Learning, and Statistics
- Selected Business Analytics, Data Science and ML articles
- How to Automatically Determine the Number of Clusters in your Data
- Fascinating New Results in the Theory of Randomness
- Hire a Data Scientist | Search DSC | Find a Job
- Post a Blog | Forum Questions

## You need to be a member of Data Science Central to add comments!

Join Data Science Central