Full Stack Web Development Internship Program
- 29k Enrolled Learners
- Weekend/Weekday
- Live Class
The era of Deep Learning and Machine Learning is at its peak. It is going to create 2.3 Million Jobs by 2020. With new frameworks coming up every month, TensorFlow and Theano have been there for a while and have gained a good amount of popularity as well. So in this Theano vs TensorFlow article, I’ll be discussing the following topics:
Theano can be defined as a library for Scientific Computing. It was developed by the Université de Montréal and has been available since 2007.
It allows you to define, optimize, and evaluate mathematical expressions involving multi-dimensional arrays efficiently. It can run on both CPU and GPU.
TensorFlow is an open-source software library by Google Brain for dataflow programming across a range of tasks.
It is a symbolic math library that is used for machine learning applications like neural networks.
We will compare Theano vs TensorFlow based on the following Metrics:
Popularity:
Theano | TensorFlow |
Theano being an old Framework is not that popular among Data Scientists, Researchers. It was once upon a time | TensorFlow is hands down the most famous Deep Learning Framework and is used in a lot of research. |
Execution Speed:
Technology Benefits:
Theano | TensorFlow |
It supports a wide range of Operations. Theano computes the gradient when determining the Backpropagation error. You have full control over Optimizers as you have to hard code it. | TensorFlow still has to come at par with Theano. That’s not the case for TensorFlow It gives access to lots of good Optimizers out of the box. Which makes Coding Easier |
Compatibility:
Community Support:
Theano | TensorFlow |
Theano has Bigger Community Support as it came way before TensorFlow. | TensorFlow’s Online Community Support is Increasing rapidly with its Popularity. Documentation is comparatively lesser. |
Code Readability:
Let us Compare Theano vs TensorFlow based on their Code. Here I’m Taking a Basic Example Script where we will take some Phony data and initialize the best fit for that data so it can predict future data points.
Theano Code:
import theano import theano.tensor as T import numpy # Again, make 100 points in numpy x_data = numpy.float32(numpy.random.rand(2, 100)) y_data = numpy.dot([0.100, 0.200], x_data) + 0.3 # Intialise the Theano model X = T.matrix() Y = T.vector() b = theano.shared(numpy.random.uniform(-1, 1), name="b") W = theano.shared(numpy.random.uniform(-1.0, 1.0, (1, 2)), name="W") y = W.dot(X) + b # Compute the gradients WRT the mean-squared-error for each parameter cost = T.mean(T.sqr(y - Y)) gradientW = T.grad(cost=cost, wrt=W) gradientB = T.grad(cost=cost, wrt=b) updates = [[W, W - gradientW * 0.5], [b, b - gradientB * 0.5]] train = theano.function(inputs=[X, Y], outputs=cost, updates=updates, allow_input_downcast=True) for i in xrange(0, 201): train(x_data, y_data) print W.get_value(), b.get_value()
Equivalent TensorFlow Code:
import tensorflow as tf import numpy as np # Make 100 phony data points in NumPy. x_data = np.float32(np.random.rand(2, 100)) # Random input y_data = np.dot([0.100, 0.200], x_data) + 0.300 # Construct a linear model. b = tf.Variable(tf.zeros([1])) W = tf.Variable(tf.random_uniform([1, 2], -1.0, 1.0)) y = tf.matmul(W, x_data) + b # Minimize the squared errors. loss = tf.reduce_mean(tf.square(y - y_data)) optimizer = tf.train.GradientDescentOptimizer(0.5) train = optimizer.minimize(loss) # For initializing the variables. init = tf.initialize_all_variables() # Launch the graph sess = tf.Session() sess.run(init) # Fit the plane. for step in xrange(0, 201): sess.run(train) if step % 20 == 0: print step, sess.run(W), sess.run(b) # Learns best fit is W: [[0.100 0.200]], b: [0.300]
Length Wise Both the Code are almost Similar there’s not much difference. Two identically-generated NumPy arrays describing the input, and the target output. But if we have a look at the Model Initialization.
Model Initialization:
# TensorFlow b = tf.Variable(tf.zeros([1])) W = tf.Variable(tf.random_uniform([1, 2], -1.0, 1.0)) y = tf.matmul(W, x_data) + b # Theano X = T.matrix() Y = T.vector() b = theano.shared(numpy.random.uniform(-1, 1), name="b") W = theano.shared(numpy.random.uniform(-1.0, 1.0, (1, 2)), name="W") y = W.dot(X) + b
As you can see here that TensorFlow doesn’t require any Special Treatment of X and Y Variables. On the other hand, Theano requires an extra effort to make sure that the variables are Symbolic Inputs to the Function. The definition of b and W are explanatory and also nicer.
The Learning: Optimization
# Tensorflow loss = tf.reduce_mean(tf.square(y - y_data)) # (1) optimizer = tf.train.GradientDescentOptimizer(0.5) # (2) train = optimizer.minimize(loss) # (3) # Theano cost = T.mean(T.sqr(y - Y)) # (1) gradientW = T.grad(cost=cost, wrt=W) # (2) gradientB = T.grad(cost=cost, wrt=b) # (2) updates = [[W, W - gradientW * 0.5], [b, b - gradientB * 0.5]] # (2) train = theano.function(inputs=[X, Y], outputs=cost, updates=updates, allow_input_downcast=True) # (3)
For (1) the MSE is almost the same for Theano vs TensorFlow.
For (2) Defining the Optimizer is easy and simple as it gets in case of TensorFlow, but Theanno gives you a great deal of Control overt the Optimizers although its quite lengthy and increases Verification Effort.
For (3) Training Function the Code is almost Similar
Training Body:
# TensorFlow init = tf.initialize_all_variables() sess = tf.Session() sess.run(init) for step in xrange(0, 201): sess.run(train) # Theano for i in xrange(0, 201): train(x_data, y_data) print W.get_value(), b.get_value()
The code for Training is almost Identical, but Encapsulating the Graph Execution in Session Object is Conceptually Cleaner than Theano.
On a Concluding Note, it can be said that both APIs have a similar Interface. But TensorFlow is comparatively easier yo use as it provides a lot of Monitoring and Debugging Tools. Theano takes the Lead in Usability and Speed, but TensorFlow is better suited for Deployment. Paperwork or Documentation for Theano is more than TensorFlow and TensorFlow being a new Language people don’t have many resources, to begin with. Open-source deep-libraries such as Keras, Lasagne and Blocks have been built on top of Theano.
If you’re trying to extend your business in this exciting field, check out our Artificial Intelligence Course to upgrade your AI skills to the next level. It is offered in collaboration with E&ICT Academy, National Institute of Technology, Warangal. This executive Masters’ Program equips students with information about the tools, techniques, and tools they require to advance their careers.
Got a question for us? Please mention it in the comments section of “Theano vs TensorFlow” and we will get back to you.
Course Name | Date | Details |
---|---|---|
Artificial Intelligence Certification Course | Class Starts on 28th December,2024 28th December SAT&SUN (Weekend Batch) | View Details |