Advanced AI Course (51 Blogs)
AWS Global Infrastructure

Theano vs TensorFlow : A Quick Comparision of Frameworks

Last updated on Apr 18,2023 7.2K Views

Kislay Keshari
Kurt is a Big Data and Data Science Expert, working as a... Kurt is a Big Data and Data Science Expert, working as a Research Analyst at Edureka. He is keen to work with Machine Learning,...

The era of Deep Learning and Machine Learning is at its peak. It is going to create 2.3 Million Jobs by 2020. With new frameworks coming up every month, TensorFlow and Theano have been there for a while and have gained a good amount of popularity as well. So in this Theano vs TensorFlow article, I’ll be discussing the following topics:

 

What is Theano?

Theano can be defined as a library for Scientific Computing. It was developed by the Université de Montréal and has been available since 2007.

theano-logo

It allows you to define, optimize, and evaluate mathematical expressions involving multi-dimensional arrays efficiently. It can run on both CPU and GPU.

 

What is TensorFlow?

TensorFlow is an open-source software library by Google Brain for dataflow programming across a range of tasks.

tensorflow

It is a symbolic math library that is used for machine learning applications like neural networks.

 

Theano vs TensorFlow

We will compare Theano vs TensorFlow based on the following Metrics:

Popularity:

popularity-theano-vs-tensorflow

TheanoTensorFlow
Theano being an old Framework is not that popular among Data Scientists, Researchers. It was once upon a timeTensorFlow is hands down the most famous Deep Learning Framework and is used in a lot of research.

 

Execution Speed:

Speed

TheanoTensorFlow
Performs Tasks Faster than TensorFlow. Especially the single GPU Tasks run, way fast in Theano.TensorFlow’s Execution speed is Slower as compared to Theano, But in Multi-GPU Tasks it takes the Lead.

 

Technology Benefits:

Technology-Benefits

TheanoTensorFlow
It supports a wide range of Operations.

Theano computes the gradient when determining the Backpropagation error.

You have full control over Optimizers as you have to hard code it.

TensorFlow still has to come at par with Theano.

That’s not the case for TensorFlow

It gives access to lots of good Optimizers out of the box. Which makes Coding Easier

 

Compatibility:

Compatibility-Theano-vs-TensorFlow

TheanoTensorFlow
Keras an amazing Deep Learning Library is compatible with Theano. It Integrates Well.

It has Native Windows Support.

It also supports High-Level Wrappers like Lasagne.

But In case of TensorFlow, it’s not quite there yet. However, in v2.0 this won’t be the case.

Currently, TensorFlow lacks this Support.

No Support for Lasagne.

 

Community Support:

Community-Support-Theano-vs-TensorFlow

TheanoTensorFlow
 Theano has Bigger Community Support as it came way before TensorFlow.

It has More Documentation than TensorFlow

TensorFlow’s Online Community Support is Increasing rapidly with its Popularity.

Documentation is comparatively lesser.

 

Code Readability:

Code

Let us Compare Theano vs TensorFlow based on their Code. Here I’m Taking a Basic Example Script where we will take some Phony data and initialize the best fit for that data so it can predict future data points.

Theano Code:

import theano
import theano.tensor as T
import numpy

# Again, make 100 points in numpy
x_data = numpy.float32(numpy.random.rand(2, 100))
y_data = numpy.dot([0.100, 0.200], x_data) + 0.3

# Intialise the Theano model
X = T.matrix()
Y = T.vector()
b = theano.shared(numpy.random.uniform(-1, 1), name="b")
W = theano.shared(numpy.random.uniform(-1.0, 1.0, (1, 2)), name="W")
y = W.dot(X) + b 

# Compute the gradients WRT the mean-squared-error for each parameter
cost = T.mean(T.sqr(y - Y))
gradientW = T.grad(cost=cost, wrt=W)
gradientB = T.grad(cost=cost, wrt=b)
updates = [[W, W - gradientW * 0.5], [b, b - gradientB * 0.5]] 

train = theano.function(inputs=[X, Y], outputs=cost, updates=updates, allow_input_downcast=True) 

for i in xrange(0, 201):
    train(x_data, y_data)
    print W.get_value(), b.get_value()

Equivalent TensorFlow Code:

import tensorflow as tf
import numpy as np

# Make 100 phony data points in NumPy.
x_data = np.float32(np.random.rand(2, 100)) # Random input
y_data = np.dot([0.100, 0.200], x_data) + 0.300

# Construct a linear model.
b = tf.Variable(tf.zeros([1]))
W = tf.Variable(tf.random_uniform([1, 2], -1.0, 1.0))
y = tf.matmul(W, x_data) + b

# Minimize the squared errors.
loss = tf.reduce_mean(tf.square(y - y_data))
optimizer = tf.train.GradientDescentOptimizer(0.5)
train = optimizer.minimize(loss)

# For initializing the variables.
init = tf.initialize_all_variables()

# Launch the graph
sess = tf.Session()
sess.run(init)

# Fit the plane.
for step in xrange(0, 201):
    sess.run(train)
    if step % 20 == 0:
        print step, sess.run(W), sess.run(b)

# Learns best fit is W: [[0.100  0.200]], b: [0.300]

Length Wise Both the Code are almost Similar there’s not much difference. Two identically-generated NumPy arrays describing the input, and the target output. But if we have a look at the Model Initialization.

Model Initialization:

# TensorFlow
b = tf.Variable(tf.zeros([1]))
W = tf.Variable(tf.random_uniform([1, 2], -1.0, 1.0))
y = tf.matmul(W, x_data) + b

# Theano
X = T.matrix()
Y = T.vector()
b = theano.shared(numpy.random.uniform(-1, 1), name="b")
W = theano.shared(numpy.random.uniform(-1.0, 1.0, (1, 2)), name="W")
y = W.dot(X) + b

As you can see here that TensorFlow doesn’t require any Special Treatment of X and Y Variables. On the other hand, Theano requires an extra effort to make sure that the variables are Symbolic Inputs to the Function. The definition of b and W are explanatory and also nicer.

The Learning: Optimization

# Tensorflow
loss = tf.reduce_mean(tf.square(y - y_data)) # (1)
optimizer = tf.train.GradientDescentOptimizer(0.5) # (2)
train = optimizer.minimize(loss) # (3)

# Theano
cost = T.mean(T.sqr(y - Y)) # (1)
gradientW = T.grad(cost=cost, wrt=W) # (2)
gradientB = T.grad(cost=cost, wrt=b) # (2)
updates = [[W, W - gradientW * 0.5], [b, b - gradientB * 0.5]] # (2)
train = theano.function(inputs=[X, Y], outputs=cost, updates=updates, allow_input_downcast=True) # (3)

For (1) the MSE is almost the same for Theano vs TensorFlow.

For (2) Defining the Optimizer is easy and simple as it gets in case of TensorFlow, but Theanno gives you a great deal of Control overt the Optimizers although its quite lengthy and increases Verification Effort.

For (3) Training Function the Code is almost Similar

 

Training Body:

# TensorFlow
init = tf.initialize_all_variables()

sess = tf.Session()
sess.run(init)
for step in xrange(0, 201):
    sess.run(train)

# Theano
for i in xrange(0, 201):
    train(x_data, y_data)
    print W.get_value(), b.get_value()

The code for Training is almost Identical, but Encapsulating the Graph Execution in Session Object is Conceptually Cleaner than Theano.

 

Final Verdict: Theano vs TensorFlow

On a Concluding Note, it can be said that both APIs have a similar Interface. But TensorFlow is comparatively easier yo use as it provides a lot of Monitoring and Debugging Tools. Theano takes the Lead in Usability and Speed, but TensorFlow is better suited for Deployment. Paperwork or Documentation for Theano is more than TensorFlow and TensorFlow being a new Language people don’t have many resources, to begin with. Open-source deep-libraries such as Keras, Lasagne and Blocks have been built on top of Theano.

If you’re trying to extend your business in this exciting field, check out our Artificial Intelligence Course to upgrade your AI skills to the next level. It is offered in collaboration with E&ICT Academy, National Institute of Technology, Warangal. This executive Masters’ Program equips students with information about the tools, techniques, and tools they require to advance their careers. 

Got a question for us? Please mention it in the comments section of “Theano vs TensorFlow” and we will get back to you.

Comments
0 Comments

Join the discussion

Browse Categories

Subscribe to our Newsletter, and get personalized recommendations.