I am a beginner in Apache Spark. I see there is a lot of focus drawn to RDDs in Spark and the faster execution is made possible because of the addition of a caching unit.
Is it fair enough to create a whole new framework like Spark just to include a cache in MapReduce Tasks?
Since am a learner, I think I have a lot to learn but can anyone this doubt of mine?