Amazing talk, I left off Spark to move in to ML when there was only RDD, I came back and see DataFrame in Spark and I am totally confused, your video helped a lot, Thank you
I am wondering how the "type safe" feature combines with the "unstructured data" that is the nature of data in the systems that spark would be used in.
I was trying out the example you mentioned @10:46 and as i am getting compile time error, I had to rewrite the final statement as below. parsedRdd.filter( content => content._2 == "en").map(filteredContent => filteredContent._3).reduce(_+_).take(100).foreach(reducedContent => printf(s"$reducedContent._1: $reducedContent._2")) I would really appreciate if you can review above statement