Contents

Recently I found that I’m kind of over talking when I speak with others, so I want to train my conclusion of solving problem. Here are some problems I meet when I read books of Big Data.

  • No 1. Why Spark does not use map-reduce?

Disk reading is too slow to complete big data analyze.Map-reduce is a shuffle which meanings parallizing.Spark just use memory to complete parallizing a huge problem.

  • No 2. How Hadoop to append data?

Just use

Contents