Think you have a big data problem? Facebook deals with one the world’s largest datasets, and it’s important to store and process it efficiently. It’s no secret that Facebook is a big user of Hadoop, and it’s amazing how they have scaled it to handle 600TB of new data every day. The way they compress the data directly impacts storage costs. Some of their methodologies and insights are outlined in a fascinating blog post.

Read - Scaling the Facebook data warehouse to 300 PB | Engineering Blog | Facebook Code | Facebook

Leave a Comment