article thumbnail

Brief History of Data Engineering

Jesse Anderson

They created MapReduce and GFS in 2004. Doug Cutting took those papers and created Apache Hadoop in 2005. They were the first companies to commercialize open source big data technologies and pushed the marketing and commercialization of Hadoop. Hadoop was hard to program, and Apache Hive came along in 2010 to add SQL.

article thumbnail

A Prequel to Data Mesh

Towards Data Science

Image by the author 2004 to 2010 — The elephant enters the room New wave of applications emerged — Social Media, Software observability, etc. Result: Hadoop & NoSQL frameworks emerged. Result: Companies started to sell pre-configured data warehouses as products. The concept of `Data Marts` was introduced.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Evolution of the Cloud Data Platform: From Google to Ascend

Ascend.io

Back in 2004, I got to work with MapReduce at Google years before Apache Hadoop was even released, using it on a nearly daily basis to analyze user activity on web search and analyze the efficacy of user experiments. I’ve had the good fortune to work at or start companies that were breaking new ground. Big data would be a big deal.

Cloud 52
article thumbnail

Evolution of the Cloud Data Platform: From Google to Ascend

Ascend.io

Back in 2004, I got to work with MapReduce at Google years before Apache Hadoop was even released, using it on a nearly daily basis to analyze user activity on web search and analyze the efficacy of user experiments. I’ve had the good fortune to work at or start companies that were breaking new ground. Big data would be a big deal.

Cloud 52
article thumbnail

Data Analysis with Spark

Zalando Engineering

For the sake of comparison, let’s recap the Hadoop way of working: Hadoop saves intermediate states to disk and communicates over a network. In fact, in a 2004 mapReduce research paper the designer states that key-value pairs is a key choice in designing mapReduce. Provides in memory storage for cached RDD’s.

article thumbnail

Q&A with Greg Rahn – The changing Data Warehouse market

Cloudera

Greg Rahn: Toward the end of that eight-year stint, I saw this thing coming up called Hadoop and an engine called Hive. In the Hadoop world, or the big data world, most of these components are separate and modular, but yet interact together to form a system that behaves very similarly. Say, circa 2004 when I started at Oracle.

article thumbnail

Industry Interview Series- How Big Data is Transforming Business Intelligence?

ProjectPro

Solocal has taken big data to the next stage of BI by designing a novel vision of BI with the open source distributed computing framework Hadoop. It replaced its traditional BI structure by integrating big data and Hadoop."-April In BI – there is a need to use ETL on top of Hadoop as there is not much scripting.