Apache Parquet, which provides columnar storage in Hadoop, is now a top-level Apache Software Foundation (ASF)-sponsored project, paving the way for its more advanced use in the Hadoop ecosystem.
Industrial environments such as those found within large electric and gas utilities are producing massive volumes of data in real-time that is overwhelming traditional ICT architectures. Additionally, ...
While Hadoop is officially 15 years old as an Apache project, it only gained mainstream IT attention 10 years ago. Hadoop started as an open source implementation of key Google technologies used for ...
The advent of scalable analytics in the form of Hadoop and Spark seems to be moving to the end of the Technology Hype Cycle. A reasonable estimate would put the technology on the “slope of ...
The proliferation of small files in distributed file systems poses significant challenges that affect both storage efficiency and operational performance. Modern systems, such as Hadoop Distributed ...
A few weeks ago, two giants of the big data Hadoop era, Cloudera and Hortonworks, announced they would be merging. The announcement claimed it would be a “merger of equals.” It is fascinating to see ...
Cisco and NetApp Tuesday rolled out a new version of their jointly developed FlexPod converged infrastructure aimed specifically at big data workloads, the first in a series of solutions targeting ...