Skip to main content

Posts Tagged ‘Hortonworks’

10 Wishes for Hadoop’s Next 10 Years! #HS16SJ

Day 2 of Hadoop Summit 2016 ended with a big birthday bash for a big yellow elephant named Hadoop.  On account of this milestone, the conference has spent a lot of time discussing what’s next for the platform and the ecosystem, here’s the top ten things on my wishlist for the next 1o years: Assemblies. […]

Top 5 Lessons of Day 1 at Hadoop Summit #HS16SJ

Perficient is at the Hadoop Summit in San Jose, CA and we’re tracking the best of the conference. Here’s the top 5 lessons from day 1: Apache Atlas for managing your business catalog is almost ready for prime time! It is not, however, ready to be a full fledged Records Management solution (no policy management, […]

Big Data and You: DataOps

Welcome to “Big Data and You (the enterprise IT leader),” the Enterprise Content Intelligence group’s demystification of the “Big Data” . The often missing piece of the Infrastructure as code movement emerging from the DevOps space is what we think of as DataOps. Big Data technologies are uniquely poised to fill this gap because they […]

Big Data and You: What is Data Variety?

Welcome to “Big Data and You (the enterprise IT leader),” the Enterprise Content Intelligence group’s demystification of the “Big Data”.  Of the three V’s (Volume, Velocity, and Variety) of big data processing, Variety is perhaps the least understood.  The modern business landscape constantly changes due the emergence of new types of data. The ability to […]

How to Connect Hortonworks Hive from Qlikview with ODBC driver

As with most BI tools, QlikView can use Apache Hive (via ODBC connection) as the SQL access to data in Hadoop. Here we are going to talk about qlikview how to connect Hortonworks Hive via ODBC. Prerequisites 1.Those are versions of each component we installed in Hortonworks Hue HDP Hadoop Hive-Hcatalog Ambari HBase Hortonworks ODBC […]

Hadoop, Spark, Cassandra, Oh My!

Previously, I reviewed why Spark will not by itself replace Hadoop, but Spark combined with other data storage and resource management technologies creates other options for managing Big Data.  Today we will investigate how an enterprise should proceed in this new, “Hadoop is not the only option” world.  Hadoop, Spark, Cassandra, Oh My!  Open source Hadoop and […]

IBM’s Spark Investment is Evidence Big Data is Dead

  Right after I posted my blog on Spark and Hadoop, I came across this article. IBM made a big announcement that they are putting their weight behind Spark.  They are committing more than 3,500 developers and programmers to help move Spark forward. This combined with significant support from the Big 3 Hadoop distributors (HortonWorks, Cloudera, […]

Will Spark Replace Hadoop?

I have seen a number of articles asking the question of whether Apache Spark will replace Hadoop.   This is the wrong question!  It is like asking if your your DVD player will replace your entire home theater system, which is pretty absurd.  Just like a home theatre system has many components, a TV or Projector, a Receiver, […]

Hadoop’s Ever-Increasing Role

With the advent of Splice Machine and the release of Hive 0.14 we are seeing Hadoop’s role in the data center continue to grow. Both of these technologies support limited transactions against data stored in HDFS. Now, I would not suggest moving your mission-critical ERP systems to Hive or Splice Machine, but the support of […]

A little stuffed animal called Hadoop

Doug Cutting – Hadoop creator – is reported to have explained how the name for his Big Data technology came about: “The name my kid gave a stuffed yellow elephant. Short, relatively easy to spell and pronounce, meaningless, and not used elsewhere: those are my naming criteria.” The term, of course, evolved over time and […]