NetApp: All content tagged as NetApp in NoSQL databases and polyglot persistence
Steve Loughran starts with a critical look at Netapp Open solution for Hadoop paper:
Actually it is weirder than I first thought. This is still HDFS, just running on more expensive hardware. You get the (current) HDFS limitations: no native filesystem mounting, a namenode to care about, security on a par with NFS, without the cost savings of pure-SATA-no-licensing-fees. Instead you have to use RAID everywhere, which not only bumps up your cost of storage, puts you at risk of RAID controller failure and errors in the OS drivers for those controller (hence their strict rules about which Linux releases to trust). If you do follow their recommendations and rely on hardware for data integrity, you’ve cut down the probability of node-local job execution, so all FUD about replication traffic is now moot as at least 1/3 more of your tasks will be running remote -possibly even with the Fair Scheduler, which waits for a bit to see if a local slot becomes free. What they are doing then is adding some HA hardware underneath a filesystem that is designed to give strong availability out of medium availability hardware. I have seen such a design before, and thought it sucked then too. Information week says this is a response to EMC, but it looks more like NetApp’s strategy to stay relevant, and Cloudera are partnering with them as NetApp offered them money and if it sells into more “enterprise customers” then why not? With the extra hardware costs of NetApp the cloudera licenses will look better value, and clearly both NetApp and their customers are in need of the hand-holding that Cloudera can offer.
Then in a follow up post, he looks at a couple of alternatives (Lustre, GPFS, IBRIX, etc):
I’m not against running MapReduce—or the entire Hadoop stack—against alternate filesystems. There are some good cases where it makes sense. Other filesystems offer security, NFS mounting, the ability to be used by other applications and other features. HDFS is designed to scale well on “commodity” hardware, (where servers containing Xeon E5 series parts with 64GB RAM, 10GbE and 8-12 SFF HDDs are considered a subset of “commodity”).
Original title and link: A Short Incursion Into Alternate Hadoop Filesystems ( ©myNoSQL)
- DataStax Brisk: Hadoop and Hive on Cassandra
- NetApp Hadoop Shared DAS
increase throughput in Hadoop clusters via its ConnectX-2 adapters with Hadoop Direct
SnapReduce transforms SnapLogic data integration pipelines directly into MapReduce tasks, making Hadoop processing much more accessible and resulting in optimal Hadoop cluster utilization.
Greenplum HD combines the Hadoop analytics platform with Greenplum’s database technology.
Ways to look at it:
- 2 large corporations getting into Hadoop
- 2 software solutions, 3 hardware solutions
- 1 open source project, 4 commercial products or
- 4 companies wanting to make a profit from Hadoop without contributing back to the community
Original title and link: Hadoop Ecosystem: EMC, NetApp, Mellanox, SnapLogic, DataStax (NoSQL databases © myNoSQL)