ALL COVERED TOPICS

NoSQL Benchmarks NoSQL use cases NoSQL Videos NoSQL Hybrid Solutions NoSQL Presentations Big Data Hadoop MapReduce Pig Hive Flume Oozie Sqoop HDFS ZooKeeper Cascading Cascalog BigTable Cassandra HBase Hypertable Couchbase CouchDB MongoDB OrientDB RavenDB Jackrabbit Terrastore Amazon DynamoDB Redis Riak Project Voldemort Tokyo Cabinet Kyoto Cabinet memcached Amazon SimpleDB Datomic MemcacheDB M/DB GT.M Amazon Dynamo Dynomite Mnesia Yahoo! PNUTS/Sherpa Neo4j InfoGrid Sones GraphDB InfiniteGraph AllegroGraph MarkLogic Clustrix CouchDB Case Studies MongoDB Case Studies NoSQL at Adobe NoSQL at Facebook NoSQL at Twitter

NAVIGATE MAIN CATEGORIES

Close

SAS: All content tagged as SAS in NoSQL databases and polyglot persistence

Hadoop Is Not a Solution

Fiona McNeill, SAS product marketing manager, cited by Computerworld:

Hadoop is not a solution. It’s a sympton, not the cure. People are storing stuff on Hadoop just because they can.

While waiting for the SAS cure, I’m filing this for future claim chowder.

Original title and link: Hadoop Is Not a Solution (NoSQL database©myNoSQL)


What Is SAS Doing With Hadoop?

Currently there are only some promises, but I’d bet some of the SAS clients are actually looking forward to seeing these tools.

At SAS, we have a number of initiatives around Hadoop to enable SAS users to access, load, process, visualize and analyze data stored in Hadoop. In the coming months we’ll have more to share but here is a sneak peak of what is to come:

  • SAS/Access interface to Hadoop - this will enable the SAS user to analyze data stored in Hadoop, it also opens up Hadoop data to processing from SAS client software like Data Integration Studio, Enterprise Guide,and Enterprise Miner. The access engine does more than just move data into and out of Hadoop; it also will enable processing to be “pushed-down” into Hadoop.
  • SAS Data Integration Studio transformations - this is a new set of Hadoop transformations that will enable the DI Studio user to load and unload data to and from Hadoop, perform EL-T like processing with HiveQL and ET-L like processing with Pig Latin. Additionally, we are working on a Hadoop specific scoring transform that will enable models developed with Enterprise Miner to be deployed to Hadoop via DI Studio.

Original title and link: What Is SAS Doing With Hadoop? (NoSQL database©myNoSQL)

via: http://blogs.sas.com/content/datamanagement/2011/08/29/sas-hadoop-and-big-data/


Data Analytics at Work

Kelley found its Microsoft SQL-based business intelligence and data warehousing infrastructure couldn’t handle its growing data analytics requirements.

About two years ago, the company started using a new IBM Netezza Twinfin data warehousing appliance, which it supplemented with a second similar system last December. The two systems together, with software from Information Builders and MicroStrategy, form the core of Kelley’s new data warehousing and business intelligence capabilities.

Kelley is also using a variety of predictive analytics, data mining and text analytics tools from SAS Institute to help analyze the data it collects. Much of the analytics used to deliver new and used-car values, targeted advertisements, customized offers and reviews on the company’s website, KBB.com, are powered by SAS’s software.

This is an exemplary story of the value of Big Data and analytic solutions. But I’d also like to learn:

  1. how they move from Microsoft SQL BI to IBM Netezza Twinfin and SAS
  2. how have they evaluated the business value brought by integrating each of these solutions
  3. what is the return on their investment
  4. what alternative solutions have been evaluated in the process and why these particular ones were chosen

Original title and link: Data Analytics at Work (NoSQL database©myNoSQL)

via: http://www.computerworld.com/s/article/9219667/Kelley_Blue_Book_taps_data_analytics_tools_to_improve_car_valuation


The Data Processing Platform for Tomorrow

In the blue corner we have IBM with Netezza as analytic database, Cognos for BI, and SPSS for predictive analytics. In the green corner we have EMC with Greenplum and the partnership with SAS[1]. And in the open source corner we have Hadoop and R.

Update: there’s also another corner I don’t know how to color where Teradata and its recently acquired Aster Data partner with SAS.

Who is ready to bet on which of these platforms will be processing more data in the next years?


  1. GigaOm has a good article on this subject here  

Original title and link: The Data Processing Platform for Tomorrow (NoSQL databases © myNoSQL)