Oracle: All content tagged as Oracle in NoSQL databases and polyglot persistence
There’s been a lot of speculation about the announcements coming from Oracle’s OpenWorld event. A first part was revealed during the keynote in the form of an in-memory analytics appliance called Exalytics . But there’s talk about a Big Data Appliance and an Oracle NoSQL database.
Here’re my predictions
Oracle became very aggressive in selling products based on hardware, software, and services. So they’ll announce a Hadoop appliance integrated with an existing Oracle product. It could be either the Oracle Exadata or even the newly announced Exalytics.
This appliance will place Oracle in competition with all other Hadoop appliance sellers: EMC, NetApp, IBM. Also these days most of the analytics databases try to integrate with Hadoop.
Oracle already has a couple of non-relational solutions in their portfolio: BerkleyDB, TimesTen, Coherence. And they’ve already started to test the NoSQL market by announcing the MySQL and MySQL Cluster NoSQL hybrid systems.
I don’t expect Oracle NoSQL database to be a new product. Just a rebranding or repackaging of one of the above mentioned ones. Probably the TimesTen.
Oracle will invest more into integrating its line of products with Hadoop. Having both a Hadoop and an in-memory analytics appliance will make them very competitive in this space.
Oracle will extend the support for NoSQLish interfaces (memcached) to its other database products.
What are your predictions?
Original title and link: The Oracle NoSQL Database and Big Data Appliance ( ©myNoSQL)
A while ago, Sid Anand has written a series of posts on challenges of a hybrid solution: Oracle - Amazon SimpleDB. This has become now a paper which offers a much better organized and detailed view on Netflix’s transition to using a hybrid Oracle - Amazon Web Services (SimpleDB, S3) architecture.
Go read the ☞ paper if one of these applies:
- interested in Amazon SimpleDB and SimpleDB best practices
- interested in running an on-premise and cloud hybrid architecture
- interested in architecting a multi data source system
Original title and link: Paper: Netflix’s Transition to High-Availability Storage Systems (NoSQL databases © myNoSQL)
I have heard many mentioning that Oracle removed InnoDB from the MySQL classical edition version. Now, I don’t know too much about the various versions and licenses of MySQL — it looks like there are at least 5: enterprise, classical, standard, cluster carrier grade, and community — but InnoDB doesn’t seem to have been dropped from the community edition too. So, I’m not really sure this is such a big deal.
What are your thoughts on this story?
InnoDB is available under the GPL. Innostore, as a derivative work of Embedded InnoDB, is also available under the GPL. Neither Oracle nor Basho can take that away from you.
- If everyone would actually be forced to go back at using MyISAM, that would be a bit more interesting as it would mean MySQL will be less durable and consistent. (↩)
Original title and link: Oracle Drops InnoDB from MySQL Classical Edition, But Not From Community Edition (NoSQL databases © myNoSQL)
Back in January, I was writing about some solutions to integrate Oracle databases with Hadoop and also speculated that most probably more and more tools will look into providing MapReduce support.
Last week, Cloudera and Quest Software have ☞ announced a new connector allowing Hadoop to pull data from Oracle databases.
For those database administrators who’ve suppressed their jealousy, while their programming and scripting colleagues talk gleefully about the death of relational databases, this week’s news may cause them to regain some pride.
[…] The new connector is currently known as Ora-Oop.
Oracle seems to be in the first lines of this initiative as it looks like ☞ its database becomes more and more aware of systems like Hadoop. The linked article presents two ways in which Oracle can pull data out of HDFS by either accessing it directly through the FUSE driver or by triggering Hadoop to push data into Oracle queues which are further accessible from table functions. A commenter on the post has suggested a 3rd option that sounds even more interesting: using the Oracle Java support for accessing the Hadoop API.
Diagram from Oracle Blogs
While the presented solutions are only about pulling data from Hadoop and processing them in parallel using Oracle parallel processing support, I do think that sooner than later we will see solutions that will use Hadoop for processing data made accessible directly by Oracle.
Here is just a thought on how this would work:
- use some special Oracle functions to pull data from tables and push it into Hadoop accessible queues
- Hadoop (with streaming support) would pull out data from queues and process them internally
- when processing is done, Hadoop can push back data into Oracle accessible queues (as per the above solutions).
Isn’t that an interesting future?
Update: in the light of the newly granted MapReduce patent (Google), I guess it will be a bit more difficult to blame anyone for not incorporating or integrating more closely with Hadoop. What do you think?