(Please refer to the following article: Oracle 12c In-Memory Database is Out
- Hardly Anybody Notices for update on Oracle 12c databases)
Contemporary large servers are routinely configured with 2TB of RAM. It is
thus possible to fit an entire average size OLTP database in memory directly
accessible by CPU. There is a long history of academic research on how to
best utilize relatively abundant computer memory. This research is becoming
increasingly relevant as databases serving business applications are heading
towards memory centric design and implementation.
If you simply place Oracle RDBMS's files on Solid State Disk, or configure
buffer cache (SGA) large enough to contain the whole database, Oracle will
not magically become an IMDB database, nor it will perform much faster. In
order to properly utilize memory, IMDB databases require purposely
architected, confi... (more)
Machine Learning is a critical part of extracting value from Big Data.
Choosing proper model, preparing data and getting usable results on large
scale data is non-trivial exercise. Typically process consists of model
prototyping using higher level, (mostly) single machine based tool like R,
Matlab, Weka, then coding in Java or some other language for large scale
deployment. This process is fairly involved, error prone, slow and
Existing tools aiming at automating and improving this process are still
somewhat immature and wide scale Machine Learning enterprise adopti... (more)
Big Data and its most prominent technical ingredient, Machine Learning, are
all the rage these days, as IT industry is trying to convince companies
technology revolution is underway. ( "If you are not doing it, your
competitors sure are, and by the time you realize it, it will be too late" ).
Data fracking, i.e. Big Data, is 21 century new oil of that will power and
grease stalled industries and reignite growth, or so the story goes.
While advanced analytics (it comes under various names - predictive
analytics, data mining, and data science, more recently) is great and in use
Many products are available as open source or proprietary products that can
handle Big Data. Which one is best fit for this task?
Today's classic RDBMSs and tools are able to quickly load the data, process
it and present results in an easy to understand format. You can use SQL
or programmatic interface to process the data randomly or in batch; RDBMS's
keep data safe, protected against hardware and software failures.
Standards tools and products are not able to cope with Big Data requirement,
which is not dissimilar to what is involved in processing today's regular
data sets, jus... (more)
Oracle RMAN (Recovery Manager) catalog database contains backup metadata
for all enterprise wide Oracle databases. Each database backed up through
RMAN needs connection to a central backup repository database. RMAN catalog
is purged and maintained by removing obsolete backup records, crosschecking
with existing media etc. If central repository grows too big then backup
performance will suffer since it will take too long for RMAN to locate
Amazon Web Services gives us the ability to quickly and easily create either
brand new RMAN catalog database or move an existing c... (more)