Click here to close now.

Welcome!

Java Authors: Elizabeth White, Pat Romanski, Liz McMillan, Carmen Gonzalez, Kelly Murphy

Related Topics: Big Data Journal, Java, Microservices Journal, Linux, Virtualization, Cloud Expo

Big Data Journal: Blog Post

In-Memory Computing By @GridGain | @CloudExpo [#BigData]

The best way to clear the air around IMC is to start with a simple explanation of what, in fact, in-memory computing is

The Facts and Fiction of In-Memory Computing

In the last year, conversations about In-Memory Computing (IMC) have become more and more prevalent in enterprise IT circles, especially with organizations feeling the pressure to process massive quantities of data at the speed that is now being demanded by the Internet. The hype around IMC is justified: tasks that once took hours to execute are streamlined down to seconds by moving the computation and data from disk, directly to RAM. Through this simple adjustment, analytics are happening in real-time, and applications (as well as the development of applications) are working at-pace with this new standard of technology and speed.

Despite becoming both more cost-effective and accepted within enterprise computing, there are still a small handful of falsehoods that confuse even the most technical of individuals in enterprise IT.

Myth: In-memory computing is about databases, so this isn't really relevant to my business.

The best way to clear the air around IMC is to start with a simple explanation of what, in fact, in-memory computing is. While many assume that because we are talking about RAM, we are having a conversation about databases and storage, but this is not the case.

IMC, at its most basic level, is using a middleware software that allows one to store data in RAM - across a broad cluster of computers - and do any and all processing where it resides (in the memory). With traditional methods, data processing is often confined to spinning disks.

By comparison, in-memory computing speeds up this process by roughly 5,000 times. Now you can see that we're not talking about storage only - instead active, fluid data and computing.

Which brings me to another, more tangible point about computing efficiency. By incorporating in-memory, a handful of processes are streamlined in order to save time, resources, and money.

To start, in-memory requires much less hardware; the result - significantly decreased capital, operational and infrastructure overhead.

Moreover, IT departments can also significantly extend the life of existing hardware and software through the increased performance that is inherent with IMC - thus amplifying the ROI on the machines that have already been purchased.

Surprisingly, in-memory computing is not a new phenomenon. Since the inception of RAM, IMC has been viewed as reliable accelerant for high-performance computing, bringing us to the next crucial misconception about this technology.

Myth: In-memory computing is expensive, therefore not practical for my operation.

There is a reason that this is one of the most common misunderstandings about IMC, because there was a point in time where the cost of memory was once quite high. That being said, the cost of RAM has been dropping consistently, at a rate of about 30% - for the last 5 years.

Today, the price of a 1 Terabyte RAM cluster can go for anywhere between $20 and $40 thousand - including all of the CPUs, networking, etc. A few years from now that same setup will likely be available for half that price.

Regardless of the future price of RAM, which based upon current projections will likely continue to fall, the current economics have already placed this technology well within the reaches of the enterprise computing budgets that require this level of scale.

Myth: My needs are already being met by Flash.

There are three different reasons why this mentality is held by IT folks, each of which are highly misinformed. I'll start with the most common, which is the idea that your business doesn't need the Lambourgini-esque super-computing power of IMC.

The hard yet obvious reality is that if your business is in any way data-driven, you likely cannot survive without speed and agility in this department. As time goes on, the amount of data that businesses accumulate compounds with new streams and variances. This is a sink-or-swim reality.

Another myth commonly used to dispel IMC is that if businesses are able to just effectively mount RAM disk, they will get in-memory processing. Unfortunately, it's not that easy. As mentioned earlier, IMC works through middleware to effectively unlock its power.

Finally, there's the notion that one can just replace their HDDs with SSDs in order to get this super-charged performance. For SSDs - in certain situations - the performance gain that you can pull from flash storage in lieu of spinning disk is enough.

However, speed matters - and is rapidly becoming more of a requirement every day. At this point, it's like comparing apples to oranges with speed improvements of 10 to 100x over SSDs.

Myth: Memory is not durable enough to be truly sustainable.

This is another notion that for whatever reason has been both widely perpetuated - and is entirely false.

The fact is - almost all in-memory computing middleware (apart from very simplistic ones) offer one or multiple strategies for in-memory backups, durable storage backups, disk-based swap space overflow, etc.

More sophisticated vendors provide a comprehensive tiered storage approach where users can decide what portion of the overall data set is stored in RAM, local disk swap space or RDBMS/HDFS - where each tier can store progressively more data but with progressively longer latencies.

Yet another source of confusion is the difference between operational datasets and historical datasets. In-memory computing is not aimed at replacing enterprise data warehouse (EDW), backup or offline storage services - like Hadoop, for example. The goal of IMC is to improve the operational datasets that require mixed OLTP and OLAP processing and in most cases are less than 10TB in size. That is to say, in-memory computing is not "all or nothing" - and does not require that every aspect of data be housed in memory.

The in-memory computing revolution is by no means intended to obliterate disks from the enterprise. For now, the disk still serves a well-defined role for offline/backup use cases - tasks that are not the focus of IMC.

Myth: In-memory is inaccessible to my business because so few developers actually know how to use it.

Yes indeed, In-memory computing is a highly complex technology, that for now, only a few vendors have even been able to successfully develop offerings for. However, like much of high-technology, in-memory computing has entered the world of open source - bringing its capabilities and power to the fingertips of developers around the world.

Currently, with GridGain, developers have the ability to get their hands on IMC with a simple download at http://gridgain.org/.

In-memory computing is already being tapped across a broad range of functions and industries including (but not limited to) financial trading systems, online game, bioinformatics, hyper-local advertising, cognitive computing, and geospatial analysis.

By raising awareness, and bringing the capabilities of IMC to more developers and organizations - industries around the globe are poised to experience entirely new standards of speed, computing, and performance.

More Stories By Nikita Ivanov

Nikita Ivanov is founder and CEO of GridGain Systems, started in 2007 and funded by RTP Ventures and Almaz Capital. Nikita has led GridGain to develop advanced and distributed in-memory data processing technologies – the top Java in-memory computing platform starting every 10 seconds around the world today.

Nikita has over 20 years of experience in software application development, building HPC and middleware platforms, contributing to the efforts of other startups and notable companies including Adaptec, Visa and BEA Systems. Nikita was one of the pioneers in using Java technology for server side middleware development while working for one of Europe’s largest system integrators in 1996.

He is an active member of Java middleware community, contributor to the Java specification, and holds a Master’s degree in Electro Mechanics from Baltic State Technical University, Saint Petersburg, Russia.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
For years, we’ve relied too heavily on individual network functions or simplistic cloud controllers. However, they are no longer enough for today’s modern cloud data center. Businesses need a comprehensive platform architecture in order to deliver a complete networking suite for IoT environment based on OpenStack. In his session at @ThingsExpo, Dhiraj Sehgal from PLUMgrid will discuss what a holistic networking solution should really entail, and how to build a complete platform that is scalable, secure, agile and automated.
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focused on understanding how industrial data can create intelligence for industrial operations. Imagine ...
Hadoop as a Service (as offered by handful of niche vendors now) is a cloud computing solution that makes medium and large-scale data processing accessible, easy, fast and inexpensive. In his session at Big Data Expo, Kumar Ramamurthy, Vice President and Chief Technologist, EIM & Big Data, at Virtusa, will discuss how this is achieved by eliminating the operational challenges of running Hadoop, so one can focus on business growth. The fragmented Hadoop distribution world and various PaaS solutions that provide a Hadoop flavor either make choices for customers very flexible in the name of opti...
In the consumer IoT, everything is new, and the IT world of bits and bytes holds sway. But industrial and commercial realms encompass operational technology (OT) that has been around for 25 or 50 years. This grittier, pre-IP, more hands-on world has much to gain from Industrial IoT (IIoT) applications and principles. But adding sensors and wireless connectivity won’t work in environments that demand unwavering reliability and performance. In his session at @ThingsExpo, Ron Sege, CEO of Echelon, will discuss how as enterprise IT embraces other IoT-related technology trends, enterprises with i...
With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo in Silicon Valley. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be! Internet of @ThingsExpo, taking place Nov 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with 17th Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The Internet of Things (IoT) is the most profound change in personal an...
Wearable devices have come of age. The primary applications of wearables so far have been "the Quantified Self" or the tracking of one's fitness and health status. We propose the evolution of wearables into social and emotional communication devices. Our BE(tm) sensor uses light to visualize the skin conductance response. Our sensors are very inexpensive and can be massively distributed to audiences or groups of any size, in order to gauge reactions to performances, video, or any kind of presentation. In her session at @ThingsExpo, Jocelyn Scheirer, CEO & Founder of Bionolux, will discuss ho...
The true value of the Internet of Things (IoT) lies not just in the data, but through the services that protect the data, perform the analysis and present findings in a usable way. With many IoT elements rooted in traditional IT components, Big Data and IoT isn’t just a play for enterprise. In fact, the IoT presents SMBs with the prospect of launching entirely new activities and exploring innovative areas. CompTIA research identifies several areas where IoT is expected to have the greatest impact.
Every day we read jaw-dropping stats on the explosion of data. We allocate significant resources to harness and better understand it. We build businesses around it. But we’ve only just begun. For big payoffs in Big Data, CIOs are turning to cognitive computing. Cognitive computing’s ability to securely extract insights, understand natural language, and get smarter each time it’s used is the next, logical step for Big Data.
We are reaching the end of the beginning with WebRTC, and real systems using this technology have begun to appear. One challenge that faces every WebRTC deployment (in some form or another) is identity management. For example, if you have an existing service – possibly built on a variety of different PaaS/SaaS offerings – and you want to add real-time communications you are faced with a challenge relating to user management, authentication, authorization, and validation. Service providers will want to use their existing identities, but these will have credentials already that are (hopefully) i...
The 4th International Internet of @ThingsExpo, co-located with the 17th International Cloud Expo - to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA - announces that its Call for Papers is open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
17th Cloud Expo, taking place Nov 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterprises are using some form of XaaS – software, platform, and infrastructure as a service.
The Industrial Internet revolution is now underway, enabled by connected machines and billions of devices that communicate and collaborate. The massive amounts of Big Data requiring real-time analysis is flooding legacy IT systems and giving way to cloud environments that can handle the unpredictable workloads. Yet many barriers remain until we can fully realize the opportunities and benefits from the convergence of machines and devices with Big Data and the cloud, including interoperability, data security and privacy.
The Internet of Things is tied together with a thin strand that is known as time. Coincidentally, at the core of nearly all data analytics is a timestamp. When working with time series data there are a few core principles that everyone should consider, especially across datasets where time is the common boundary. In his session at Internet of @ThingsExpo, Jim Scott, Director of Enterprise Strategy & Architecture at MapR Technologies, discussed single-value, geo-spatial, and log time series data. By focusing on enterprise applications and the data center, he will use OpenTSDB as an example t...
The 17th International Cloud Expo has announced that its Call for Papers is open. 17th International Cloud Expo, to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, APM, APIs, Microservices, Security, Big Data, Internet of Things, DevOps and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal today!
SYS-CON Events announced today that MetraTech, now part of Ericsson, has been named “Silver Sponsor” of SYS-CON's 16th International Cloud Expo®, which will take place on June 9–11, 2015, at the Javits Center in New York, NY. Ericsson is the driving force behind the Networked Society- a world leader in communications infrastructure, software and services. Some 40% of the world’s mobile traffic runs through networks Ericsson has supplied, serving more than 2.5 billion subscribers.
Scott Jenson leads a project called The Physical Web within the Chrome team at Google. Project members are working to take the scalability and openness of the web and use it to talk to the exponentially exploding range of smart devices. Nearly every company today working on the IoT comes up with the same basic solution: use my server and you'll be fine. But if we really believe there will be trillions of these devices, that just can't scale. We need a system that is open a scalable and by using the URL as a basic building block, we open this up and get the same resilience that the web enjoys.
Converging digital disruptions is creating a major sea change - Cisco calls this the Internet of Everything (IoE). IoE is the network connection of People, Process, Data and Things, fueled by Cloud, Mobile, Social, Analytics and Security, and it represents a $19Trillion value-at-stake over the next 10 years. In her keynote at @ThingsExpo, Manjula Talreja, VP of Cisco Consulting Services, will discuss IoE and the enormous opportunities it provides to public and private firms alike. She will share what businesses must do to thrive in the IoE economy, citing examples from several industry sector...
Container frameworks, such as Docker, provide a variety of benefits, including density of deployment across infrastructure, convenience for application developers to push updates with low operational hand-holding, and a fairly well-defined deployment workflow that can be orchestrated. Container frameworks also enable a DevOps approach to application development by cleanly separating concerns between operations and development teams. But running multi-container, multi-server apps with containers is very hard. You have to learn five new and different technologies and best practices (libswarm, sy...
SYS-CON Events announced today that DragonGlass, an enterprise search platform, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. After eleven years of designing and building custom applications, OpenCrowd has launched DragonGlass, a cloud-based platform that enables the development of search-based applications. These are a new breed of applications that utilize a search index as their backbone for data retrieval. They can easily adapt to new data sets and provide access to both structured and unstruc...
All major researchers estimate there will be tens of billions devices - computers, smartphones, tablets, and sensors - connected to the Internet by 2020. This number will continue to grow at a rapid pace for the next several decades. With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo, June 9-11, 2015, at the Javits Center in New York City. Learn what is going on, contribute to the discussions, and ensure that your enterprise is as "IoT-Ready" as it can be