Virtualization and Cloud executives share their predictions for 2017. Read them in this 9th annual VMblog.com series exclusive.
Contributed by George Teixeira, President and CEO, DataCore Software
Parallel Processing Software Will be a 'Productivity Disrupter' and Game Changer in 2017
With so much computing power still sitting idle - despite
all of the incredible technology advancements that have occurred - in 2017, the
time is right for parallel processing software to go mainstream and unleash the
immense processing power of today's multicore systems to positively disrupt the
economic and productivity impact of what computing can do and where it can be
applied.
New
software innovations will make 2017 a breakout year for parallel processing. The
key is that the software has to become simple to use and non-disruptive to
applications to allow it to move from specialized use cases to general
application usage. By doing so, the impact of this will be massive because application
performance, enterprise workloads and greater consolidation densities on
virtual platforms and in cloud computing that have been stifled by the growing
gap between compute and I/O will no longer be held back. This will be realized
with new parallel I/O software technologies now available that are easy to use,
require no changes to the applications and are capable of fully leveraging the
power of multicores to dramatically increase productivity and overcome the I/O
bottleneck that has been holding back our industry; this is the catalyst of
change.
Parallel
processing software can now go beyond the realm of specialized uses such as HPC
and areas like genomics that have focused primarily on computation, and impact
the broader world of applications that require real-time responses and
interactions. This includes mainstream applications and storage that drive
business transactions, cloud computing, databases, data analytics, as well as
the interactive worlds of machine learning and the Internet of Things (IoT).
The real
driver of change is the economic and productivity disruption. Today, many new applications
such as analytics are not practical because they require hundreds if not
thousands of servers to get the job done; yet each server is becoming capable
of supporting hundreds of multi-threading computing cores, all available to
drive workloads that until now have sat there idle, waiting for work to do. We
are ushering in an era where one server will do the work of 10 -- or 100
servers -- of the past. This will be the result of parallel processing software
that unlocks the full utilization of multicores, leading to a revolution in
productivity and making a new world of applications affordable to mainstream IT
in 2017.
The
Impact on Real-time Analytics and Big Data Performance will be Profound
The
combination of faster response times and the multiplying impact on productivity
through parallelization will fuel the next step forward in ‘real-time'
analytics, big data and database performance. DataCore sees this as the next
step forward in 2017. Our background in parallel processing, real-time I/O and
software-defined storage has made our company uniquely well positioned to take
advantage of the next big challenge in a world that requires the rate and
amount of interactions and transactions to happen at a far faster pace with
much faster response times.
The ability
to do more work by doing it in parallel -- and to react quickly -- is the key.
DataCore sees itself as helping to drive the step function change needed to
make real-time analytics and big data performance practical and affordable. The
implications on productivity and business decision making based on insights
from data in areas such as financial, banking, retail, fraud detection,
healthcare, and genomics, as well as machine learning and Internet of Things
type applications, will be profound.
The Microsoft Impact Arrives: Azure Stack,
Hybrid Cloud, Windows and SQL Server 2016
The success
and growth of Microsoft's Azure Cloud has already become evident, however the
real impact is the larger strategy of how Microsoft has worked to reconcile the
world of on-premise and cloud computing. Microsoft was one of the first
cloud vendors to recognize that the world is not just public clouds but that it
will continue to be a mix of on-premise and cloud. Microsoft's
Azure Stack continues to advance in making it seamless to get the benefits of cloud-like
computing whether in the cloud or within a private cloud. It has become the
model for hybrid cloud computing. Likewise, Microsoft continues to further integrate
its Windows and server solutions to work more seamlessly with cloud
capabilities.
While
Windows and Azure get most of the attention, one of the most dramatic changes
at Microsoft has been how it has reinvented and transformed its database offerings
into a true big data and analytics platform for the future. It is time to take
another look at SQL Server 2016; it is far more powerful and capable, and now
deals with all types of data. As a platform, it is primed to work with Microsoft's
large eco-system of marketplace partners, including DataCore with its parallel
processing innovations, to redefine what is possible in the enterprise, the
cloud, and with big data performance and real-time analytic use cases for
traditional business applications, as well as new developing use cases in
machine learning, cognitive computing and the Internet of Things.
Storage has Transformed;
It's Servers + Software-Defined Infrastructure!
We are the midst of an inevitable and
increasing trend in which servers are defining what storage is. Escalating this
trend DataCore used parallel I/O software technologies to power off-the-shelf
multicore servers to drive the world's fastest storage systems in terms of
performance, lowest latencies and best price-performance. Traditional storage
systems can no longer keep up and are on the decline, and as a result, are
increasingly being replaced by commodity servers and software-defined
infrastructure solutions that can leverage their power to solve the growing
data storage problem. The storage function and associated data services are now
being driven by software and becoming another "application workload" running on
these cost-efficient server platforms, and this wave of flexible server-based
storage systems are already having a disruptive industry impact.
Marketed as server-SANs, virtual SANs,
web-scale, scale-out and hyper-converged systems, they are a collection of
standard off-the-shelf servers, flash cards and disk drives - but it is the
software that truly defines their value differentiation. Storage has become a
server game. Parallel processing software and the ability to leverage multicore
server technology is the major game-changer. In combination with software-defined
infrastructure, it will lead to a productivity revolution and further solidify "servers
as the new storage." For additional information, see the following report: http://wikibon.com/server-san-readies-for-enterprise-and-cloud-domination/
What's Beyond Flash?
Remember when flash was the next big thing? Now it's here.
What is the next step -- how do we go faster and do more with less? The answer
is obvious; if flash is now here and yet performance and productivity are still
an issue for many enterprise applications especially database use cases, then
we need to parallelize the I/O processing. Why? It multiplies what can be done as a result of
many compute engines working in parallel to process and remove bottlenecks and
queuing delays higher up in the stack, near the application, so we avoid as
much device level I/O as possible and drive performance and response times far
beyond any single device level optimization that flash/SSD alone can deliver.
The power of the ‘many' far exceed what only ‘one' can do - combining flash and
parallel I/O enables users to drive more applications faster, do more work and
open up applications and use cases that have been previously impossible to do.
Going Beyond Hyper-Convergence: Hyper-Productivity
is the Real Objective
As 2017
progresses, hyper-converged software will continue to grow in popularity
but to cement its success, users will need to be able take full advantage of
its productivity promise. The incredible power of parallel processing software will
enable users to take advantage of what their hardware and software can do (see
this video from ESG
as an example).
Hyper-converged
systems today are in essence a server plus a software-defined infrastructure,
but often they are severely restricted in terms of performance and use cases
and too often lack needed flexibility and a path for integration within the
larger IT environment (for instance not supporting fibre channel, which often
is key to enterprise and database connectivity). Powerful software-defined storage
technologies that can do parallel I/O effectively provide a higher level of flexibility
and leverage the power of multicore servers so fewer nodes are needed to get
the work done, making them more cost-effective. Likewise, the software can incorporate
existing flash and disk storage without creating additional silos; migrate and
manage data across the entire storage infrastructure; and effectively utilize
data stored in the cloud.
Data infrastructures including hyper-converged systems can
all benefit from these advances through advanced parallel I/O software
technologies that can dramatically increase their productivity by untapping the
power that lies within standard multicore servers. While hyper-converged has
become the buzzword of the day, let's remember the real objective is to achieve
the most productivity at the lowest cost, therefore better utilization of one's
storage and servers to drive applications is the key.
The Next Giant Leap
Forward - Leveraging the Multiplier Impact of Parallel Processing on
Productivity
This
combination of powerful software and servers will drive greater functionality,
more automation, and comprehensive services to productively manage and store
data across the entire data infrastructure. It will lead to a new era where the
benefits of multicore parallel processing can be applied universally. These
advances (which are already before us) are key to solving the problems caused
by slow I/O and inadequate response times that have been responsible for
holding back application workload performance and cost savings from
consolidation. The advances in multicore
processing, parallel processing software and software-defined infrastructure,
collectively, are fundamental to achieving the next giant leap forward in
business productivity.
##
About the Author
George Teixeira creates and executes the overall strategic direction and vision for DataCore Software.
Mr. Teixeira co-founded the company and has served as CEO and President
of DataCore Software since 1998. Prior to that time, Mr. Teixeira served
in a number of executive management positions including Worldwide VP of
Marketing and GM of the Product Business Group at Encore Computer
Corporation, where he also played a major role as the team leader of OEM
marketing and sales to Amdahl, IBM, and DEC. His work culminated in the
$185 million sale of Encore's storage control business to Sun
Microsystems in 1997. He also held a number of senior management
positions at the Computer Systems Division of Gould Electronics.