Virtualization Technology News and Information
White Papers
RSS
White Papers Search Results
Showing 17 - 29 of 29 white papers, page 2 of 2.
How Data Temperature Drives Data Placement Decisions and What to Do About It
In this white paper, learn (1) how the relative proportion of hot, warm, and cooler data changes over time, (2) new machine learning (ML) techniques that sense the cooling temperature of data throughout its half-life, and (3) the role of artificial intelligence (AI) in migrating data to the most cost-effective tier.

The emphasis on fast flash technology concentrates much attention on hot, frequently accessed data. However, budget pressures preclude consuming such premium-priced capacity when the access frequency diminishes. Yet many organizations do just that, unable to migrate effectively to lower cost secondary storage on a regular basis.
In this white paper, explore:

•    How the relative proportion of hot, warm, and cooler data changes over time
•    New machine learning (ML) techniques that sense the cooling temperature of data throughout its half-life
•    The role of artificial intelligence (AI) in migrating data to the most cost-effective tier.

Frost & Sullivan Best Practices in Storage Management 2019
This analyst report examines the differentiation of Tintri Global Center in storage management, resulting in receiving this award for product leadership in 2019. Businesses need planning tools to help them handle data growth, new workloads, decommissioning old hardware, and more. The Tintri platform provides both the daily management tasks required to streamline storage management, as well as the forward-looking insights to help businesses plan accordingly. Today Tintri technology is different
This analyst report examines the differentiation of Tintri Global Center in storage management, resulting in receiving this award for product leadership in 2019.  Businesses need planning tools to help them handle data growth, new workloads, decommissioning old hardware, and more. The Tintri platform provides both the daily management tasks required to streamline storage management, as well as the forward-looking insights to help businesses plan accordingly.  Today Tintri technology is differentiated by its level of abstraction—the ability to take every action on individual virtual machines.  Hypervisor administrators and staff members associated with architecting, deploying and managing virtual machines will want to dig into this document to understand how Tintri can save them the majority of their management effort and greatly reduce operating expense.
NexentaStor Adds NAS Capabilities to HCI or Block Storage Systems
Companies are adopting new enterprise architecture options (virtualized environments, block-only storage, HCI) to improve performance & simplify deployments. However, over time there is a need to expand the workloads- challenges arise due to not having file-based storage services. This white paper provides insight on how Nexenta by DDN enables these modern architectures to flourish with simplicity to help you grow your business by providing complementary NAS & hybrid public cloud capabilities.
Companies are adopting new enterprise architecture options (virtualized environments, block-only storage, HCI) to improve performance & simplify deployments. However, over time there is a need to expand the workloads- challenges arise due to not having file-based storage services. This white paper provides insight on how Nexenta by DDN enables these modern architectures to flourish with simplicity to help you grow your business by providing complementary NAS & hybrid public cloud capabilities.
How to Develop a Multi-cloud Management Strategy
Increasingly, organizations are looking to move workloads into the cloud. The goal may be to leverage cloud resources for Dev/Test, or they may want to “lift and shift” an application to the cloud and run it natively. In order to enable these various cloud options, it is critical that organizations develop a multi-cloud data management strategy.

The primary goal of a multi-cloud data management strategy is to supply data, either via copying or moving data to the various multi-cloud use cases. A key enabler of this movement is the data management software applications. In theory, data protection applications can perform both of the copy and move functions. A key consideration is how the multi-cloud data management experience is unified. In most cases, data protection applications ignore the user experience of each cloud and use their proprietary interface as the unifying entity, which increases complexity.

There are a variety of reasons organizations may want to leverage multiple clouds. The first use case is to use public cloud storage as a backup mirror to an on-premises data protection process. Using public cloud storage as a backup mirror enables the organization to automatically off-site data. It also sets up many of the more advanced use cases.

Another use case is using the cloud for disaster recovery.

Another use case is “Lift and Shift,” which means the organization wants to run the application in the cloud natively. Initial steps in the “lift and shift” use case are similar to Dev/Test, but now the workload is storing unique data in the cloud.

Multi-cloud is a reality now for most organizations and managing the movement of data between these clouds is critical.

The SysAdmin Guide to Azure Infrastructure as a Service
If you're used to on-premises infrastructures, cloud platforms can seem daunting. But it doesn't need to be. This eBook written by the veteran IT consultant and trainer Paul Schnackenburg, covers all aspects of setting up and maintaining a high-performing Azure IaaS environment, including: • VM sizing and deployment • Migration • Storage and networking • Security and identity • Infrastructure as code and more!

The cloud computing era is well and truly upon us, and knowing how to take advantage of the benefits of this computing paradigm while maintaining security, manageability, and cost control are vital skills for any IT professional in 2020 and beyond. And its importance is only getting greater.

In this eBook, we’re going to focus on Infrastructure as a Service (IaaS) on Microsoft’s Azure platform - learning how to create VMs, size them correctly, manage storage, networking, and security, along with backup best practices. You’ll also learn how to operate groups of VMs, deploy resources based on templates, managing security and automate your infrastructure. If you currently have VMs in your own datacenter and are looking to migrate to Azure, we’ll also teach you that.

If you’re new to the cloud (or have experience with AWS/GCP but not Azure), this book will cover the basics as well as more advanced skills. Given how fast things change in the cloud, we’ll cover the why (as well as the how) so that as features and interfaces are updated, you’ll have the theoretical knowledge to effectively adapt and know how to proceed.

You’ll benefit most from this book if you actively follow along with the tutorials. We will be going through terms and definitions as we go – learning by doing has always been my preferred way of education. If you don’t have access to an Azure subscription, you can sign up for a free trial with Microsoft. This will give you 30 days 6 to use $200 USD worth of Azure resources, along with 12 months of free resources. Note that most of these “12 months” services aren’t related to IaaS VMs (apart from a few SSD based virtual disks and a small VM that you can run for 750 hours a month) so be sure to get everything covered on the IaaS side before your trial expires. There are also another 25 services that have free tiers “forever”.

Now you know what’s in store, let’s get started!

Evaluator Group Report on Liqid Composable Infrastructure
In this report from Eric Slack, Senior Analyst at the Evaluator Group, learn how Liqid’s software-defined platform delivers comprehensive, multi-fabric composable infrastructure for the industry’s widest array of data center resources.
Composable Infrastructures direct-connect compute and storage resources dynamically—using virtualized networking techniques controlled by software. Instead of physically constructing a server with specific internal devices (storage, NICs, GPUs or FPGAs), or cabling the appropriate device chassis to a server, composable enables the virtual connection of these resources at the device level as needed, when needed.

Download this report from Eric Slack, Senior Analyst at the Evaluator Group to learn how Liqid’s software-defined platform delivers comprehensive, multi-fabric composable infrastructure for the industry’s widest array of data center resources.
LQD4500 Gen4x16 NVMe SSD Performance Report
The LQD4500 is the World’s Fastest SSD. The Liqid Element LQD4500 PCIe Add-in-Card (AIC), aka the “Honey Badger" for its fierce, lightning fast data speeds, delivers Gen-4 PCIe performance with up to 4m IOPS, 24 GB/s throughput, ultra-low transactional latency of just 20 µs, in capacities up to 32TB. What does all that mean in real-world testing? Download the full LQD4500 performance report and learn how the ultra-high performance and density Gen-4 AIC can supercharge data for your most demand
The LQD4500 is the World’s Fastest SSD.

The Liqid Element LQD4500 PCIe Add-in-Card (AIC), aka the “Honey Badger" for its fierce, lightning fast data speeds, delivers Gen-4 PCIe performance with up to 4m IOPS, 24 GB/s throughput, ultra-low transactional latency of just 20 µs, in capacities up to 32TB.

The document will contain test results and performance measurements for the Liqid LQD4500 Gen4x16 NVMe SSD. The performance test reports include Sequential, Random, and Latency measurement on the LQD4500 high performance storage device. The following data has been measured in a Linux OS environment and results are taken per the SNIA enterprise performance test specification standards. The below results are steady state after sufficient device preconditioning.

Download the full LQD4500 performance report and learn how the ultra-high performance and density Gen-4 AIC can supercharge data for your most demanding applications.
ESG - DataCore vFilO: Visibility and Control of Unstructured Data for the Modern, Digital Business
Organizations that want to succeed in the digital economy must contend with the cost and complexity introduced by the conventional segregation of multiple file system silos and separate object storage repositories. Fortunately, they can look to DataCore vFilO software for help. DataCore employs innovative techniques to combine diverse unstructured data resources to achieve unprecedented visibility, control, and flexibility.
DataCore’s new vFilO software shares important traits with its existing SANsymphony software-defined block storage platform. Both technologies are certainly enterprise class (highly agile, available, and performant). But each solution exhibits those traits in its own manner, taking the varying requirements for block, file, and object data into account. That’s important at a time when a lot of companies are maintaining hundreds to thousands of terabytes of unstructured data spread across many file servers, other NAS devices, and object storage repositories both onsite and in the cloud. The addition of vFilO to its product portfolio will allow DataCore to position itself in a different, even more compelling way now. DataCore is able to offer a “one-two punch”—namely, one of the best block storage SDS solutions in SANsymphony, and now one of the best next-generation SDS solutions for file and object data in vFilO. Together, vFilO and SANsymphony will put DataCore in a really strong position to support any IT organization looking for better ways to overcome end-users’ file-sharing/access difficulties, keep hardware costs low … and maximize the value of corporate data to achieve success in a digital age.
ESG Showcase - DataCore vFilO: NAS Consolidation Means Freedom from Data Silos
File and object data are valuable tools that help organizations gain market insights, improve operations, and fuel revenue growth. However, success in utilizing all of that data depends on consolidating data silos. Replacing an existing infrastructure is often expensive and impractical, but DataCore vFilO software offers an intelligent, powerful option—an alternative, economically appealing way to consolidate and abstract existing storage into a single, efficient, capable ecosystem of readily-se

Companies have NAS systems all over the place—hardware-centric devices that make data difficult to migrate and leverage to support the business. It’s natural that companies would desire to consolidate those systems, and vFilO is a technology that could prove to be quite useful as an assimilation tool. Best of all, there’s no need to replace everything. A business can modernize its IT environment and finally achieve a unified view, plus gain more control and efficiency via the new “data layer” sitting on top of the hardware. When those old silos finally disappear, employees will discover they can find whatever information they need by examining and searching what appears to be one big catalog for a large pool of resources.

And for IT, the capacity-balancing capability should have especially strong appeal. With it, file and object data can shuffle around and be balanced for efficiency without IT or anyone needing to deal with silos. Today, too many organizations still perform capacity balancing work manually—putting some files on a different NAS system because the first one started running out of room. It’s time for those days to end. DataCore, with its 20-year history offering SANsymphony, is a vendor in a great position to deliver this new type of solution, one that essentially virtualizes NAS and object systems and even includes keyword search capabilities to help companies use their data to become stronger, more competitive, and more profitable.

ESG Showcase - Time to Better Leverage Your File and Object Data
The need to handle unstructured data according to business relevance is becoming urgent. Modern datarelated demands have begun surpassing what traditional file and object storage architectures can achieve. Businesses today need an unstructured data storage environment that lets end-users easily and flexibly access the benefits of the file- and object-based information they need to do their jobs.
It’s time for a “Nirvana” between file and object: The marketplace has long been debating which of the two unstructured data type models is the “right one.” But data is data. If you remove the traditional constraints and challenges associated with those two particular technologies, it becomes possible to imagine a type of data service that supports both. That kind of solution would give an organization the benefits of a global multi-site namespace (i.e., object style), including the ability to address a piece of data regardless of where it is in a folder hierarchy, or what device it’s stored on. And it would offer the rich metadata that object storage is known for. This new kind of solution would also deliver the performance and ease of use that file systems are known for. DataCore says it is providing such a solution, called vFilO. To vFilO, data is just data, where metadata is decoupled but tightly aligned with the data, and data’s location is independent of data access needs or metadata location. It bridges both worlds by providing a system that has all the benefits of file and object—without any of the limitations. It is a next-generation system built on a different paradigm, promising to bridge the world of file and object. It just might be the answer IT departments need. No longer required to pick between two separate systems, IT organizations can find great value in a consolidated, global, unified storage system that can deliver on the business’s needs for unstructured data.
The Time is Now for File Virtualization
DataCore’s vFilO is a distributed files and object storage virtualization solution that can consume storage from a variety of providers, including NFS or SMB file servers, most NAS systems, and S3 Object Storage systems, including S3-based public cloud providers. Once vFilO integrates these various storage systems into its environment, it presents users with a logical file system and abstracts it from the actual physical location of data.

DataCore vFilO is a top-tier file virtualization solution. Not only can it serve as a global file system, IT can also add new NAS systems or file servers to the environment without having to remap users of the new hardware. vFilO supports live migration of data between the storage systems it has assimilated and leverages the capabilities of the global file system and the software’s policy-driven data management to move older data to less expensive storage automatically; either high capacity NAS or an object storage system. vFilO also transparently moves data from NFS/SMB to object storage. If the user needs access to this data in the future, they access it like they always have. To them, the data has not moved.

The ROI of File virtualization is powerful, but it has struggled to gain adoption in the data center. File Virtualization needs to be explained, and explaining it takes time. vFilO more than meets the requirements to qualify as a top tier file virtualization solution. DataCore has the advantage of over 10,000 customers that are much more likely to be receptive to the concept since they have already embraced block storage virtualization with SANSymphony. Building on its customer base as a beachhead, DataCore can then expand File Virtualization’s reach to new customers, who, because of the changing state of unstructured data, may finally be receptive to the concept. At the same time, these new file virtualization customers may be amenable to virtualizing block storage, and it may open up new doors for SANSymphony.

IDC: DataCore SDS: Enabling Speed of Business and Innovation with Next-Gen Building Blocks
DataCore solutions include and combine block, file, object, and HCI software offerings that enable the creation of a unified storage system which integrates more functionalities such as data protection, replication, and storage/device management to eliminate complexity. They also converge primary and secondary storage environments to give a unified view, predictive analytics and actionable insights. DataCore’s newly engineered SDS architecture make it a key player in the modern SDS solutions spa
The enterprise IT infrastructure market is undergoing a once-in-a-generation change due to ongoing digital transformation initiatives and the onslaught of applications and data. The need for speed, agility, and efficiency is pushing demand for modern datacenter technologies that can lower costs while providing new levels of scale, quality, and operational efficiency. This has driven strong demand for next-generation solutions such as software-defined storage/networking/compute, public cloud infrastructure as a service, flash-based storage systems, and hyperconverged infrastructure. Each of these solutions offers enterprise IT departments a way to rethink how they deploy, manage, consume, and refresh IT infrastructure. These solutions represent modern infrastructure that can deliver the performance and agility required for both existing virtualized workloads and next-generation applications — applications that are cloud-native, highly dynamic, and built using containers and microservices architectures. As we enter the next phase of datacenter modernization, businesses need to leverage newer capabilities enabled by software-defined storage that help them eliminate management complexities, overcome data fragmentation and growth challenges, and become a data-driven organization to propel innovation. As enterprises embark on their core datacenter modernization initiatives with compelling technologies, they should evaluate enterprise-grade solutions that redefine storage and data architectures designed for the demands of the digital-native economy. Digital transformation is a technology-based business strategy that is becoming increasingly imperative for success. However, unless infrastructure provisioning evolves to suit new application requirements, IT will not be viewed as a business enabler. IDC believes that those organizations that do not leverage proven technologies such as SDS to evolve their datacenters truly risk losing their competitive edge.
IDC: SaaS Backup and Recovery: Simplified Data Protection Without Compromise
Although the majority of organizations have a "cloud first" strategy, most also continue to manage onsite applications and the backup infrastructure associated with them. However, many are moving away from backup specialists and instead are leaving the task to virtual infrastructure administrators or other IT generalists. Metallic represents Commvault's direct entry into one of the fastest-growing segments of the data protection market. Its hallmarks are simplicity and flexibility of deployment

Metallic is a new SaaS backup and recovery solution based on Commvault's data protection software suite, proven in the marketplace for more than 20 years. It is designed specifically for the needs of medium-scale enterprises but is architected to grow with them based on data growth, user growth, or other requirements. Metallic initially offers either monthly or annual subscriptions through reseller partners; it will be available through cloud service providers and managed service providers over time. The initial workload use cases for Metallic include virtual machine (VM), SQL Server, file server, MS Office 365, and endpoint device recovery support; the company expects to add more use cases and supported workloads as the solution evolves.

Metallic is designed to offer flexibility as one of the service's hallmarks. Aspects of this include:

  • On-demand infrastructure: Metallic manages the cloud-based infrastructure components and software for the backup environment, though the customer will still manage any of its own on- premise infrastructure. This environment will support on-premise, cloud, and hybrid workloads. IT organizations are relieved of the daily task of managing the infrastructure components and do not have to worry about upgrades, OS or firmware updates and the like, for the cloud infrastructure, so people can repurpose that time saved toward other activities.
  • Metallic offers preconfigured plans designed to have users up and running in approximately 15 minutes, eliminating the need for a proof-of-concept test. These preconfigured systems have Commvault best practices built into the design, or organizations can configure their own.
  • Partner-delivered services: Metallic plans to go to market with resellers that can offer a range of services on top of the basic solution's capabilities. These services will vary by provider and will give users a variety of choices when selecting a provider to match the services offered with the organization's needs.
  • "Bring your own storage": Among the flexible options of Metallic, including VM and file or SQL database use cases, users can deploy their own storage, either on-premise or in the cloud, while utilizing the backup/recovery services of Metallic. The company refers to this option as "SaaS Plus."
top25