Virtualization Technology News and Information
White Papers
RSS
White Papers Search Results
Showing 49 - 64 of 79 white papers, page 4 of 5.
The Case for Converged Application & Infrastructure Performance Monitoring
Read this white paper and learn how you can combine and correlate performance insights from the application (code, SQL, logs) and the underlying hardware infrastructure (server, network, virtualization, storage, etc.)

One of the toughest problems facing enterprise IT teams today is troubleshooting slow applications. When a user complains of slowness in application access, all hell breaks loose, and the blame game begins: app owners, developers and IT ops teams enter into endless war room sessions to figure out what went wrong and where. Have you also been in this situation before?

Read this white paper by Larry Dragich, and learn how you can combine and correlate performance insights from the application (code, SQL, logs) and the underlying hardware infrastructure (server, network, virtualization, storage, etc.) in order to:

  • Proactively detect user experience issues before your customers are impacted
  • Trace business transactions and isolate the cause of application slowness
  • Get code-level visibility to identify inefficient application code and slow database queries
  • Automatically map application dependencies within the infrastructure to pinpoint the root cause of the problem
Achieve centralized visibility of all your applications and infrastructure and easily diagnose the root cause of performance slowdowns.
PrinterLogic and IGEL Enable Healthcare Organizations to Deliver Better Patient Outcomes
Healthcare professionals need to print effortlessly and reliably to nearby or appropriate printers within virtual environments, and PrinterLogic and IGEL can help make that an easy, reliable process—all while efficiently maintaining the protection of confidential patient information.

Many organizations have turned to virtualizing user endpoints to help reduce capital and operational expenses while increasing security. This is especially true within healthcare, where hospitals, clinics, and urgent care centers seek to offer the best possible patient outcomes while adhering to a variety of mandated patient security and information privacy requirements.

With the movement of desktops and applications into the secure data center or cloud, the need for reliable printing of documents, some very sensitive in nature, remains a constant that can be challenging when desktops are virtual but the printing process remains physical. Directing print jobs to the correct printer with the correct physical access rights in the correct location while ensuring compliance with key healthcare mandates like the General Data Protection Regulation (GDPR) and the Healthcare Insurance Portability and Accountability Act (HIPAA) is critical.

Healthcare IT needs to keep pace with these requirements and the ongoing printing demands of healthcare. Medical professionals need to print effortlessly and reliably to nearby or appropriate printers within virtual environments, and PrinterLogic and IGEL can help make that an easy, reliable process—all while efficiently maintaining the protection of confidential patient information. By combining PrinterLogic’s enterprise print management software with centrally managed direct IP printing and IGEL’s software-defined thin client endpoint management, healthcare organizations can:

  • Reduce capital and operational costs
  • Support virtual desktop infrastructure (VDI) and electronic medical records (EMR) systems effectively
  • Centralize and simplify print management
  • Add an essential layer of security from the target printer all the way to the network edge
UD Pocket Saves the Day After Malware Cripple’s Hospital’s Mission-Critical PCs
IGEL Platinum Partner A2U had endpoints within the healthcare organization’s finance department up and running within a few hours following the potentially crippling cyberattack, thanks to the innovative micro thin client.

A2U, an IGEL Platinum Partner, recently experienced a situation where one of its large, regional healthcare clients was hit by a cyberattack. “Essentially, malware entered the client’s network via a computer and began replicating like wildfire,” recalls A2U Vice President of Sales, Robert Hammond.

During the cyberattack, a few hundred of the hospital’s PCs were affected. Among those were 30 endpoints within the finance department that the healthcare organization deemed mission critical due to the volume of daily transactions between patients, insurance companies, and state and county agencies for services rendered. “It was very painful from a business standpoint not to be able to conduct billing and receiving, not to mention payroll,” said Hammond.

Prior to this particular incident, A2U had received demo units of the IGEL UD Pocket, a revolutionary micro thin client that can transform x86-compatible PCs and laptops into IGEL OS-powered desktops.

“We had been having a discussion with this client about re-imaging their PCs, but their primary concern was maintaining the integrity of the data that was already on the hardware,” continued Hammond. “HIPAA and other regulations meant that they needed to preserve the data and keep it secure, and we thought that the IGEL UD Pocket could be the answer to this problem. We didn’t see why it wouldn’t work, but we needed to test our theory.”

When the malware attack hit, that opportunity came sooner, rather than later for A2U. “We plugged the UD Pocket into one of the affected machines and were able to bypass the local hard drive, installing the Linux-based IGEL OS on the system without impacting existing data,” said Hammond. “It was like we had created a ‘Linux bubble’ that protected the machine, yet created an environment that allowed end users to quickly return to productivity.”

Working with the hospital’s IT team, it only took a few hours for A2U to get the entire finance department back online. “They were able to start billing the very next day,” added Hammond.

Top 10 strategies to manage cost and continuously optimize AWS
The great cloud migration has upended decades of established architecture patterns, operating principles, and governance models. Without any controls in place, cloud spend inevitably rises faster than anticipated and often gets overlooked until it gets out of control. With its granular and short-term billing cycles, the cloud requires a degree of financial discipline that is unfamiliar to most traditional IT departments. Faced with having to provide short-term forecasts and justify them agains

The public cloud has unleashed an unprecedented wave of creativity and agility for the modern enterprise. A great cloud migration has upended decades of established architecture patterns, operating principles, and governance models. However, without any replacement for these traditional controls in place, cloud spend inevitably rises faster than anticipated. If not addressed early in the cycle, this is often overlooked until it gets out of control.

Over the course of a few decades, we have created a well-established model of IT spending; to arrive at economies of scale, procurement is centralized and typically happens at three- to five-year intervals, with all internal customers forecasting and pooling their needs. This created a natural tendency for individual project owners to overprovision resources as insurance against unexpected demand. As a result, the corporate data center today is where the two famous laws of technology meet:

  • Moore’s Law ensures that capacity increases to meet demand
  • Parkinson’s Law ensures that demand rises to meet capacity

With its granular and short-term billing cycles, the cloud requires a degree of financial discipline that is unfamiliar to most traditional IT departments. Faced with having to provide short-term forecasts and justify them against actual spend, they need to evolve their governance models to support these new patterns.

Having a well thought out AWS strategy is crucial to your long-term cloud gains. Taking the time to understand and pick the right instances for your apps is well worth the time and effort as it will directly impact your AWS pricing and bill.

We hope these ten strategies help inform and support you as you navigate the sometimes-turbulent waters of cloud transition. They are here for you to consult and rely on as best practices and cost-saving opportunities.

Given the virtually uncountable number of combinations, we have tried to identify the most practical and reliable ways to optimize your deployment at all stages and empower your end-users while insulating them from temptations, assumptions and habits that can cost you some unpleasant surprises when the bill arrives.

Gartner Market Guide for IT Infrastructure Monitoring Tools
With the onset of more modular and cloud-centric architectures, many organizations with disparate monitoring tools are reassessing their monitoring landscape. According to Gartner, hybrid IT (especially with IaaS subscription) enterprises must adopt more holistic IT infrastructure monitoring tools (ITIM) to gain visibility into their IT landscapes.

With the onset of more modular and cloud-centric architectures, many organizations with disparate monitoring tools are reassessing their monitoring landscape. According to Gartner, hybrid IT (especially with IaaS subscription) enterprises must adopt more holistic IT infrastructure monitoring tools (ITIM) to gain visibility into their IT landscapes.

The guide provides insight into the IT infrastructure monitoring tool market and providers as well as key findings and recommendations.

Get the 2018 Gartner Market Guide for IT Infrastructure Monitoring Tools to see:

  • The ITIM market definition, direction and analysis
  • A list of representative ITIM vendors
  • Recommendations for adoption of ITIM platforms

Key Findings Include:

  • ITIM tools are helping organizations simplify and unify monitoring across domains within a single tool, eliminating the problems of multitool integration.
  • ITIM tools are allowing infrastructure and operations (I&O) leaders to scale across hybrid infrastructures and emerging architectures (such as containers and microservices).
  • Metrics and data acquired by ITIM tools are being used to derive context enabling visibility for non-IT teams (for example, line of business [LOB] and app owners) to help achieve optimization targets.
Overcome the Data Protection Dilemma - Vembu
Selecting a high-priced legacy backup application that protects an entire IT environment or adopting a new age solution that focuses on protecting a particular area of an environment is a dilemma for every IT professional. Read this whitepaper to overcome the data protection dilemma with Vembu.
IT professionals face a dilemma while selecting a backup solution for their environment. Selecting a legacy application that protects their entire environment means that they have to tolerate high pricing and live with software that does not fully exploit the capabilities of modern IT environment.

On the other hand, they can adopt solutions that focus on a particular area of an IT environment and limited just to that environment. These solutions have a relatively small customer base which means the solution has not been vetted as the legacy applications. Vembu is a next-generation company that provides the capabilities of the new class of backup solutions while at the same time providing completeness of platform coverage, similar to legacy applications.
Understanding Windows Server Hyper-V Cluster Configuration, Performance and Security
The Windows Server Hyper-V Clusters are definitely an important option when trying to implement High Availability to critical workloads of a business. Guidelines on how to get started with things like deployment, network configurations to some of the industries best practices on performance, security, and storage management are something that any IT admin would not want to miss. Get started with reading this white paper that discusses the same through scenarios on a production field and helps yo
How do you increase the uptime of your critical workloads? How do you start setting up a Hyper-V Cluster in your organization? What are the Hyper-V design and networking configuration best practices? These are some of the questions you may have when you have large environments with many Hyper-V deployments. It is very essential for IT administrators to build disaster-ready Hyper-V Clusters rather than thinking about troubleshooting them in their production workloads. This whitepaper will help you in deploying a Hyper-V Cluster in your infrastructure by providing step-by-step configuration and consideration guides focussing on optimizing the performance and security of your setup.
VMware vSphere 6.7 Update 1 Upgrade and Security Configuration
Most businesses hugely invest in tackling the security vulnerabilities of their data centers. VMware vSphere 6.7 Upgrade 1 tackles it head-on with its functionalities that aligns with both the legacy and the modern technology capabilities. Read this white paper to know how you can maximize the security posture of vSphere workloads on production environments.
Security is a top concern when it comes to addressing data protection complexities for business-critical systems. VMware vSphere 6.7 Upgrade 1 can be the right fit for your data centers when it comes to resolving security vulnerabilities thereby helping you to take your IT infrastructure to the next level. While there are features that align with the legacy security standards, there are some of the best newly announced functionalities in vSphere 6.7 like Virtual TPM 2.0 and virtualization-based security capabilities that will help you enhance your current security measures for your production workloads. Get started with reading this white paper to know more on how you can implement a solution of this kind in your data centers.
vSphere Troubleshooting Guide
Troubleshooting complex virtualization technology is something all VMware users will have to face at some point. It requires an understanding of how various components fit together and finding a place to start is not easy. Thankfully, VMware vExpert Ryan Birk is here to help with this eBook preparing you for any problems you may encounter along the way.

This eBook explains how to identify problems with vSphere and how to solve them. Before we begin, we need to start off with an introduction to a few things that will make life easier. We’ll start with a troubleshooting methodology and how to gather logs. After that, we’ll break this eBook into the following sections: Installation, Virtual Machines, Networking, Storage, vCenter/ESXi and Clustering.

ESXi and vSphere problems arise from many different places, but they generally fall into one of these categories: Hardware issues, Resource contention, Network attacks, Software bugs, and Configuration problems.

A typical troubleshooting process contains several tasks: 1. Define the problem and gather information. 2. Identify what is causing the problem. 3. Fix the problem, implement a fix.

One of the first things you should try to do when experiencing a problem with a host, is try to reproduce the issue. If you can find a way to reproduce it, you have a great way to validate that the issue is resolved when you do fix it. It can be helpful as well to take a benchmark of your systems before they are implemented into a production environment. If you know HOW they should be running, it’s easier to pinpoint a problem.

You should decide if it’s best to work from a “Top Down” or “Bottom Up” approach to determine the root cause. Guest OS Level issues typically cause a large amount of problems. Let’s face it, some of the applications we use are not perfect. They get the job done but they utilize a lot of memory doing it.

In terms of virtual machine level issues, is it possible that you could have a limit or share value that’s misconfigured? At the ESXi Host Level, you could need additional resources. It’s hard to believe sometimes, but you might need another host to help with load!

Once you have identified the root cause, you should assess the impact of the problem on your day to day operations. When and what type of fix should you implement? A short-term one or a long-term solution? Assess the impact of your solution on daily operations. Short-term solution: Implement a quick workaround. Long-term solution: Reconfiguration of a virtual machine or host.

Now that the basics have been covered, download the eBook to discover how to put this theory into practice!

Mastering vSphere – Best Practices, Optimizing Configurations & More
Do you regularly work with vSphere? If so, this free eBook is for you. Learn how to leverage best practices for the most popular features contained within the vSphere platform and boost your productivity using tips and tricks learnt direct from an experienced VMware trainer and highly qualified professional. In this eBook, vExpert Ryan Birk shows you how to master: Advanced Deployment Scenarios using Auto-Deploy Shared Storage Performance Monitoring and Troubleshooting Host Network configurati

If you’re here to gather some of the best practices surrounding vSphere, you’ve come to the right place! Mastering vSphere: Best Practices, Optimizing Configurations & More, the free eBook authored by me, Ryan Birk, is the product of many years working with vSphere as well as teaching others in a professional capacity. In my extensive career as a VMware consultant and teacher (I’m a VMware Certified Instructor) I have worked with people of all competence levels and been asked hundreds - if not thousands - of questions on vSphere. I was approached to write this eBook to put that experience to use to help people currently working with vSphere step up their game and reach that next level. As such, this eBook assumes readers already have a basic understanding of vSphere and will cover the best practices for four key aspects of any vSphere environment.

The best practices covered here will focus largely on management and configuration solutions so should remain relevant for quite some time. However, with that said, things are constantly changing in IT, so I would always recommend obtaining the most up-to-date information from VMware KBs and official documentation especially regarding specific versions of tools and software updates. This eBook is divided into several sections, and although I would advise reading the whole eBook as most elements relate to others, you might want to just focus on a certain area you’re having trouble with. If so, jump to the section you want read about.

Before we begin, I want to note that in a VMware environment, it’s always best to try to keep things simple. Far too often I have seen environments be thrown off the tracks by trying to do too much at once. I try to live by the mentality of “keeping your environment boring” – in other words, keeping your host configurations the same, storage configurations the same and network configurations the same. I don’t mean duplicate IP addresses, but the hosts need identical port groups, access to the same storage networks, etc. Consistency is the name of the game and is key to solving unexpected problems down the line. Furthermore, it enables smooth scalability - when you move from a single host configuration to a cluster configuration, having the same configurations will make live migrations and high availability far easier to configure without having to significantly re-work the entire infrastructure. Now the scene has been set, let’s get started!

Forrester: Monitoring Containerized Microservices - Elevate Your Metrics
As enterprises continue to rapidly adopt containerized microservices, infrastructure and operations (I&O) teams need to address the growing complexities of monitoring these highly dynamic and distributed applications. The scale of these environments can pose tremendous monitoring challenges. This report will guide I&O leaders in what to consider when developing their technology and metric strategies for monitoring microservices and container-based applications.
As enterprises continue to rapidly adopt containerized microservices, infrastructure and operations (I&O) teams need to address the growing complexities of monitoring these highly dynamic and distributed applications. The scale of these environments can pose tremendous monitoring challenges. This report will guide I&O leaders in what to consider when developing their technology and metric strategies for monitoring microservices and container-based applications.
How to Get the Most Out of Windows Admin Center
Windows Admin Center is the future of Windows and Windows Server management. Are you using it to its full potential? In this free eBook, Microsoft Cloud and Datacenter Management MVP, Eric Siron, has put together a 70+ page guide on what Windows Admin Center brings to the table, how to get started, and how to squeeze as much value out of this incredible free management tool from Microsoft. This eBook covers: - Installation - Getting Started - Full UI Analysis - Security - Managing Extensions

Each version of Windows and Windows Server showcases new technologies. The advent of PowerShell marked a substantial step forward in managing those features. However, the built-in graphical Windows management tools have largely stagnated - the same basic Microsoft Management Console (MMC) interfaces had remained since Windows Server 2000. Microsoft tried out multiple overhauls over the years to the built-in Server Manager console but gained little traction. Until Windows Admin Center.

WHAT IS WINDOWS ADMIN CENTER?
Windows Admin Center (WAC) represents a modern turn in Windows and Windows Server system management. From its home page, you establish a list of the networked Windows and Windows Server computers to manage. From there, you can connect to an individual system to control components such as hardware drivers. You can also use it to manage Windows roles, such as Hyper-V.

On the front-end, Windows Admin Center is presented through a sleek HTML 5 web interface. On the back-end, it leverages PowerShell extensively to control the systems within your network. The entire package runs on a single system, so you don’t need a complicated infrastructure to support it. In fact, you can run it locally on your Windows 10 workstation if you want. If you require more resiliency, you can run Windows Admin Center as a role on a Microsoft Failover Cluster.

WHY WOULD I USE WINDOWS ADMIN CENTER?
In the modern era of Windows management, we have shifted to a greater reliance on industrial-strength tools like PowerShell and Desired State Configuration. However, we still have servers that require individualized attention and infrequently utilized resources. WAC gives you a one-stop hub for dropping in on any system at any time and work with almost any of its facets.

ABOUT THIS EBOOK
This eBook has been written by Microsoft Cloud & Datacenter Management MVP Eric Siron. Eric has worked in IT since 1998, designing, deploying, and maintaining server, desktop, network, and storage systems. He has provided all levels of support for businesses ranging from single-user through enterprises with thousands of seats. He has achieved numerous Microsoft certifications and was a Microsoft Certified Trainer for four years. Eric is also a seasoned technology blogger and has amassed a significant following through his top-class work on the Altaro Hyper-V Dojo.

Reducing Data Center Infrastructure Costs with Software-Defined Storage
Download this white paper to learn how software-defined storage can help reduce data center infrastructure costs, including guidelines to help you structure your TCO analysis comparison.

With a software-based approach, IT organizations see a better return on their storage investment. DataCore’s software-defined storage provides improved resource utilization, seamless integration of new technologies, and reduced administrative time - all resulting in lower CAPEX and OPEX, yielding a superior TCO.

A survey of 363 DataCore customers found that over half of them (55%) achieved positive ROI within the first year of deployment, and 21% were able to reach positive ROI in less than 6 months.

Download this white paper to learn how software-defined storage can help reduce data center infrastructure costs, including guidelines to help you structure your TCO analysis comparison.

Preserve Proven Business Continuity Practices Despite Inevitable Changes in Your Data Storage
Download this solution brief and get insights on how to avoid spending time and money reinventing BC/DR plans every time your storage infrastructure changes.
Nothing in Business Continuity circles ranks higher in importance than risk reduction. Yet the risk of major disruptions to business continuity practices looms ever larger today, mostly due to the troubling dependencies on the location, topology and suppliers of data storage.

Download this solution brief and get insights on how to avoid spending time and money reinventing BC/DR plans every time your storage infrastructure changes. 
How Data Temperature Drives Data Placement Decisions and What to Do About It
In this white paper, learn (1) how the relative proportion of hot, warm, and cooler data changes over time, (2) new machine learning (ML) techniques that sense the cooling temperature of data throughout its half-life, and (3) the role of artificial intelligence (AI) in migrating data to the most cost-effective tier.

The emphasis on fast flash technology concentrates much attention on hot, frequently accessed data. However, budget pressures preclude consuming such premium-priced capacity when the access frequency diminishes. Yet many organizations do just that, unable to migrate effectively to lower cost secondary storage on a regular basis.
In this white paper, explore:

•    How the relative proportion of hot, warm, and cooler data changes over time
•    New machine learning (ML) techniques that sense the cooling temperature of data throughout its half-life
•    The role of artificial intelligence (AI) in migrating data to the most cost-effective tier.

All-Flash Array Buying Considerations: The Long-Term Advantages of Software-Defined Storage
In this white paper, analysts from the Enterprise Strategy Group (ESG) provide insights into (1) the modern data center challenge, (2) buying considerations before your next flash purchase, and (3) the value of storage infrastructure independence and how to obtain it with software-defined storage.
All-flash technology is the way of the future. Performance matters, and flash is fast—and it is getting even faster with the advent of NVMe and SCM technologies. IT organizations are going to continue to increase the amount of flash storage in their shops for this simple reason.

However, this also introduces more complexity into the modern data center. In the real world, blindly deploying all-flash everywhere is costly, and it doesn’t solve management/operational silo problems. In the Enterprise Strategy Group (ESG) 2018 IT spending intentions survey, 68% of IT decision makers said that IT is more complex today than it was just two years ago. In this white paper, ESG discusses:

•    The modern data center challenge
•    Buying considerations before your next flash purchase
•    The value of storage infrastructure independence and how to obtain it with software-defined storage

top25