Virtualization Technology News and Information
White Papers
RSS
White Papers Search Results
Showing 33 - 48 of 71 white papers, page 3 of 5.
VMware vSphere 6.7 Update 1 Upgrade and Security Configuration
Most businesses hugely invest in tackling the security vulnerabilities of their data centers. VMware vSphere 6.7 Upgrade 1 tackles it head-on with its functionalities that aligns with both the legacy and the modern technology capabilities. Read this white paper to know how you can maximize the security posture of vSphere workloads on production environments.
Security is a top concern when it comes to addressing data protection complexities for business-critical systems. VMware vSphere 6.7 Upgrade 1 can be the right fit for your data centers when it comes to resolving security vulnerabilities thereby helping you to take your IT infrastructure to the next level. While there are features that align with the legacy security standards, there are some of the best newly announced functionalities in vSphere 6.7 like Virtual TPM 2.0 and virtualization-based security capabilities that will help you enhance your current security measures for your production workloads. Get started with reading this white paper to know more on how you can implement a solution of this kind in your data centers.
vSphere Troubleshooting Guide
Troubleshooting complex virtualization technology is something all VMware users will have to face at some point. It requires an understanding of how various components fit together and finding a place to start is not easy. Thankfully, VMware vExpert Ryan Birk is here to help with this eBook preparing you for any problems you may encounter along the way.

This eBook explains how to identify problems with vSphere and how to solve them. Before we begin, we need to start off with an introduction to a few things that will make life easier. We’ll start with a troubleshooting methodology and how to gather logs. After that, we’ll break this eBook into the following sections: Installation, Virtual Machines, Networking, Storage, vCenter/ESXi and Clustering.

ESXi and vSphere problems arise from many different places, but they generally fall into one of these categories: Hardware issues, Resource contention, Network attacks, Software bugs, and Configuration problems.

A typical troubleshooting process contains several tasks: 1. Define the problem and gather information. 2. Identify what is causing the problem. 3. Fix the problem, implement a fix.

One of the first things you should try to do when experiencing a problem with a host, is try to reproduce the issue. If you can find a way to reproduce it, you have a great way to validate that the issue is resolved when you do fix it. It can be helpful as well to take a benchmark of your systems before they are implemented into a production environment. If you know HOW they should be running, it’s easier to pinpoint a problem.

You should decide if it’s best to work from a “Top Down” or “Bottom Up” approach to determine the root cause. Guest OS Level issues typically cause a large amount of problems. Let’s face it, some of the applications we use are not perfect. They get the job done but they utilize a lot of memory doing it.

In terms of virtual machine level issues, is it possible that you could have a limit or share value that’s misconfigured? At the ESXi Host Level, you could need additional resources. It’s hard to believe sometimes, but you might need another host to help with load!

Once you have identified the root cause, you should assess the impact of the problem on your day to day operations. When and what type of fix should you implement? A short-term one or a long-term solution? Assess the impact of your solution on daily operations. Short-term solution: Implement a quick workaround. Long-term solution: Reconfiguration of a virtual machine or host.

Now that the basics have been covered, download the eBook to discover how to put this theory into practice!

Mastering vSphere – Best Practices, Optimizing Configurations & More
Do you regularly work with vSphere? If so, this free eBook is for you. Learn how to leverage best practices for the most popular features contained within the vSphere platform and boost your productivity using tips and tricks learnt direct from an experienced VMware trainer and highly qualified professional. In this eBook, vExpert Ryan Birk shows you how to master: Advanced Deployment Scenarios using Auto-Deploy Shared Storage Performance Monitoring and Troubleshooting Host Network configurati

If you’re here to gather some of the best practices surrounding vSphere, you’ve come to the right place! Mastering vSphere: Best Practices, Optimizing Configurations & More, the free eBook authored by me, Ryan Birk, is the product of many years working with vSphere as well as teaching others in a professional capacity. In my extensive career as a VMware consultant and teacher (I’m a VMware Certified Instructor) I have worked with people of all competence levels and been asked hundreds - if not thousands - of questions on vSphere. I was approached to write this eBook to put that experience to use to help people currently working with vSphere step up their game and reach that next level. As such, this eBook assumes readers already have a basic understanding of vSphere and will cover the best practices for four key aspects of any vSphere environment.

The best practices covered here will focus largely on management and configuration solutions so should remain relevant for quite some time. However, with that said, things are constantly changing in IT, so I would always recommend obtaining the most up-to-date information from VMware KBs and official documentation especially regarding specific versions of tools and software updates. This eBook is divided into several sections, and although I would advise reading the whole eBook as most elements relate to others, you might want to just focus on a certain area you’re having trouble with. If so, jump to the section you want read about.

Before we begin, I want to note that in a VMware environment, it’s always best to try to keep things simple. Far too often I have seen environments be thrown off the tracks by trying to do too much at once. I try to live by the mentality of “keeping your environment boring” – in other words, keeping your host configurations the same, storage configurations the same and network configurations the same. I don’t mean duplicate IP addresses, but the hosts need identical port groups, access to the same storage networks, etc. Consistency is the name of the game and is key to solving unexpected problems down the line. Furthermore, it enables smooth scalability - when you move from a single host configuration to a cluster configuration, having the same configurations will make live migrations and high availability far easier to configure without having to significantly re-work the entire infrastructure. Now the scene has been set, let’s get started!

How to Get the Most Out of Windows Admin Center
Windows Admin Center is the future of Windows and Windows Server management. Are you using it to its full potential? In this free eBook, Microsoft Cloud and Datacenter Management MVP, Eric Siron, has put together a 70+ page guide on what Windows Admin Center brings to the table, how to get started, and how to squeeze as much value out of this incredible free management tool from Microsoft. This eBook covers: - Installation - Getting Started - Full UI Analysis - Security - Managing Extensions

Each version of Windows and Windows Server showcases new technologies. The advent of PowerShell marked a substantial step forward in managing those features. However, the built-in graphical Windows management tools have largely stagnated - the same basic Microsoft Management Console (MMC) interfaces had remained since Windows Server 2000. Microsoft tried out multiple overhauls over the years to the built-in Server Manager console but gained little traction. Until Windows Admin Center.

WHAT IS WINDOWS ADMIN CENTER?
Windows Admin Center (WAC) represents a modern turn in Windows and Windows Server system management. From its home page, you establish a list of the networked Windows and Windows Server computers to manage. From there, you can connect to an individual system to control components such as hardware drivers. You can also use it to manage Windows roles, such as Hyper-V.

On the front-end, Windows Admin Center is presented through a sleek HTML 5 web interface. On the back-end, it leverages PowerShell extensively to control the systems within your network. The entire package runs on a single system, so you don’t need a complicated infrastructure to support it. In fact, you can run it locally on your Windows 10 workstation if you want. If you require more resiliency, you can run Windows Admin Center as a role on a Microsoft Failover Cluster.

WHY WOULD I USE WINDOWS ADMIN CENTER?
In the modern era of Windows management, we have shifted to a greater reliance on industrial-strength tools like PowerShell and Desired State Configuration. However, we still have servers that require individualized attention and infrequently utilized resources. WAC gives you a one-stop hub for dropping in on any system at any time and work with almost any of its facets.

ABOUT THIS EBOOK
This eBook has been written by Microsoft Cloud & Datacenter Management MVP Eric Siron. Eric has worked in IT since 1998, designing, deploying, and maintaining server, desktop, network, and storage systems. He has provided all levels of support for businesses ranging from single-user through enterprises with thousands of seats. He has achieved numerous Microsoft certifications and was a Microsoft Certified Trainer for four years. Eric is also a seasoned technology blogger and has amassed a significant following through his top-class work on the Altaro Hyper-V Dojo.

Reducing Data Center Infrastructure Costs with Software-Defined Storage
Download this white paper to learn how software-defined storage can help reduce data center infrastructure costs, including guidelines to help you structure your TCO analysis comparison.

With a software-based approach, IT organizations see a better return on their storage investment. DataCore’s software-defined storage provides improved resource utilization, seamless integration of new technologies, and reduced administrative time - all resulting in lower CAPEX and OPEX, yielding a superior TCO.

A survey of 363 DataCore customers found that over half of them (55%) achieved positive ROI within the first year of deployment, and 21% were able to reach positive ROI in less than 6 months.

Download this white paper to learn how software-defined storage can help reduce data center infrastructure costs, including guidelines to help you structure your TCO analysis comparison.

How Data Temperature Drives Data Placement Decisions and What to Do About It
In this white paper, learn (1) how the relative proportion of hot, warm, and cooler data changes over time, (2) new machine learning (ML) techniques that sense the cooling temperature of data throughout its half-life, and (3) the role of artificial intelligence (AI) in migrating data to the most cost-effective tier.

The emphasis on fast flash technology concentrates much attention on hot, frequently accessed data. However, budget pressures preclude consuming such premium-priced capacity when the access frequency diminishes. Yet many organizations do just that, unable to migrate effectively to lower cost secondary storage on a regular basis.
In this white paper, explore:

•    How the relative proportion of hot, warm, and cooler data changes over time
•    New machine learning (ML) techniques that sense the cooling temperature of data throughout its half-life
•    The role of artificial intelligence (AI) in migrating data to the most cost-effective tier.

PowerCLI - The Aspiring Automator's Guide
Automation is awesome but don't just settle for using other people's scripts. Learn how to create your own scripts and take your vSphere automation game to the next level! Written by VMware vExpert Xavier Avrillier, this free eBook presents a use-case approach to learning how to automate tasks in vSphere environments using PowerCLI. We start by covering the basics of installation, set up, and an overview of PowerCLI terms. From there we move into scripting logic and script building with step-by

Scripting and PowerCLI are words that most people working with VMware products know pretty well and have used once or twice. Everyone knows that scripting and automation are great assests to have in your toolbox. The problem usually is that getting into scripting appears daunting to many people who feel like the learning curve is just too steep, and they usually don't know where to start. The good thing is you don't need to learn everything straight away to start working with PowerShell and PowerCLI. Once you have the basics down and have your curiosity tickled, you’ll learn what you need as you go, a lot faster than you thought you would!

ABOUT POWERCLI

Let's get to know PowerCLI a little better before we start getting our hands dirty in the command prompt. If you are reading this you probably already know what PowerCLI is about or have a vague idea of it, but it’s fine you don’t. After a while working with it, it becomes second nature, and you won't be able to imagine life without it anymore! Thanks to VMware's drive to push automation, the product's integration with all of their components has significantly improved over the years, and it has now become a critical part of their ecosystem.

WHAT IS PowerCLI?

Contrary to what many believe, PowerCLI is not in fact a stand-alone software but rather a command-line and scripting tool built on Windows PowerShell for managing and automating vSphere environments. It used to be distributed as an executable file to install on a workstation. Previously, an icon was generated that would essentially launch PowerShell and load the PowerCLI snap-ins in the session. This behavior changed back in version 6.5.1 when the executable file was removed and replaced by a suite of PowerShell modules to install from within the prompt itself. This new deployment method is preferred because these modules are now part of Microsoft’s Official PowerShell Gallery. 7 These modules provide the means to interact with the components of a VMware environment and offer more than 600 cmdlets! The below command returns a full list of VMware-Associated Cmdlets!

Why Network Verification Requires a Mathematical Model
Learn how verification can be used in key IT processes and workflows, why a mathematical model is required and how it works; as well as example use cases from the Forward Enterprise platform.
Network verification is a rapidly emerging technology that is a key part of Intent Based Networking (IBN). Verification can help avoid outages, facilitate compliance processes and accelerate change windows. Full-feature verification solutions require an underlying mathematical model of network behavior to analyze and reason about policy objectives and network designs. A mathematical model, as opposed to monitoring or testing live traffic, can perform exhaustive and definitive analysis of network implementations and behavior, including proving network isolation or security rules.

In this paper, we will describe how verification can be used in key IT processes and workflows, why a mathematical model is required and how it works, as well as example use cases from the Forward Enterprise platform. This will also clarify what requirements a mathematical model must meet and how to evaluate alternative products.
ESG Report: Verifying Network Intent with Forward Enterprise
This ESG Technical Review documents hands-on validation of Forward Enterprise, a solution developed by Forward Networks to help organizations save time and resources when verifying that their IT networks can deliver application traffic consistently in line with network and security policies. The review examines how Forward Enterprise can reduce network downtime, ensure compliance with policies, and minimize adverse impact of configuration changes on network behavior.
ESG research recently uncovered that 66% of organizations view their IT environments as more or significantly more complex than they were two years ago. The complexity will most likely increase, since 46% of organizations anticipate their network infrastructure spending to exceed that of 2018 as they upgrade and expand their networks.

Large enterprise and service provider networks consist of multiple device types—routers, switches, firewalls, and load balancers—with proprietary operating systems (OS) and different configuration rules. As organizations support more applications and users, their networks will grow and become more complex, making it more difficult to verify and manage correctly implemented policies across the entire network. Organizations have also begun to integrate public cloud services with their on-premises networks, adding further network complexity to manage end-to-end policies.

With increasing network complexity, organizations cannot easily confirm that their networks are operating as intended when they implement network and security policies. Moreover, when considering a fix to a service-impact issue or a network update, determining how it may impact other applications negatively or introduce service-affecting issues becomes difficult. To assess adherence to policies or the impact of any network change, organizations have typically relied on disparate tools and material—network topology diagrams, device inventories, vendor-dependent management systems, command line (CLI) commands, and utilities such as “ping” and “traceroute.” The combination of these tools cannot provide a reliable and holistic assessment of network behavior efficiently.

Organizations need a vendor-agnostic solution that enables network operations to automate the verification of network implementations against intended policies and requirements, regardless of the number and types of devices, operating systems, traffic rules, and policies that exist. The solution must represent the topology of the entire network or subsets of devices (e.g., in a region) quickly and efficiently. It should verify network implementations from prior points in time, as well as proposed network changes prior to implementation. Finally, the solution must also enable organizations to quickly detect issues that affect application delivery or violate compliance requirements.
Forward Networks ROI Case Study
See how a large financial services business uses Forward Enterprise to achieve significant ROI with process improvements in trouble ticket resolution, audit-related fixes and change windows.
Because Forward Enterprise automates the intelligent analysis of network designs, configurations and state, we provide an immediate and verifiable return on investment (ROI) in terms of accelerating key IT processes and reducing manhours of highly skilled engineers in troubleshooting and testing the network.

In this paper, we will quantify the ROI of a large financial services firm and document the process improvements that led to IT cost savings and a more agile network. In this analysis, we will look at process improvements in trouble ticket resolution, audit-related fixes and acceleration of network updates and change windows. We will explore each of these areas in more detail, along with the input assumptions for the calculations, but for this financial services customer, the following benefits were achieved, resulting in an annualized net savings of over $3.5 million.
Restoring Order to Virtualization Chaos
Get Tighter Control of a Mixed VM Environment and Meet Your Data Protection SLAs Is virtualization bringing you the promised benefits of increased IT agility and reduced operating costs, or is virtualization just adding more chaos and complexity? Getting a grip on the prismatic environment of virtualized platforms – whether on-premises, in-cloud, or in some hybrid combination – is key to realizing virtualization’s benefits. To truly achieve better IT productivity, reduce costs, and meet ever mo
Get Tighter Control of a Mixed VM Environment and Meet Your Data Protection SLAs

Is virtualization bringing you the promised benefits of increased IT agility and reduced operating costs, or is virtualization just adding more chaos and complexity? Getting a grip on the prismatic environment of virtualized platforms – whether on-premises, in-cloud, or in some hybrid combination – is key to realizing virtualization’s benefits. To truly achieve better IT productivity, reduce costs, and meet ever more stringent service level agreements (SLAs), you need to create order out of virtualization chaos.

We’ll examine ways in which IT executives can more effectively manage a hybrid virtual machine (VM) environment, and more importantly, how to deliver consistent data protection and recovery across all virtualized platforms. The goal is to control complexity and meet your SLAs, regardless of VM container. In so doing, you will control your VMs, instead of allowing their chaos to control you!
How to seamlessly and securely transition to hybrid cloud
Find out how to optimize your hybrid cloud workload through system hardening, incident detection, active defense and mitigation, quarantining and more. Plus, learn how to ensure protection and performance in your environment through an ideal hybrid cloud workload protection solution.

With digital transformation a constantly evolving reality for the modern organization, businesses are called upon to manage complex workloads across multiple public and private clouds—in addition to their on-premises systems.

The upside of the hybrid cloud strategy is that businesses can benefit from both lowered costs and dramatically increased agility and flexibility. The problem, however, is maintaining a secure environment through challenges like data security, regulatory compliance, external threats to the service provider, rogue IT usage and issues related to lack of visibility into the provider’s infrastructure.

Find out how to optimize your hybrid cloud workload through system hardening, incident detection, active defense and mitigation, quarantining and more. Plus, learn how to ensure protection and performance in your environment through an ideal hybrid cloud workload protection solution that:


•    Provides the necessary level of protection for different workloads
•    Delivers an essential set of technologies
•    Is structured as a comprehensive, multi-layered solution
•    Avoids performance degradation for services or users
•    Supports compliance by satisfying a range of regulation requirements
•    Enforces consistent security policies through all parts of hybrid infrastructure
•    Enables ongoing audit by integrating state of security reports
•    Takes account of continuous infrastructure changes

Office 365 / Microsoft 365: The Essential Companion Guide
Office 365 and Microsoft 365 contain truly powerful applications that can significantly boost productivity in the workplace. However, there’s a lot on offer so we’ve put together a comprehensive companion guide to ensure you get the most out of your investment! This free 85-page eBook, written by Microsoft Certified Trainer Paul Schnackenburg, covers everything from basic descriptions, to installation, migration, use-cases, and best practices for all features within the Office/Microsoft 365 sui

Welcome to this free eBook on Office 365 and Microsoft 365 brought to you by Altaro Software. We’re going to show you how to get the most out of these powerful cloud packages and improve your business. This book follows an informal reference format providing an overview of the most powerful applications of each platform’s feature set in addition to links directing to supporting information and further reading if you want to dig further into a specific topic. The intended audience for this book is administrators and IT staff who are either preparing to migrate to Office/Microsoft 365 or who have already migrated and who need to get the lay of the land. If you’re a developer looking to create applications and services on top of the Microsoft 365 platform, this book is not for you. If you’re a business decision-maker, rather than a technical implementer, this book will give you a good introduction to what you can expect when your organization has been migrated to the cloud and ways you can adopt various services in Microsoft 365 to improve the efficiency of your business.

THE BASICS

We’ll cover the differences (and why one might be more appropriate for you than the other) in more detail later but to start off let’s just clarify what each software package encompasses in a nutshell. Office 365 (from now on referred to as O365) 7 is an email collaboration and a host of other services provided as a Software as a Service (SaaS) whereas Microsoft 365 (M365) is Office 365 plus Azure Active Directory Premium, Intune – cloud-based management of devices and security and Windows 10 Enterprise. Both are per user-based subscription services that require no (or very little) infrastructure deployments on-premises.

How to Develop a Multi-cloud Management Strategy
Increasingly, organizations are looking to move workloads into the cloud. The goal may be to leverage cloud resources for Dev/Test, or they may want to “lift and shift” an application to the cloud and run it natively. In order to enable these various cloud options, it is critical that organizations develop a multi-cloud data management strategy.

The primary goal of a multi-cloud data management strategy is to supply data, either via copying or moving data to the various multi-cloud use cases. A key enabler of this movement is the data management software applications. In theory, data protection applications can perform both of the copy and move functions. A key consideration is how the multi-cloud data management experience is unified. In most cases, data protection applications ignore the user experience of each cloud and use their proprietary interface as the unifying entity, which increases complexity.

There are a variety of reasons organizations may want to leverage multiple clouds. The first use case is to use public cloud storage as a backup mirror to an on-premises data protection process. Using public cloud storage as a backup mirror enables the organization to automatically off-site data. It also sets up many of the more advanced use cases.

Another use case is using the cloud for disaster recovery.

Another use case is “Lift and Shift,” which means the organization wants to run the application in the cloud natively. Initial steps in the “lift and shift” use case are similar to Dev/Test, but now the workload is storing unique data in the cloud.

Multi-cloud is a reality now for most organizations and managing the movement of data between these clouds is critical.

Multi-cloud Data Protection-as-a-service: The HYCU Protégé Platform
Multi-cloud environments are here to stay and will keep on growing in diversity, use cases, and, of course, size. Data growth is not stopping anytime soon, only making the problem more acute. HYCU has taken a very different approach from many traditional vendors by selectively delivering deeply integrated solutions to the platforms they protect, and is now moving to the next challenge of unification and simplification with Protégé, calling it a data protection-as-a-service platform.

There are a number of limitations today keeping organizations from not only lifting and shifting from one cloud to another but also migrating across clouds. Organizations need the flexibility to leverage multiple clouds and move applications and workloads around freely, whether for data reuse or for disaster recovery. This is where the HYCU Protégé platform comes in. HYCU Protégé is positioned as a complete multi-cloud data protection and disaster recovery-as-a-service solution. It includes a number of capabilities that make it relevant and notable compared with other approaches in the market:

  • It was designed for multi-cloud environments, with a “built-for-purpose” approach to each workload and environment, leveraging APIs and platform expertise.
  • It is designed as a one-to-many cross-cloud disaster recovery topology rather than a one-to-one cloud or similarly limited topology.
  • It is designed for the IT generalist. It’s easy to use, it includes dynamic provisioning on-premises and in the cloud, and it can be deployed without impacting production systems. In other words, no need to manually install hypervisors or agents.
  • It is application-aware and will automatically discover and configure applications. Additionally, it supports distributed applications with shared storage. 
Normal 0 false false false EN-US X-NONE X-NONE
Data Protection as a Service - Simplify Your Backup and Disaster Recovery
Data protection is a catch-all term that encompasses a number of technologies, business practices and skill sets associated with preventing the loss, corruption or theft of data. The two primary data protection categories are backup and disaster recovery (DR) — each one providing a different type, level and data protection objective. While managing each of these categories occupies a significant percentage of the IT budget and systems administrator’s time, it doesn’t have to. Data protection can
Simplify Your Backup and Disaster Recovery

Today, there are an ever-growing number of threats to businesses and uptime is crucial. Data protection has never been a more important function of IT. As data center complexity and demand for new resources increases, the difficulty of providing effective and cost-efficient data protection increases as well.

Luckily, data protection can now be provided as a service.

Get this white paper to learn:
  • How data protection service providers enable IT teams to focus on business objectives
  • The difference, and importance, of cloud-based backup and disaster recovery
  • Why cloud-based backup and disaster recovery are required for complete protection
top25