Virtualization Technology News and Information
White Papers
RSS
White Papers Search Results
Showing 17 - 32 of 55 white papers, page 2 of 4.
How to Get the Most Out of Windows Admin Center
Windows Admin Center is the future of Windows and Windows Server management. Are you using it to its full potential? In this free eBook, Microsoft Cloud and Datacenter Management MVP, Eric Siron, has put together a 70+ page guide on what Windows Admin Center brings to the table, how to get started, and how to squeeze as much value out of this incredible free management tool from Microsoft. This eBook covers: - Installation - Getting Started - Full UI Analysis - Security - Managing Extensions

Each version of Windows and Windows Server showcases new technologies. The advent of PowerShell marked a substantial step forward in managing those features. However, the built-in graphical Windows management tools have largely stagnated - the same basic Microsoft Management Console (MMC) interfaces had remained since Windows Server 2000. Microsoft tried out multiple overhauls over the years to the built-in Server Manager console but gained little traction. Until Windows Admin Center.

WHAT IS WINDOWS ADMIN CENTER?
Windows Admin Center (WAC) represents a modern turn in Windows and Windows Server system management. From its home page, you establish a list of the networked Windows and Windows Server computers to manage. From there, you can connect to an individual system to control components such as hardware drivers. You can also use it to manage Windows roles, such as Hyper-V.

On the front-end, Windows Admin Center is presented through a sleek HTML 5 web interface. On the back-end, it leverages PowerShell extensively to control the systems within your network. The entire package runs on a single system, so you don’t need a complicated infrastructure to support it. In fact, you can run it locally on your Windows 10 workstation if you want. If you require more resiliency, you can run Windows Admin Center as a role on a Microsoft Failover Cluster.

WHY WOULD I USE WINDOWS ADMIN CENTER?
In the modern era of Windows management, we have shifted to a greater reliance on industrial-strength tools like PowerShell and Desired State Configuration. However, we still have servers that require individualized attention and infrequently utilized resources. WAC gives you a one-stop hub for dropping in on any system at any time and work with almost any of its facets.

ABOUT THIS EBOOK
This eBook has been written by Microsoft Cloud & Datacenter Management MVP Eric Siron. Eric has worked in IT since 1998, designing, deploying, and maintaining server, desktop, network, and storage systems. He has provided all levels of support for businesses ranging from single-user through enterprises with thousands of seats. He has achieved numerous Microsoft certifications and was a Microsoft Certified Trainer for four years. Eric is also a seasoned technology blogger and has amassed a significant following through his top-class work on the Altaro Hyper-V Dojo.

ESG Report: Verifying Network Intent with Forward Enterprise
This ESG Technical Review documents hands-on validation of Forward Enterprise, a solution developed by Forward Networks to help organizations save time and resources when verifying that their IT networks can deliver application traffic consistently in line with network and security policies. The review examines how Forward Enterprise can reduce network downtime, ensure compliance with policies, and minimize adverse impact of configuration changes on network behavior.
ESG research recently uncovered that 66% of organizations view their IT environments as more or significantly more complex than they were two years ago. The complexity will most likely increase, since 46% of organizations anticipate their network infrastructure spending to exceed that of 2018 as they upgrade and expand their networks.

Large enterprise and service provider networks consist of multiple device types—routers, switches, firewalls, and load balancers—with proprietary operating systems (OS) and different configuration rules. As organizations support more applications and users, their networks will grow and become more complex, making it more difficult to verify and manage correctly implemented policies across the entire network. Organizations have also begun to integrate public cloud services with their on-premises networks, adding further network complexity to manage end-to-end policies.

With increasing network complexity, organizations cannot easily confirm that their networks are operating as intended when they implement network and security policies. Moreover, when considering a fix to a service-impact issue or a network update, determining how it may impact other applications negatively or introduce service-affecting issues becomes difficult. To assess adherence to policies or the impact of any network change, organizations have typically relied on disparate tools and material—network topology diagrams, device inventories, vendor-dependent management systems, command line (CLI) commands, and utilities such as “ping” and “traceroute.” The combination of these tools cannot provide a reliable and holistic assessment of network behavior efficiently.

Organizations need a vendor-agnostic solution that enables network operations to automate the verification of network implementations against intended policies and requirements, regardless of the number and types of devices, operating systems, traffic rules, and policies that exist. The solution must represent the topology of the entire network or subsets of devices (e.g., in a region) quickly and efficiently. It should verify network implementations from prior points in time, as well as proposed network changes prior to implementation. Finally, the solution must also enable organizations to quickly detect issues that affect application delivery or violate compliance requirements.
Lift and Shift Backup and Disaster Recovery Scenario for Google Cloud: Step by Step Guide
There are many new challenges, and reasons, to migrate workloads to the cloud. Especially for public cloud, like Google Cloud Platform. Whether it is for backup, disaster recovery, or production in the cloud, you should be able to leverage the cloud platform to solve your technology challenges. In this step-by-step guide, we outline how GCP is positioned to be one of the easiest cloud platforms for app development. And, the critical role data protection as-as-service (DPaaS) can play.

There are many new challenges, and reasons, to migrate workloads to the cloud.

For example, here are four of the most popular:

  • Analytics and Machine learning (ML) are everywhere. Once you have your data in a cloud platform like Google Cloud Platform, you can leverage their APIs to run analytics and ML on everything.
  • Kubernetes is powerful and scalable, but transitioning legacy apps to Kubernetes can be daunting.
  • SAP HANA is a secret weapon. With high mem instances in the double digit TeraBytes migrating SAP to a cloud platform is easier than ever.
  • Serverless is the future for application development. With CloudSQL, Big Query, and all the other serverless solutions, cloud platforms like GCP are well positioned to be the easiest platform for app development.

Whether it is for backup, disaster recovery, or production in the cloud, you should be able to leverage the cloud platform to solve your technology challenges. In this step-by-step guide, we outline how GCP is positioned to be one of the easiest cloud platforms for app development. And, the critical role data protection as-as-service (DPaaS) can play.

How to seamlessly and securely transition to hybrid cloud
Find out how to optimize your hybrid cloud workload through system hardening, incident detection, active defense and mitigation, quarantining and more. Plus, learn how to ensure protection and performance in your environment through an ideal hybrid cloud workload protection solution.

With digital transformation a constantly evolving reality for the modern organization, businesses are called upon to manage complex workloads across multiple public and private clouds—in addition to their on-premises systems.

The upside of the hybrid cloud strategy is that businesses can benefit from both lowered costs and dramatically increased agility and flexibility. The problem, however, is maintaining a secure environment through challenges like data security, regulatory compliance, external threats to the service provider, rogue IT usage and issues related to lack of visibility into the provider’s infrastructure.

Find out how to optimize your hybrid cloud workload through system hardening, incident detection, active defense and mitigation, quarantining and more. Plus, learn how to ensure protection and performance in your environment through an ideal hybrid cloud workload protection solution that:


•    Provides the necessary level of protection for different workloads
•    Delivers an essential set of technologies
•    Is structured as a comprehensive, multi-layered solution
•    Avoids performance degradation for services or users
•    Supports compliance by satisfying a range of regulation requirements
•    Enforces consistent security policies through all parts of hybrid infrastructure
•    Enables ongoing audit by integrating state of security reports
•    Takes account of continuous infrastructure changes

Office 365 / Microsoft 365: The Essential Companion Guide
Office 365 and Microsoft 365 contain truly powerful applications that can significantly boost productivity in the workplace. However, there’s a lot on offer so we’ve put together a comprehensive companion guide to ensure you get the most out of your investment! This free 85-page eBook, written by Microsoft Certified Trainer Paul Schnackenburg, covers everything from basic descriptions, to installation, migration, use-cases, and best practices for all features within the Office/Microsoft 365 sui

Welcome to this free eBook on Office 365 and Microsoft 365 brought to you by Altaro Software. We’re going to show you how to get the most out of these powerful cloud packages and improve your business. This book follows an informal reference format providing an overview of the most powerful applications of each platform’s feature set in addition to links directing to supporting information and further reading if you want to dig further into a specific topic. The intended audience for this book is administrators and IT staff who are either preparing to migrate to Office/Microsoft 365 or who have already migrated and who need to get the lay of the land. If you’re a developer looking to create applications and services on top of the Microsoft 365 platform, this book is not for you. If you’re a business decision-maker, rather than a technical implementer, this book will give you a good introduction to what you can expect when your organization has been migrated to the cloud and ways you can adopt various services in Microsoft 365 to improve the efficiency of your business.

THE BASICS

We’ll cover the differences (and why one might be more appropriate for you than the other) in more detail later but to start off let’s just clarify what each software package encompasses in a nutshell. Office 365 (from now on referred to as O365) 7 is an email collaboration and a host of other services provided as a Software as a Service (SaaS) whereas Microsoft 365 (M365) is Office 365 plus Azure Active Directory Premium, Intune – cloud-based management of devices and security and Windows 10 Enterprise. Both are per user-based subscription services that require no (or very little) infrastructure deployments on-premises.

How to Develop a Multi-cloud Management Strategy
Increasingly, organizations are looking to move workloads into the cloud. The goal may be to leverage cloud resources for Dev/Test, or they may want to “lift and shift” an application to the cloud and run it natively. In order to enable these various cloud options, it is critical that organizations develop a multi-cloud data management strategy.

The primary goal of a multi-cloud data management strategy is to supply data, either via copying or moving data to the various multi-cloud use cases. A key enabler of this movement is the data management software applications. In theory, data protection applications can perform both of the copy and move functions. A key consideration is how the multi-cloud data management experience is unified. In most cases, data protection applications ignore the user experience of each cloud and use their proprietary interface as the unifying entity, which increases complexity.

There are a variety of reasons organizations may want to leverage multiple clouds. The first use case is to use public cloud storage as a backup mirror to an on-premises data protection process. Using public cloud storage as a backup mirror enables the organization to automatically off-site data. It also sets up many of the more advanced use cases.

Another use case is using the cloud for disaster recovery.

Another use case is “Lift and Shift,” which means the organization wants to run the application in the cloud natively. Initial steps in the “lift and shift” use case are similar to Dev/Test, but now the workload is storing unique data in the cloud.

Multi-cloud is a reality now for most organizations and managing the movement of data between these clouds is critical.

Multi-cloud Data Protection-as-a-service: The HYCU Protégé Platform
Multi-cloud environments are here to stay and will keep on growing in diversity, use cases, and, of course, size. Data growth is not stopping anytime soon, only making the problem more acute. HYCU has taken a very different approach from many traditional vendors by selectively delivering deeply integrated solutions to the platforms they protect, and is now moving to the next challenge of unification and simplification with Protégé, calling it a data protection-as-a-service platform.

There are a number of limitations today keeping organizations from not only lifting and shifting from one cloud to another but also migrating across clouds. Organizations need the flexibility to leverage multiple clouds and move applications and workloads around freely, whether for data reuse or for disaster recovery. This is where the HYCU Protégé platform comes in. HYCU Protégé is positioned as a complete multi-cloud data protection and disaster recovery-as-a-service solution. It includes a number of capabilities that make it relevant and notable compared with other approaches in the market:

  • It was designed for multi-cloud environments, with a “built-for-purpose” approach to each workload and environment, leveraging APIs and platform expertise.
  • It is designed as a one-to-many cross-cloud disaster recovery topology rather than a one-to-one cloud or similarly limited topology.
  • It is designed for the IT generalist. It’s easy to use, it includes dynamic provisioning on-premises and in the cloud, and it can be deployed without impacting production systems. In other words, no need to manually install hypervisors or agents.
  • It is application-aware and will automatically discover and configure applications. Additionally, it supports distributed applications with shared storage. 
Normal 0 false false false EN-US X-NONE X-NONE
The SysAdmin Guide to Azure Infrastructure as a Service
If you're used to on-premises infrastructures, cloud platforms can seem daunting. But it doesn't need to be. This eBook written by the veteran IT consultant and trainer Paul Schnackenburg, covers all aspects of setting up and maintaining a high-performing Azure IaaS environment, including: • VM sizing and deployment • Migration • Storage and networking • Security and identity • Infrastructure as code and more!

The cloud computing era is well and truly upon us, and knowing how to take advantage of the benefits of this computing paradigm while maintaining security, manageability, and cost control are vital skills for any IT professional in 2020 and beyond. And its importance is only getting greater.

In this eBook, we’re going to focus on Infrastructure as a Service (IaaS) on Microsoft’s Azure platform - learning how to create VMs, size them correctly, manage storage, networking, and security, along with backup best practices. You’ll also learn how to operate groups of VMs, deploy resources based on templates, managing security and automate your infrastructure. If you currently have VMs in your own datacenter and are looking to migrate to Azure, we’ll also teach you that.

If you’re new to the cloud (or have experience with AWS/GCP but not Azure), this book will cover the basics as well as more advanced skills. Given how fast things change in the cloud, we’ll cover the why (as well as the how) so that as features and interfaces are updated, you’ll have the theoretical knowledge to effectively adapt and know how to proceed.

You’ll benefit most from this book if you actively follow along with the tutorials. We will be going through terms and definitions as we go – learning by doing has always been my preferred way of education. If you don’t have access to an Azure subscription, you can sign up for a free trial with Microsoft. This will give you 30 days 6 to use $200 USD worth of Azure resources, along with 12 months of free resources. Note that most of these “12 months” services aren’t related to IaaS VMs (apart from a few SSD based virtual disks and a small VM that you can run for 750 hours a month) so be sure to get everything covered on the IaaS side before your trial expires. There are also another 25 services that have free tiers “forever”.

Now you know what’s in store, let’s get started!

Why Should Enterprises Move to a True Composable Infrastructure Solution?
IT Infrastructure needs are constantly fluctuating in a world where powerful emerging software applications such as artificial intelligence can create, transform, and remodel markets in a few months or even weeks. While the public cloud is a flexible solution, it doesn’t solve every data center need—especially when businesses need to physically control their data on premises. Download this report to see how composable infrastructure helps you deploy faster, effectively utilize existing hardwar

IT Infrastructure needs are constantly fluctuating in a world where powerful emerging software applications such as artificial intelligence can create, transform, and remodel markets in a few months or even weeks. While the public cloud is a flexible solution, it doesn’t solve every data center need—especially when businesses need to physically control their data on premises. This leads to overspend— purchasing servers and equipment to meet peak demand at all times. The result? Expensive equipment sitting idle during non-peak times.

For years, companies have wrestled with overspend and underutilization of equipment, but now businesses can reduce cap-ex and rein in operational expenditures for underused hardware with software-defined composable infrastructure. With a true composable infrastructure solution, businesses realize optimal performance of IT resources while improving business agility. In addition, composable infrastructure allows organizations to take better advantage of the most data-intensive applications on existing hardware while preparing for future, disaggregated growth.

Download this report to see how composable infrastructure helps you deploy faster, effectively utilize existing hardware, rein in capital expenses, and more.

Make the Move: Linux Desktops with Cloud Access Software
Gone are the days where hosting Linux desktops on-premises is the only way to ensure uncompromised customization, choice and control. You can host Linux desktops & applications remotely and visualize them to further security, flexibility and performance. Learn why IT teams are virtualizing Linux.

Make the Move: Linux Remote Desktops Made Easy

Securely run Linux applications and desktops from the cloud or your data center.

Download this guide and learn...

  • Why organizations are virtualizing Linux desktops & applications
  • How different industries are leveraging remote Linux desktops & applications
  • What your organization can do to begin this journey


Key Considerations for Configuring Virtual Desktops For Remote Work
At any time, organizations worldwide and individuals can be forced to work from home. Learn about a sustainable solution to enable your remote workforce quickly and easily and gain tips to enhance your business continuity strategy when it comes to employee computing resources.

Assess what you already have

If you have a business continuity plan or a disaster recovery plan in place, that’s a good place to start. This scenario may not fit the definition of disaster that you originally intended, but it can serve to help you test your plan in a more controlled fashion that can benefit both your current situation by giving you a head start, and your overall plan by revealing gaps that would be more problematic in a more urgent or catastrophic environment with less time to prepare and implement.

Does your plan include access to remote desktops in a data center or the cloud? If so, and you already have a service in place ready to transition or expand, you’re well on your way.

Read the guide to learn what it takes for IT teams to set up staff to work effectively from home with virtual desktop deployments. Learn how to get started, if you’re new to VDI or if you already have an existing remote desktop scenario but are looking for alternatives.

The State of Multicloud: Virtual Desktop Deployments
Download this free 15-page report to understand the key differences and benefits to the many cloud deployment models and the factors that are driving tomorrow’s decisions.

The future of compute is in the cloud

Flexible, efficient, and economical, the cloud is no longer a question - it's the answer.

IT professionals that once considered if or when to migrate to the cloud are now talking about how. Earlier this year, we reached out to thousands of IT professionals to learn more about how.

Private Cloud, On-Prem, Public Cloud, Hybrid, Multicloud - each of these deployment models offers unique advantages and challenges. We asked IT decision-makers how they are currently leveraging the cloud and how they plan to grow.

Survey respondents overwhelmingly believed in the importance of a hybrid or multicloud strategy, regardless of whether they had actually implemented one themselves.

The top reasons for moving workloads between clouds

  • Cost Savings
  • Disaster Recovery
  • Data Center Location
  • Availability of Virtual Machines/GPUs
ESG - DataCore vFilO: Visibility and Control of Unstructured Data for the Modern, Digital Business
Organizations that want to succeed in the digital economy must contend with the cost and complexity introduced by the conventional segregation of multiple file system silos and separate object storage repositories. Fortunately, they can look to DataCore vFilO software for help. DataCore employs innovative techniques to combine diverse unstructured data resources to achieve unprecedented visibility, control, and flexibility.
DataCore’s new vFilO software shares important traits with its existing SANsymphony software-defined block storage platform. Both technologies are certainly enterprise class (highly agile, available, and performant). But each solution exhibits those traits in its own manner, taking the varying requirements for block, file, and object data into account. That’s important at a time when a lot of companies are maintaining hundreds to thousands of terabytes of unstructured data spread across many file servers, other NAS devices, and object storage repositories both onsite and in the cloud. The addition of vFilO to its product portfolio will allow DataCore to position itself in a different, even more compelling way now. DataCore is able to offer a “one-two punch”—namely, one of the best block storage SDS solutions in SANsymphony, and now one of the best next-generation SDS solutions for file and object data in vFilO. Together, vFilO and SANsymphony will put DataCore in a really strong position to support any IT organization looking for better ways to overcome end-users’ file-sharing/access difficulties, keep hardware costs low … and maximize the value of corporate data to achieve success in a digital age.
The Time is Now for File Virtualization
DataCore’s vFilO is a distributed files and object storage virtualization solution that can consume storage from a variety of providers, including NFS or SMB file servers, most NAS systems, and S3 Object Storage systems, including S3-based public cloud providers. Once vFilO integrates these various storage systems into its environment, it presents users with a logical file system and abstracts it from the actual physical location of data.

DataCore vFilO is a top-tier file virtualization solution. Not only can it serve as a global file system, IT can also add new NAS systems or file servers to the environment without having to remap users of the new hardware. vFilO supports live migration of data between the storage systems it has assimilated and leverages the capabilities of the global file system and the software’s policy-driven data management to move older data to less expensive storage automatically; either high capacity NAS or an object storage system. vFilO also transparently moves data from NFS/SMB to object storage. If the user needs access to this data in the future, they access it like they always have. To them, the data has not moved.

The ROI of File virtualization is powerful, but it has struggled to gain adoption in the data center. File Virtualization needs to be explained, and explaining it takes time. vFilO more than meets the requirements to qualify as a top tier file virtualization solution. DataCore has the advantage of over 10,000 customers that are much more likely to be receptive to the concept since they have already embraced block storage virtualization with SANSymphony. Building on its customer base as a beachhead, DataCore can then expand File Virtualization’s reach to new customers, who, because of the changing state of unstructured data, may finally be receptive to the concept. At the same time, these new file virtualization customers may be amenable to virtualizing block storage, and it may open up new doors for SANSymphony.

IDC: SaaS Backup and Recovery: Simplified Data Protection Without Compromise
Although the majority of organizations have a "cloud first" strategy, most also continue to manage onsite applications and the backup infrastructure associated with them. However, many are moving away from backup specialists and instead are leaving the task to virtual infrastructure administrators or other IT generalists. Metallic represents Commvault's direct entry into one of the fastest-growing segments of the data protection market. Its hallmarks are simplicity and flexibility of deployment

Metallic is a new SaaS backup and recovery solution based on Commvault's data protection software suite, proven in the marketplace for more than 20 years. It is designed specifically for the needs of medium-scale enterprises but is architected to grow with them based on data growth, user growth, or other requirements. Metallic initially offers either monthly or annual subscriptions through reseller partners; it will be available through cloud service providers and managed service providers over time. The initial workload use cases for Metallic include virtual machine (VM), SQL Server, file server, MS Office 365, and endpoint device recovery support; the company expects to add more use cases and supported workloads as the solution evolves.

Metallic is designed to offer flexibility as one of the service's hallmarks. Aspects of this include:

  • On-demand infrastructure: Metallic manages the cloud-based infrastructure components and software for the backup environment, though the customer will still manage any of its own on- premise infrastructure. This environment will support on-premise, cloud, and hybrid workloads. IT organizations are relieved of the daily task of managing the infrastructure components and do not have to worry about upgrades, OS or firmware updates and the like, for the cloud infrastructure, so people can repurpose that time saved toward other activities.
  • Metallic offers preconfigured plans designed to have users up and running in approximately 15 minutes, eliminating the need for a proof-of-concept test. These preconfigured systems have Commvault best practices built into the design, or organizations can configure their own.
  • Partner-delivered services: Metallic plans to go to market with resellers that can offer a range of services on top of the basic solution's capabilities. These services will vary by provider and will give users a variety of choices when selecting a provider to match the services offered with the organization's needs.
  • "Bring your own storage": Among the flexible options of Metallic, including VM and file or SQL database use cases, users can deploy their own storage, either on-premise or in the cloud, while utilizing the backup/recovery services of Metallic. The company refers to this option as "SaaS Plus."
Ten Topics to Discuss with Your Cloud Provider
Find the “just right” cloud for your business. For this paper, we will focus on existing applications (vs. new application services) that require high levels of performance and security, but that also enable customers to meet specific cost expectations.

Choosing the right cloud service for your organization, or for your target customer if you are a managed service provider, can be time consuming and effort intensive. For this paper, we will focus on existing applications (vs. new application services) that require high levels of performance and security, but that also enable customers to meet specific cost expectations.

Topics covered include:

  • Global access and availability
  • Cloud management
  • Application performance
  • Security and compliance
  • And more!