When deploying Red Hat Virtualization with as much automation as possible, building a custom boot ISO can be a useful tool. While the default installer for the Red Hat Virtualization Host (RHVH) is easy enough, sometimes, it pays to trim down the process. Augmenting the default install process for RHVH is pretty straight forward. In the following article I will walk you through the steps and describe the details of the RHVH installer.
While performance benchmarks are often application or industry specific they can also provide useful insights that are widely applicable. Risk analytics applications used in financial services industries have performance characteristics similar to many technical computing applications. These applications are large, compute intensive, and take full advantage of parallel processing and compute accelerators.
As many of you are aware, Red Hat Virtualization (RHV) provides ways to access its functionality via REST APIs or SDKs (Python, Java, Ruby), and recently Ansible has also joined this growing list.
Red Hat Virtualization 4.1 and later now includes support for the Ansible automation tool. Ansible can be used to configure systems, deploy software, and perform rolling updates. Ansible provides modules that allow you to
Over the years, the SuperComputing conference has become a focal point for many global supercomputing sites to showcase their capabilities and compete for a placement on the coveted Top500 list. Many powerful supercomputers and new technological advances are showcased during the conference, making it perhaps no surprise that Red Hat is planning to be at SuperComputing17 next week to demonstrate our latest high-performance computing (HPC) solutions (booth #1763).
In our first post of discussing Red Hat’s multi-architecture strategy, we focused on the disruptive nature of enabling new and rapidly-evolving architectures and how this enablement necessitates a different set of product requirements to fulfill our vision of providing a consistent and familiar experience to our customers across multiple hardware architectures. While we have been working with many original equipment manufacturers (OEMs) on x86_64-based servers for years, we have seen interest from our customer base in delivering parity across multiple architectures, including IBM Power Little Endian (ppc64le) and ARMv8-A (aarch64).
With the release of Red Hat Enterprise Linux Atomic Host 7.4, we are excited to jointly announce official support for VMware’s ESXi and vCenter. While it’s always been possible to run Atomic Host in a VMware environment, Red Hat Enterprise Linux Atomic Host now supports VMware’s guest agent and can be listed as a supported guest OS. The Red Hat Atomic Image for vSphere now embeds the guest agent to provide the out-of-the-box experience and integration that end users expect.
Enrolling a client system into Identity Management (IdM) can be done with a single command, namely: ipa-client-install. This command will configure SSSD, Kerberos, Certmonger and other elements of the system to work with IdM. The important result is that the system will get an identity and key so that it can securely connect to IdM and perform its operations. However, to get the identity and key, the system should
In this article I want to talk about a runC container which I want to migrate around the world while clients stay connected to the application.
In my previous Checkpoint/Restore In Userspace (CRIU) articles I introduced CRIU (From Checkpoint/Restore to Container Migration) and in the follow-up I gave an example how to use it in combination with containers (Container Live Migration Using runC and CRIU). Recently Christian Horn published an additional article about CRIU which is also a good starting point.
In my container I am running Xonotic. Xonotic calls itself ‘The Free and Fast Arena Shooter’. The part that is running in the container is the server part of the game to which multiple clients can connect to play together. In this article the client is running on my local system while the server and its container is live migrated around the world.
I recently had the pleasure of linking up with one of my favorite Red Hat colleagues (David “Pinky” Pinkerton) from Australia while we were both in Southeast Asia for a Red Hat event. We both have a propensity for KVM and Red Hat Virtualization (RHV) in particular, and he brought up a fantastic topic – truly segregated networks to support other security requirements. The reason came up because he had a “high security” client that needed to keep different traffic types separated within RHV, as the VMs were used to scan live malware. And that is why I made the comment about the (justifiably) paranoid.
Let’s take a look. |
In the previous post I talked about Smart Card Support in Red Hat Enterprise Linux. In this article I will drill down into how to select the right deployment architecture depending on your constraints, requirements and availability of the smart card related functionality in different versions of Red Hat Enterprise Linux.