When building anything substantial, such as a house or bridge, you start by laying down a solid foundation. Nothing changes this aspect of building brick by brick when you move from traditional constructions to application development and architecting your supporting infrastructure. Throw in Cloud terminology and you might think that the principles of a solid foundation are a bit flighty, but nothing is further from the truth.
When looking to manage an organization’s journey into their digital future, CIOs are dealing with a lot of challenges. Challenges that they face on the road to digital transformation can be daunting as first glance, but must be embraced to properly navigate the road to success.
What if I told you that you can have your Red Hat Enterprise Linux (RHEL) based Cloud infrastructure, with Red Hat Virtualization, OpenStack, OpenShift and CloudForms all setup before you have to stop for lunch?
Would you be surprised?
Could you do that today?
In most cases I am betting your answer would be not possible, not even on your best day. Not to worry, the solution is here and it’s called the QuickStart Cloud Installer (QCI).
There have been countless advances in technology in the last few years; both in general and at Red Hat. To list just the ones specific to Red Hat could actually boggle the mind. Arguably, some of the biggest advances have come more in the form of “soft” skills. Namely, Red Hat has become really good at listening – not only to our own customers but to our competitors’ customers as well. This is no more apparent than in our approach to applying a self-service catalog to virtualization. Specifically, pairing Red Hat Enterprise Virtualization (RHEV) with CloudForms for the purpose of streamlining and automation of virtual machine provisioning.
In our first post defending the pet container, we looked at the challenge of complexity facing modern software stacks and one way that containers address this challenge through aggregation. In essence, the Docker “wrapper” consolidates the next level of the stack, much like RPM did at the component level, but aggregation is just the beginning of what the project provides.
If we take a step back and look at the Docker project in context, there are four aspects that contribute to its exceptional popularity:
it simplifies the way users interact with the kernel, for features we have come to call Linux containers;
it’s a tool and format for aggregate packaging of software stacks to be deployed into containers;
it is a model for layering generations of changes on top of each other in a single inheritance model;
Cloud conversations are evolving at a seemingly ever increasing pace. In my experience, nearly all “…what is the cloud?” type conversations have long since past. In fact, for some organizations, private and public clouds are now central to daily business operations. For the both the early and late majority, however, their (usually large) install base of traditional applications makes the cloud far from reality. These organizations tend to have significant investments in proprietary virtualization, management, and operations technologies, and it’s not a given that these applications are cloud ready (today). While many proprietary technology vendors offer re-packaged versions of existing products to create a thin veil of “cloudiness” – this style of cloud enablement usually comes at a heavy price
The rapid rise of Linux containers as an enterprise-ready technology in 2015, thanks in no small part to the technology provided by the Docker project, should come as no surprise: Linux containers offer a broad array of benefits to the enterprise, from greater application portability and scalability to the ability to fully leverage the benefits of composite applications.
But these benefits aside, Linux containers can, if IT security procedures are not followed, also cause serious harm to mission-critical operations. As Red Hat’s Lars Herrmann has pointed out, containers aren’t exactly transparent when it comes to seeing and understanding all of their internal code. This means that tools and technologies to actually see inside a container are critical to enterprises that want to deploy Linux containers in mission-critical scenarios.
With the release of Red Hat Enterprise Linux 6.7, we’re happy to also announce general availability of Red Hat Access Insights, a new hosted service from Red Hat designed to help customers proactively identify and resolve issues that could impact business operations.
Given the complexity and scale of modern IT environments, we recognize that it can be increasingly complicated to monitor, maintain, and secure enterprise infrastructure. By tapping into Red Hat’s collective experience
As many specialists in the security world know – the RSA Security Conference is one of the biggest security conferences in North America. This year it was once again held in San Francisco at the Moscone Center. Every year the conference gets bigger and bigger, bringing in more and more people and companies from all over the world.
If you attended – you may have noticed that Red Hat had a booth this year. Located in the corner of the main expo floor (not far from some of the “big guys” like: IBM, Microsoft, EMC, CA Technologies, and Oracle) we were in a great location – receiving no shortage of traffic. In fact, despite staffing the booth with six Red Hatters we didn’t have any “down time” – everyone seemed to be interested in what Red Hat has to offer in security.
Over the course of the conference I made a few interesting observations…
Although these days we contribute extensively to Red Hat’s cloud offerings, Red Hat Enterprise Linux remains a core responsibility as the building block for our ecosystem of customers and partners, plus much of Red Hat’s growing product portfolio.
Prior to beginning efforts on Red Hat Enterprise Linux 7 in earnest