This post is brought to you by Command Line Heroes, an original podcast from Red Hat.
My journey, as one might say, in search of the Holy Grail or the great unicorn called DevOps, began well over 20 years ago; yet I never knew it at the time… Actually, it began in 1984 when I was 13 and got my own first computer, a Commodore VIC-20. It wasn’t the first computer that I had ever used but it was mine. I pushed that system with 3 ½ KB of RAM to its limits. Technology has grown by leaps and bounds in such a short timespan. Since those days, I’ve worn many hats. I’ve owned my own company, I’ve helped tech-edit books, and I’ve been pretty active in the open source community. Workshops, social media, MeetUps, Red Hat user groups (RHUGs), virtualization technology user groups (VTUGs), etc. have all allowed me to share and learn at the same time.
Fast forward 10 years, I’m a sergeant in the Army. Sometimes, when in garrison (which was rare), I got tasked out to a bit of side work on UNIX systems simply because nobody knew how to use them. These puppies were coupled together with many pre-internet technologies.
Continue reading “From Manual to Automated DevOps: One Man’s Journey”
Previously in The Quest for Operations Intelligence, the focus was placed on what can be delivered with log aggregation and how to improve it. A conclusion was that to have full situational awareness on IT, you would need logs, metrics, configuration and events information correlated for easy one stop analysis when problems arise.
While we talked about logs, metrics and configuration in depth, we left events at the time without any sort of definition. What are events and what can we use them for in our quest for operations happiness?
Continue reading “Events and Monitoring Supercharging your Operational Intelligence”
As covered in the previous article, The Quest for Operations Intelligence, we have very high expectations from any modern Cloud architecture applications deployed on Red Hat hybrid cloud solutions.
No matter how much support is put into place, the customer needs to be able to operate their hybrid clouds.
After taking a look a correlating all of the available data we reached a conclusion in the previous article that we needed to do something more
Continue reading “Ops Happiness – Harness Data for Operations Intelligence”
We are continuing our series on how Red Hat Keeps the Lights on with Red Hat Virtualization. Please read our previous blog post if you missed any of the series. In this blog post, I will highlight how Red Hat IT uses Red Hat Fuse and JBoss A-MQ to integrate our internal systems that support Accounting, Consulting, Engineering, Finance, Legal, Marketing, Operations, Sales and Training departments. Our internal Enterprise Service Bus (ESB) flows over 100,000 messages a day between these systems with wildly differing interchange and data formats. Our entire ESB deployment runs on virtual machines backed by Red Hat Virtualization. This deployment enables us to scale on demand to meet the changing needs of our business needs and integrated systems.
Continue reading “Red Hat IT Integrates Fuse and A-MQ on Red Hat Virtualization”
Red Hat IT makes extensive use of our own product offerings to effectively manage and to scale our large IT infrastructure. Red Hat Virtualization plays a key role in Red Hat’s overall IT infrastructure, as mentioned in a recent blog by the head of our IT Platform Operations team, Anderson Silva: Red Hat Keeps the Lights on with Red Hat Virtualization
Continue reading “Red Hat IT runs OpenShift Container Platform on Red Hat Virtualization and Ansible”
In our third and final installment (see: part one & part two), let’s take a look at some high-level use cases for Linux containers as well as finally (finally) defending what I like to call “pet” containers. From a general perspective, we see three repeated high-level use cases for containerizing applications:
- The fully orchestrated, multi-container application as you would create in OpenShift via the Red Hat Container Development Kit;
- Loosely orchestrated containers that don’t use advanced features like application templates and Kubernetes; and
- Pet containers.
Continue reading “In Defense of the Pet Container, Part 3: Puppies, Kittens and… Containers”
In our first post defending the pet container, we looked at the challenge of complexity facing modern software stacks and one way that containers address this challenge through aggregation. In essence, the Docker “wrapper” consolidates the next level of the stack, much like RPM did at the component level, but aggregation is just the beginning of what the project provides.
If we take a step back and look at the Docker project in context, there are four aspects that contribute to its exceptional popularity:
- it simplifies the way users interact with the kernel, for features we have come to call Linux containers;
- it’s a tool and format for aggregate packaging of software stacks to be deployed into containers;
- it is a model for layering generations of changes on top of each other in a single inheritance model;
- it adds a transport for these aggregate packages.
Continue reading “In Defense of the Pet Container, Part 2: Wrappers, Aggregates and Models… Oh My!”
It’s been just over three years since Solomon Hykes presented the world with the (so far) most creative way to use the tar command: the Docker project. Not only does the project combine existing container-technologies and make them easier to use, but its well-timed introduction drove an unprecedented rate of adoption for new technology.
Did people run containers before the Docker project? Yes, but it was harder to do so. The broader community was favoring LXC, and Red Hat was working on a libvirt-based model for Red Hat Enterprise Linux. With OpenShift 2, Red Hat had already been running containers in production for several years – both in an online PaaS as well as on-premise for enterprise customers. The model pre-Docker however was fundamentally different from what we are seeing today: rather than enabling completely independent runtimes inside the containers, the approach in
Continue reading “In Defense of the Pet Container, Part 1: Prelude – The Only Constant is Complexity”
The recent release of Red Hat Cloud Suite marked a new milestone for Red Hat and our customers. First, it is the first in what will become a family of suites. Second, it enables enterprise IT to transform their application development and operations toward an agile, innovation center based on hybrid cloud and devops technologies. Curating a broad set of open source technologies, Red Hat Cloud Suite offers a turnkey Cloud solution with a container-based app-development platform, private-cloud infrastructure, and a common management framework. Specifically, Red Hat Cloud Suite includes
Continue reading “Cloud Solutions Made Simple”
Some time ago, two different projects were started in the open source community, namely: Ipsilon and Keycloak. These projects were started by groups with different backgrounds and different perspectives. In the beginning, it seemed like these two projects would have very little in common… though both aimed to include
Continue reading “Red Hat Federation Story: Ipsilon & Keycloak… a “Clash of the Titans””