Given the recent massive spike in interest in Linux Containers, you could be forgiven for wondering, “Why now?”. It has been argued that the increasingly prevalent cloud computing model more closely resembles hosting providers than traditional enterprise IT, and that containers are a perfect match for this model.
Despite the sudden ubiquity of container technology, like so much in the world of open source software, containerization depends on a long series of previous innovations, especially in the operating system. “One cannot resist an idea whose time has come.” Containers are such an idea, one that has been a long time coming.
The promise of Docker is that it simplifies application deployment, allows greater application density on hosts, and features a portable format that offers unparalleled flexibility over standard packaging. But one thing Docker doesn’t get you is the simplicity of `yum install foo` to install an application. Nor can Docker define or process a directed graph of container orchestration dependencies. We aim to change that.
In a recent blog post on the appc spec, I mentioned Project Atomic’s evolving Nulecule [pronounced: noo-le-kyul] spec as an attempt to move beyond the current limitations of the container model. Let’s dig a bit deeper into that.
Continue reading “The Atomic App Concept…”It All Starts When a Nulecule Comes Out of its Nest””
Linux containers have been getting a lot of hype recently, and it’s easy to understand why. Delivering applications to meet the demands of the business is challenging and containers are disrupting traditional application development and deployment models, enabling businesses to explore new, better ways to deliver products and services.
New innovations like the Docker image format and Kubernetes give you a simpler way to quickly create, package, assemble, and distribute applications. But with hype comes misunderstandings and misconceptions.
By Sayan Saha, Sr. Manager, Product Management, Storage & Data Business, Red Hat.
This week Red Hat announced the general availability of Red Hat Enterprise Linux Atomic Host – a host environment optimized to run containerized applications with a minimal footprint. Red Hat Enterprise Linux Atomic Host simplifies maintenance using image-based update and rollback and includes orchestration toolsets such as kubernetes for managing containers across a cluster of hosts. The new Red Hat Enterprise Linux Atomic Host inherits the industry-leading hardware ecosystem, reliability, stability and security the industry has come to expect from Red Hat Enterprise Linux.
What this means for Red Hat Storage customers
This announcement is significant for Red Hat Storage customers on multiple fronts. Workloads running in containers require persistent storage for application code and data. Given the rapid growth in the number of containers within today’s IT shops, software-defined storage has an advantage over traditional storage…
In November we announced Red Hat Enterprise Linux 7 Atomic Host Public Beta, a small footprint, container host based on Red Hat Enterprise Linux 7. It provides a stable host platform, optimized for running application containers, and brings a number of application software packaging and deployment benefits to customers.
What are the top 7 reasons to deploy containers on Red Hat Enterprise Linux 7 Atomic Host?
Continue reading “Top 7 Reasons to Use Red Hat Enterprise Linux Atomic Host”
Red Hat Enterprise Linux 7 Atomic Host Beta is an operating platform that is optimized and minimized to run containers. It packages key components of Red Hat Enterprise Linux 7 such as SELinux, systemd, and tuned with the kernel to facilitate running containers in a secure and optimized manner. It also offers Kubernetes and Docker to facilitate the rapid creation, deployment, and orchestration of containers – simplifying the life cycle management of applications and systems.
Containers allow users to put application and all of their runtime dependencies into secure packages that are both easy to deploy and easy to manage. Containers are also portable and images of a given container can be copied and replicated to other systems. Since containers are isolated from each other and are isolated from the host OS, libraries and application binaries can be updated individually without affecting other containers or the host OS (and vice versa).
The following video (below) mirrors the demo as presented
Continue reading “Performance Testing Red Hat Enterprise Linux 7 Atomic Host Beta on Amazon EC2”
Developers and system administrators need better ways to deliver applications with increased speed and flexibility. Linux Containers, when used as an open source application packaging and delivery technology, meet this need by combining lightweight application isolation with the flexibility of an image-based deployment method. Red Hat has been working hard to make container technologies safer and easier to consume for the enterprise. Yesterday, at AWS re:Invent, we continued to make progress by offering attendees a chance to dive deep and develop skills for working with containers on AWS at a technical bootcamp.
This full-day, in-person training session provided a chance for developers and system administrators to learn first-hand from Red Hat knowledge experts and gain skills to deploy container-based applications with AWS. Content included instructor-led presentations and practical exercises, with several hands-on labs.
Through a series of labs
Continue reading “AWS re:Invent Bootcamp Attendees Learn How to Accelerate Development with Linux Containers”