Western Union is six months into a migration of 900 to 1,200 applications that run across a 3,900-core server fleet. The decision to move came during a period of re-invention at Western Union, a 175-year-old company that is currently working to become more customer-focused and therefore is open to new suppliers to help reach that goal.
The most dangerous assumption in quality engineering right now is that you can validate an autonomous testing agent the same way you validated a deterministic application. When your systems can reason, adapt, and make decisions on their own, that linear validation model collapses.
Red Hat AI Enterprise provides a foundation for modern AI workloads, including AI life-cycle management, high-performance inference at scale, agentic AI innovation, integrated observability and performance modeling, and trustworthy AI and continuous evaluation. Tools are provided for dynamic resource scaling, monitoring, and security.
The foundation for modernization rests on a consistent environment that can host virtual machines and cloud native applications side by side. Red Hat OpenShift and Red Hat OpenStack Platform together create a unified telco cloud that supports high throughput network functions and the container platforms used by modern application teams.
The vulnerabilities exploit a confused deputy attack. An unauthorized user can manipulate a privileged process to perform actions on their behalf, without having the necessary rights themselves. Specifically, attackers abuse tools such as Sudo or Postfix to modify AppArmor profiles via pseudo-files such as /sys/kernel/security/apparmor/.load and .replace.
AI Armor provides dynamic runtime security and relies on a central policy engine in the Universal Management Suite (UMS) to meet compliance requirements, ensuring that organizations can manage their security effectively.
You may have noticed that many European Union (EU) governments and agencies, worried about ceding control to untrustworthy US companies, have been embracing digital sovereignty. Those bodies are turning to running their own cloud and services instead of relying on, say, Microsoft 365 or Google Workspace. If you prize your privacy and want to control your own services, you can take that approach as well.
In order to use agents or in order to use AI in IT operations, all of your systems need to be interconnected and what interconnects all of your systems is an automation platform. Interconnecting systems is only a piece of the puzzle though. There is also some well-founded concern about the autonomous AI systems we are moving towards. AI agents may make decisions and inferences, but enterprises remain hesitant to allow direct execution on production systems.
A North American manufacturer spent most of 2024 and early 2025 doing what many innovative enterprises did: aggressively standardizing on the public cloud by using data lakes, analytics, CI/CD, and even a good chunk of ERP integration. The board liked the narrative because it sounded like simplification, and simplification sounded like savings. Then generative AI arrived, not as a lab toy but as a mandate. "Put copilots everywhere," leadership said. "Start with maintenance, then procurement, then the call center, then engineering change orders."
Ring the bells, sound the trumpet, the Linux 6.19 kernel has arrived. Linus Torvalds announced that "6.19 is out as expected -- just as the US prepares to come to a complete standstill later today, watching the latest batch of televised commercials." Because while the big news in Linux circles might be a new Linux release, Torvalds recognizes that for many people, the "big news [was] some random sporting event." American football, what can you do?
For the longest time, Linux was considered to be geared specifically for developers and computer scientists. Modern distributions are far more general purpose now -- but that doesn't mean there aren't certain distros that are also ideal platforms for developers. What makes a distribution right for developers? Although I consider app compatibility, stability, and flexibility to be essential attributes for most any Linux distribution, developers also need the right tools
Almost a quarter of those surveyed said they had experienced a container-related security incident in the past year. The bottleneck is rarely in detecting vulnerabilities, but mainly in what happens next. Weeks or months can pass between the discovery of a problem and the actual implementation of a solution. During that period, applications continued to run with known risks, making organizations vulnerable, reports The Register.
I've had several incarnations of the self-hosted home lab for decades. At one point, I had a small server farm of various machines that were either too old to serve as desktops or that people simply no longer wanted. I'd grab those machines, install Linux on them, and use them for various server purposes. Here are two questions you should ask yourself:
I recently wrote about my migration away from VirtualBox to KVM/Virt-Machine for my virtual machine needs. I've found those tools to be far superior (albeit with a bit more of a learning curve) than VirtualBox. Since then, however, I've found another method of working with KVM (the Linux kernel virtual machine technology), one that not only allows me to create and manage virtual machines on my local computer, but also from any machine on my LAN. That tool is Cockpit, which makes managing your Linux machines considerably easier.
The updates are installed onto a different (and isolated) system image or subvolume. Once the update finishes successfully, you can switch to the new system by rebooting. Again, if the update isn't 100% successful, it will not happen. And because this all occurs on a separate partition (or image), you don't have to worry about it affecting your system's current state.
Percona recently announced OpenEverest, an open-source platform for automated database provisioning and management that supports multiple database technologies. Launched initially as Percona Everest, OpenEverest can be hosted on any Kubernetes infrastructure, in the cloud, or on-premises. The main goal of the project is to avoid vendor lock-in while still providing an automated private DBaaS. Built on top of Kubernetes operators, it aims to avoid complex deployments that depend on a single cloud provider's technology.