The latest release of Anypoint Platform includes major upgrades to our logging service for applications running on CloudHub (now a Platform service). With this release, users will be able to retain more log information per application, using a globally distributed infrastructure, and access the logs via an updated UI. Additionally, users who do hybrid management of on-premises servers and applications will now be able to set up High Availability (HA) for their on-prem servers using the new clustering feature.
One of our New Year resolutions is to create more content around specific subject areas. The aim of MuleSoft Blueprints is to provide a complete guide on a specific topic in a way that can be easily read and kept for reference.
To get you started we have a preview of the Load Balancing Mule for Availability and Scalability. This blueprint will introduce the necessary concepts and show how to achieve availability and scale with JMS,
When building Mule architectures a company will often need to run several instances of Mule ESB: Some on QA, some on staging, and on production, perhaps some instances running locally and some others in another continent. Managing Clusters of Mule Servers, keeping track of what application is running where, and knowing what is the health of those instances at a glance, or even being warned when something wrong happens… That is Mule Enterprise Console job!
Mule 3.2 introduced out-of-the-box HA clustering and reliability patterns. Want learn how to get the most out of these new capabilities?
Mike Schilling, MuleSoft’s Lead Architect, shares his in depth knowledge about Mule 3.2’s HA clusters. If you are looking learn the leading tips and tricks for designing and managing clusters, as well as the applications running on the cluster, you should plan to attend MuleSoft’s upcoming webinar titled: High Availability for Integration Applications.
In Mule 3.2 a group of stand-alone Mule instances can be configured to act as a cluster. One or more applications runs in each instance – or node – and the cluster processes requests as if a single unit. A node goes down, the application is still running; the more nodes, the more throughput. And the more nodes, the greater the headache. How many Putty sessions are you already running, let alone a group of new sessions to manage all those nodes?
Mule 3.2 is right around the corner and it is shaping up as the best Mule release ever.
Some highlights include:
- High availability clustering for mission critical environments
- A business event analyzer to gain deep visibility into business events for root cause analysis and compliance
- Drools integration for business rules and complex event processing
Last week, we did a webinar with Terracotta on scaling and managing your web applications running on Tomcat. During the webinar, I did a short demo that showed a Terracotta profile, which makes it super easy to setup the environment required for web applications that use HTTP session clustering using Terracotta.
I previously blogged about how to create a server profile and apply it to servers here. In this case,
We’re pleased to announce the immediate availability of our newest release of Tcat Server 6. This new release includes many fixes, in addition to bundling the Apache Software Foundation’s official release binaries of the newest Tomcat release, version 6.0.26.
Here is a summary of the changes and fixes that are included in the new version of Tomcat, since our last release of Tcat Server 6: