Set Up Inter-Microservice Communication Using RabbitMQ On App Platform For LKE

Create a Linode account to try this guide with a $100 credit.
This credit will be applied to any valid services used during your first 60 days.
Beta Notice
The Akamai App Platform is now available as a limited beta. It is not recommended for production workloads. To register for the beta, visit the Betas page in the Cloud Manager and click the Sign Up button next to the Akamai App Platform Beta.

Introduction

Asynchronous messaging is a common microservice architecture pattern used to decouple inter-service communication. Akamai App Platform uses RabbitMQ to provide an integrated messaging and streaming broker. RabbitMQ is a widely-adopted, open source message broker that uses AMQP (Advanced Message Queuing Protocol) to communicate with producers (apps that send messages) and consumers (apps that receive messages).

This guide provides steps for creating a RabbitMQ cluster on App Platform, as well as building and deploying an example Python application configured to send messages to a RabbitMQ queue using a fanout exchange. The example app consists of a static website built to send messages through the application server to a connected RabbitMQ cluster.

What is a Fanout Exchange?

A “fanout” exchange is a message protocol method that relays messages to all bound queues. The result in this architecture is RabbitMQ sending all messages to all connected clients.

The fanout exchange protocol is built into RabbitMQ and can be used to route one message to multiple users, where a common use case would be sending a single notification to all customers. A consideration for this use case is customer preference; some customers may wish to receive information via email while others prefer to receive messages over SMS or social media network.

To address this, RabbitMQ allows you to bind, or link, each service - email, SMS, social media - to a fanout exchange, thus allowing you to send one message to reach all services. The means that when a message is delivered, queues bound to the exchange receive it and can then process the message to each connected client without needing to send multiple messages to reach each service.

Diagram

  1. A RabbitMQ cluster is created using the RabbitMQ-Cluster Helm chart in the App Platform Catalog.

  2. A Tekton pipeline is created by App Platform that pulls the source code from a GitHub repository. A Buildpacks task is used to build the image and push it to the private registry in Harbor.

  3. An Argo CD application is created by App Platform to deploy the image built by Buildpacks.

  4. The example Python web application sends messages to RabbitMQ’s fanout exchange and binds to the exchange to receive the message.

  5. The Istio virtual service and ingress are created by App Platform to expose the example Python application to the public internet via the NGINX Ingress Controller.

Components

Infrastructure

  • Linode Kubernetes Engine (LKE): LKE is Akamai’s managed Kubernetes service, enabling you to deploy containerized applications without needing to build out and maintain your own Kubernetes cluster.

  • App Platform for LKE: A Kubernetes-based platform that combines developer and operations-centric tools, automation, self-service, and management of containerized application workloads. App Platform for LKE streamlines the application lifecycle from development to delivery and connects numerous CNCF (Cloud Native Computing Foundation) technologies in a single environment, allowing you to construct a bespoke Kubernetes architecture.

Software

  • RabbitMQ: An open source alternative message broker that uses queue-based messaging to provide customization and control over message routing and delivery patterns.

  • Harbor: An open source registry that helps store and manage container images and OCI artifacts like Helm charts. Harbor uses RBAC (role-based access control) for secure artifact and policy management.

  • Argo CD: An open source, declarative, continuous delivery tool for Kubernetes. Argo CD uses the GitOps workflow to continuously monitor applications and compare their current states against desired states in a Git repository.

  • Cloud Native Buildpacks: A buildpack is software used to take application source code and transform it into containerized images for building applications.

  • Istio: An open source service mesh used for securing, connecting, and monitoring microservices.

  • Ingress NGINX Controller: When creating a Service in App Platform, an ingress is created using NGINX’s Ingress Controller to allow public access to internal services.

Prerequisites

  • A Cloud Manager account is required to use Akamai’s cloud computing services, including LKE.

  • Enrollment into the Akamai App Platform’s beta program.

  • An provisioned and configured LKE cluster with App Platform enabled and auto-scaling turned on. An LKE cluster consisting of 3 Dedicated Compute Instances is sufficient for the deployment in this guide to run, but additional resources may be required during the configuration of your App Platform architecture.

    To ensure sufficient resources are available, it is recommended that node pool auto-scaling for your LKE cluster is enabled after deployment. Make sure to set the max number of nodes higher than your minimum. This may result in higher billing costs.

To learn more about provisioning a LKE cluster with App Platform, see our Getting Started with App Platform for LKE guide.

Set Up Infrastructure

Once your LKE cluster with App Platform has been fully deployed, sign in to the App Platform web UI using the platform-admin account, or another account that uses the platform-admin role.

Enable RabbitMQ and Harbor

  1. Select view > platform in the top bar.

  2. Select Apps from the left menu.

  3. Enable the RabbitMQ and Harbor apps by hovering over each app icon and clicking the power on button. It may take a few minutes for the apps to enable.

    Enabled apps move up and appear in color towards the top of the available app list.

    Using Object Storage With Harbor
    Optionally, Harbor can be configured to use object storage when first activated. While not necessary for the demo in this tutorial, it is recommended for production environments.

Create a New Team

Teams are isolated tenants on the platform to support Development and DevOps teams, projects, or even DTAP (Development, Testing, Acceptance, Production). A Team gets access to the Console, including access to self-service features and all shared apps available on the platform.

When working in the context of an admin-level Team, users can create and access resources in any namespace. When working in the context of a non-admin Team, users can only create and access resources used in that Team’s namespace.

  1. Select view > platform.

  2. Select Teams in the left menu.

  3. Click Create Team.

  4. Provide a Name for the Team. Keep all other default values, and click Submit. This guide uses the Team name demo.

Create a RabbitMQ Cluster with Workloads

A Workload is a self-service feature for creating Kubernetes resources using Helm charts from the Catalog.

  1. Switch to your newly created team view by selecting view > team and team > demo in the top bar. You can switch back to team admin as needed by selecting view > team and team > admin.

  2. Select Workloads from the left menu, and then click Create Workload.

  3. Select the Rabbitmq-Cluster Helm chart from the Catalog.

  4. Click Values.

  5. Provide a name for the Workload. This guide uses the Workload name rabbitmq-demo.

  6. Continue with the rest of the default values, and click Submit. The Workload may take a few minutes to become ready.

  7. A Workload is ready when there is a green checkmark in the Status column:

Build the App

This guide uses an example Python chat app to send messages to all connected clients. The steps below pull the example app code from the Linode Git repository and deploy it using the “Build” feature in App Platform.

The example app in this guide is not meant for production workloads, and steps may vary depending on the app you are using.

  1. Select view > team and team > demo in the top bar.

  2. Select Builds, and click Create Build.

  3. Provide a name for the Build. This is the same name used for the image stored in the private Harbor registry of your Team. This guide uses the Build name rmq-example-app with the tag latest.

  4. Select the Mode Buildpacks.

  5. To use the example Python messaging app, provide the following GitHub repository URL:

    https://github.com/linode/apl-examples.git
  6. Set the Buildpacks path to rabbitmq-python.

  7. Click Submit. The build may take a few minutes to be ready.

    Make sure auto-scaling is enabled on your cluster
    When a build is created, each task in the pipeline runs in a pod, which requires a certain amount of CPU and memory resources. To ensure the sufficient number of resources are available, it is recommended that auto-scaling for your LKE cluster is enabled prior to creating the build.

Check the Build Status

The backend status of the build can be checked from the PipelineRuns section of the Tekton app interface:

  1. Select Apps from the left menu, and open the Tekton app.

  2. Click PipelineRuns in the left menu.

  3. Click the link of your build in the list of available Pipelines.

  4. A successful build is denoted with a green check mark and Completed status:

Copy the Image Repository

Once successfully built, copy the image repository link so that you can create a Workload for deploying the app in the next step.

  1. Select Builds to view the status of your build.

  2. When ready, use the “copy” button in the Repository column to copy the repository URL link to your clipboard.

Deploy the App

  1. Select view > team and team > demo in the top bar.

  2. Select Workloads, and click Create Workload.

  3. Select the K8s-Deployment Helm chart from the Catalog.

  4. Click Values.

  5. Provide a name for the Workload. This guide uses the Workload name rmq-example-app.

  6. Click on Values.

  7. Add or edit the following chart values. You may need to uncomment or add additional lines in the Values configuration:

    • Update <image-repo-link> under image > repository, where <image-repo-link> is the repository URL link from the rmq-example-app build in the previous section.

    • Update the name references under each secretKeyRef entry. The format is <workload-name>-rabbitmq-cluster-default-user, where <workload-name> is the name of the RabbitMQ cluster Workload, rabbitmq-demo.

    image:
      repository: <image-repo-link>
      pullPolicy: IfNotPresent
      tag: latest
    env:
      - name: NOTIFIER_RABBITMQ_HOST
        valueFrom:
          secretKeyRef:
            name: <workload-name>-rabbitmq-cluster-default-user
            key: host
      - name: NOTIFIER_RABBITMQ_USER
        valueFrom:
          secretKeyRef:
            name: <workload-name>-rabbitmq-cluster-default-user
            key: username
      - name: NOTIFIER_RABBITMQ_PASSWORD
        valueFrom:
          secretKeyRef:
            name: <workload-name>-rabbitmq-cluster-default-user
            key: password
    
  8. Click Submit. It may take a few minutes for the Workload to be ready.

Allow Traffic to the RabbitMQ Cluster

In order for the RabbitMQ Cluster to be accessible, a Network Policy must be created.

A Network Policy in App Platform is a self-service method of controlling traffic to and from your deployment. Ingress (inbound) policies control access to internal Team pods, and egress (outbound) policies control traffic to external endpoints.

  1. Select Network Policies from the left menu, and click Create NetPol.

  2. Provide a name for the policy. This guide uses the name rabbitmq-example.

  3. Select Rule type ingress using the following values. Make sure to replace rabbitmq-demo with the name you used for the RabbitMQ cluster Workload:

    • Selector label name: otomi.io/app

    • Selector label value: rabbitmq-demo-rabbitmq-cluster

  4. Set the Mode to AllowOnly.

  5. To give permissions to your Team, set the namespace to team-demo or your created team name.

  6. Optionally, you can limit the in-cluster exposure to the app by configuring Workload-level authentication.

    To do this, set the Selector label name to otomi.io/app and the Selector label value to the name of your app’s Workload (i.e. rmq-example-app).

  7. Click Submit.

Expose the App

Create a service to expose the rmq-example-app application to external traffic.

  1. Select Services in the left menu, and click Create Service.

  2. In the Name dropdown menu, select the rmq-example-app service.

  3. Under Exposure, select External.

  4. Click Submit. The service may take a few minutes to be ready.

Access the Demo App

  1. In the list of Services, click the URL for the rmq-example-app application. This should bring you to the live chat demo application in your browser.

  2. Enter an example message (i.e. Hello World!), and click Send:

  3. Once sent, you should your message display in the messages list. This acts as confirmation that your message was received.

More Information

You may wish to consult the following resources for additional information on this topic. While these are provided in the hope that they will be useful, please note that we cannot vouch for the accuracy or timeliness of externally hosted materials.

This page was originally published on


Your Feedback Is Important

Let us know if this guide was helpful to you.


Join the conversation.
Read other comments or post your own below. Comments must be respectful, constructive, and relevant to the topic of the guide. Do not post external links or advertisements. Before posting, consider if your comment would be better addressed by contacting our Support team or asking on our Community Site.
The Disqus commenting system for Linode Docs requires the acceptance of Functional Cookies, which allow us to analyze site usage so we can measure and improve performance. To view and create comments for this article, please update your Cookie Preferences on this website and refresh this web page. Please note: You must have JavaScript enabled in your browser.