Reading Time: 13 minutes

We often expose the proxy APIs that connect applications to their backend APIs. With a proxy API, the application continues to run without issue and continue to call and connect to the backend API while a is editing it. Exposing a proxy API also protects the backend API from the world, shielding the real IP address.

The advantage of using a proxy is having a layer of separation to ensure that any attacks against our API are stopped well before anyone interacts with our main servers. This creates extra protection for our existing APIs.

latest report
Learn why we are the Leaders in API management and iPaaS

API Gateway acts as a proxy server that is dedicated to hosting the proxy applications and for Lightning Connect to gather all the existing backend APIs together which are either hosted in our on-premises standalone server or CloudHub.

The advantage of API Gateway is that it automatically creates an application that can proxy the backend API from the URL that the backend API exposes, and we do not need to write any code for it. Not only that, through API Manager, we can implement various runtime policies on HTTP/HTTPS endpoints to govern our proxy API.

API Gateway solution is also very versatile, as it can be implemented both in CloudHub as well as in on-premises.

Here we will demonstrate implementing a proxy application on an on-premises Data Gateway server that will connect/communicate with a backend API that is deployed on an on-premises standalone server with a runtime policy.

Flow

First, we start the demonstration by creating a simple application in our studio and deploying it on our on-premises standalone server. We then create our proxy application in the API Manager interface, apply the required policies on it, and finally deploy the proxy application to an on-premises API Gateway in our system.

Creating and deploying application:-

So, let's create a sample application in our studio as follows:-

pic2

Once it is finished, we create a Mule application deployable zip file as follows:-

pic3

After the zip file is created we will copy the zip file in the {MULE_HOME}/apps folder and deploy the application in the on-premises standalone server as follows:-

pic4

Now, if we test our application on a REST client like Postman, we can see the following responses with the following URL http://localhost:9091/testapp :-

pic5

This means our backend application is ready and running successfully on an on-premises standalone server!

Creating Proxy application for Standalone Data Gateway server:-

Now as our backend service is ready and running on an on-premises standalone server, we don't want to expose this API URL to the world. Instead, we can deploy a proxy service to our on-premises proxy server which we can then finally expose to the outside world. Additionally, we should apply some policies or rules with this proxy URL that is exposed to the client.

We will be configuring the proxy application, and it's policies via the API Manager interface. To begin the process, we need to log in to our Anypoint Platform account to access the interface.

Creating a Proxy application:

We need to go to the API Manager and create an API as follows:-

pic6

After we create an API, we will see that API Manager provided an option to configure our endpoints:-

pic7

As we select the configure endpoint option, we will find options like implementation URI asking for a backend service URL and down the line port and path for the proxy application:-

pic8

Adding on-premises DataGateway server to CoudHub:-

In the server section we can add our on-premises Data Gateway server with the Runtime Manager interface:-

pic9

Now we move to the /bin folder of our on-premises Data Gateway server located in our system and add the above-given command in command prompt and execute it as follows:-

pic10

Now, we will find that our Data Gateway is successfully added to our Runtime Manager interface as below:-

pic11

So, if we start our on-premises Data Gateway server in our system it will reflect in our Runtime Manager interface as below:-

pic12

Applying policies to our proxy application:-

Now, back to the API Manager where we configured our proxy application, we will find an option to apply different policies on it.

A policy is a mechanism/rule for enforcing filters on traffic. These filters are used to control things like authentication, access, allotted consumption, and SLAs. There can be custom policies or some pre-built policies like Rate Limiting, Throttling, OAuth 2, Basic HTTP Authentication, and etc… for APIs:-

pic13

We are going to apply the Rate Limiting policy on our proxy application as shown above.

The Rate Limiting policy specifies the maximum value for the number of messages processed per period and rejects any messages beyond the maximum. Therefore, it will apply rate limiting to all API calls regardless of the source and thus control our proxy API.

pic14

So, here we will configure two requests for our proxy API per minute as we can see above.

Deploying proxy application from API Manager interface to on-premises server directly:-

After applying the Rate Limiting policy, we will deploy the proxy application from API Manager interface to directly our on-premises Data Gateway server by selecting Deploy Proxy option we can see below:-

pic15

We need to select our on-premises Data Gateway proxy server which is just registered with the API Manager interface and than we need to click Deploy Proxy button as shown below:-

pic16

The API Manager interface will directly deploy our proxy application to our on-premises Data Gateway server located in our system and will show the status as follows:-

pic17

We can also see the change in our system in the Gateway server console:-

pic18

Testing our proxy API:-

In the final stage, we will now test our proxy application deployed in our proxy server as well as the policy/rule applied on it.

So, we will hit the proxy URL:- http://localhost:8081/proxyapp on a REST client like Postman as follows and see the result:-

pic19

And voila! We are getting the response from our backend API!

Testing the rule applied on proxy API:-

As we already implemented a Rate Limiting policy on our proxy API, if we hit more than two requests per minute we will get the following restriction below:-

pic21

Adding Version:-

API Manager has a major feature of adding versions to our proxy APIs. This helps to maintain backward compatibility across different API versions. We can also add different versions of our proxy APIs, each of which will be compatible with former versions of the backend API:-

pic22

Conclusion:-

As you can see, we were able to easily create a proxy application for our backend application and apply various rules over it without writing a single line of code! API Gateway acts as a proxy server brilliantly and helps us create and control our proxy application without putting additional effort in writing more code. Moreover, the Runtime Manager interface can add and control our on-premises server and can deploy applications directly from there with minimum effort! Another important aspect of most gateway vendors are is that they force us to transit our traffic through *their* premises while API Gateway doesn't. In this case, it doesn't require API traffic to flow through a third party in the cloud, but instead only goes directly through on-premises servers. The cloud aspect is only for management and control.

Hope you liked the article, and please feel free to share your feedback and experiences in the comment section below.