How the Cloud Improves (or Hinders) Application Performance
June 14, 2017

Josh Gray
Cedexis

Writing elegant, effective, flawless code is totally in the hands of the developer. When you accidentally drop in an infinitely recursive loop that takes your whole app down … well, you kick yourself, fix it, and pretty much just hope nobody else saw it.
 
You may even get to have a say in the stack it runs on to ensure there's nothing in there that could get in the way of smooth operation. Nobody loves finding out that the Linux instance is old and buggy, or that somebody forgot to update JQuery to the latest release — but, with a little interrogation and tenacity, things like that can be found and fixed.
 
The big hurdle remaining just outside your grasp, though, is the network your app ultimately flies across: the Internet itself. When everyone in Boston decides to watch the series finale of The Leftovers the moment HBO drops it, and the peering connections between all the local ISPs and HBO's CDN get slammed, your app can suffer if it coincidentally shares these linkages regardless of how much effort you put into making it hyper-efficient.
 
And given the efforts we put in these days to deconstructing monolithic applications, and using distributed microservices to make us more agile, the potential for app performance to take a nosedive because of unseen (and unanticipated) network congestion and outages is only getting greater. If even one of the dozens of APIs you call as the app opens is blocked by network problems, the entire user experience can fall apart.
 
There is help at hand, though, in the form of new ways to program network awareness directly into your code. There are many different words, acronyms, and marketing nonsense naming this, but let's call it Delivery as Code (DAC). Think of it as Infrastructure as Code, but instantiated and refined within the application rather than across the network.
 
The core principle of DAC is pretty simple, actually: don't try to hit any microservice at a particular location. Rather, hit a dynamic traffic manager endpoint that will automatically calculate which of several options to use for that microservice and the best route to get to it. Think of it like hitting a load balancer, instead of a specific server, inside a LAN, except that it's designed to work from the client over the unmanaged Internet, as well as within internally-managed data centers.
 
How could the traffic manager actually know which endpoint to direct your request to? In a perfect world, the answer is that it is part of a broader platform that doesn't just direct traffic — it also measures, analyzes, and acts on a broad swath of user experience data.

For instance, a system that tracks availability, latency, and throughput across and between networks, CDNs, and ISPs, will have a clear sense of where congestion exists, and where the clearest paths between your app and its desired APIs are.

Configured correctly, the system would also take into account all the possible endpoints you have within your own network: that way, it would have the option of directing requests between your own building and the farthest reaches of the Internet, based on an objective virtual map of Internet traffic.
 
By embedding DAC instructions within your code, then, you can remove the uncertainty of network delivery fluctuations from your development process. You can know that each microservice or API call will automatically be directed to the most efficient endpoint — and if your app is running slow it's because of something other than network congestion (ouch).

Josh Gray is Chief Architect at Cedexis
Share this

Industry News

April 25, 2024

JFrog announced a new machine learning (ML) lifecycle integration between JFrog Artifactory and MLflow, an open source software platform originally developed by Databricks.

April 25, 2024

Copado announced the general availability of Test Copilot, the AI-powered test creation assistant.

April 25, 2024

SmartBear has added no-code test automation powered by GenAI to its Zephyr Scale, the solution that delivers scalable, performant test management inside Jira.

April 24, 2024

Opsera announced that two new patents have been issued for its Unified DevOps Platform, now totaling nine patents issued for the cloud-native DevOps Platform.

April 23, 2024

mabl announced the addition of mobile application testing to its platform.

April 23, 2024

Spectro Cloud announced the achievement of a new Amazon Web Services (AWS) Competency designation.

April 22, 2024

GitLab announced the general availability of GitLab Duo Chat.

April 18, 2024

SmartBear announced a new version of its API design and documentation tool, SwaggerHub, integrating Stoplight’s API open source tools.

April 18, 2024

Red Hat announced updates to Red Hat Trusted Software Supply Chain.

April 18, 2024

Tricentis announced the latest update to the company’s AI offerings with the launch of Tricentis Copilot, a suite of solutions leveraging generative AI to enhance productivity throughout the entire testing lifecycle.

April 17, 2024

CIQ launched fully supported, upstream stable kernels for Rocky Linux via the CIQ Enterprise Linux Platform, providing enhanced performance, hardware compatibility and security.

April 17, 2024

Redgate launched an enterprise version of its database monitoring tool, providing a range of new features to address the challenges of scale and complexity faced by larger organizations.

April 17, 2024

Snyk announced the expansion of its current partnership with Google Cloud to advance secure code generated by Google Cloud’s generative-AI-powered collaborator service, Gemini Code Assist.

April 16, 2024

Kong announced the commercial availability of Kong Konnect Dedicated Cloud Gateways on Amazon Web Services (AWS).

April 16, 2024

Pegasystems announced the general availability of Pega Infinity ’24.1™.