- Newest
- Most votes
- Most comments
What's missing here is "what is the application" (i.e. is it web-based or something else?) and "where are you measuring the latency from" - because it's going to be variable depending on where the requests come from.
The short answer is "host the application in the region which is closest to your users". The longer answer is that there are many thing you can do to reduce latency but in the end you can't beat the speed of light. One answer might be to host in multiple regions but that increases the level of complexity required in your application.
If your application is web-based, consider using CloudFront. Because there are hundreds of points of presence around the world and it provides caching for HTTP objects.
If your application is not web-based you might consider Global Accelerator which uses the AWS global network to carry traffic from the points of presence to the nearest region where your application is hosted.
Relevant content
- asked 4 months ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated 2 years ago
- AWS OFFICIALUpdated 3 months ago
Application is the binance futures trading api endpoint, https://fapi.binance.com I'm measuring latency between raw python get request and response time
P.S. I don't need to beat the speed of light, CloudShell latencies are just fine for me, because in optimized scenario they will be reduced down to ~4 ms, the question is how do i get these unoptimized cloudshell latencies on my EC2 servers
CloudShell is (under the hood) a compute-based service running in an AWS region. You should be able to get similar latency by using EC2 in the region that you're using CloudShell.