High latency observed calling feature store and SageMaker service from EKS deployed Java application #4137
tushar-meesho
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
We have recently done an integration with AWS SageMaker and FeatureStore for one of our use-case. On Cloudwatch, we could see the p99: 15-20ms for sagemaker and p99: 30-40ms for feature-store.
But we are observing the high latencies while using JAVA AWS sdk 2.0, to call to feature-store, somewhat around. p99: 150ms for sagemaker and p99: 100ms for feature-store.
We have the containerised spring-boot application running on EKS and here we are calling FS and SM.
Sample code :-
feature-store call
Sagemaker call:
Actual
SageMaker
call method and instrumentation over it.Using AWS java-sdk2 for SageMaker and feature-store.
On cloudwatch for Sagemaker we are using: p99 of ModelLatency + OverheadLatency metric as stated in the docs : docs.aws.amazon.com/sagemaker/latest/dg/…
.And For feature-store we are using Latency metrics as mentioned.
We are benchmarking our application latency from these metrics, but there's a significant difference in the latencies observed on the application side.
This is how we are measuring the end to end p99 latency on application level for sagemaker and feature-store:
histogram_quantile(0.99, sum(rate(resilience4j_circuitbreaker_calls_seconds_bucket{name="awsSageMaker"}[1m])) by (le))
. and same query with method name awsFeatureMaker as well.Could this be because of different VPC(EKS pods and SageMaker both are in the same region though AP_SOUTHEAST_1), or some other form of network latency, etc ? Is there a way to figure the cause of this ?
Beta Was this translation helpful? Give feedback.
All reactions