By using AWS re:Post, you agree to the AWS re:Post Terms of Use

Questions tagged with ML Ops with Amazon SageMaker and Kubernetes

Kubernetes is an open source system used to automate the deployment, scaling, and management of containerized applications. Kubeflow Pipelines is a workflow manager that offers an interface to manage and schedule machine learning (ML) workflows on a Kubernetes cluster. Using open source tools offers flexibility and standardization, but requires time and effort to set up infrastructure, provision notebook environments for data scientists, and stay up-to-date with the latest deep learning framework versions.

Content language: English

Select up to 5 tags to filter
Sort by most recent

Browse through the questions and answers listed below or filter and sort to narrow down your results.

15 results
Wanted to check if AWS supports GPU inferencing via serverless compute (dynamic loading), since I don't want to spend $1,5/h for EC2 instance, which my client will use not more than 5 minutes per mont...
3
answers
0
votes
66
views
asked 21 days ago
Hello, I am trying to serve a model using SageMaker Endpoint. I am using Triton Inference Server as a framework, I know that I can enable Triton's gRPC protocol communication by setting the `SAGEMA...
1
answers
0
votes
38
views
asked 22 days ago
I was trying to create a sagemaker project using the template "MLOps template for model building, training, and deployment with third-party Git repositories using Jenkins". But I kept getting the err...
2
answers
0
votes
32
views
asked a month ago
Hello AWS team! I am trying to run a suite of inference recommendation jobs leveraging NVIDIA Triton Inference Server on a set of GPU instances (ml.g5.12xlarge, ml.g5.8xlarge, ml.g5.16xlarge) as well...
1
answers
0
votes
674
views
asked 5 months ago
Hello, I am trying to run a suite of inference recommendation jobs on a set of GPU instances (ml.g5.12xlarge, ml.g5.8xlarge, ml.g5.16xlarge) as well as AWS Inferentia machines (ml.inf2.2xlarge, ml.in...
1
answers
0
votes
385
views
asked 5 months ago
Hi, How would one go about designing a serverless ML application in AWS? Currently, our project is using the [serverless framework](https://www.serverless.com/) and lambda functions to accomplish thi...
1
answers
0
votes
465
views
asked a year ago
I want to create a training step in sagemaker pipeline, and use custom processor like below. But instead of python code I want to use java code in the place of [code = 'src/processing.py' ]. Is it po...
1
answers
0
votes
425
views
asked a year ago
I am trying to build a architecture for custom anomaly ai on AWS for my startup. Please let me know if my way of thinking is correct or not 1. Data Ingestion: Ingesting the data into AWS S3 in JSON fo...
1
answers
0
votes
378
views
asked 2 years ago
Calling the sagemaker model endpoint with contentType `application/octet-stream` which is also being captured in Data Capture Logs. What would be the ideal way to transform the data such that model mo...
1
answers
0
votes
638
views
asked 2 years ago
based on aws docs/examples (https://docs.aws.amazon.com/sagemaker/latest/dg/model-registry-version.html), one can create/register model that is generated by your training pipeline. first we need to cr...
1
answers
0
votes
405
views
asked 2 years ago
Hi, I'm working on an end-to-end ml project which, for the moment, goes from training (it takes already processed train/val/test data from an S3 bucket) to deploy, passing through hyperparameter tun...
1
answers
0
votes
341
views
asked 2 years ago
I cant save neuron model after compile the model into an AWS Neuron optimized TorchScript. My code: ``` import tensorflow # to workaround a protobuf version conflict issue import torch import torch....
1
answers
0
votes
467
views
asked 2 years ago
  • 1
  • 2
  • 12 / page