All Content tagged with ML Ops with Amazon SageMaker and Kubernetes

Kubernetes is an open source system used to automate the deployment, scaling, and management of containerized applications. Kubeflow Pipelines is a workflow manager that offers an interface to manage and schedule machine learning (ML) workflows on a Kubernetes cluster. Using open source tools offers flexibility and standardization, but requires time and effort to set up infrastructure, provision notebook environments for data scientists, and stay up-to-date with the latest deep learning framework versions.

Content language: English

Select tags to filter
Sort by most recent
16 results
Wanted to check if AWS supports GPU inferencing via serverless compute (dynamic loading), since I don't want to spend $1,5/h for EC2 instance, which my client will use not more than 5 minutes per mont...
3
answers
0
votes
294
views
asked 4 months ago
Hello, I am trying to serve a model using SageMaker Endpoint. I am using Triton Inference Server as a framework, I know that I can enable Triton's gRPC protocol communication by setting the `SAGEMA...
1
answers
0
votes
213
views
asked 4 months ago
I was trying to create a sagemaker project using the template "MLOps template for model building, training, and deployment with third-party Git repositories using Jenkins". But I kept getting the err...
2
answers
0
votes
55
views
asked 5 months ago
Hello AWS team! I am trying to run a suite of inference recommendation jobs leveraging NVIDIA Triton Inference Server on a set of GPU instances (ml.g5.12xlarge, ml.g5.8xlarge, ml.g5.16xlarge) as well...
1
answers
0
votes
719
views
asked 9 months ago
Hello, I am trying to run a suite of inference recommendation jobs on a set of GPU instances (ml.g5.12xlarge, ml.g5.8xlarge, ml.g5.16xlarge) as well as AWS Inferentia machines (ml.inf2.2xlarge, ml.in...
1
answers
0
votes
461
views
asked 9 months ago
Hi, How would one go about designing a serverless ML application in AWS? Currently, our project is using the [serverless framework](https://www.serverless.com/) and lambda functions to accomplish thi...
1
answers
0
votes
500
views
asked 2 years ago
I want to create a training step in sagemaker pipeline, and use custom processor like below. But instead of python code I want to use java code in the place of [code = 'src/processing.py' ]. Is it po...
1
answers
0
votes
456
views
asked 2 years ago
I am trying to build a architecture for custom anomaly ai on AWS for my startup. Please let me know if my way of thinking is correct or not 1. Data Ingestion: Ingesting the data into AWS S3 in JSON fo...
1
answers
0
votes
413
views
asked 2 years ago
Calling the sagemaker model endpoint with contentType `application/octet-stream` which is also being captured in Data Capture Logs. What would be the ideal way to transform the data such that model mo...
1
answers
0
votes
666
views
asked 2 years ago
based on aws docs/examples (https://docs.aws.amazon.com/sagemaker/latest/dg/model-registry-version.html), one can create/register model that is generated by your training pipeline. first we need to cr...
1
answers
0
votes
424
views
asked 2 years ago
Hi, I'm working on an end-to-end ml project which, for the moment, goes from training (it takes already processed train/val/test data from an S3 bucket) to deploy, passing through hyperparameter tun...
1
answers
0
votes
363
views
asked 2 years ago
  • 1
  • 2
  • 12 / page