Confirm and Proceed
View More
View Less
System Message
An unknown error has occurred and your request could not be completed. Please contact support.
Reserved - Scan in at least 10 minutes before the beginning of the session.
This has been added to your Planner. Please note: This is not a reserved seat.
Waitlisted - You may be assigned a reserved seat if one becomes available.

Please be sure to check the session schedule for any repeats of this session. In order to search for repeats of this session, please type the Session ID into the search bar at the top of the page.
Personal Calendar
Conference Event
There aren't any available sessions at this time.
Conflict Found
This session is already scheduled at another time. Would you like to...
Please enter a maximum of {0} characters.
{0} remaining of {1} character maximum.
Please enter a maximum of {0} words.
{0} remaining of {1} word maximum.
must be 50 characters or less.
must be 40 characters or less.
Session Summary
We were unable to load the map image.
This has not yet been assigned to a map.
Search Catalog
Replies ()
New Post
Microblog Thread
Post Reply
Your session timed out.
Meeting Summary

ARC302-R1 - [REPEAT 1] AI/ML-powered microservices

Session Description

Companies across industries want to use machine learning models in real time. To do this, customers need to use machine learning models in conjunction with their microservices architectures where high throughput and low latency are paramount. In this session, we cover how Amazon SageMaker enables low-latency real-time inferencing and review considerations needed to achieve satisfactory performance. We discuss how Amazon SageMaker enables large-scale batch inferencing. We then detail how to make these batch inferences available in low-latency environments. We conclude with practical tips and tricks and provide demos that show these techniques in action.

Additional Information
Chalk Talk
300 - Advanced
Please note that session information is subject to change.
Session Schedule