DeveloperWeek Global 2020 DeveloperWeek Global 2020

Wednesday, June 17, 2020

ML/AI Service Mesh Made Easy With API Management
Join on Hopin
Rakesh Talanki
Rakesh Talanki
Google, Principal Architect
Kaz Sato
Kaz Sato
Google, Staff Developer Advocate

The digital transformation in next decade will be empowered by what we call it as "ML/AI Service Mesh". Even though many companies are now generating features from raw data and extracting business insights with ML models, the challenge has been to share the valuable asset for internal and external consumption at scale. Each project or department in enterprises are siloed in the most of ML/AI projects; building features from raw data, training ML models, extracting embeddings, building prediction microservices, and use it internally. There is no standardized way to share the valuable assets and microservices with cross-functional groups and divisions.

API management is the missing link for building the service mesh quickly. By introducing a standardized and established way of securing services, enabling service discovery and observability, Operations teams don't have to spend much resources on exposing the assets to enable the ML/AI Service Mesh across the enterprise. This approach will democratize the ML assets for faster and scalable enterprise-wide consumption.

Solution: AI Platform + Apigee Edge
In this session, we will take a ML model built in the Cloud Machine learning engine and look at ways on how to consume this model from an internal consumer and an external consumer perspective. We will use Apigeeā€™s API Management solution to expose the models. We will also touch upon how to build "ML/AI Service Mesh" where enterprises can build a collection of microservices that exposes the features.

The demo will provide:
- Serving predictions with scalability, performance, and availability in mind
- Authentication, authorization services depending on who the user is
- Managing the life cycle of API keys
- Granting access to your ML APIs with an approval process
- Rolling out new model versions as models are updated
- Self-service consumption using Portal without any DevOps involved
- Monitoring and Analyzing Analytics
- Monetizing the ML Models