Loading the player...

From siloed ML to ML service mesh using API Management

  • The digital transformation in next decade will be empowered by what we call it as "ML/AI Service Mesh". Even though many companies are now generating features from raw data and extracting business insights with ML models, the challenge has been to share the valuable asset for internal and external consumption at scale. Each project or department in enterprises are siloed in the most of ML/AI projects; building features from raw data, training ML models, extracting embeddings, building prediction microservices, and use it internally. There is no standardized way to share the valuable assets and microservices with cross-functional groups and divisions.
    API management is the missing link for building the service mesh quickly. By introducing a standardized and established way of securing services, enabling service discovery, and observability, Operations teams don’t have to spend much resources on exposing the assets to enable the ML/AI Service Mesh across the enterprise. This approach democratizes the ML assets for faster and scalable enterprise-wide consumption.
    Solution: AI Platform + Apigee Edge
    In this session, learn from a ML model built in the Cloud Machine learning engine and look at ways on how to consume this model from an internal consumer and an external consumer perspective. We use Apigee’s API Management solution to expose the models. This video also touches upon how to build "ML/AI Service Mesh" where enterprises can build a collection of microservices that exposes the features.
    The demo provides:
    - Serving predictions with scalability, performance, and availability in mind
    - Authentication, authorization services depending on who the user is
    - Managing the life cycle of API keys
    - Granting access to your ML APIs with an approval process
    - Rolling out new model versions as models are updated
    - Self-service consumption using Portal without any DevOps involved
    - Monitoring and Analyzing Analytics
    - Monetizing the ML Models
    Speakers: Kaz Sato, Rakesh Talanki
    Google Cloud Next ’20: OnAir → goo.gle/next2020
    Subscribe to the GCP Channel → goo.gle/GCP
    event: Google Cloud Next 2020; re_ty: Publish; product: Cloud - General; fullname: Kaz Sato;

    Category : Openshift


    0 Comments and 0 replies
Privacy policyAccept and close