MLOps / AIOps
Tuesday, October 25, 2022
PRO Workshop (AI): KEYNOTE: Hugging Face -- Hyperproductive Machine Learning with Transformers and Hugging Face
According to the latest State of AI report, "transformers have emerged as a general-purpose architecture for ML. Not just for Natural Language Processing, but also Speech, Computer Vision or even protein structure prediction." Indeed, the Transformer architecture has proven very efficient on a wide variety of Machine Learning tasks. But how can we keep up with the frantic pace of innovation? Do we really need expert skills to leverage these state-of-the-art models? Or is there a shorter path to creating business value in less time?
In this code-level talk, we'll gradually build and deploy a Machine Learning application based on Transformers models. Along the way, you'll learn about the portfolio of open source and commercial Hugging Face solutions, how they can help you become hyper-productive in order to deliver high-quality Machine Learning solutions faster than ever before.
Wednesday, October 26, 2022
PRO TALK (AI): ML Drift Monitoring : What to Observe, How to Analyze & When to Act
Deploying a new ML model in production successfully is a great achievement, but also is the beginning of a persistent challenge to keep them performing at expected levels. Models in product will drift and decay, and the value provided by them to the business will drop. ML drift monitoring is a challenging tasks, from identifying the right data to collect, the right metrics to compute, the right trends to analyze and the right actions to take. This session will explore the process of model drift monitoring, from model instrumentation to determining the next-best-action. Real life challenges will be explored and best practices and recommendations will be discussed.
Thursday, October 27, 2022
OPEN TALK (AI): Shift Left Strategy to Enable Autonomous Data Science
Data Science is hard, achieving ROI from your AI projects is even harder. Data Scientists spend more time wrangling data and slinging models to software and devops engineers than time developing and analyzing their ML models. The solution is to enable a culture shift similar to the DevOps movement where developers manage software quality in production - data scientists should manage ML model performance in production environments. Dedicated ML Engineers are helping to bridge this transition, but they struggle with the tools and automations required to enable scale with autonomy.
Join Manish Modh, Founder & CEO of Andromeda 360 AI on this journey to envision a world of autonomous data science and how Data Scientists and ML Engineers are empowered to own the development, deployment, operations, and performance of their machine learning use cases. Experience the challenges data science teams face today and why most AI projects fail. Learn the art of the possible that leverages all of the wisdom gathered over 20 years of technology evolution from Big Data, Cloud, DevSecOps, AI/ML, and Edge computing
OPEN TALK (AI): Scaling AIaaS: from DALL-E to Uber
As companies begin to embrace AI in key parts of their businesses, they want to explore and scale AI at minimal costs. However developing in-house AI-based solutions for every problem is a complex process and requires huge capital investment. The industry is now embracing AI as a service wherein third party tools can fill in the gaps. In this talk, Daniel will walk through the current landscape, trends, and technical challenges. He will also feature a few customer stories and a proposed modular solution to help your team jumpstart on this journey.
OPEN TALK (AI): Level Up Your Data Lake - to ML and Beyond
A data lake is primarily two things: an object store and the objects being stored. Even with the most basic setup, data lakes are capable of supporting BI, Machine Learning, and operational analytics use cases. This flexibility speaks to the strength of object stores, particularly their flexibility in integrating with a diverse set of data processing engines.
As data lakes exploded in adoption, a number of improvements were made to the first architectures. The first and most obvious improvement was to file formats, which led to the development of analytics-optimized formats like parquet, and eventually modern table formats.
An even newer improvement has been the emergence of data source control tools that bring new levels of manageability across an entire lake! In this talk, we'll cover how to incorporate these technologies into your data lake, and how they simplify workflows critical to ML experimentation, deployment of datasets, and more!
OPEN TALK (AI): Reducing Latency and Resource Consumption for Offline Feature Generation
Personalization is one of the key pillars of Netflix as it enables each member to experience the vast collection of content tailored to their interests. Our personalization system is powered by various machine learning models. We constantly innovate by adding new features to our personalization models and running A/B tests to improve recommendations for our members. We also continue to see that providing larger training sets to our models helps make better predictions. Our ML fact store has enabled us to provide larger training sets where the training set spans over a long time window. While a great success, the ML fact store architecture has its limitations. For example, features computed while generating recommendations must be recomputed by offline feature generation pipelines. This talk is about those limitations and how we enhanced our architecture to run optimized offline feature generation pipelines.
Wednesday, November 2, 2022
[#VIRTUAL] PRO TALK (AI): ML Drift Monitoring : What to Observe, How to Analyze & When to Act
Join on HopinDeploying a new ML model in production successfully is a great achievement, but also is the beginning of a persistent challenge to keep them performing at expected levels. Models in product will drift and decay, and the value provided by them to the business will drop. ML drift monitoring is a challenging tasks, from identifying the right data to collect, the right metrics to compute, the right trends to analyze and the right actions to take. This session will explore the process of model drift monitoring, from model instrumentation to determining the next-best-action. Real life challenges will be explored and best practices and recommendations will be discussed.
[#VIRTUAL] PRO Workshop (AI): Deploying Machine Learning Models with Pulsar Functions
Join on HopinIn this talk I will present a technique for deploying machine learning models to provide real-time predictions using Apache Pulsar Functions. In order to provide a prediction in real-time, the model usually receives a single data point from the caller, and is expected to provide an accurate prediction within a few milliseconds.
Throughout this talk, I will demonstrate the steps required to deploy a fully-trained ML that predicts the delivery time for a food delivery service based upon real-time traffic information, the customer's location, and the restaurant that will be fulfilling the order.
[#VIRTUAL] OPEN TALK (AI): The Enterprise Ready Feature Store: Scaling your Feature Store for Real-time AI/ML
Join on HopinNo longer considered a new concept, ML Feature Stores have existed for several years now, becoming the cornerstone of MLOps platforms. Today, with the rise of Real-time AI and the wide span of AI/ML use cases they enable, It's no wonder then that some companies are already outgrowing their existing Feature Stores. This talk is both for those who are new to Feature Stores and those looking to scale or upgrade their existing implementation. It will explore how to make sure your Feature Store is both future proof and enterprise-ready across supported ML feature types, advanced functionalities as well as infrastructure and operational considerations required to cost-effectively deliver real-time AI/ML use cases with low latency at scale. This talk will cover a range of approaches including building your own feature store, using open source products such as Feast of Feathr, or opting for a commercial Feature Store implementation. Each option will be considered also in the context of the rise of real-time AI and the specific challenges that it creates.
Thursday, November 3, 2022
[#VIRTUAL] PRO TALK (AI): Avoid Mistakes Building AI Products
Based on Gartner's research, 85% of AI projects fail. In this talk, we show the most typical mistakes made by the managers, developers, and data scientists that might make the product fail. We base on ten case studies of products that failed and explain the reasons for each fail. On the other hand, we show how to avoid such mistakes by introducing a few lifecycle changes that make an AI product more probable to succeed.
[#VIRTUAL] OPEN TALK (AI): Shift Left Strategy to Enable Autonomous Data Science
Join on HopinData Science is hard, achieving ROI from your AI projects is even harder. Data Scientists spend more time wrangling data and slinging models to software and devops engineers than time developing and analyzing their ML models. The solution is to enable a culture shift similar to the DevOps movement where developers manage software quality in production - data scientists should manage ML model performance in production environments. Dedicated ML Engineers are helping to bridge this transition, but they struggle with the tools and automations required to enable scale with autonomy.
Join Manish Modh, Founder & CEO of Andromeda 360 AI on this journey to envision a world of autonomous data science and how Data Scientists and ML Engineers are empowered to own the development, deployment, operations, and performance of their machine learning use cases. Experience the challenges data science teams face today and why most AI projects fail. Learn the art of the possible that leverages all of the wisdom gathered over 20 years of technology evolution from Big Data, Cloud, DevSecOps, AI/ML, and Edge computing
[#VIRTUAL] OPEN TALK (AI): Scaling AIaaS: from DALL-E to Uber
Join on HopinAs companies begin to embrace AI in key parts of their businesses, they want to explore and scale AI at minimal costs. However developing in-house AI-based solutions for every problem is a complex process and requires huge capital investment. The industry is now embracing AI as a service wherein third party tools can fill in the gaps. In this talk, Daniel will walk through the current landscape, trends, and technical challenges. He will also feature a few customer stories and a proposed modular solution to help your team jumpstart on this journey.
[#VIRTUAL] OPEN TALK (AI): Level Up Your Data Lake - to ML and Beyond
Join on HopinA data lake is primarily two things: an object store and the objects being stored. Even with the most basic setup, data lakes are capable of supporting BI, Machine Learning, and operational analytics use cases. This flexibility speaks to the strength of object stores, particularly their flexibility in integrating with a diverse set of data processing engines.
As data lakes exploded in adoption, a number of improvements were made to the first architectures. The first and most obvious improvement was to file formats, which led to the development of analytics-optimized formats like parquet, and eventually modern table formats.
An even newer improvement has been the emergence of data source control tools that bring new levels of manageability across an entire lake! In this talk, we'll cover how to incorporate these technologies into your data lake, and how they simplify workflows critical to ML experimentation, deployment of datasets, and more!
[#VIRTUAL] OPEN TALK (AI): Reducing Latency and Resource Consumption for Offline Feature Generation
Join on HopinPersonalization is one of the key pillars of Netflix as it enables each member to experience the vast collection of content tailored to their interests. Our personalization system is powered by various machine learning models. We constantly innovate by adding new features to our personalization models and running A/B tests to improve recommendations for our members. We also continue to see that providing larger training sets to our models helps make better predictions. Our ML fact store has enabled us to provide larger training sets where the training set spans over a long time window. While a great success, the ML fact store architecture has its limitations. For example, features computed while generating recommendations must be recomputed by offline feature generation pipelines. This talk is about those limitations and how we enhanced our architecture to run optimized offline feature generation pipelines.