Deploying large-scale and real-time predictions
using streaming endpoints with Apache Kafka
Industry
Gaming/Entertainment
Resources
Use-case
Real-time Streaming Endpoints
Computer Vision
TOOLS
With Intel® Tiber™ AI Studio we were able to increase our model throughput by up to 50% and on average by 30% when comparing to RESTful APIs. Intel® Tiber™ AI Studio also allows us to monitor our models in production, set alerts and retrain with high-level automation ML pipelines
Avi Gabay
Director of Architecture at Playtika
About Playtika
Playtika is a Game-Entertainment company that has been leading with top-grossing titles for over five straight years. Playtika has been notorious for its strategic use of data for monetization and game improvement and its investment in machine learning.
Overview
Playtika is a leading Game-Entertainment company that provides audiences around the world with a wide variety of games based on quality, and personalized content. Playtika uses massive amounts of data to reshape the gaming landscape by tailoring UX based on in-game actions. With over 10 million daily active users (DAU), 10 billion daily events and over 9TB of daily processed data, Playtika is able to provide its scientists the data they need to create a continuously changing and adaptive game environment for its users based on their ingame behavior, earning them top-grossing games for the past 5 years.
Challenges
Scaling production of real-time machine learning
Playtika’s large-scale ML processes gather millions of users and thousands of features every minute. The engineering team deployed through batch, and web services, but neither solution could scale to Playtikas needs, or produce predictions in real-time. REST APIs in their ML Pipelines caused service exhaustion, client starvation, handling failures and retries, and performance tuning of bulk size for batch partitioning.
Playtika’s event-driven ecosystem required a solution that could support real-time streaming of their production models and scale without downtime to ensure the best player experience. Playtika’s models also required various integrations including Airflow and Spark processes which would gather features for millions of players and then send requests and events for the model to predict. As demand to answer business needs with ML models increased, Playtika’s AI Research Engineering teams found themselves in need of a robust infrastructure that scales horizontally and can handle bursts, peaks, and fast creation of new ML Pipelines. They needed a solution that integrates quickly and can easily handle all types of deployments including batch, REST API and Kafka streams.
That is when they reached out to Intel® Tiber™ AI Studio as an all purpose deployment solution, that could scale and monitor all models in production.
Solution
One-click streaming endpoints with built-in monitoring and MLOps
Intel® Tiber™ AI Studio AI OS was the perfect solution to handle Playtika’s experiments, scaling and deployments. With Intel® Tiber™ AI Studio, Playtika was able to execute real time predictions with advanced model monitoring features (logs, a/b testing, canary deployment, integration with flows, continual learning and more). Intel® Tiber™ AI Studio organizes every stage of Playtika’s data-science projects, including research, information collection, model development, and model optimization at scale and is used to unify their ML workflow. Using Intel® Tiber™ AI Studio’s MLOps solutions, Playtika was able to bridge the work between their Data Scientists and ML Data Engineers and continuously write, train and deploy machine learning models to various stages (i.e. staging, production) in one click. Intel® Tiber™ AI Studio delivered a scalable solution for streaming endpoints with Apache Kafka, leading to a massive increase in successful throughput, and little to no latency. Intel® Tiber™ AI Studio streaming endpoints provided Playtika with:
Event-at-a-time processing (not microbatch) with millisecond latency
Exactly once processing
Distributed processing and fault-tolerance with fast failover
Reprocessing capabilities so you can recalculate output when your code changes
Enabled Kubernetes backed autoscaling for Kafka s
Results
Intel® Tiber™ AI Studio delivered a simple, and quick solution to Playtika’s unique ML production challenges. Intel® Tiber™ AI Studio reduced technical debt in Playtika’s workflow, and connected data scientists from engineering teams. Intel® Tiber™ AI Studio MLOps solutions, enabled Playtika’s engineering team to easily deploy, update and monitor ML in production to ensure peak performance, and reduced complex triggering and scheduling as data came in. Their data scientists are able to visualize results and business impact in real-time in a unified platform. With multiple deployment options, Intel® Tiber™ AI Studio offered a one click deployment solution which resulted in frictionless deployment solutions for every model in production. Intel® Tiber™ AI Studio enabled Playtika to instantly deploy with Kafka and easily integrated into their existing system. Playtika’s ML services perform better than ever, with native integration with Kubernetes and Apache Kafka allowing them to successfully handle any spike in incoming demand and predict and handle workloads and scale consistently and linearly by adding more pods. Intel® Tiber™ AI Studio streaming endpoints solution increases successful throughput, and reduces latency/errors to zero compared to RESTful API’s.
Increased performance by 40%
Gained up to 50% increase in successful throughput
Reduced latency and error rates to zero
Handled real-time predictions with zero downtime and zero batching
Maximized model computing performance
Reduced divergence from 5TPS to 0.5TPS compared to Web Services
Gain better model computing performance with single thread/process run on single CPU