Apply Conf 2022
29. Intelligent Customer Preference Engine

29. Intelligent Customer Preference Engine with Real-time ML systems, Manoj & Praveen, Walmart Lab

https://www.youtube.com/watch?v=WPKWRXU0bOQ&ab_channel=Tecton (opens in a new tab)

  • Customers are increasingly omnichannel

    We need to understand behaviour across the whole channels

    Personalization plays a key role throughout the customer journey, and also after the purchase in case of recommendation

    Dozen of recommenders in play in the background, to make baskets very quickly. Adding each item can be time-consuming, so personalisation is right from the home page.

    Product discovery is another recommender use case.

    Aside from product impressions, we also prioritize non-product impressions like banners, to show that we’re to help find items they’re looking for.

  • 2 types of recommendations: baskets and checkout/last-call (candies, batteries etc).

    For last-call, we don’t want customers going back to the discovery again, but go ahead and checkout

  • Flywheel

    Screen Shot 2022-05-31 at 16.23.25.png

  • 1:1 micro intent: for a given time, for a given customer

    ex of micro intent: Party supplies, then ice cream, then gift cards

    detecting micro-intent with the inference engine

    Screen Shot 2022-05-31 at 16.24.49.png

  • Features

    Screen Shot 2022-05-31 at 16.26.40.png

    Variable time windows for online predictions

  • Ranking using some context

    If a user already has an iPhone and search for AirPods, we’ll propose Apple ones first.

  • Recommender flow

    Screen Shot 2022-05-31 at 16.30.43.png

    Screen Shot 2022-05-31 at 16.33.37.png

  • Feature and model store

    Model store allows to A/B test and to iterate quickly on different models

    Screen Shot 2022-05-31 at 16.35.05.png

  • Online inferencing platform

    Multi-model, DS can bring their framework and model on it and spin up a microservice easily

    Can tie different models together in an ensemble fashion

    Inference graph is very high-level: expose human-readable API to our clients. Most of the time, embedding and transformations are only understandable by the fews who built the model. Given a feature, it can bundle features together using the feature store.

    Post-processing is a filtering mechanism of the recall set, response to the client

    Real-time platform is both async and real-time for different use-cases.

    Screen Shot 2022-05-31 at 16.37.44.png

  • Q&A

    • Does the feature store contain pre-computed features for online input? Or differents need to speed up inference.

      Yes precomputed. Feature store has many purposes: saving compute costs, storing final and intermediate features for reusability purposes.