The Machine Learning (ML) Service provides a common runtime for evaluating machine learning models on device. The service wraps the TensorFlow Lite runtime and provides infrastructure for deployment of trained models. Chromium communicates with ML Service via a Mojo Interface.
You need to provide your trained models to ML Service by following these instructions. You can then load and use your model from Chromium using the client library provided at //chromeos/services/machine_learning/public/cpp/.
The following metrics are currently recorded by the daemon process in order to understand its resource costs in the wild:
TODO(alanlxl): Additional metrics to be added, ideally per-model, to understand levels of performance that clients are getting: