In what follows it is assumed that the architecture has evolved into a streaming architecture as outlined in the inference pipeline options that is required for real-time inference at a rate of 10fps. As such any discussion associated with cold start should be associated with a sliding window of data that is used to store/update the nominal/anomalous vectors to/from the datastore.In high throughput pipelines, a calibration phase is run before the steady state and the management plane software can operate this pipeline is turned on.
Introduction
The cold start problem is a problem inherent in any anomaly detection system since these systems require at various degrees prior knowledge about nominal products. Such data are either not present or are not useful at all times as described next, making it challenging for the system to accurately identify anomalies, as it has not yet established a required baseline of nominal vectors in its vector store. To address this problem we need to consider the following requirements:- Each hyper-edge compute node must be able to receive from an edge or centralized datastore a set of nominal vectors that are representative of the nominal data distribution that is specific to the manufacturing machine. The loading of the nominal vectors must be done during the device provisioning stage obviously in a secure manner.
- Each hyper-edge compute node must also be able to update and version control in the vectorstore the latest nominal data to account for new product skews or raw materials batches that can cause a distributional shift in .
- The hyper-edge compute node must be able to self-detect when new nominal vectors are outliers and ignore those. This is especially true given that the manufacturing machine has itself its own cold start phase during which it is calibrated and during which the nominal data is not representative of the steady state. The manufacturing machine cold start and the anomaly detection cold start should be considered two separate processes.

