Predictive Coding
A Novel Hybrid AI Training Approach Increases Flexibility in Changing Situations
Researchers from Ghent University and Imec have created a revolutionary training process that enables deep neural networks to effectively adapt to changing real-world situations, marking a significant step towards the development of more resilient and dependable artificial intelligence. The novel method helps AI models retain accuracy even in the face of erratic data changes by combining two potent machine learning techniques: predictive coding and backpropagation.
When used outside of the lab, deep neural networks, the technology underlying many contemporary AI applications, frequently perform poorly. When the input data changes, their performance can suffer a great deal, which is a common problem in dynamic contexts. Domain shift is a problem where a model’s accuracy drastically declines due to factors like changes in lighting, sensor drift, or other environmental adjustments. The dependability of AI in crucial applications has been hampered by this issue, particularly on devices with limited resources like embedded systems or smartphones.
This “domain adaptation” issue is addressed by the novel hybrid approach, which combines the unique advantages of two well-established methodologies. This two-step procedure provides a strong yet effective way to modify AI on-device.
You can also read Delta.g Receives £4.6M Seed Funding for Quantum Sensing
A Two-Step Solution: Combining Power and Efficiency
The industry-standard Backpropagation algorithm is used to train a deep neural network offline as the first step in the approach. Backpropagation has a solid performance baseline and is renowned for its capacity to train extremely accurate models. Through this first stage, the network is guaranteed to learn reliable and efficient feature representations from a sizable dataset.
The second stage starts after this high-performing model is established. After that, the model transitions to an online fine-tuning and adaptability strategy based on predictive coding. The innovation is found here. A more computationally efficient learning method that uses error-driven updates and local calculations is called predictive coding. The model employs Predictive Coding to modify its internal parameters and restore lost accuracy when it comes across new data that is different from its initial training set, such as photos with extra noise or inverted colors.
For devices at the “edge,” like smartphones or sensors, which have limited energy and processing power, this hybrid approach works especially well. With Predictive Coding, “efficient on-device domain adaptation” is possible without transferring massive volumes of data to the cloud, as opposed to a resource-intensive full retraining of the model. For real-time applications where continuous adaptation is required, this makes it an attractive choice.
You can also read Uncut Gem platform: NV Centre Diamond Magnetometry platform
Putting the Theory to the Test
Cardoni and Leroux carried out in-depth tests utilizing the MNIST and CIFAR-10 standard image classification datasets to confirm their methodology. They used a variety of network topologies to train their models, such as reduced versions of the popular VGG network and Multi-Layer Perceptron’s. The researchers then added several forms of noise, including color inversions, rotations, and random noise patterns, to the test images to mimic real-world data shifts.
The outcomes showed how successful the hybrid approach was. After being trained using Backpropagation, the models were able to adjust to the new, changed data distributions and achieve their high performance levels through the application of Predictive Coding. The procedure guarantees the stability and reproducibility of their results, including the particular hyperparameters utilized for both training methods, such as learning rates and weight decay. To improve the models’ initial resilience, data normalization and augmentation methods like random cropping and horizontal flipping were also used.
Iteratively minimizing an energy function, which is the total of the prediction errors at every network layer, is the fundamental step in the Predictive Coding adaptation process. The model successfully learns from the new data distribution by modifying layer activities to lower this error, bringing its internal predictions into line with the inputs and intended outputs.
Future Directions and Broader Implications
It is a potential development in the field of continuous learning, increasing the viability of deploying AI systems in the uncertain real world. There are still difficulties, especially when adjusting to deeper and more intricate network designs. The present solution also depends on supervised learning, which necessitates labelled data for adaptation a luxury that isn’t always accessible in practical situations.
In order to increase the technique’s adaptability, future research will look into unsupervised learning and self-supervised learning strategies. Future research will also concentrate on testing their approach’s training durations and energy efficiency on real embedded and neuromorphic hardware specialized processors made to resemble the structure of the brain, which may greatly benefit from Predictive Coding’s local, error-driven nature.
This study lays the road for the next generation of intelligent systems, ones that can continuously learn and evolve on their own, making AI more robust, dependable, and accessible than ever before by developing a computationally efficient method for domain adaptation.
You can also read Diraq Solves Key Obstacle In Silicon Quantum Chip Production