PoPI: A Machine Learning Consensus Mechanism for Blockchain IoT

by Priyanka Patel

The rapid expansion of the Internet of Things (IoT) has created a fundamental architectural tension. While we are surrounding ourselves with billions of connected sensors and actuators, the “brains” of these operations remain stubbornly centralized in cloud servers. This dependency creates a precarious single point of failure and drives up operational costs as networks scale, leaving critical infrastructure vulnerable to outages, and bottlenecks.

Blockchain technology has long been proposed as the antidote, offering a decentralized ledger that removes the need for a central authority. However, the “consensus mechanisms” that allow blockchains to agree on data—such as the energy-intensive mining used by Bitcoin—are typically far too demanding for the modest batteries and limited processing power of a remote environmental sensor or a wearable health monitor.

A new approach called Proof of Periodic Inference (PoPI) aims to resolve this deadlock. By utilizing a machine learning-based consensus mechanism for blockchain-enabled IoT systems, PoPI replaces constant, heavy computation with a streamlined, periodic selection process. This allows resource-constrained devices to maintain a secure, decentralized network without draining their batteries in a matter of hours.

The system shifts the burden of proof from raw computational power to a predictive suitability score. Instead of every node fighting to produce every block, PoPI uses a supervised machine learning model to rank nodes based on their current health—considering factors like battery levels, network bandwidth, and historical reliability—and selects a group of producers to handle the workload over a defined period.

Solving the IoT Resource Gap

For most IoT devices, energy is the primary constraint. A sensor placed in a remote forest or a medical implant cannot be recharged daily. Traditional consensus protocols often overlook these dynamic conditions, treating every node as if it has a constant power supply and a stable connection. When a selected node unexpectedly shuts down due to power failure, the entire network can experience significant latency.

PoPI addresses this by introducing a rotating set of “supervisor nodes.” These supervisors act as coordinators, collecting “interest messages” from active devices. These messages contain real-time hardware and network status, ensuring that only nodes with the actual capacity to produce a block are chosen. To maintain decentralization, the role of the supervisor is temporary and rotates among the network’s most reliable participants.

Fig 1. High-level illustration of the workflow of our proposed mechanism.

By grouping producers rather than picking a single winner for every block, the system drastically reduces the number of times the machine learning model must run. The size of this producer group is scaled to the square root of the number of active nodes, balancing the need for speed with the necessity of security. To prevent targeted attacks, the actual order in which these group members produce blocks is randomized using the hash of the previous block, meaning a node only knows It’s “up next” at the very last moment.

Fairness and the ‘Starvation’ Problem

One of the persistent risks in decentralized systems is the “rich get richer” phenomenon, where a few high-capacity nodes dominate the consensus process, leading to a new form of centralization. PoPI implements a set of “fair participation” features to ensure the network remains truly distributed.

The mechanism tracks “Starvation Duration”—a metric that increases every time a capable node expresses interest but is not selected. As a node’s starvation duration grows, its probability of being chosen for the next cycle increases. This is balanced by a “Longevity Probability” function, which rewards nodes for their long-term presence in the system without allowing established nodes to permanently block newcomers.

These fairness metrics are fed into the ML model alongside performance data. If a new node joins the network and behaves honestly, the system gradually increases its suitability score. Conversely, if a node fails to produce a block or provides false state information, the model penalizes its score, reducing its future chances of selection.

thumbnail
Fig 3. Illustration of the duties of supervisor nodes during their term.

Measuring Performance in Dynamic Networks

To test the efficacy of this approach, PoPI was compared against five other state-of-the-art mechanisms, including Proof of Elapsed Time (PoET) and the Proof of Evolutionary Model (PoEM). The results indicate that the periodic grouping strategy significantly outperforms linear selection models, particularly as the network grows.

In simulations involving 1,000 nodes, PoPI maintained a block latency of 0.25 seconds, while competing models like PoEM saw latency rise to 0.37 seconds. More importantly, PoPI achieved a throughput of 610 transactions per second (tps) at that scale—roughly 33% higher than PoEM. This efficiency is largely attributed to the fact that the system does not need to perform expensive ML inferences for every single block.

Performance Comparison at 1,000 Nodes
Mechanism Block Latency (s) Throughput (tps) Computational Overhead
PoPI 0.25 610 Lowest
PoEM 0.37 465 High
CE-PBFT 0.50 364 Medium
PoET 0.89 85 Low

From an energy perspective, PoPI allows non-group nodes to remain in low-power states for longer periods. In a 1,000-node scenario, the system ensured that 56% of all nodes produced at least one block, a significantly higher participation rate than trust-based models like DT-DPoS, where only 11% of nodes were ever selected.

Security and the ‘Black Box’ Advantage

A common critique of ML-based selection is predictability. If an attacker can predict which node will be the next block producer, they can launch a targeted Distributed Denial of Service (DDoS) attack to halt the network. PoPI mitigates this by treating the ML model as a “black box.”

Since the model weights are updated incrementally after every production cycle based on real-world success and failure, the parameters change constantly. The dynamic features (like current battery and bandwidth) are transmitted only to the rotating supervisors and are not globally accessible. This makes it mathematically improbable for an external adversary to reconstruct the inference inputs and predict the next set of producers.

To prevent the supervisors themselves from becoming a point of corruption, all data exchanges—including the “interest messages” and the final inference results—are recorded as digitally signed transactions on the blockchain. This creates an immutable audit trail, allowing any node in the network to verify that the selection process was conducted fairly and based on authentic data.

While the current framework has been validated in custom simulation environments, the next phase of development involves testing PoPI against real-world IoT traces and production-grade deployments. Researchers are also exploring the leverage of federated learning to further decentralize the model training process, reducing the reliance on any single coordinating server.

We invite you to share your thoughts on the future of decentralized IoT in the comments below.

You may also like

Leave a Comment