Integrating AI inference into wireless sensing edge networks presents notable challenges due to limited resources, changing environments, and diverse devices. In this study, we proposed a novel resource allocation framework that enhances energy efficiency, reduces latency, and ensures fairness across distributed edge nodes for AI inference. The framework models a multi-objective optimization problem that reflects the interdependence of computation, communication, and energy at each device. We also develop a decentralized algorithm based on dual decomposition and projected gradient ascent, by using local data. The extensive simulations demonstrate that our proposed method reduces the average inference latency by 31.4% and energy consumption by 27.8% compared to the greedy and round-robin techniques. The system utility is improved by up to 59.2%, and fairness, measured using Jain’s index, remains within 8% of the ideal. Additionally, throughput analysis further confirms that our approach gains up to 49 tasks/sec, outperforming existing strategies by more than 40%. These findings show that the resource-aware AI inference approach is scalable, energy-efficient, and appropriate for real-time use in multi-user wireless edge networks.
扫码关注我们
求助内容:
应助结果提醒方式:
