Edge AI Inference in Heterogeneous Constrained Computing: Feasibility and Opportunities

2023 IEEE 28th International Workshop on Computer Aided Modeling and Design of Communication Links and Networks (CAMAD)(2023)

引用 0|浏览15
暂无评分
摘要
The network edge's role in Artificial Intelligence (AI) inference processing is rapidly expanding, driven by a plethora of applications seeking computational advantages. These applications strive for data-driven efficiency, leveraging robust AI capabilities and prioritizing real-time responsiveness. However, as demand grows, so does system complexity. The proliferation of AI inference accelerators showcases innovation but also underscores challenges, particularly the varied software and hardware configurations of these devices. This diversity, while advantageous for certain tasks, introduces hurdles in device integration and coordination. In this paper, our objectives are three-fold. Firstly, we outline the requirements and components of a framework that accommodates hardware diversity. Next, we assess the impact of device heterogeneity on AI inference performance, identifying strategies to optimize outcomes without compromising service quality. Lastly, we shed light on the prevailing challenges and opportunities in this domain, offering insights for both the research community and industry stakeholders.
更多
查看译文
关键词
Edge Artificial Intelligence,Artificial Intelligence Inference,Service Quality,Components Of The Framework,Role Of Artificial Intelligence,Artificial Intelligence Capabilities,Special Case,Performance Metrics,Object Detection,Unmanned Aerial Vehicles,Language Model,Artificial Intelligence Applications,Edge Computing,Height Images,Artificial Intelligence Models,Health Checks,Hardware Platform,Frame Size,Edge Nodes,Edge Devices,Network Latency,Control Plane,Inference Task,Code Generation,Specific Use Case,Acceleration Components,Resource Utilization,1-minute Intervals,Quality Of Service Requirements,Distributed Edge
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要