Edge AI Inference Optimization: Quantization and Pruning on Resource Constrained Platforms
Keywords:
Edge AI Optimization, Neural Network Quantization, Model Pruning, Embedded Deep Learning, Resource-Constrained InferenceAbstract
The deployment of sophisticated artificial intelligence models on resource-constrained embedded systemspresents fundamental challenges in balancing computational efficiency with accuracy preservation. Contemporary edge devices, including
References
Yiwen Guo, et al., "Dynamic Network Surgery for Efficient DNNs," arxiv, 2016. Available: https://arxiv.org/abs/1608.04493
Downloads
Published
2025-12-11
How to Cite
Ishan Pardesi. (2025). Edge AI Inference Optimization: Quantization and Pruning on Resource Constrained Platforms . Journal of Computational Analysis and Applications (JoCAAA), 34(12), 385–393. Retrieved from https://www.eudoxuspress.com/index.php/pub/article/view/4367
Issue
Section
Articles


