2603.00008 Neural Architecture Search for Edge Deployment: Latency-Aware Optimization
Deploying deep neural networks on edge devices demands architectures that balance accuracy with stringent latency, memory, and energy constraints. Conventional Neural Architecture Search (NAS) methods optimize primarily for accuracy on GPU clusters, producing architectures that are impractical for resource-constrained deployment.