Image Classification

MobileNet V4

Use case : Image classification

Model description

MobileNetV4 represents the latest evolution in the MobileNet family, introducing Universal Inverted Bottleneck (UIB) blocks that unify various efficient convolution designs. It achieves state-of-the-art accuracy-efficiency trade-offs on mobile hardware.

The architecture features a flexible UIB block design accommodating various operations, optimized through Neural Architecture Search for multiple hardware platforms. It includes Mobile MQA Attention as an efficient attention mechanism for mobile deployment, providing enhanced feature extraction with improved capacity per FLOP.

MobileNetV4 is ideal for state-of-the-art mobile vision applications requiring the latest architectural improvements, though it shows quantization sensitivity (~10% drop) that should be considered for INT8 deployment.

(source: https://arxiv.org/abs/2404.10518)

The model is quantized to int8 using ONNX Runtime and exported for efficient deployment.

Network information

Network Information Value
Framework Torch
MParams ~3.67 M
Quantization Int8
Provenance https://github.com/huggingface/pytorch-image-models
Paper https://arxiv.org/abs/2404.10518

Network inputs / outputs

For an image resolution of NxM and P classes

Input Shape Description
(1, N, M, 3) Single NxM RGB image with UINT8 values between 0 and 255
Output Shape Description
(1, P) Per-class confidence for P classes in FLOAT32

Recommended platforms

Platform Supported Recommended
STM32L0 [] []
STM32L4 [] []
STM32U5 [] []
STM32H7 [] []
STM32MP1 [] []
STM32MP2 [] []
STM32N6 [x] [x]

Performances

Metrics

  • Measures are done with default STEdgeAI Core configuration with enabled input / output allocated option.
  • All the models are trained from scratch on imagenet dataset

Reference NPU memory footprint on Imagenet dataset (see Accuracy for details on dataset)

Model Dataset Format Resolution Series Internal RAM (KiB) External RAM (KiB) Weights Flash (KiB) STEdgeAI Core version
mobilenetv4small_pt_224 Imagenet Int8 224×224×3 STM32N6 539 0 3760.53 3.0.0

Reference NPU inference time on Imagenet dataset (see Accuracy for details on dataset)

Model Dataset Format Resolution Board Execution Engine Inference time (ms) Inf / sec STEdgeAI Core version
mobilenetv4small_pt_224 Imagenet Int8 224×224×3 STM32N6570-DK NPU/MCU 13.74 72.78 3.0.0

Accuracy with Imagenet dataset

Dataset details: link Number of classes: 1000. To perform the quantization, we calibrated the activations with a random subset of the training set. For the sake of simplicity, the accuracy reported here was estimated on the 50000 labelled images of the validation set.

Model Format Resolution Top 1 Accuracy
mobilenetv4small_pt Float 224x224x3 74.33 %
mobilenetv4small_pt Int8 224x224x3 64.24 %

Retraining and Integration in a simple example:

Please refer to the stm32ai-modelzoo-services GitHub here

References

[1] - Dataset: Imagenet (ILSVRC 2012) — https://www.image-net.org/

[2] - Model: MobileNetV4 — https://arxiv.org/abs/2404.10518

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Paper for STMicroelectronics/mobilenetv4_pt