nexaml commited on
Commit
1a8091e
·
verified ·
1 Parent(s): 0cafe07

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -0
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # ConvNeXt-Tiny
2
+
3
+ Run **ConvNeXt-Tiny** on Qualcomm NPU with [nexaSDK](https://sdk.nexa.ai).
4
+
5
+ ## Quickstart
6
+
7
+ 1. **Install nexaSDK** and create a free account at [sdk.nexa.ai](https://sdk.nexa.ai)
8
+ 2. **Activate your device** with your access token:
9
+
10
+ ```bash
11
+ nexa config set license '<access_token>'
12
+ ```
13
+ 3. Run the model locally in one line:
14
+
15
+ ```bash
16
+ nexa infer NexaAI/convnext-tiny-npu
17
+ ```
18
+
19
+ ## Model Description
20
+ **ConvNeXt-Tiny** is a lightweight convolutional neural network (CNN) developed by Meta AI, designed to modernize traditional ConvNet architectures with design principles inspired by Vision Transformers (ViTs).
21
+ With around **28 million parameters**, it achieves competitive ImageNet performance while remaining efficient for on-device and edge inference.
22
+
23
+ ConvNeXt-Tiny brings transformer-like accuracy to a purely convolutional design — combining modern architectural updates with the efficiency of classical CNNs.
24
+
25
+ ## Features
26
+ - **High-accuracy Image Classification**: Pretrained on ImageNet-1K with strong top-1 accuracy.
27
+ - **Flexible Backbone**: Commonly used as a feature extractor for detection, segmentation, and multimodal systems.
28
+ - **Optimized for Efficiency**: Compact model size enables fast inference and low latency on CPUs, GPUs, and NPUs.
29
+ - **Modernized CNN Design**: Adopts ViT-inspired improvements such as layer normalization, larger kernels, and inverted bottlenecks.
30
+ - **Scalable Family**: Part of the ConvNeXt suite (Tiny, Small, Base, Large, XLarge) for different compute and accuracy trade-offs.
31
+
32
+ ## Use Cases
33
+ - Real-time image recognition on edge or mobile devices
34
+ - Vision backbone for multimodal and perception models
35
+ - Visual search, tagging, and recommendation systems
36
+ - Transfer learning and fine-tuning for domain-specific tasks
37
+ - Efficient deployment in production or research environments
38
+
39
+ ## Inputs and Outputs
40
+ **Input:**
41
+ - RGB image tensor (usually `3 × 224 × 224`)
42
+ - Normalized using ImageNet mean and standard deviation
43
+
44
+ **Output:**
45
+ - 1000-dimensional logits for ImageNet class probabilities
46
+ - Optional intermediate feature maps when used as a backbone
47
+
48
+
49
+ ## License
50
+ - All NPU-related components of this project — including code, models, runtimes, and configuration files under the src/npu/ and models/npu/ directories — are licensed under the Creative Commons Attribution–NonCommercial 4.0 International (CC BY-NC 4.0) license.
51
+ - Commercial licensing or usage rights must be obtained through a separate agreement. For inquiries regarding commercial use, please contact `[email protected]`