fangmingguo commited on
Commit
4e210f4
·
verified ·
1 Parent(s): 946af25

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +65 -3
README.md CHANGED
@@ -1,3 +1,65 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ base_model:
6
+ - tianweiy/CenterPoint
7
+ pipeline_tag: object-detection
8
+ tags:
9
+ - Axera
10
+ - NPU
11
+ - Pulsar2
12
+ - CenterPoint
13
+ - 3D-Object-Detection
14
+ - LiDAR
15
+ ---
16
+
17
+ # CenterPoint on Axera NPU
18
+
19
+ This repository contains the [CenterPoint](https://arxiv.org/abs/2006.11275) model converted for high-performance inference on the Axera NPU. CenterPoint is a center-based framework for 3D object detection and tracking that represents objects as points, significantly simplifying the detection pipeline while achieving state-of-the-art performance on LiDAR point clouds.
20
+
21
+ This version is optimized with **w8a16** quantization and is compatible with **Pulsar2 version 4.2**.
22
+
23
+ ## Convert Tools Links
24
+
25
+ For model conversion and deployment guidance:
26
+ - [AXera Platform GitHub Repo](https://github.com/AXERA-TECH/centerpoint.axera): Sample code and optimization guides for Axera NPU.
27
+ - [Pulsar2 Documentation](https://pulsar2-docs.readthedocs.io/en/latest/pulsar2/introduction.html): Guide for converting ONNX models to `.axmodel`.
28
+
29
+ ## Support Platforms
30
+
31
+ - **AX650**
32
+ - [M4N-Dock (爱芯派Pro)](https://wiki.sipeed.com/hardware/zh/maixIV/m4ndock/m4ndock.html)
33
+ - [M.2 Accelerator card](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html)
34
+
35
+ | Chips | Model Variant | NPU3 Latency (Per Frame) |
36
+ |---|---|---|---|
37
+ | AX650 | CenterPoint-Pillar | TBD |
38
+
39
+ ## How to Use
40
+
41
+ CenterPoint requires 3D point cloud inputs (typically LiDAR data in `.bin` or `.pcd` format).
42
+
43
+ ### Prerequisites
44
+
45
+ 1. **Environment:** Ensure you have the required Python environment activated with the following core packages installed:
46
+ * **NPU Runtime:** `axengine` (PyAXEngine)
47
+ * **Core Libraries:** `numba` , `opencv-python` and `tqdm`, and.
48
+
49
+ 2. **Model/Data:** Ensure the compiled `.axmodel`, `inference_config.json`, and input data (`inference_data/`) are available on the host.
50
+
51
+ ### Inference Command
52
+
53
+ Run the inference script by providing the compiled model, configuration, and data directory.
54
+
55
+ ```bash
56
+ python inference_axmodel.py compiled.axmodel inference_config.json inference_data/ --output-dir inference_results
57
+ ```
58
+
59
+ ### Inference with AX650 Host
60
+
61
+ ### Results
62
+ The model generates a 3D detection map with bounding boxes oriented in 3D space. Results are saved as images and videos which visualize the ego-vehicle, point cloud data, and detected objects.
63
+
64
+ ### Example Visualization
65
+