Wanli
commited on
Commit
·
af3dd88
1
Parent(s):
f8da712
update and modify documentation (#161)
Browse files- README.md +4 -0
- benchmark/README.md +1 -1
- models/face_detection_yunet/README.md +3 -3
- models/facial_expression_recognition/README.md +1 -1
- models/handpose_estimation_mediapipe/README.md +5 -4
- models/human_segmentation_pphumanseg/README.md +1 -1
- models/license_plate_detection_yunet/README.md +1 -1
- models/object_detection_nanodet/README.md +1 -1
- models/object_detection_yolox/README.md +1 -1
- models/object_tracking_dasiamrpn/README.md +1 -1
- models/palm_detection_mediapipe/README.md +6 -5
- models/person_detection_mediapipe/README.md +3 -3
- models/person_reid_youtureid/README.md +1 -1
- models/person_reid_youtureid/demo.py +4 -4
- models/qrcode_wechatqrcode/README.md +1 -1
- models/text_detection_db/README.md +1 -1
- models/text_recognition_crnn/README.md +1 -1
- tools/eval/README.md +2 -2
README.md
CHANGED
@@ -82,6 +82,10 @@ Some examples are listed below. You can find more in the directory of each model
|
|
82 |
|
83 |

|
84 |
|
|
|
|
|
|
|
|
|
85 |
### QR Code Detection and Parsing with [WeChatQRCode](./models/qrcode_wechatqrcode/)
|
86 |
|
87 |

|
|
|
82 |
|
83 |

|
84 |
|
85 |
+
### Person Detection with [MP-PersonDet](./models/person_detection_mediapipe)
|
86 |
+
|
87 |
+

|
88 |
+
|
89 |
### QR Code Detection and Parsing with [WeChatQRCode](./models/qrcode_wechatqrcode/)
|
90 |
|
91 |

|
benchmark/README.md
CHANGED
@@ -79,7 +79,7 @@ Benchmark is done with latest `opencv-python==4.7.0.72` and `opencv-contrib-pyth
|
|
79 |
| [YoutuReID](../models/person_reid_youtureid) | Person Re-Identification | 128x256 | 30.39 | 625.56 | 11117.07 | 195.67 | 898.23 | 14886.02 | 90.07 | 44.61 | 5.58 | --- |
|
80 |
| [MP-PalmDet](../models/palm_detection_mediapipe) | Palm Detection | 192x192 | 6.29 | 86.83 | 872.09 | 38.03 | 142.23 | 1191.81 | 83.20 | 33.81 | 5.17 | --- |
|
81 |
| [MP-HandPose](../models/handpose_estimation_mediapipe) | Hand Pose Estimation | 224x224 | 4.68 | 43.57 | 460.56 | 20.27 | 80.67 | 636.22 | 40.10 | 19.47 | 6.27 | --- |
|
82 |
-
| [MP-PersonDet](
|
83 |
|
84 |
\*: Models are quantized in per-channel mode, which run slower than per-tensor quantized models on NPU.
|
85 |
|
|
|
79 |
| [YoutuReID](../models/person_reid_youtureid) | Person Re-Identification | 128x256 | 30.39 | 625.56 | 11117.07 | 195.67 | 898.23 | 14886.02 | 90.07 | 44.61 | 5.58 | --- |
|
80 |
| [MP-PalmDet](../models/palm_detection_mediapipe) | Palm Detection | 192x192 | 6.29 | 86.83 | 872.09 | 38.03 | 142.23 | 1191.81 | 83.20 | 33.81 | 5.17 | --- |
|
81 |
| [MP-HandPose](../models/handpose_estimation_mediapipe) | Hand Pose Estimation | 224x224 | 4.68 | 43.57 | 460.56 | 20.27 | 80.67 | 636.22 | 40.10 | 19.47 | 6.27 | --- |
|
82 |
+
| [MP-PersonDet](../models/person_detection_mediapipe) | Person Detection | 224x224 | 13.88 | 98.52 | 1326.56 | 46.07 | 191.41 | 1835.97 | 56.69 | --- | 16.45 | --- |
|
83 |
|
84 |
\*: Models are quantized in per-channel mode, which run slower than per-tensor quantized models on NPU.
|
85 |
|
models/face_detection_yunet/README.md
CHANGED
@@ -28,7 +28,7 @@ Run the following command to try the demo:
|
|
28 |
# detect on camera input
|
29 |
python demo.py
|
30 |
# detect on an image
|
31 |
-
python demo.py --input /path/to/image
|
32 |
|
33 |
# get help regarding various parameters
|
34 |
python demo.py --help
|
@@ -40,13 +40,13 @@ Install latest OpenCV and CMake >= 3.24.0 to get started with:
|
|
40 |
|
41 |
```shell
|
42 |
# A typical and default installation path of OpenCV is /usr/local
|
43 |
-
cmake -B build -D OPENCV_INSTALLATION_PATH
|
44 |
cmake --build build
|
45 |
|
46 |
# detect on camera input
|
47 |
./build/demo
|
48 |
# detect on an image
|
49 |
-
./build/demo -i=/path/to/image
|
50 |
# get help messages
|
51 |
./build/demo -h
|
52 |
```
|
|
|
28 |
# detect on camera input
|
29 |
python demo.py
|
30 |
# detect on an image
|
31 |
+
python demo.py --input /path/to/image -v
|
32 |
|
33 |
# get help regarding various parameters
|
34 |
python demo.py --help
|
|
|
40 |
|
41 |
```shell
|
42 |
# A typical and default installation path of OpenCV is /usr/local
|
43 |
+
cmake -B build -D OPENCV_INSTALLATION_PATH=/path/to/opencv/installation .
|
44 |
cmake --build build
|
45 |
|
46 |
# detect on camera input
|
47 |
./build/demo
|
48 |
# detect on an image
|
49 |
+
./build/demo -i=/path/to/image -v
|
50 |
# get help messages
|
51 |
./build/demo -h
|
52 |
```
|
models/facial_expression_recognition/README.md
CHANGED
@@ -22,7 +22,7 @@ Results of accuracy evaluation on [RAF-DB](http://whdeng.cn/RAF/model1.html).
|
|
22 |
Run the following command to try the demo:
|
23 |
```shell
|
24 |
# recognize the facial expression on images
|
25 |
-
python demo.py --input /path/to/image
|
26 |
```
|
27 |
|
28 |
### Example outputs
|
|
|
22 |
Run the following command to try the demo:
|
23 |
```shell
|
24 |
# recognize the facial expression on images
|
25 |
+
python demo.py --input /path/to/image -v
|
26 |
```
|
27 |
|
28 |
### Example outputs
|
models/handpose_estimation_mediapipe/README.md
CHANGED
@@ -10,7 +10,7 @@ This model is converted from TFlite to ONNX using following tools:
|
|
10 |
|
11 |
**Note**:
|
12 |
- The int8-quantized model may produce invalid results due to a significant drop of accuracy.
|
13 |
-
- Visit https://
|
14 |
|
15 |
## Demo
|
16 |
|
@@ -19,7 +19,7 @@ Run the following commands to try the demo:
|
|
19 |
# detect on camera input
|
20 |
python demo.py
|
21 |
# detect on an image
|
22 |
-
python demo.py -i /path/to/image
|
23 |
```
|
24 |
|
25 |
### Example outputs
|
@@ -32,6 +32,7 @@ All files in this directory are licensed under [Apache 2.0 License](./LICENSE).
|
|
32 |
|
33 |
## Reference
|
34 |
|
35 |
-
- MediaPipe Handpose: https://
|
36 |
-
- MediaPipe hands model and model card: https://
|
|
|
37 |
- Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
|
|
|
10 |
|
11 |
**Note**:
|
12 |
- The int8-quantized model may produce invalid results due to a significant drop of accuracy.
|
13 |
+
- Visit https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands for models of larger scale.
|
14 |
|
15 |
## Demo
|
16 |
|
|
|
19 |
# detect on camera input
|
20 |
python demo.py
|
21 |
# detect on an image
|
22 |
+
python demo.py -i /path/to/image -v
|
23 |
```
|
24 |
|
25 |
### Example outputs
|
|
|
32 |
|
33 |
## Reference
|
34 |
|
35 |
+
- MediaPipe Handpose: https://developers.google.com/mediapipe/solutions/vision/hand_landmarker
|
36 |
+
- MediaPipe hands model and model card: https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands
|
37 |
+
- Handpose TFJS:https://github.com/tensorflow/tfjs-models/tree/master/handpose
|
38 |
- Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
|
models/human_segmentation_pphumanseg/README.md
CHANGED
@@ -10,7 +10,7 @@ Run the following command to try the demo:
|
|
10 |
# detect on camera input
|
11 |
python demo.py
|
12 |
# detect on an image
|
13 |
-
python demo.py --input /path/to/image
|
14 |
|
15 |
# get help regarding various parameters
|
16 |
python demo.py --help
|
|
|
10 |
# detect on camera input
|
11 |
python demo.py
|
12 |
# detect on an image
|
13 |
+
python demo.py --input /path/to/image -v
|
14 |
|
15 |
# get help regarding various parameters
|
16 |
python demo.py --help
|
models/license_plate_detection_yunet/README.md
CHANGED
@@ -12,7 +12,7 @@ Run the following command to try the demo:
|
|
12 |
# detect on camera input
|
13 |
python demo.py
|
14 |
# detect on an image
|
15 |
-
python demo.py --input /path/to/image
|
16 |
# get help regarding various parameters
|
17 |
python demo.py --help
|
18 |
```
|
|
|
12 |
# detect on camera input
|
13 |
python demo.py
|
14 |
# detect on an image
|
15 |
+
python demo.py --input /path/to/image -v
|
16 |
# get help regarding various parameters
|
17 |
python demo.py --help
|
18 |
```
|
models/object_detection_nanodet/README.md
CHANGED
@@ -12,7 +12,7 @@ Run the following command to try the demo:
|
|
12 |
# detect on camera input
|
13 |
python demo.py
|
14 |
# detect on an image
|
15 |
-
python demo.py --input /path/to/image
|
16 |
```
|
17 |
Note:
|
18 |
- image result saved as "result.jpg"
|
|
|
12 |
# detect on camera input
|
13 |
python demo.py
|
14 |
# detect on an image
|
15 |
+
python demo.py --input /path/to/image -v
|
16 |
```
|
17 |
Note:
|
18 |
- image result saved as "result.jpg"
|
models/object_detection_yolox/README.md
CHANGED
@@ -18,7 +18,7 @@ Run the following command to try the demo:
|
|
18 |
# detect on camera input
|
19 |
python demo.py
|
20 |
# detect on an image
|
21 |
-
python demo.py --input /path/to/image
|
22 |
```
|
23 |
Note:
|
24 |
- image result saved as "result.jpg"
|
|
|
18 |
# detect on camera input
|
19 |
python demo.py
|
20 |
# detect on an image
|
21 |
+
python demo.py --input /path/to/image -v
|
22 |
```
|
23 |
Note:
|
24 |
- image result saved as "result.jpg"
|
models/object_tracking_dasiamrpn/README.md
CHANGED
@@ -15,7 +15,7 @@ Run the following command to try the demo:
|
|
15 |
# track on camera input
|
16 |
python demo.py
|
17 |
# track on video input
|
18 |
-
python demo.py --input /path/to/video
|
19 |
|
20 |
# get help regarding various parameters
|
21 |
python demo.py --help
|
|
|
15 |
# track on camera input
|
16 |
python demo.py
|
17 |
# track on video input
|
18 |
+
python demo.py --input /path/to/video -v
|
19 |
|
20 |
# get help regarding various parameters
|
21 |
python demo.py --help
|
models/palm_detection_mediapipe/README.md
CHANGED
@@ -4,11 +4,11 @@ This model detects palm bounding boxes and palm landmarks, and is converted from
|
|
4 |
|
5 |
- TFLite model to ONNX: https://github.com/onnx/tensorflow-onnx
|
6 |
- simplified by [onnx-simplifier](https://github.com/daquexian/onnx-simplifier)
|
7 |
-
- SSD Anchors are generated from [GenMediaPipePalmDectionSSDAnchors](https://github.com/VimalMollyn/GenMediaPipePalmDectionSSDAnchors)
|
8 |
|
|
|
9 |
|
10 |
**Note**:
|
11 |
-
- Visit https://
|
12 |
|
13 |
## Demo
|
14 |
|
@@ -18,7 +18,7 @@ Run the following commands to try the demo:
|
|
18 |
# detect on camera input
|
19 |
python demo.py
|
20 |
# detect on an image
|
21 |
-
python demo.py -i /path/to/image
|
22 |
|
23 |
# get help regarding various parameters
|
24 |
python demo.py --help
|
@@ -34,6 +34,7 @@ All files in this directory are licensed under [Apache 2.0 License](./LICENSE).
|
|
34 |
|
35 |
## Reference
|
36 |
|
37 |
-
- MediaPipe Handpose: https://
|
38 |
-
- MediaPipe hands model and model card: https://
|
|
|
39 |
- Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
|
|
|
4 |
|
5 |
- TFLite model to ONNX: https://github.com/onnx/tensorflow-onnx
|
6 |
- simplified by [onnx-simplifier](https://github.com/daquexian/onnx-simplifier)
|
|
|
7 |
|
8 |
+
SSD Anchors are generated from [GenMediaPipePalmDectionSSDAnchors](https://github.com/VimalMollyn/GenMediaPipePalmDectionSSDAnchors)
|
9 |
|
10 |
**Note**:
|
11 |
+
- Visit https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands for models of larger scale.
|
12 |
|
13 |
## Demo
|
14 |
|
|
|
18 |
# detect on camera input
|
19 |
python demo.py
|
20 |
# detect on an image
|
21 |
+
python demo.py -i /path/to/image -v
|
22 |
|
23 |
# get help regarding various parameters
|
24 |
python demo.py --help
|
|
|
34 |
|
35 |
## Reference
|
36 |
|
37 |
+
- MediaPipe Handpose: https://developers.google.com/mediapipe/solutions/vision/hand_landmarker
|
38 |
+
- MediaPipe hands model and model card: https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands
|
39 |
+
- Handpose TFJS:https://github.com/tensorflow/tfjs-models/tree/master/handpose
|
40 |
- Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
|
models/person_detection_mediapipe/README.md
CHANGED
@@ -15,7 +15,7 @@ Run the following commands to try the demo:
|
|
15 |
# detect on camera input
|
16 |
python demo.py
|
17 |
# detect on an image
|
18 |
-
python demo.py -i /path/to/image
|
19 |
|
20 |
# get help regarding various parameters
|
21 |
python demo.py --help
|
@@ -30,6 +30,6 @@ python demo.py --help
|
|
30 |
All files in this directory are licensed under [Apache 2.0 License](LICENSE).
|
31 |
|
32 |
## Reference
|
33 |
-
- MediaPipe Pose: https://google.
|
34 |
-
- MediaPipe pose model and model card: https://
|
35 |
- BlazePose TFJS: https://github.com/tensorflow/tfjs-models/tree/master/pose-detection/src/blazepose_tfjs
|
|
|
15 |
# detect on camera input
|
16 |
python demo.py
|
17 |
# detect on an image
|
18 |
+
python demo.py -i /path/to/image -v
|
19 |
|
20 |
# get help regarding various parameters
|
21 |
python demo.py --help
|
|
|
30 |
All files in this directory are licensed under [Apache 2.0 License](LICENSE).
|
31 |
|
32 |
## Reference
|
33 |
+
- MediaPipe Pose: https://developers.google.com/mediapipe/solutions/vision/pose_landmarker
|
34 |
+
- MediaPipe pose model and model card: https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#pose
|
35 |
- BlazePose TFJS: https://github.com/tensorflow/tfjs-models/tree/master/pose-detection/src/blazepose_tfjs
|
models/person_reid_youtureid/README.md
CHANGED
@@ -11,7 +11,7 @@ Note:
|
|
11 |
Run the following command to try the demo:
|
12 |
|
13 |
```shell
|
14 |
-
python demo.py --query_dir /path/to/query --gallery_dir /path/to/gallery
|
15 |
|
16 |
# get help regarding various parameters
|
17 |
python demo.py --help
|
|
|
11 |
Run the following command to try the demo:
|
12 |
|
13 |
```shell
|
14 |
+
python demo.py --query_dir /path/to/query --gallery_dir /path/to/gallery -v
|
15 |
|
16 |
# get help regarding various parameters
|
17 |
python demo.py --help
|
models/person_reid_youtureid/demo.py
CHANGED
@@ -43,10 +43,10 @@ parser.add_argument('--topk', type=int, default=10,
|
|
43 |
help='Top-K closest from gallery for each query.')
|
44 |
parser.add_argument('--model', '-m', type=str, default='person_reid_youtu_2021nov.onnx',
|
45 |
help='Path to the model.')
|
46 |
-
parser.add_argument('--save', '-s',
|
47 |
-
help='
|
48 |
-
parser.add_argument('--vis', '-v',
|
49 |
-
help='
|
50 |
args = parser.parse_args()
|
51 |
|
52 |
def readImageFromDirectory(img_dir, w=128, h=256):
|
|
|
43 |
help='Top-K closest from gallery for each query.')
|
44 |
parser.add_argument('--model', '-m', type=str, default='person_reid_youtu_2021nov.onnx',
|
45 |
help='Path to the model.')
|
46 |
+
parser.add_argument('--save', '-s', action='store_true',
|
47 |
+
help='Usage: Specify to save file with results (i.e. bounding box, confidence level). Invalid in case of camera input.')
|
48 |
+
parser.add_argument('--vis', '-v', action='store_true',
|
49 |
+
help='Usage: Specify to open a new window to show results. Invalid in case of camera input.')
|
50 |
args = parser.parse_args()
|
51 |
|
52 |
def readImageFromDirectory(img_dir, w=128, h=256):
|
models/qrcode_wechatqrcode/README.md
CHANGED
@@ -15,7 +15,7 @@ Run the following command to try the demo:
|
|
15 |
# detect on camera input
|
16 |
python demo.py
|
17 |
# detect on an image
|
18 |
-
python demo.py --input /path/to/image
|
19 |
|
20 |
# get help regarding various parameters
|
21 |
python demo.py --help
|
|
|
15 |
# detect on camera input
|
16 |
python demo.py
|
17 |
# detect on an image
|
18 |
+
python demo.py --input /path/to/image -v
|
19 |
|
20 |
# get help regarding various parameters
|
21 |
python demo.py --help
|
models/text_detection_db/README.md
CHANGED
@@ -17,7 +17,7 @@ Run the following command to try the demo:
|
|
17 |
# detect on camera input
|
18 |
python demo.py
|
19 |
# detect on an image
|
20 |
-
python demo.py --input /path/to/image
|
21 |
|
22 |
# get help regarding various parameters
|
23 |
python demo.py --help
|
|
|
17 |
# detect on camera input
|
18 |
python demo.py
|
19 |
# detect on an image
|
20 |
+
python demo.py --input /path/to/image -v
|
21 |
|
22 |
# get help regarding various parameters
|
23 |
python demo.py --help
|
models/text_recognition_crnn/README.md
CHANGED
@@ -42,7 +42,7 @@ Run the demo detecting English:
|
|
42 |
# detect on camera input
|
43 |
python demo.py
|
44 |
# detect on an image
|
45 |
-
python demo.py --input /path/to/image
|
46 |
|
47 |
# get help regarding various parameters
|
48 |
python demo.py --help
|
|
|
42 |
# detect on camera input
|
43 |
python demo.py
|
44 |
# detect on an image
|
45 |
+
python demo.py --input /path/to/image -v
|
46 |
|
47 |
# get help regarding various parameters
|
48 |
python demo.py --help
|
tools/eval/README.md
CHANGED
@@ -19,9 +19,9 @@ Supported datasets:
|
|
19 |
- [ImageNet](#imagenet)
|
20 |
- [WIDERFace](#widerface)
|
21 |
- [LFW](#lfw)
|
22 |
-
- [ICDAR](#
|
23 |
- [IIIT5K](#iiit5k)
|
24 |
-
- [Mini Supervisely](#
|
25 |
|
26 |
## ImageNet
|
27 |
|
|
|
19 |
- [ImageNet](#imagenet)
|
20 |
- [WIDERFace](#widerface)
|
21 |
- [LFW](#lfw)
|
22 |
+
- [ICDAR](#icdar2003)
|
23 |
- [IIIT5K](#iiit5k)
|
24 |
+
- [Mini Supervisely](#mini-supervisely)
|
25 |
|
26 |
## ImageNet
|
27 |
|