Wanli commited on
Commit
af3dd88
·
1 Parent(s): f8da712

update and modify documentation (#161)

Browse files
README.md CHANGED
@@ -82,6 +82,10 @@ Some examples are listed below. You can find more in the directory of each model
82
 
83
  ![handpose estimation](models/handpose_estimation_mediapipe/examples/mphandpose_demo.webp)
84
 
 
 
 
 
85
  ### QR Code Detection and Parsing with [WeChatQRCode](./models/qrcode_wechatqrcode/)
86
 
87
  ![qrcode](./models/qrcode_wechatqrcode/examples/wechat_qrcode_demo.gif)
 
82
 
83
  ![handpose estimation](models/handpose_estimation_mediapipe/examples/mphandpose_demo.webp)
84
 
85
+ ### Person Detection with [MP-PersonDet](./models/person_detection_mediapipe)
86
+
87
+ ![person det](./models/person_detection_mediapipe/examples/mppersondet_demo.webp)
88
+
89
  ### QR Code Detection and Parsing with [WeChatQRCode](./models/qrcode_wechatqrcode/)
90
 
91
  ![qrcode](./models/qrcode_wechatqrcode/examples/wechat_qrcode_demo.gif)
benchmark/README.md CHANGED
@@ -79,7 +79,7 @@ Benchmark is done with latest `opencv-python==4.7.0.72` and `opencv-contrib-pyth
79
  | [YoutuReID](../models/person_reid_youtureid) | Person Re-Identification | 128x256 | 30.39 | 625.56 | 11117.07 | 195.67 | 898.23 | 14886.02 | 90.07 | 44.61 | 5.58 | --- |
80
  | [MP-PalmDet](../models/palm_detection_mediapipe) | Palm Detection | 192x192 | 6.29 | 86.83 | 872.09 | 38.03 | 142.23 | 1191.81 | 83.20 | 33.81 | 5.17 | --- |
81
  | [MP-HandPose](../models/handpose_estimation_mediapipe) | Hand Pose Estimation | 224x224 | 4.68 | 43.57 | 460.56 | 20.27 | 80.67 | 636.22 | 40.10 | 19.47 | 6.27 | --- |
82
- | [MP-PersonDet](./models/person_detection_mediapipe) | Person Detection | 224x224 | 13.88 | 98.52 | 1326.56 | 46.07 | 191.41 | 1835.97 | 56.69 | --- | 16.45 | --- |
83
 
84
  \*: Models are quantized in per-channel mode, which run slower than per-tensor quantized models on NPU.
85
 
 
79
  | [YoutuReID](../models/person_reid_youtureid) | Person Re-Identification | 128x256 | 30.39 | 625.56 | 11117.07 | 195.67 | 898.23 | 14886.02 | 90.07 | 44.61 | 5.58 | --- |
80
  | [MP-PalmDet](../models/palm_detection_mediapipe) | Palm Detection | 192x192 | 6.29 | 86.83 | 872.09 | 38.03 | 142.23 | 1191.81 | 83.20 | 33.81 | 5.17 | --- |
81
  | [MP-HandPose](../models/handpose_estimation_mediapipe) | Hand Pose Estimation | 224x224 | 4.68 | 43.57 | 460.56 | 20.27 | 80.67 | 636.22 | 40.10 | 19.47 | 6.27 | --- |
82
+ | [MP-PersonDet](../models/person_detection_mediapipe) | Person Detection | 224x224 | 13.88 | 98.52 | 1326.56 | 46.07 | 191.41 | 1835.97 | 56.69 | --- | 16.45 | --- |
83
 
84
  \*: Models are quantized in per-channel mode, which run slower than per-tensor quantized models on NPU.
85
 
models/face_detection_yunet/README.md CHANGED
@@ -28,7 +28,7 @@ Run the following command to try the demo:
28
  # detect on camera input
29
  python demo.py
30
  # detect on an image
31
- python demo.py --input /path/to/image
32
 
33
  # get help regarding various parameters
34
  python demo.py --help
@@ -40,13 +40,13 @@ Install latest OpenCV and CMake >= 3.24.0 to get started with:
40
 
41
  ```shell
42
  # A typical and default installation path of OpenCV is /usr/local
43
- cmake -B build -D OPENCV_INSTALLATION_PATH /path/to/opencv/installation .
44
  cmake --build build
45
 
46
  # detect on camera input
47
  ./build/demo
48
  # detect on an image
49
- ./build/demo -i=/path/to/image
50
  # get help messages
51
  ./build/demo -h
52
  ```
 
28
  # detect on camera input
29
  python demo.py
30
  # detect on an image
31
+ python demo.py --input /path/to/image -v
32
 
33
  # get help regarding various parameters
34
  python demo.py --help
 
40
 
41
  ```shell
42
  # A typical and default installation path of OpenCV is /usr/local
43
+ cmake -B build -D OPENCV_INSTALLATION_PATH=/path/to/opencv/installation .
44
  cmake --build build
45
 
46
  # detect on camera input
47
  ./build/demo
48
  # detect on an image
49
+ ./build/demo -i=/path/to/image -v
50
  # get help messages
51
  ./build/demo -h
52
  ```
models/facial_expression_recognition/README.md CHANGED
@@ -22,7 +22,7 @@ Results of accuracy evaluation on [RAF-DB](http://whdeng.cn/RAF/model1.html).
22
  Run the following command to try the demo:
23
  ```shell
24
  # recognize the facial expression on images
25
- python demo.py --input /path/to/image
26
  ```
27
 
28
  ### Example outputs
 
22
  Run the following command to try the demo:
23
  ```shell
24
  # recognize the facial expression on images
25
+ python demo.py --input /path/to/image -v
26
  ```
27
 
28
  ### Example outputs
models/handpose_estimation_mediapipe/README.md CHANGED
@@ -10,7 +10,7 @@ This model is converted from TFlite to ONNX using following tools:
10
 
11
  **Note**:
12
  - The int8-quantized model may produce invalid results due to a significant drop of accuracy.
13
- - Visit https://google.github.io/mediapipe/solutions/models.html#hands for models of larger scale.
14
 
15
  ## Demo
16
 
@@ -19,7 +19,7 @@ Run the following commands to try the demo:
19
  # detect on camera input
20
  python demo.py
21
  # detect on an image
22
- python demo.py -i /path/to/image
23
  ```
24
 
25
  ### Example outputs
@@ -32,6 +32,7 @@ All files in this directory are licensed under [Apache 2.0 License](./LICENSE).
32
 
33
  ## Reference
34
 
35
- - MediaPipe Handpose: https://github.com/tensorflow/tfjs-models/tree/master/handpose
36
- - MediaPipe hands model and model card: https://google.github.io/mediapipe/solutions/models.html#hands
 
37
  - Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
 
10
 
11
  **Note**:
12
  - The int8-quantized model may produce invalid results due to a significant drop of accuracy.
13
+ - Visit https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands for models of larger scale.
14
 
15
  ## Demo
16
 
 
19
  # detect on camera input
20
  python demo.py
21
  # detect on an image
22
+ python demo.py -i /path/to/image -v
23
  ```
24
 
25
  ### Example outputs
 
32
 
33
  ## Reference
34
 
35
+ - MediaPipe Handpose: https://developers.google.com/mediapipe/solutions/vision/hand_landmarker
36
+ - MediaPipe hands model and model card: https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands
37
+ - Handpose TFJS:https://github.com/tensorflow/tfjs-models/tree/master/handpose
38
  - Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
models/human_segmentation_pphumanseg/README.md CHANGED
@@ -10,7 +10,7 @@ Run the following command to try the demo:
10
  # detect on camera input
11
  python demo.py
12
  # detect on an image
13
- python demo.py --input /path/to/image
14
 
15
  # get help regarding various parameters
16
  python demo.py --help
 
10
  # detect on camera input
11
  python demo.py
12
  # detect on an image
13
+ python demo.py --input /path/to/image -v
14
 
15
  # get help regarding various parameters
16
  python demo.py --help
models/license_plate_detection_yunet/README.md CHANGED
@@ -12,7 +12,7 @@ Run the following command to try the demo:
12
  # detect on camera input
13
  python demo.py
14
  # detect on an image
15
- python demo.py --input /path/to/image
16
  # get help regarding various parameters
17
  python demo.py --help
18
  ```
 
12
  # detect on camera input
13
  python demo.py
14
  # detect on an image
15
+ python demo.py --input /path/to/image -v
16
  # get help regarding various parameters
17
  python demo.py --help
18
  ```
models/object_detection_nanodet/README.md CHANGED
@@ -12,7 +12,7 @@ Run the following command to try the demo:
12
  # detect on camera input
13
  python demo.py
14
  # detect on an image
15
- python demo.py --input /path/to/image
16
  ```
17
  Note:
18
  - image result saved as "result.jpg"
 
12
  # detect on camera input
13
  python demo.py
14
  # detect on an image
15
+ python demo.py --input /path/to/image -v
16
  ```
17
  Note:
18
  - image result saved as "result.jpg"
models/object_detection_yolox/README.md CHANGED
@@ -18,7 +18,7 @@ Run the following command to try the demo:
18
  # detect on camera input
19
  python demo.py
20
  # detect on an image
21
- python demo.py --input /path/to/image
22
  ```
23
  Note:
24
  - image result saved as "result.jpg"
 
18
  # detect on camera input
19
  python demo.py
20
  # detect on an image
21
+ python demo.py --input /path/to/image -v
22
  ```
23
  Note:
24
  - image result saved as "result.jpg"
models/object_tracking_dasiamrpn/README.md CHANGED
@@ -15,7 +15,7 @@ Run the following command to try the demo:
15
  # track on camera input
16
  python demo.py
17
  # track on video input
18
- python demo.py --input /path/to/video
19
 
20
  # get help regarding various parameters
21
  python demo.py --help
 
15
  # track on camera input
16
  python demo.py
17
  # track on video input
18
+ python demo.py --input /path/to/video -v
19
 
20
  # get help regarding various parameters
21
  python demo.py --help
models/palm_detection_mediapipe/README.md CHANGED
@@ -4,11 +4,11 @@ This model detects palm bounding boxes and palm landmarks, and is converted from
4
 
5
  - TFLite model to ONNX: https://github.com/onnx/tensorflow-onnx
6
  - simplified by [onnx-simplifier](https://github.com/daquexian/onnx-simplifier)
7
- - SSD Anchors are generated from [GenMediaPipePalmDectionSSDAnchors](https://github.com/VimalMollyn/GenMediaPipePalmDectionSSDAnchors)
8
 
 
9
 
10
  **Note**:
11
- - Visit https://google.github.io/mediapipe/solutions/models.html#hands for models of larger scale.
12
 
13
  ## Demo
14
 
@@ -18,7 +18,7 @@ Run the following commands to try the demo:
18
  # detect on camera input
19
  python demo.py
20
  # detect on an image
21
- python demo.py -i /path/to/image
22
 
23
  # get help regarding various parameters
24
  python demo.py --help
@@ -34,6 +34,7 @@ All files in this directory are licensed under [Apache 2.0 License](./LICENSE).
34
 
35
  ## Reference
36
 
37
- - MediaPipe Handpose: https://github.com/tensorflow/tfjs-models/tree/master/handpose
38
- - MediaPipe hands model and model card: https://google.github.io/mediapipe/solutions/models.html#hands
 
39
  - Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
 
4
 
5
  - TFLite model to ONNX: https://github.com/onnx/tensorflow-onnx
6
  - simplified by [onnx-simplifier](https://github.com/daquexian/onnx-simplifier)
 
7
 
8
+ SSD Anchors are generated from [GenMediaPipePalmDectionSSDAnchors](https://github.com/VimalMollyn/GenMediaPipePalmDectionSSDAnchors)
9
 
10
  **Note**:
11
+ - Visit https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands for models of larger scale.
12
 
13
  ## Demo
14
 
 
18
  # detect on camera input
19
  python demo.py
20
  # detect on an image
21
+ python demo.py -i /path/to/image -v
22
 
23
  # get help regarding various parameters
24
  python demo.py --help
 
34
 
35
  ## Reference
36
 
37
+ - MediaPipe Handpose: https://developers.google.com/mediapipe/solutions/vision/hand_landmarker
38
+ - MediaPipe hands model and model card: https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#hands
39
+ - Handpose TFJS:https://github.com/tensorflow/tfjs-models/tree/master/handpose
40
  - Int8 model quantized with rgb evaluation set of FreiHAND: https://lmb.informatik.uni-freiburg.de/resources/datasets/FreihandDataset.en.html
models/person_detection_mediapipe/README.md CHANGED
@@ -15,7 +15,7 @@ Run the following commands to try the demo:
15
  # detect on camera input
16
  python demo.py
17
  # detect on an image
18
- python demo.py -i /path/to/image
19
 
20
  # get help regarding various parameters
21
  python demo.py --help
@@ -30,6 +30,6 @@ python demo.py --help
30
  All files in this directory are licensed under [Apache 2.0 License](LICENSE).
31
 
32
  ## Reference
33
- - MediaPipe Pose: https://google.github.io/mediapipe/solutions/pose
34
- - MediaPipe pose model and model card: https://google.github.io/mediapipe/solutions/models.html#pose
35
  - BlazePose TFJS: https://github.com/tensorflow/tfjs-models/tree/master/pose-detection/src/blazepose_tfjs
 
15
  # detect on camera input
16
  python demo.py
17
  # detect on an image
18
+ python demo.py -i /path/to/image -v
19
 
20
  # get help regarding various parameters
21
  python demo.py --help
 
30
  All files in this directory are licensed under [Apache 2.0 License](LICENSE).
31
 
32
  ## Reference
33
+ - MediaPipe Pose: https://developers.google.com/mediapipe/solutions/vision/pose_landmarker
34
+ - MediaPipe pose model and model card: https://github.com/google/mediapipe/blob/master/docs/solutions/models.md#pose
35
  - BlazePose TFJS: https://github.com/tensorflow/tfjs-models/tree/master/pose-detection/src/blazepose_tfjs
models/person_reid_youtureid/README.md CHANGED
@@ -11,7 +11,7 @@ Note:
11
  Run the following command to try the demo:
12
 
13
  ```shell
14
- python demo.py --query_dir /path/to/query --gallery_dir /path/to/gallery
15
 
16
  # get help regarding various parameters
17
  python demo.py --help
 
11
  Run the following command to try the demo:
12
 
13
  ```shell
14
+ python demo.py --query_dir /path/to/query --gallery_dir /path/to/gallery -v
15
 
16
  # get help regarding various parameters
17
  python demo.py --help
models/person_reid_youtureid/demo.py CHANGED
@@ -43,10 +43,10 @@ parser.add_argument('--topk', type=int, default=10,
43
  help='Top-K closest from gallery for each query.')
44
  parser.add_argument('--model', '-m', type=str, default='person_reid_youtu_2021nov.onnx',
45
  help='Path to the model.')
46
- parser.add_argument('--save', '-s', type=str2bool, default=False,
47
- help='Set true to save results. This flag is invalid when using camera.')
48
- parser.add_argument('--vis', '-v', type=str2bool, default=True,
49
- help='Set true to open a window for result visualization. This flag is invalid when using camera.')
50
  args = parser.parse_args()
51
 
52
  def readImageFromDirectory(img_dir, w=128, h=256):
 
43
  help='Top-K closest from gallery for each query.')
44
  parser.add_argument('--model', '-m', type=str, default='person_reid_youtu_2021nov.onnx',
45
  help='Path to the model.')
46
+ parser.add_argument('--save', '-s', action='store_true',
47
+ help='Usage: Specify to save file with results (i.e. bounding box, confidence level). Invalid in case of camera input.')
48
+ parser.add_argument('--vis', '-v', action='store_true',
49
+ help='Usage: Specify to open a new window to show results. Invalid in case of camera input.')
50
  args = parser.parse_args()
51
 
52
  def readImageFromDirectory(img_dir, w=128, h=256):
models/qrcode_wechatqrcode/README.md CHANGED
@@ -15,7 +15,7 @@ Run the following command to try the demo:
15
  # detect on camera input
16
  python demo.py
17
  # detect on an image
18
- python demo.py --input /path/to/image
19
 
20
  # get help regarding various parameters
21
  python demo.py --help
 
15
  # detect on camera input
16
  python demo.py
17
  # detect on an image
18
+ python demo.py --input /path/to/image -v
19
 
20
  # get help regarding various parameters
21
  python demo.py --help
models/text_detection_db/README.md CHANGED
@@ -17,7 +17,7 @@ Run the following command to try the demo:
17
  # detect on camera input
18
  python demo.py
19
  # detect on an image
20
- python demo.py --input /path/to/image
21
 
22
  # get help regarding various parameters
23
  python demo.py --help
 
17
  # detect on camera input
18
  python demo.py
19
  # detect on an image
20
+ python demo.py --input /path/to/image -v
21
 
22
  # get help regarding various parameters
23
  python demo.py --help
models/text_recognition_crnn/README.md CHANGED
@@ -42,7 +42,7 @@ Run the demo detecting English:
42
  # detect on camera input
43
  python demo.py
44
  # detect on an image
45
- python demo.py --input /path/to/image
46
 
47
  # get help regarding various parameters
48
  python demo.py --help
 
42
  # detect on camera input
43
  python demo.py
44
  # detect on an image
45
+ python demo.py --input /path/to/image -v
46
 
47
  # get help regarding various parameters
48
  python demo.py --help
tools/eval/README.md CHANGED
@@ -19,9 +19,9 @@ Supported datasets:
19
  - [ImageNet](#imagenet)
20
  - [WIDERFace](#widerface)
21
  - [LFW](#lfw)
22
- - [ICDAR](#ICDAR2003)
23
  - [IIIT5K](#iiit5k)
24
- - [Mini Supervisely](#mini_supervisely)
25
 
26
  ## ImageNet
27
 
 
19
  - [ImageNet](#imagenet)
20
  - [WIDERFace](#widerface)
21
  - [LFW](#lfw)
22
+ - [ICDAR](#icdar2003)
23
  - [IIIT5K](#iiit5k)
24
+ - [Mini Supervisely](#mini-supervisely)
25
 
26
  ## ImageNet
27