Ryan Lee
commited on
Commit
·
1528081
1
Parent(s):
a8f2672
Fix spelling, spacing, and unused variables (#236)
Browse files* Fix spelling, spacing, and unused variables
* Fixed typo in 3 more models
* Added open back
* Remove setBackendAndTarget functions from C++ demos, which are unused functionality.
- models/face_detection_yunet/demo.cpp +0 -7
- models/image_classification_mobilenet/demo.cpp +1 -1
- models/object_detection_nanodet/demo.cpp +2 -7
- models/object_detection_nanodet/demo.py +1 -1
- models/object_detection_yolox/demo.cpp +2 -10
- models/object_detection_yolox/demo.py +1 -1
- models/person_detection_mediapipe/demo.cpp +1 -10
- models/pose_estimation_mediapipe/demo.cpp +2 -18
- models/text_detection_ppocr/demo.cpp +1 -1
- models/text_recognition_crnn/demo.cpp +1 -1
- models/text_recognition_crnn/demo.py +1 -1
models/face_detection_yunet/demo.cpp
CHANGED
|
@@ -31,13 +31,6 @@ public:
|
|
| 31 |
model = cv::FaceDetectorYN::create(model_path_, "", input_size_, conf_threshold_, nms_threshold_, top_k_, backend_id_, target_id_);
|
| 32 |
}
|
| 33 |
|
| 34 |
-
void setBackendAndTarget(int backend_id, int target_id)
|
| 35 |
-
{
|
| 36 |
-
backend_id_ = backend_id;
|
| 37 |
-
target_id_ = target_id;
|
| 38 |
-
model = cv::FaceDetectorYN::create(model_path_, "", input_size_, conf_threshold_, nms_threshold_, top_k_, backend_id_, target_id_);
|
| 39 |
-
}
|
| 40 |
-
|
| 41 |
/* Overwrite the input size when creating the model. Size format: [Width, Height].
|
| 42 |
*/
|
| 43 |
void setInputSize(const cv::Size& input_size)
|
|
|
|
| 31 |
model = cv::FaceDetectorYN::create(model_path_, "", input_size_, conf_threshold_, nms_threshold_, top_k_, backend_id_, target_id_);
|
| 32 |
}
|
| 33 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
/* Overwrite the input size when creating the model. Size format: [Width, Height].
|
| 35 |
*/
|
| 36 |
void setInputSize(const cv::Size& input_size)
|
models/image_classification_mobilenet/demo.cpp
CHANGED
|
@@ -69,7 +69,7 @@ int main(int argc, char** argv)
|
|
| 69 |
else
|
| 70 |
cap.open(0);
|
| 71 |
if (!cap.isOpened())
|
| 72 |
-
CV_Error(Error::StsError, "Cannot
|
| 73 |
Mat frame, blob;
|
| 74 |
static const std::string kWinName = model;
|
| 75 |
int nbInference = 0;
|
|
|
|
| 69 |
else
|
| 70 |
cap.open(0);
|
| 71 |
if (!cap.isOpened())
|
| 72 |
+
CV_Error(Error::StsError, "Cannot open video or file");
|
| 73 |
Mat frame, blob;
|
| 74 |
static const std::string kWinName = model;
|
| 75 |
int nbInference = 0;
|
models/object_detection_nanodet/demo.cpp
CHANGED
|
@@ -46,7 +46,8 @@ public:
|
|
| 46 |
{
|
| 47 |
this->strides = { 8, 16, 32, 64 };
|
| 48 |
this->net = readNet(modelPath);
|
| 49 |
-
|
|
|
|
| 50 |
this->project = Mat::zeros(1, this->regMax + 1, CV_32F);
|
| 51 |
for (size_t i = 0; i <= this->regMax; ++i)
|
| 52 |
{
|
|
@@ -57,12 +58,6 @@ public:
|
|
| 57 |
this->generateAnchors();
|
| 58 |
}
|
| 59 |
|
| 60 |
-
void setBackendAndTarget(Backend bId, Target tId)
|
| 61 |
-
{
|
| 62 |
-
this->net.setPreferableBackend(bId);
|
| 63 |
-
this->net.setPreferableTarget(tId);
|
| 64 |
-
}
|
| 65 |
-
|
| 66 |
Mat preProcess(const Mat& inputImage)
|
| 67 |
{
|
| 68 |
Image2BlobParams paramNanodet;
|
|
|
|
| 46 |
{
|
| 47 |
this->strides = { 8, 16, 32, 64 };
|
| 48 |
this->net = readNet(modelPath);
|
| 49 |
+
this->net.setPreferableBackend(bId);
|
| 50 |
+
this->net.setPreferableTarget(tId);
|
| 51 |
this->project = Mat::zeros(1, this->regMax + 1, CV_32F);
|
| 52 |
for (size_t i = 0; i <= this->regMax; ++i)
|
| 53 |
{
|
|
|
|
| 58 |
this->generateAnchors();
|
| 59 |
}
|
| 60 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 61 |
Mat preProcess(const Mat& inputImage)
|
| 62 |
{
|
| 63 |
Image2BlobParams paramNanodet;
|
models/object_detection_nanodet/demo.py
CHANGED
|
@@ -148,7 +148,7 @@ if __name__=='__main__':
|
|
| 148 |
img = vis(preds, image, letterbox_scale)
|
| 149 |
|
| 150 |
if args.save:
|
| 151 |
-
print('
|
| 152 |
cv.imwrite('result.jpg', img)
|
| 153 |
|
| 154 |
if args.vis:
|
|
|
|
| 148 |
img = vis(preds, image, letterbox_scale)
|
| 149 |
|
| 150 |
if args.save:
|
| 151 |
+
print('Results saved to result.jpg\n')
|
| 152 |
cv.imwrite('result.jpg', img)
|
| 153 |
|
| 154 |
if args.vis:
|
models/object_detection_yolox/demo.cpp
CHANGED
|
@@ -61,14 +61,6 @@ public:
|
|
| 61 |
this->generateAnchors();
|
| 62 |
}
|
| 63 |
|
| 64 |
-
void setBackendAndTarget(dnn::Backend bId, dnn::Target tId)
|
| 65 |
-
{
|
| 66 |
-
this->backendId = bId;
|
| 67 |
-
this->targetId = tId;
|
| 68 |
-
this->net.setPreferableBackend(this->backendId);
|
| 69 |
-
this->net.setPreferableTarget(this->targetId);
|
| 70 |
-
}
|
| 71 |
-
|
| 72 |
Mat preprocess(Mat img)
|
| 73 |
{
|
| 74 |
Mat blob;
|
|
@@ -137,7 +129,7 @@ public:
|
|
| 137 |
boxesXYXY[r].height = boxes_xyxy.at<float>(r, 3);
|
| 138 |
}
|
| 139 |
|
| 140 |
-
vector<
|
| 141 |
NMSBoxesBatched(boxesXYXY, maxScores, maxScoreIdx, this->confThreshold, this->nmsThreshold, keep);
|
| 142 |
Mat candidates(int(keep.size()), 6, CV_32FC1);
|
| 143 |
int row = 0;
|
|
@@ -282,7 +274,7 @@ int main(int argc, char** argv)
|
|
| 282 |
else
|
| 283 |
cap.open(0);
|
| 284 |
if (!cap.isOpened())
|
| 285 |
-
CV_Error(Error::StsError, "Cannot
|
| 286 |
Mat frame, inputBlob;
|
| 287 |
double letterboxScale;
|
| 288 |
|
|
|
|
| 61 |
this->generateAnchors();
|
| 62 |
}
|
| 63 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
Mat preprocess(Mat img)
|
| 65 |
{
|
| 66 |
Mat blob;
|
|
|
|
| 129 |
boxesXYXY[r].height = boxes_xyxy.at<float>(r, 3);
|
| 130 |
}
|
| 131 |
|
| 132 |
+
vector<int> keep;
|
| 133 |
NMSBoxesBatched(boxesXYXY, maxScores, maxScoreIdx, this->confThreshold, this->nmsThreshold, keep);
|
| 134 |
Mat candidates(int(keep.size()), 6, CV_32FC1);
|
| 135 |
int row = 0;
|
|
|
|
| 274 |
else
|
| 275 |
cap.open(0);
|
| 276 |
if (!cap.isOpened())
|
| 277 |
+
CV_Error(Error::StsError, "Cannot open video or file");
|
| 278 |
Mat frame, inputBlob;
|
| 279 |
double letterboxScale;
|
| 280 |
|
models/object_detection_yolox/demo.py
CHANGED
|
@@ -120,7 +120,7 @@ if __name__=='__main__':
|
|
| 120 |
img = vis(preds, image, letterbox_scale)
|
| 121 |
|
| 122 |
if args.save:
|
| 123 |
-
print('
|
| 124 |
cv.imwrite('result.jpg', img)
|
| 125 |
|
| 126 |
if args.vis:
|
|
|
|
| 120 |
img = vis(preds, image, letterbox_scale)
|
| 121 |
|
| 122 |
if args.save:
|
| 123 |
+
print('Results saved to result.jpg\n')
|
| 124 |
cv.imwrite('result.jpg', img)
|
| 125 |
|
| 126 |
if args.vis:
|
models/person_detection_mediapipe/demo.cpp
CHANGED
|
@@ -43,14 +43,6 @@ public:
|
|
| 43 |
this->anchors = getMediapipeAnchor();
|
| 44 |
}
|
| 45 |
|
| 46 |
-
void setBackendAndTarget(dnn::Backend bId, dnn::Target tId)
|
| 47 |
-
{
|
| 48 |
-
this->backendId = bId;
|
| 49 |
-
this->targetId = tId;
|
| 50 |
-
this->net.setPreferableBackend(this->backendId);
|
| 51 |
-
this->net.setPreferableTarget(this->targetId);
|
| 52 |
-
}
|
| 53 |
-
|
| 54 |
pair<Mat, Size> preprocess(Mat img)
|
| 55 |
{
|
| 56 |
Mat blob;
|
|
@@ -237,10 +229,9 @@ int main(int argc, char** argv)
|
|
| 237 |
backendTargetPairs[backendTargetid].first, backendTargetPairs[backendTargetid].second);
|
| 238 |
//! [Open a video file or an image file or a camera stream]
|
| 239 |
if (!cap.isOpened())
|
| 240 |
-
CV_Error(Error::StsError, "Cannot
|
| 241 |
|
| 242 |
static const std::string kWinName = "MPPersonDet Demo";
|
| 243 |
-
int nbInference = 0;
|
| 244 |
while (waitKey(1) < 0)
|
| 245 |
{
|
| 246 |
cap >> frame;
|
|
|
|
| 43 |
this->anchors = getMediapipeAnchor();
|
| 44 |
}
|
| 45 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
pair<Mat, Size> preprocess(Mat img)
|
| 47 |
{
|
| 48 |
Mat blob;
|
|
|
|
| 229 |
backendTargetPairs[backendTargetid].first, backendTargetPairs[backendTargetid].second);
|
| 230 |
//! [Open a video file or an image file or a camera stream]
|
| 231 |
if (!cap.isOpened())
|
| 232 |
+
CV_Error(Error::StsError, "Cannot open video or file");
|
| 233 |
|
| 234 |
static const std::string kWinName = "MPPersonDet Demo";
|
|
|
|
| 235 |
while (waitKey(1) < 0)
|
| 236 |
{
|
| 237 |
cap >> frame;
|
models/pose_estimation_mediapipe/demo.cpp
CHANGED
|
@@ -45,14 +45,6 @@ public:
|
|
| 45 |
this->anchors = getMediapipeAnchor();
|
| 46 |
}
|
| 47 |
|
| 48 |
-
void setBackendAndTarget(dnn::Backend bId, dnn::Target tId)
|
| 49 |
-
{
|
| 50 |
-
this->backendId = bId;
|
| 51 |
-
this->targetId = tId;
|
| 52 |
-
this->net.setPreferableBackend(this->backendId);
|
| 53 |
-
this->net.setPreferableTarget(this->targetId);
|
| 54 |
-
}
|
| 55 |
-
|
| 56 |
pair<Mat, Size> preprocess(Mat img)
|
| 57 |
{
|
| 58 |
Mat blob;
|
|
@@ -124,7 +116,7 @@ public:
|
|
| 124 |
{
|
| 125 |
rotBoxes[i] = Rect2d(Point2d(boxes.at<float>(i, 0), boxes.at<float>(i, 1)), Point2d(boxes.at<float>(i, 2), boxes.at<float>(i, 3)));
|
| 126 |
}
|
| 127 |
-
vector<
|
| 128 |
NMSBoxes(rotBoxes, score, this->scoreThreshold, this->nmsThreshold, keep, 1.0f, this->topK);
|
| 129 |
if (keep.size() == 0)
|
| 130 |
return Mat();
|
|
@@ -179,14 +171,6 @@ public:
|
|
| 179 |
this->personBoxEnlargeFactor = 1.25;
|
| 180 |
}
|
| 181 |
|
| 182 |
-
void setBackendAndTarget(dnn::Backend bId, dnn::Target tId)
|
| 183 |
-
{
|
| 184 |
-
this->backendId = bId;
|
| 185 |
-
this->targetId = tId;
|
| 186 |
-
this->net.setPreferableBackend(this->backendId);
|
| 187 |
-
this->net.setPreferableTarget(this->targetId);
|
| 188 |
-
}
|
| 189 |
-
|
| 190 |
tuple<Mat, Mat, float, Mat, Size> preprocess(Mat image, Mat person)
|
| 191 |
{
|
| 192 |
/***
|
|
@@ -567,7 +551,7 @@ int main(int argc, char** argv)
|
|
| 567 |
MPPose poseEstimator(model, confThreshold, backendTargetPairs[backendTargetid].first, backendTargetPairs[backendTargetid].second);
|
| 568 |
//! [Open a video file or an image file or a camera stream]
|
| 569 |
if (!cap.isOpened())
|
| 570 |
-
CV_Error(Error::StsError, "Cannot
|
| 571 |
|
| 572 |
static const std::string kWinName = "MPPose Demo";
|
| 573 |
while (waitKey(1) < 0)
|
|
|
|
| 45 |
this->anchors = getMediapipeAnchor();
|
| 46 |
}
|
| 47 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
pair<Mat, Size> preprocess(Mat img)
|
| 49 |
{
|
| 50 |
Mat blob;
|
|
|
|
| 116 |
{
|
| 117 |
rotBoxes[i] = Rect2d(Point2d(boxes.at<float>(i, 0), boxes.at<float>(i, 1)), Point2d(boxes.at<float>(i, 2), boxes.at<float>(i, 3)));
|
| 118 |
}
|
| 119 |
+
vector<int> keep;
|
| 120 |
NMSBoxes(rotBoxes, score, this->scoreThreshold, this->nmsThreshold, keep, 1.0f, this->topK);
|
| 121 |
if (keep.size() == 0)
|
| 122 |
return Mat();
|
|
|
|
| 171 |
this->personBoxEnlargeFactor = 1.25;
|
| 172 |
}
|
| 173 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 174 |
tuple<Mat, Mat, float, Mat, Size> preprocess(Mat image, Mat person)
|
| 175 |
{
|
| 176 |
/***
|
|
|
|
| 551 |
MPPose poseEstimator(model, confThreshold, backendTargetPairs[backendTargetid].first, backendTargetPairs[backendTargetid].second);
|
| 552 |
//! [Open a video file or an image file or a camera stream]
|
| 553 |
if (!cap.isOpened())
|
| 554 |
+
CV_Error(Error::StsError, "Cannot open video or file");
|
| 555 |
|
| 556 |
static const std::string kWinName = "MPPose Demo";
|
| 557 |
while (waitKey(1) < 0)
|
models/text_detection_ppocr/demo.cpp
CHANGED
|
@@ -124,7 +124,7 @@ int main(int argc, char** argv)
|
|
| 124 |
else
|
| 125 |
cap.open(0);
|
| 126 |
if (!cap.isOpened())
|
| 127 |
-
CV_Error(Error::StsError, "Cannot
|
| 128 |
Mat originalImage;
|
| 129 |
static const std::string kWinName = modelName;
|
| 130 |
while (waitKey(1) < 0)
|
|
|
|
| 124 |
else
|
| 125 |
cap.open(0);
|
| 126 |
if (!cap.isOpened())
|
| 127 |
+
CV_Error(Error::StsError, "Cannot open video or file");
|
| 128 |
Mat originalImage;
|
| 129 |
static const std::string kWinName = modelName;
|
| 130 |
while (waitKey(1) < 0)
|
models/text_recognition_crnn/demo.cpp
CHANGED
|
@@ -224,7 +224,7 @@ int main(int argc, char** argv)
|
|
| 224 |
else
|
| 225 |
cap.open(0);
|
| 226 |
if (!cap.isOpened())
|
| 227 |
-
CV_Error(Error::StsError, "Cannot
|
| 228 |
Mat originalImage;
|
| 229 |
static const std::string kWinName = modelPath;
|
| 230 |
while (waitKey(1) < 0)
|
|
|
|
| 224 |
else
|
| 225 |
cap.open(0);
|
| 226 |
if (!cap.isOpened())
|
| 227 |
+
CV_Error(Error::StsError, "Cannot open video or file");
|
| 228 |
Mat originalImage;
|
| 229 |
static const std::string kWinName = modelPath;
|
| 230 |
while (waitKey(1) < 0)
|
models/text_recognition_crnn/demo.py
CHANGED
|
@@ -106,7 +106,7 @@ if __name__ == '__main__':
|
|
| 106 |
|
| 107 |
# Save results if save is true
|
| 108 |
if args.save:
|
| 109 |
-
print('
|
| 110 |
cv.imwrite('result.jpg', original_image)
|
| 111 |
|
| 112 |
# Visualize results in a new window
|
|
|
|
| 106 |
|
| 107 |
# Save results if save is true
|
| 108 |
if args.save:
|
| 109 |
+
print('Results saved to result.jpg\n')
|
| 110 |
cv.imwrite('result.jpg', original_image)
|
| 111 |
|
| 112 |
# Visualize results in a new window
|