From d9d3d5be605912bce86bc816cba46f696482d660 Mon Sep 17 00:00:00 2001 From: Yuantao Feng Date: Thu, 8 Jun 2023 17:59:23 +0800 Subject: [PATCH] Add back example output images as lfs tracked files (#166) * lfs track images now * add back example images * modify paths * modify paths part 2 * correct paths in each model's readme --- .gitattributes | 8 ++++- README.md | 30 +++++++++---------- models/face_detection_yunet/README.md | 4 +-- .../example_outputs/largest_selfie.jpg | 3 ++ .../example_outputs/yunet_demo.gif | 3 ++ .../facial_expression_recognition/README.md | 2 +- .../example_outputs/selfie.jpg | 3 ++ .../handpose_estimation_mediapipe/README.md | 4 +-- .../example_outputs/hand_keypoints.png | 3 ++ .../example_outputs/mphandpose_demo.webp | 3 ++ .../human_segmentation_pphumanseg/README.md | 4 +-- .../example_outputs/messi.jpg | 3 ++ .../example_outputs/pphumanseg_demo.gif | 3 ++ .../license_plate_detection_yunet/README.md | 2 +- .../example_outputs/lpd_yunet_demo.gif | 3 ++ .../example_outputs/result-1.jpg | 3 ++ .../example_outputs/result-2.jpg | 3 ++ .../example_outputs/result-3.jpg | 3 ++ .../example_outputs/result-4.jpg | 3 ++ models/object_detection_nanodet/README.md | 6 ++-- .../example_outputs/1_res.jpg | 3 ++ .../example_outputs/2_res.jpg | 3 ++ .../example_outputs/3_res.jpg | 3 ++ .../example_outputs/WebCamR.gif | 3 ++ models/object_detection_yolox/README.md | 6 ++-- .../example_outputs/1_res.jpg | 3 ++ .../example_outputs/2_res.jpg | 3 ++ .../example_outputs/3_res.jpg | 3 ++ models/object_tracking_dasiamrpn/README.md | 2 +- .../example_outputs/dasiamrpn_demo.gif | 3 ++ models/palm_detection_mediapipe/README.md | 2 +- .../example_outputs/mppalmdet_demo.gif | 3 ++ models/person_detection_mediapipe/README.md | 2 +- .../example_outputs/mppersondet_demo.webp | 3 ++ models/pose_estimation_mediapipe/README.md | 2 +- .../example_outputs/mpposeest_demo.webp | 3 ++ .../example_outputs/pose_landmarks.png | 3 ++ models/qrcode_wechatqrcode/README.md | 2 +- .../example_outputs/wechat_qrcode_demo.gif | 3 ++ models/text_detection_db/README.md | 4 +-- .../example_outputs/gsoc.jpg | 3 ++ .../example_outputs/mask.jpg | 3 ++ models/text_recognition_crnn/README.md | 4 +-- .../example_outputs/CRNNCTC.gif | 3 ++ .../example_outputs/demo.jpg | 3 ++ 45 files changed, 132 insertions(+), 39 deletions(-) create mode 100644 models/face_detection_yunet/example_outputs/largest_selfie.jpg create mode 100644 models/face_detection_yunet/example_outputs/yunet_demo.gif create mode 100644 models/facial_expression_recognition/example_outputs/selfie.jpg create mode 100644 models/handpose_estimation_mediapipe/example_outputs/hand_keypoints.png create mode 100644 models/handpose_estimation_mediapipe/example_outputs/mphandpose_demo.webp create mode 100644 models/human_segmentation_pphumanseg/example_outputs/messi.jpg create mode 100644 models/human_segmentation_pphumanseg/example_outputs/pphumanseg_demo.gif create mode 100644 models/license_plate_detection_yunet/example_outputs/lpd_yunet_demo.gif create mode 100644 models/license_plate_detection_yunet/example_outputs/result-1.jpg create mode 100644 models/license_plate_detection_yunet/example_outputs/result-2.jpg create mode 100644 models/license_plate_detection_yunet/example_outputs/result-3.jpg create mode 100644 models/license_plate_detection_yunet/example_outputs/result-4.jpg create mode 100644 models/object_detection_nanodet/example_outputs/1_res.jpg create mode 100644 models/object_detection_nanodet/example_outputs/2_res.jpg create mode 100644 models/object_detection_nanodet/example_outputs/3_res.jpg create mode 100644 models/object_detection_nanodet/example_outputs/WebCamR.gif create mode 100644 models/object_detection_yolox/example_outputs/1_res.jpg create mode 100644 models/object_detection_yolox/example_outputs/2_res.jpg create mode 100644 models/object_detection_yolox/example_outputs/3_res.jpg create mode 100644 models/object_tracking_dasiamrpn/example_outputs/dasiamrpn_demo.gif create mode 100644 models/palm_detection_mediapipe/example_outputs/mppalmdet_demo.gif create mode 100644 models/person_detection_mediapipe/example_outputs/mppersondet_demo.webp create mode 100644 models/pose_estimation_mediapipe/example_outputs/mpposeest_demo.webp create mode 100644 models/pose_estimation_mediapipe/example_outputs/pose_landmarks.png create mode 100644 models/qrcode_wechatqrcode/example_outputs/wechat_qrcode_demo.gif create mode 100644 models/text_detection_db/example_outputs/gsoc.jpg create mode 100644 models/text_detection_db/example_outputs/mask.jpg create mode 100644 models/text_recognition_crnn/example_outputs/CRNNCTC.gif create mode 100644 models/text_recognition_crnn/example_outputs/demo.jpg diff --git a/.gitattributes b/.gitattributes index 62891cec..62dddc62 100644 --- a/.gitattributes +++ b/.gitattributes @@ -13,4 +13,10 @@ *.weights filter=lfs diff=lfs merge=lfs -text # ONNX -*.onnx filter=lfs diff=lfs merge=lfs -text \ No newline at end of file +*.onnx filter=lfs diff=lfs merge=lfs -text + +# Images +*.jpg filter=lfs diff=lfs merge=lfs -text +*.gif filter=lfs diff=lfs merge=lfs -text +*.png filter=lfs diff=lfs merge=lfs -text +*.webp filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md index 7b75fad2..a19d2ced 100644 --- a/README.md +++ b/README.md @@ -50,61 +50,61 @@ Some examples are listed below. You can find more in the directory of each model ### Face Detection with [YuNet](./models/face_detection_yunet/) -![largest selfie](./models/face_detection_yunet/examples/largest_selfie.jpg) +![largest selfie](./models/face_detection_yunet/example_outputs/largest_selfie.jpg) ### Facial Expression Recognition with [Progressive Teacher](./models/facial_expression_recognition/) -![fer demo](./models/facial_expression_recognition/examples/selfie.jpg) +![fer demo](./models/facial_expression_recognition/example_outputs/selfie.jpg) ### Human Segmentation with [PP-HumanSeg](./models/human_segmentation_pphumanseg/) -![messi](./models/human_segmentation_pphumanseg/examples/messi.jpg) +![messi](./models/human_segmentation_pphumanseg/example_outputs/messi.jpg) ### License Plate Detection with [LPD_YuNet](./models/license_plate_detection_yunet/) -![license plate detection](./models/license_plate_detection_yunet/examples/lpd_yunet_demo.gif) +![license plate detection](./models/license_plate_detection_yunet/example_outputs/lpd_yunet_demo.gif) ### Object Detection with [NanoDet](./models/object_detection_nanodet/) & [YOLOX](./models/object_detection_yolox/) -![nanodet demo](./models/object_detection_nanodet/samples/1_res.jpg) +![nanodet demo](./models/object_detection_nanodet/example_outputs/1_res.jpg) -![yolox demo](./models/object_detection_yolox/samples/3_res.jpg) +![yolox demo](./models/object_detection_yolox/example_outputs/3_res.jpg) ### Object Tracking with [DaSiamRPN](./models/object_tracking_dasiamrpn/) -![webcam demo](./models/object_tracking_dasiamrpn/examples/dasiamrpn_demo.gif) +![webcam demo](./models/object_tracking_dasiamrpn/example_outputs/dasiamrpn_demo.gif) ### Palm Detection with [MP-PalmDet](./models/palm_detection_mediapipe/) -![palm det](./models/palm_detection_mediapipe/examples/mppalmdet_demo.gif) +![palm det](./models/palm_detection_mediapipe/example_outputs/mppalmdet_demo.gif) ### Hand Pose Estimation with [MP-HandPose](models/handpose_estimation_mediapipe/) -![handpose estimation](models/handpose_estimation_mediapipe/examples/mphandpose_demo.webp) +![handpose estimation](models/handpose_estimation_mediapipe/example_outputs/mphandpose_demo.webp) ### Person Detection with [MP-PersonDet](./models/person_detection_mediapipe) -![person det](./models/person_detection_mediapipe/examples/mppersondet_demo.webp) +![person det](./models/person_detection_mediapipe/example_outputs/mppersondet_demo.webp) ### Pose Estimation with [MP-Pose](models/pose_estimation_mediapipe) -![pose_estimation](models/pose_estimation_mediapipe/examples/mpposeest_demo.webp) +![pose_estimation](models/pose_estimation_mediapipe/example_outputs/mpposeest_demo.webp) ### QR Code Detection and Parsing with [WeChatQRCode](./models/qrcode_wechatqrcode/) -![qrcode](./models/qrcode_wechatqrcode/examples/wechat_qrcode_demo.gif) +![qrcode](./models/qrcode_wechatqrcode/example_outputs/wechat_qrcode_demo.gif) ### Chinese Text detection [DB](./models/text_detection_db/) -![mask](./models/text_detection_db/examples/mask.jpg) +![mask](./models/text_detection_db/example_outputs/mask.jpg) ### English Text detection [DB](./models/text_detection_db/) -![gsoc](./models/text_detection_db/examples/gsoc.jpg) +![gsoc](./models/text_detection_db/example_outputs/gsoc.jpg) ### Text Detection with [CRNN](./models/text_recognition_crnn/) -![crnn_demo](./models/text_recognition_crnn/examples/CRNNCTC.gif) +![crnn_demo](./models/text_recognition_crnn/example_outputs/CRNNCTC.gif) ## License diff --git a/models/face_detection_yunet/README.md b/models/face_detection_yunet/README.md index 4097ac9b..1eb32854 100644 --- a/models/face_detection_yunet/README.md +++ b/models/face_detection_yunet/README.md @@ -53,9 +53,9 @@ cmake --build build ### Example outputs -![webcam demo](./examples/yunet_demo.gif) +![webcam demo](./example_outputs/yunet_demo.gif) -![largest selfie](./examples/largest_selfie.jpg) +![largest selfie](./example_outputs/largest_selfie.jpg) ## License diff --git a/models/face_detection_yunet/example_outputs/largest_selfie.jpg b/models/face_detection_yunet/example_outputs/largest_selfie.jpg new file mode 100644 index 00000000..fe494914 --- /dev/null +++ b/models/face_detection_yunet/example_outputs/largest_selfie.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab8413ad9bb4f53068f4fb63c6747e5989991dd02241c923d5595b614ecf2bf6 +size 1147146 diff --git a/models/face_detection_yunet/example_outputs/yunet_demo.gif b/models/face_detection_yunet/example_outputs/yunet_demo.gif new file mode 100644 index 00000000..099beab6 --- /dev/null +++ b/models/face_detection_yunet/example_outputs/yunet_demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db90459c308b14dd423014eabf3253f5f6147fbe7906e81429a7a88c8dbe7b8c +size 661072 diff --git a/models/facial_expression_recognition/README.md b/models/facial_expression_recognition/README.md index d8fc7719..f5d1415c 100644 --- a/models/facial_expression_recognition/README.md +++ b/models/facial_expression_recognition/README.md @@ -29,7 +29,7 @@ python demo.py --input /path/to/image -v Note: Zoom in to to see the recognized facial expression in the top-left corner of each face boxes. -![fer demo](./examples/selfie.jpg) +![fer demo](./example_outputs/selfie.jpg) ## License diff --git a/models/facial_expression_recognition/example_outputs/selfie.jpg b/models/facial_expression_recognition/example_outputs/selfie.jpg new file mode 100644 index 00000000..5a74c3d3 --- /dev/null +++ b/models/facial_expression_recognition/example_outputs/selfie.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3f8148169fe993afd0164200335a24301f1221a45535d7a938a0d133f2149ac +size 1233078 diff --git a/models/handpose_estimation_mediapipe/README.md b/models/handpose_estimation_mediapipe/README.md index a6c36019..d43ca313 100644 --- a/models/handpose_estimation_mediapipe/README.md +++ b/models/handpose_estimation_mediapipe/README.md @@ -2,7 +2,7 @@ This model estimates 21 hand keypoints per detected hand from [palm detector](../palm_detection_mediapipe). (The image below is referenced from [MediaPipe Hands Keypoints](https://github.com/tensorflow/tfjs-models/tree/master/hand-pose-detection#mediapipe-hands-keypoints-used-in-mediapipe-hands)) -![MediaPipe Hands Keypoints](./examples/hand_keypoints.png) +![MediaPipe Hands Keypoints](./example_outputs/hand_keypoints.png) This model is converted from TFlite to ONNX using following tools: - TFLite model to ONNX: https://github.com/onnx/tensorflow-onnx @@ -24,7 +24,7 @@ python demo.py -i /path/to/image -v ### Example outputs -![webcam demo](./examples/mphandpose_demo.webp) +![webcam demo](./example_outputs/mphandpose_demo.webp) ## License diff --git a/models/handpose_estimation_mediapipe/example_outputs/hand_keypoints.png b/models/handpose_estimation_mediapipe/example_outputs/hand_keypoints.png new file mode 100644 index 00000000..b3541222 --- /dev/null +++ b/models/handpose_estimation_mediapipe/example_outputs/hand_keypoints.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24ce896420149a7b9dc0c05e547681e3027b4a65ce3a0a6c73685ba1cc535496 +size 159592 diff --git a/models/handpose_estimation_mediapipe/example_outputs/mphandpose_demo.webp b/models/handpose_estimation_mediapipe/example_outputs/mphandpose_demo.webp new file mode 100644 index 00000000..1bb1441a --- /dev/null +++ b/models/handpose_estimation_mediapipe/example_outputs/mphandpose_demo.webp @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7873ed63f8c19a5961b53d866d71f0fe767f137e4ea6dbb6756a03b93b2b6a0 +size 1656194 diff --git a/models/human_segmentation_pphumanseg/README.md b/models/human_segmentation_pphumanseg/README.md index f10214d0..a9ca23c1 100644 --- a/models/human_segmentation_pphumanseg/README.md +++ b/models/human_segmentation_pphumanseg/README.md @@ -18,9 +18,9 @@ python demo.py --help ### Example outputs -![webcam demo](./examples/pphumanseg_demo.gif) +![webcam demo](./example_outputs/pphumanseg_demo.gif) -![messi](./examples/messi.jpg) +![messi](./example_outputs/messi.jpg) --- Results of accuracy evaluation with [tools/eval](../../tools/eval). diff --git a/models/human_segmentation_pphumanseg/example_outputs/messi.jpg b/models/human_segmentation_pphumanseg/example_outputs/messi.jpg new file mode 100644 index 00000000..6a152708 --- /dev/null +++ b/models/human_segmentation_pphumanseg/example_outputs/messi.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92fff88f42578a1bb46755ac11530fdcbb7b5f9a06ab478a45b1995feb1cd4e6 +size 62004 diff --git a/models/human_segmentation_pphumanseg/example_outputs/pphumanseg_demo.gif b/models/human_segmentation_pphumanseg/example_outputs/pphumanseg_demo.gif new file mode 100644 index 00000000..122ab0d4 --- /dev/null +++ b/models/human_segmentation_pphumanseg/example_outputs/pphumanseg_demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a386278ce825418231a371a0a2990f63ab0dc976bf03164517d9491150d34400 +size 548204 diff --git a/models/license_plate_detection_yunet/README.md b/models/license_plate_detection_yunet/README.md index 36ea80c7..c69e7820 100644 --- a/models/license_plate_detection_yunet/README.md +++ b/models/license_plate_detection_yunet/README.md @@ -19,7 +19,7 @@ python demo.py --help ### Example outputs -![lpd](./examples/lpd_yunet_demo.gif) +![lpd](./example_outputs/lpd_yunet_demo.gif) ## License diff --git a/models/license_plate_detection_yunet/example_outputs/lpd_yunet_demo.gif b/models/license_plate_detection_yunet/example_outputs/lpd_yunet_demo.gif new file mode 100644 index 00000000..f62dc8f1 --- /dev/null +++ b/models/license_plate_detection_yunet/example_outputs/lpd_yunet_demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e878ac62d49fca94f7eccaa5ac0b60e97508ef8225744a6a898f1bc833cee314 +size 300669 diff --git a/models/license_plate_detection_yunet/example_outputs/result-1.jpg b/models/license_plate_detection_yunet/example_outputs/result-1.jpg new file mode 100644 index 00000000..6f371d3b --- /dev/null +++ b/models/license_plate_detection_yunet/example_outputs/result-1.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff4f66031aa7ac82f1e218791c89e0655f3bdaf226a2b7272f0d48b5a62cb083 +size 58506 diff --git a/models/license_plate_detection_yunet/example_outputs/result-2.jpg b/models/license_plate_detection_yunet/example_outputs/result-2.jpg new file mode 100644 index 00000000..0b876b35 --- /dev/null +++ b/models/license_plate_detection_yunet/example_outputs/result-2.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6a18a5593bdd3794d7c484a9eb1a97f418d0daa5a35938d092e805a10c2df44 +size 55650 diff --git a/models/license_plate_detection_yunet/example_outputs/result-3.jpg b/models/license_plate_detection_yunet/example_outputs/result-3.jpg new file mode 100644 index 00000000..47f0ba0b --- /dev/null +++ b/models/license_plate_detection_yunet/example_outputs/result-3.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e9c8cc2ff8272075b73c1352fc93fb5b802737d2a89eefee47859f9737e5640 +size 63523 diff --git a/models/license_plate_detection_yunet/example_outputs/result-4.jpg b/models/license_plate_detection_yunet/example_outputs/result-4.jpg new file mode 100644 index 00000000..f9afaf44 --- /dev/null +++ b/models/license_plate_detection_yunet/example_outputs/result-4.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6257486bd6e08c6c1fd80874ad7cc8be2d1ed06e288d16670a04b9b8acb18530 +size 52606 diff --git a/models/object_detection_nanodet/README.md b/models/object_detection_nanodet/README.md index be39a831..f0adcc9b 100644 --- a/models/object_detection_nanodet/README.md +++ b/models/object_detection_nanodet/README.md @@ -22,13 +22,13 @@ Note: Here are some of the sample results that were observed using the model, -![test1_res.jpg](./samples/1_res.jpg) -![test2_res.jpg](./samples/2_res.jpg) +![test1_res.jpg](./example_outputs/1_res.jpg) +![test2_res.jpg](./example_outputs/2_res.jpg) Check [benchmark/download_data.py](../../benchmark/download_data.py) for the original images. Video inference result, -![WebCamR.gif](./samples/WebCamR.gif) +![WebCamR.gif](./example_outputs/WebCamR.gif) ## Model metrics: diff --git a/models/object_detection_nanodet/example_outputs/1_res.jpg b/models/object_detection_nanodet/example_outputs/1_res.jpg new file mode 100644 index 00000000..642fd18c --- /dev/null +++ b/models/object_detection_nanodet/example_outputs/1_res.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:920fb925f17720c68476fe77b396b87504081be0372662d33df0c0dcf9fc9562 +size 128531 diff --git a/models/object_detection_nanodet/example_outputs/2_res.jpg b/models/object_detection_nanodet/example_outputs/2_res.jpg new file mode 100644 index 00000000..1949805b --- /dev/null +++ b/models/object_detection_nanodet/example_outputs/2_res.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0138234ef80f63ecb51f51f651248a0157f5dda81d5b3fe390cbec42951bf99 +size 419826 diff --git a/models/object_detection_nanodet/example_outputs/3_res.jpg b/models/object_detection_nanodet/example_outputs/3_res.jpg new file mode 100644 index 00000000..675f320d --- /dev/null +++ b/models/object_detection_nanodet/example_outputs/3_res.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdf23ae1ceb2cde982c83763d74ea8317fceb3c5a091331cd5c7d39a08dda840 +size 114182 diff --git a/models/object_detection_nanodet/example_outputs/WebCamR.gif b/models/object_detection_nanodet/example_outputs/WebCamR.gif new file mode 100644 index 00000000..3b4dd944 --- /dev/null +++ b/models/object_detection_nanodet/example_outputs/WebCamR.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:169d6f224a20bef4c7a7c889d1b4e9168adf114da981edc3f9c087b7a3ec40ad +size 4814729 diff --git a/models/object_detection_yolox/README.md b/models/object_detection_yolox/README.md index 42f7dd0e..f6683575 100644 --- a/models/object_detection_yolox/README.md +++ b/models/object_detection_yolox/README.md @@ -29,9 +29,9 @@ Note: Here are some of the sample results that were observed using the model (**yolox_s.onnx**), -![1_res.jpg](./samples/1_res.jpg) -![2_res.jpg](./samples/2_res.jpg) -![3_res.jpg](./samples/3_res.jpg) +![1_res.jpg](./example_outputs/1_res.jpg) +![2_res.jpg](./example_outputs/2_res.jpg) +![3_res.jpg](./example_outputs/3_res.jpg) Check [benchmark/download_data.py](../../benchmark/download_data.py) for the original images. diff --git a/models/object_detection_yolox/example_outputs/1_res.jpg b/models/object_detection_yolox/example_outputs/1_res.jpg new file mode 100644 index 00000000..aab8c964 --- /dev/null +++ b/models/object_detection_yolox/example_outputs/1_res.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e03d69d6e1420bb84f3426b7d5f607cdfc86e522e33ee646cfc970e9ff53d9ea +size 124300 diff --git a/models/object_detection_yolox/example_outputs/2_res.jpg b/models/object_detection_yolox/example_outputs/2_res.jpg new file mode 100644 index 00000000..a13b5f3e --- /dev/null +++ b/models/object_detection_yolox/example_outputs/2_res.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a54924474eeb1c2d8cbb4f2245e003c28aff8256eebfa1bf2653f4af41eaa66 +size 402200 diff --git a/models/object_detection_yolox/example_outputs/3_res.jpg b/models/object_detection_yolox/example_outputs/3_res.jpg new file mode 100644 index 00000000..42ac0a1f --- /dev/null +++ b/models/object_detection_yolox/example_outputs/3_res.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e26d3b5cb7ac326fd2d431611f397ad710d210a22109fa6d876dacee262db63c +size 119379 diff --git a/models/object_tracking_dasiamrpn/README.md b/models/object_tracking_dasiamrpn/README.md index 054d3404..339072dc 100644 --- a/models/object_tracking_dasiamrpn/README.md +++ b/models/object_tracking_dasiamrpn/README.md @@ -23,7 +23,7 @@ python demo.py --help ### Example outputs -![webcam demo](./examples/dasiamrpn_demo.gif) +![webcam demo](./example_outputs/dasiamrpn_demo.gif) ## License diff --git a/models/object_tracking_dasiamrpn/example_outputs/dasiamrpn_demo.gif b/models/object_tracking_dasiamrpn/example_outputs/dasiamrpn_demo.gif new file mode 100644 index 00000000..dde59c67 --- /dev/null +++ b/models/object_tracking_dasiamrpn/example_outputs/dasiamrpn_demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a03746ec263e63c15d5b5bd8a6adb9a2691f63c415d9a8dcfd58991070b6fb9 +size 911768 diff --git a/models/palm_detection_mediapipe/README.md b/models/palm_detection_mediapipe/README.md index 211774d1..75de371e 100644 --- a/models/palm_detection_mediapipe/README.md +++ b/models/palm_detection_mediapipe/README.md @@ -26,7 +26,7 @@ python demo.py --help ### Example outputs -![webcam demo](./examples/mppalmdet_demo.gif) +![webcam demo](./example_outputs/mppalmdet_demo.gif) ## License diff --git a/models/palm_detection_mediapipe/example_outputs/mppalmdet_demo.gif b/models/palm_detection_mediapipe/example_outputs/mppalmdet_demo.gif new file mode 100644 index 00000000..98dae4d4 --- /dev/null +++ b/models/palm_detection_mediapipe/example_outputs/mppalmdet_demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4a6e6ff306117f575807ea05bf06c67190bf16fe2d315873acd0824a678dfaf +size 2178521 diff --git a/models/person_detection_mediapipe/README.md b/models/person_detection_mediapipe/README.md index 78c0da4b..a3b80472 100644 --- a/models/person_detection_mediapipe/README.md +++ b/models/person_detection_mediapipe/README.md @@ -23,7 +23,7 @@ python demo.py --help ### Example outputs -![webcam demo](examples/mppersondet_demo.webp) +![webcam demo](./example_outputs/mppersondet_demo.webp) ## License diff --git a/models/person_detection_mediapipe/example_outputs/mppersondet_demo.webp b/models/person_detection_mediapipe/example_outputs/mppersondet_demo.webp new file mode 100644 index 00000000..7cc4ec96 --- /dev/null +++ b/models/person_detection_mediapipe/example_outputs/mppersondet_demo.webp @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c2aeb6b5f2afa91063c737f983cf7e46e8096decd8476cc7817c0f8523d22e1 +size 708710 diff --git a/models/pose_estimation_mediapipe/README.md b/models/pose_estimation_mediapipe/README.md index 8a874b9b..a59977a7 100644 --- a/models/pose_estimation_mediapipe/README.md +++ b/models/pose_estimation_mediapipe/README.md @@ -22,7 +22,7 @@ python demo.py -i /path/to/image -v ### Example outputs -![webcam demo](examples/mpposeest_demo.webp) +![webcam demo](./example_outputs/mpposeest_demo.webp) ## License diff --git a/models/pose_estimation_mediapipe/example_outputs/mpposeest_demo.webp b/models/pose_estimation_mediapipe/example_outputs/mpposeest_demo.webp new file mode 100644 index 00000000..2e43b190 --- /dev/null +++ b/models/pose_estimation_mediapipe/example_outputs/mpposeest_demo.webp @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f95c6e80fa90dd22b06a88b95d8dac512e52192d8367ea6b5f576bd667df3d4c +size 1564162 diff --git a/models/pose_estimation_mediapipe/example_outputs/pose_landmarks.png b/models/pose_estimation_mediapipe/example_outputs/pose_landmarks.png new file mode 100644 index 00000000..50441c72 --- /dev/null +++ b/models/pose_estimation_mediapipe/example_outputs/pose_landmarks.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c93063a83adff4db00c02aec8cf04d7444ae9169956c8ec67ee2351adbcd8c0f +size 123013 diff --git a/models/qrcode_wechatqrcode/README.md b/models/qrcode_wechatqrcode/README.md index 380a27e6..f310b48d 100644 --- a/models/qrcode_wechatqrcode/README.md +++ b/models/qrcode_wechatqrcode/README.md @@ -23,7 +23,7 @@ python demo.py --help ### Example outputs -![webcam demo](./examples/wechat_qrcode_demo.gif) +![webcam demo](./example_outputs/wechat_qrcode_demo.gif) ## License diff --git a/models/qrcode_wechatqrcode/example_outputs/wechat_qrcode_demo.gif b/models/qrcode_wechatqrcode/example_outputs/wechat_qrcode_demo.gif new file mode 100644 index 00000000..1980e19b --- /dev/null +++ b/models/qrcode_wechatqrcode/example_outputs/wechat_qrcode_demo.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef1aa6f9b78320b3e3d6032648261dcfe250db332a58455787c88a87711a8b58 +size 1785414 diff --git a/models/text_detection_db/README.md b/models/text_detection_db/README.md index c55b6467..52a5ec68 100644 --- a/models/text_detection_db/README.md +++ b/models/text_detection_db/README.md @@ -25,9 +25,9 @@ python demo.py --help ### Example outputs -![mask](./examples/mask.jpg) +![mask](./example_outputs/mask.jpg) -![gsoc](./examples/gsoc.jpg) +![gsoc](./example_outputs/gsoc.jpg) ## License diff --git a/models/text_detection_db/example_outputs/gsoc.jpg b/models/text_detection_db/example_outputs/gsoc.jpg new file mode 100644 index 00000000..58d29c89 --- /dev/null +++ b/models/text_detection_db/example_outputs/gsoc.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6c852b20c3b187d3eefc7e0d4e89a89ec96637dfc544f8169bcfe4981ce8143 +size 314342 diff --git a/models/text_detection_db/example_outputs/mask.jpg b/models/text_detection_db/example_outputs/mask.jpg new file mode 100644 index 00000000..5f36556c --- /dev/null +++ b/models/text_detection_db/example_outputs/mask.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d5b84065442652e94a78fbcf11f210668862f205dad52e7fbf1642a5371898d +size 121326 diff --git a/models/text_recognition_crnn/README.md b/models/text_recognition_crnn/README.md index c3329d06..50c40a47 100644 --- a/models/text_recognition_crnn/README.md +++ b/models/text_recognition_crnn/README.md @@ -62,9 +62,9 @@ python demo.py --help ### Examples -![CRNNCTC](./examples/CRNNCTC.gif) +![CRNNCTC](./example_outputs/CRNNCTC.gif) -![demo](./examples/demo.jpg) +![demo](./example_outputs/demo.jpg) ## License diff --git a/models/text_recognition_crnn/example_outputs/CRNNCTC.gif b/models/text_recognition_crnn/example_outputs/CRNNCTC.gif new file mode 100644 index 00000000..09689aaf --- /dev/null +++ b/models/text_recognition_crnn/example_outputs/CRNNCTC.gif @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad60d87b58f365d168ae4d444dc27306e6d379ed16dbe82b44f443a43f4e65db +size 5249246 diff --git a/models/text_recognition_crnn/example_outputs/demo.jpg b/models/text_recognition_crnn/example_outputs/demo.jpg new file mode 100644 index 00000000..35ae4184 --- /dev/null +++ b/models/text_recognition_crnn/example_outputs/demo.jpg @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93b5838416d9d131f7a0fe3f00addfce0ed984052c15f69a8904d553066aa0aa +size 39430