forked from quic/ai-hub-models
-
Notifications
You must be signed in to change notification settings - Fork 0
/
demo.py
91 lines (78 loc) · 2.79 KB
/
demo.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
# ---------------------------------------------------------------------
# Copyright (c) 2024 Qualcomm Innovation Center, Inc. All rights reserved.
# SPDX-License-Identifier: BSD-3-Clause
# ---------------------------------------------------------------------
import argparse
import numpy as np
from PIL import Image
from qai_hub_models.models.mediapipe_face.app import MediaPipeFaceApp
from qai_hub_models.models.mediapipe_face.model import (
MODEL_ASSET_VERSION,
MODEL_ID,
MediaPipeFace,
)
from qai_hub_models.utils.args import add_output_dir_arg
from qai_hub_models.utils.asset_loaders import CachedWebModelAsset, load_image
from qai_hub_models.utils.camera_capture import capture_and_display_processed_frames
from qai_hub_models.utils.display import display_or_save_image
INPUT_IMAGE_ADDRESS = CachedWebModelAsset.from_asset_store(
MODEL_ID, MODEL_ASSET_VERSION, "face.jpeg"
)
# Run Mediapipe Face landmark detection end-to-end on a sample image or camera stream.
# The demo will display output with the predicted landmarks & bounding boxes drawn.
def main(is_test: bool = False):
# Demo parameters
parser = argparse.ArgumentParser()
parser.add_argument(
"--image",
type=str,
default=None,
help="image file path or URL",
)
parser.add_argument(
"--camera",
type=int,
default=0,
help="Camera Input ID",
)
parser.add_argument(
"--score-threshold",
type=float,
default=0.75,
help="Score threshold for NonMaximumSuppression",
)
parser.add_argument(
"--iou-threshold",
type=float,
default=0.3,
help="Intersection over Union (IoU) threshold for NonMaximumSuppression",
)
add_output_dir_arg(parser)
print(
"Note: This readme is running through torch, and not meant to be real-time without dedicated ML hardware."
)
print("Use Ctrl+C in your terminal to exit.")
args = parser.parse_args([] if is_test else None)
if is_test:
args.image = INPUT_IMAGE_ADDRESS
# Load app
app = MediaPipeFaceApp(
MediaPipeFace.from_pretrained(),
args.score_threshold,
args.iou_threshold,
)
print("Model and App Loaded")
if args.image:
image = load_image(args.image).convert("RGB")
pred_image = app.predict_landmarks_from_image(image)
out_image = Image.fromarray(pred_image[0], "RGB")
if not is_test:
display_or_save_image(out_image, args.output_dir)
else:
def frame_processor(frame: np.ndarray) -> np.ndarray:
return app.predict_landmarks_from_image(frame)[0] # type: ignore
capture_and_display_processed_frames(
frame_processor, "QAIHM Mediapipe Face Demo", args.camera
)
if __name__ == "__main__":
main()