Our new model ZOO works with DepthAI V3. Find out more in our documentation.
0 Likes
Model Details
Model Description
Head pose estimation model is a lightweight, simple, and handmade convolutional neural network. The model works on cropped-face images and predicts how the face is oriented. This let us detect where the specific person is looking. It outputs 3 values yaw, pitch and roll.
Developed by: OpenVINO
Shared by:
Model type: Computer Vision
License:
Resources for more information:
Training Details
Training Data
The training dataset is not provided. For the evaluation dataset dataset is used.
Testing Details
Metrics
The get more information about evaluation, please check
Angle
Mean ± std. of absolute error
Yaw
5.4 ± 4.4
Pitch
5.5 ± 5.3
Roll
4.6 ± 5.6
Technical Specifications
Input/Output Details
Input:
Name: data
Info: NCHW BGR un-normalized image
Output:
Name: Multiple (please consult NN archive config.json)
Info: The three predicted values - yaw, pitch, roll.
Model Architecture
Head pose estimation network is based on simple, handmade CNN architecture. Angle regression layers are convolutions + ReLU + batch normalization + fully connected layer with one output.
* Benchmarked with , using 2 threads (and the DSP runtime in balanced mode for RVC4).
* Parameters and FLOPs are obtained from the package.
Quantization
The model is only supported in floating-point 16 and it is fast enough. It can still run on DSP.
Utilization
Models converted for RVC Platforms can be used for inference on OAK devices.
DepthAI pipelines are used to define the information flow linking the device, inference model, and the output parser (as defined in model head(s)).
Below, we present the most crucial utilization steps for the particular model.
Please consult the docs for more information.