Key Points Detection
Description
Key points detection is a tool that allows the user to detect key points in images using a pre-trained model. It is useful for various applications such as determining objects' orientation, pose estimation, and gesture recognition.
Settings
Model Folder Path
The model folder path is the path to the folder containing the model files. The model files are the files that are used to detect key points in the images. The model folder path is must be located on the server's machine.
See Model Folder Structure for more information about how to structure the model folder.
Supported Architectures
Here are the supported frameworks and architectures:
ONNX TensorFlow 2
- MoveNet Lightning
- Single Pose
- Multi Pose
- MoveNet Thunder
- Single Pose
- CenterNet
- 512x512
- EfficientNet B1
Label file
The label file is file containing the labels that will be used to label the key points. This is in JSON format.
Below is an example of a label file:
{
"names": [
"nose",
"left_eye",
"right_eye",
"left_ear",
"right_ear",
"left_shoulder",
"right_shoulder",
"left_elbow",
"right_elbow",
"left_wrist",
"right_wrist",
"left_hip",
"right_hip",
"left_knee",
"right_knee",
"left_ankle",
"right_ankle"
],
"connections": [
[0, 1],
[0, 2],
[1, 3],
[2, 4],
[0, 5],
[0, 6],
[5, 7],
[7, 9],
[6, 8],
[8, 10],
[5, 6],
[5, 11],
[6, 12],
[11, 12],
[11, 13],
[13, 15],
[12, 14],
[14, 16]
],
"colours": [
[3, 16, 255],
[255, 81, 0],
[3, 16, 255],
[255, 81, 0],
[3, 16, 255],
[255, 81, 0],
[3, 16, 255],
[3, 16, 255],
[255, 81, 0],
[255, 81, 0],
[0, 255, 196],
[3, 16, 255],
[255, 81, 0],
[0, 255, 196],
[3, 16, 255],
[3, 16, 255],
[255, 81, 0],
[255, 81, 0]
]
}
Where names
is the list of key point names, connections
is the list of connections between the key points, and colours
or colors
is the list of colours for each connection. The names
, connections
, and colours
or colors
are corresponding to each other.
For example, the first element in the names
list is nose
, the first element in the connections
list is [0, 1]
, and the first element in the colours
or colors
list is [3, 16, 255]
. This means that the nose
key point is connected to the left_eye
key point, and the colour of the connection is #FF1003
(red).
Architecture
The architecture is the type of model that will be used to segment the images. See Supported Architecture for more information.
Bounding boxes Confidence Threshold
The bounding boxes confidence threshold is the minimum confidence score a bounding box should have to be considered valid. Bounding boxes with confidence scores below this threshold will be discarded. A higher confidence threshold will result in fewer bounding boxes, but with higher accuracy, whereas a lower threshold will result in more bounding boxes, but with potentially lower accuracy.
Key points Confidence Threshold
The key points confidence threshold is the minimum confidence score a key point should have to be considered valid. Key points with confidence scores below this threshold will be discarded. A higher confidence threshold will result in fewer key points, but with higher accuracy, whereas a lower threshold will result in more key points, but with potentially lower accuracy.
Advanced Settings
Whether to enabled advanced settings. The advanced settings allow the user to have more control over the model parameters.
Please use the advanced settings with caution as changing the parameters could result in unexpected behaviour.
To RGB
Advanced SettingsConvert from BGR
colours (default in Zene) to RGB
colours.This is useful when using models trained on images with different channel orders.
Scale
Advanced SettingsThe scale is the scale factor that is applied to the input image. The scale factor is used to normalise the input image before it is passed to the model. The scale factor is calculated by dividing the input image by the scale factor. A higher scale factor will result in a smaller input image, while a lower scale factor will result in a larger input image.
Mean
Advanced SettingsThe mean is the mean value that is subtracted from the input image. The mean value is used to normalise the input image before it is passed to the model. The mean value is calculated by subtracting the input image by the mean value. A higher mean value will result in a darker input image, while a lower mean value will result in a brighter input image.
Input Pixel (width and height)
Advanced SettingsThe input pixel is the size of the input image. The input image is the image that is passed to the model. The user can specify the width and height of the input image.
Display Results
Overlay Results
Whether to draw the results on top of the image frame.
Draw Labels
Whether to draw the results on top of the image frame.