When running the FPENet deepstream example application on a camera input, the confidence on each landmark seems to be in-between 0.2 and 0.35
When the facial landmarks seem to be very accurate it goes to the 0.3 side, when turning profile and the hidden points are all over the place, it drops to 0.2.
Is this expected or is there something wrong ? I was expecting the values to fluctuate between 0.0 and 1.0
(the orange bar on the left indicates the average confidence. ignore the blue graph)
My question is : if the output of the FPENet from the confidence output layer when running in deepstream as secondary model is influenced (multiplied) by the confidence of the primary model.
In the model card I can see that the output will give:
N X 1 keypoint confidence.
and a pixel accuracy
The region keypoint pixel error is the mean euclidean error in pixel location prediction as compared to the ground truth. We bucketize and average the error per face region (eyes, mouth, chin, etc.). Metric- Region keypoints pixel error
* All keypoints: 6.1
* Eyes region: 3.33
* Mouth region: 2.96
But not what the keypoint confidence range is in the pre-trained model.