Leveraging Heteroscedastic Aleatoric Uncertainties for Robust Real-Time LiDAR 3D Object Detection
Last updated
Last updated
Link of paper: https://arxiv.org/abs/1809.05590
Proposes: Real time - LiDAR 3D object Detection - alearotric uncertainties.
They also uses probabilistic 2 stage detector and model aleatoric uncertainties using additional auxiliary layers along with boudning box and classification layers.
They transformed 3D point cloud to the bird's eye view and then pass it as the input to the network.
They used same base architecture as Faster RCNN.
Added extra fully connected layers to output uncertainties, which is pretty standard.
They also predict uncertainties for RPN network. Although the RPN predicted undercertainties is not been used further in the architecure i.e either with NMS or postprocessing, but it is used for RPN regression loss attenuation.
They regress instead of for numerical stability.
Aleatoric undertainties are self contained in the softmax scores for classification.
Aleatoric uncertainty increases loss robustness against noist data. i.e. if some one input have large then it will factor down the loss contribution in the total loss. Hence, it is said the using uncertainty helps to igmore the noisy label.
Modelling uncertainties in both of RPN and fast rcnn head gets better performance than uncertainty in only one of them.
The RPN and Fast RCNN head uncertainties are highly correlated.
Far and occluded objects have higher uncertainties then near and easy objects.
According to hypothesis, the objects which are occluded or distance should have higher uncertainty. Now, we should test that the variance which we learnt and call uncertainty, support this hypothesis or not?
No need to explicitely model classic aleatoric unceratinties, as they are self-contained from the softmax scores which follow the categorical distribution. ? hecen , softmax scores can be used as measure of uncertainty for classification.
So instead of ingoring noisy labels, shouldn't it be something like that we can extract as much information from noisy samples as well.
We should try to use focal loss with two stage, which is opposite to ignoring outliers which have high variance.