"examples/pytorch/vscode:/vscode.git/clone" did not exist on "3e2f94ed20ad50de0907050b2b29f1303b01339c"
[Feature] Support ScanNet semantic segmentation dataset (#390)
* remove max_num_point in ScanNet data preprocessing * add config file for ScanNet semantic segmentation dataset * modify NormalizePointsColor in pipeline * add visualization function for semantic segmentation * add ignore_index to semantic segmentation visualization function * add ignore_index to semantic segmentation evaluation function * fix ignore_index bug in semantic segmentation evaluation function * add test function to check ignore_index assignment in PointSegClassMapping * fix slicing bug in BasePoints class and add unittest * add IndoorPatchPointSample class for indoor semantic segmentation data loading and add unittest * modify LoadPointsFromFile class and its unittest to support point color loading * fix data path in unittest * add setter function for coord and attributes of BasePoint and modify unittest * modify color normalization operation to work on BasePoint class * add unittest for ScanNet semantic segmentation data loading pipeline * fix ignore_index bug in seg_eval function * add ScanNet semantic segmentation dataset and unittest * modify config file for ScanNet semantic segmentation * fix visualization function and modify unittest * fix a typo in seg_eval.py * raise exception when semantic mask is not provided in train/eval data loading * support custom computation of label weight for loss calculation * modify seg_eval function to be more efficient * fix small bugs & change variable names for clarity & add more cases to unittest * move room index resampling and label weight computation to data pre-processing * add option allowing user to determine whether to sub-sample point clouds * fix typos & change .format to f-string & fix link in comment * save all visualizations into .obj format for consistency * infer num_classes from label2cat in eval_seg function * add pre-computed room index and label weight for ScanNet dataset * replace .ply with .obj in unittests and documents * add TODO in case data is on ceph * add base dataset for all semantic segmentation tasks & add ScanNet dataset inheriting from base dataset * rename class for consistency * fix minor typos in comment * move Custom3DSegDataset to a new file * modify BasePoint setter function to enable attribute adding * add unittest for NormalizePointsColor and fix small bugs * fix unittest for BasePoints * modify ScanNet data pre-processing scripts * change ignore_idx to -1 in seg_eval function * remove sliding inference from PatchSample function and modify unittest * remove PatchSample from scannet seg test_pipeline
Showing
Please register or sign in to comment