add carla_dataset

This commit is contained in:
freeneuro 2022-04-20 15:40:59 +08:00 committed by GitHub
parent 1e268451fc
commit 0e5eac09a9
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
1 changed files with 10 additions and 10 deletions

View File

@ -1,11 +1,11 @@
# The meaning for each folder in carla_dataset
+ masks folder contains all masks file, including train mask and test mask
+ train folder contain all train file, which ended with .npy. Each file contains sampled image, and the corresponding sampled information(i.e., location and rotation information of vehicle and camera).
+ test folder as the same to train folder
+ train_new folder contain all sample images extracted from each file of train folder.
+ test_new folder as the same to train_new folder
+ train_label_new folder which contain the label (refer to example file in train_label_new folder) of each training file
+ test_label_new folder as the same to train_label_new folder
## Note that
1. After you download the dataset and place them into each folder, you should create a yaml in data folder, see data/carala.yaml for example.
# The meaning for each folder in carla_dataset
+ masks folder contains all masks file, including train mask and test mask
+ train folder contain all train file, which ended with .npy. Each file contains sampled image, and the corresponding sampled information(i.e., location and rotation information of vehicle and camera).
+ test folder as the same to train folder
+ train_new folder contain all sample images extracted from each file of train folder.
+ test_new folder as the same to train_new folder
+ train_label_new folder which contain the label (refer to example file in train_label_new folder) of each training file
+ test_label_new folder as the same to train_label_new folder
## Note that
1. After you download the dataset and place them into each folder, you should create a yaml in data folder, see data/carala.yaml for example.
2. To run compute the adversarial loss, you need to get the label(annotation) for each image. In our implementation, for simplicity, we used the Faster RCNN to inference each rendered image(due to the character of rendered image, we can get relatively accuracy label) and get the corresponding label.