From 2454972098006103db4841be1676275a9ccd903e Mon Sep 17 00:00:00 2001 From: Daniel Santos-Olivan Date: Fri, 31 Jul 2020 01:14:52 +0200 Subject: [PATCH] Refactored documentation --- Docs/bp_library.md | 2 +- ...oud.png => semantic_lidar_point_cloud.png} | Bin Docs/python_api.md | 108 +++++++++--------- Docs/ref_sensors.md | 61 +++++----- PythonAPI/docs/sensor.yml | 2 +- PythonAPI/docs/sensor_data.yml | 12 +- PythonAPI/util/raycast_sensor_testing.py | 48 ++++---- 7 files changed, 117 insertions(+), 116 deletions(-) rename Docs/img/{rawlidar_point_cloud.png => semantic_lidar_point_cloud.png} (100%) diff --git a/Docs/bp_library.md b/Docs/bp_library.md index 1f7405ff3..a02bdb6e6 100755 --- a/Docs/bp_library.md +++ b/Docs/bp_library.md @@ -155,7 +155,7 @@ Check out the [introduction to blueprints](core_actors.md). - `rotation_frequency` (_Float_)_ – Modifiable_ - `sensor_tick` (_Float_)_ – Modifiable_ - `upper_fov` (_Float_)_ – Modifiable_ -- **sensor.lidar.ray_cast_raw** +- **sensor.lidar.ray_cast_semantic** - **Attributes:** - `channels` (_Int_)_ – Modifiable_ - `lower_fov` (_Float_)_ – Modifiable_ diff --git a/Docs/img/rawlidar_point_cloud.png b/Docs/img/semantic_lidar_point_cloud.png similarity index 100% rename from Docs/img/rawlidar_point_cloud.png rename to Docs/img/semantic_lidar_point_cloud.png diff --git a/Docs/python_api.md b/Docs/python_api.md index 52acfc4bb..2c49b8985 100644 --- a/Docs/python_api.md +++ b/Docs/python_api.md @@ -1055,58 +1055,6 @@ Retrieves the number of points sorted by channel that are generated by this meas --- -## carla.LidarRawDetection -Data contained inside a [carla.LidarRawMeasurement](#carla.LidarRawMeasurement). Each of these represents one of the points in the cloud with its location and its asociated intensity. - -

Instance Variables

-- **point** (_[carla.Location](#carla.Location)_) -Point in xyz coordinates. -- **cos_inc_angle** (_float_) -Cosine of the incident angle between the ray and the normal of the hit object. -- **object_idx** (_uint_) -Carla index of the hitted actor. -- **object_tag** (_uint_) -Semantic tag of the hitted component. - -

Methods

- -
Dunder methods
-
- **\__str__**(**self**) - ---- - -## carla.LidarRawMeasurement -
Inherited from _[carla.SensorData](#carla.SensorData)_

Class that defines the raw lidar data retrieved by a sensor.lidar.ray_cast_raw. This essentially simulates a rotating lidar using ray-casting. Learn more about this [here](ref_sensors.md#rawlidar-raycast-sensor). - -

Instance Variables

-- **channels** (_int_) -Number of lasers shot. -- **horizontal_angle** (_float_) -Horizontal angle the Lidar is rotated at the time of the measurement (in radians). -- **raw_data** (_bytes_) -Received list of raw detection points. Each point consists in a 3D-xyz data plus cosine of the incident angle, the idx of the hit actor and its semantic tag. - -

Methods

-- **save_to_disk**(**self**, **path**) -Saves the point cloud to disk as a .ply file describing data from 3D scanners. The files generated are ready to be used within [MeshLab](http://www.meshlab.net/), an open source system for processing said files. Just take into account that axis may differ from Unreal Engine and so, need to be reallocated. - - **Parameters:** - - `path` (_str_) - -
Getters
-
- **get_point_count**(**self**, **channel**) -Retrieves the number of points sorted by channel that are generated by this measure. Sorting by channel allows to identify the original channel for every point. - - **Parameters:** - - `channel` (_int_) - -
Dunder methods
-
- **\__getitem__**(**self**, **pos**=int) -- **\__iter__**(**self**) -- **\__len__**(**self**) -- **\__setitem__**(**self**, **pos**=int, **detection**=[carla.LidarRawDetection](#carla.LidarRawDetection)) -- **\__str__**(**self**) - ---- - ## carla.Light This class exposes the lights that exist in the scene, except for vehicle lights. The properties of a light can be queried and changed at will. Lights are automatically turned on when the simulator enters night mode (sun altitude is below zero). @@ -1737,6 +1685,58 @@ Sets the log level. --- +## carla.SemanticLidarDetection +Data contained inside a [carla.SemanticLidarMeasurement](#carla.SemanticLidarMeasurement). Each of these represents one of the points in the cloud with its location and its asociated intensity. + +

Instance Variables

+- **point** (_[carla.Location](#carla.Location)_) +Point in xyz coordinates. +- **cos_inc_angle** (_float_) +Cosine of the incident angle between the ray and the normal of the hit object. +- **object_idx** (_uint_) +Carla index of the hitted actor. +- **object_tag** (_uint_) +Semantic tag of the hitted component. + +

Methods

+ +
Dunder methods
+
- **\__str__**(**self**) + +--- + +## carla.SemanticLidarMeasurement +
Inherited from _[carla.SensorData](#carla.SensorData)_

Class that defines the semantic lidar data retrieved by a sensor.lidar.ray_cast_semantic. This essentially simulates a rotating lidar using ray-casting. Learn more about this [here](ref_sensors.md#semanticlidar-raycast-sensor). + +

Instance Variables

+- **channels** (_int_) +Number of lasers shot. +- **horizontal_angle** (_float_) +Horizontal angle the Lidar is rotated at the time of the measurement (in radians). +- **raw_data** (_bytes_) +Received list of raw detection points. Each point consists in a 3D-xyz data plus cosine of the incident angle, the idx of the hit actor and its semantic tag. + +

Methods

+- **save_to_disk**(**self**, **path**) +Saves the point cloud to disk as a .ply file describing data from 3D scanners. The files generated are ready to be used within [MeshLab](http://www.meshlab.net/), an open source system for processing said files. Just take into account that axis may differ from Unreal Engine and so, need to be reallocated. + - **Parameters:** + - `path` (_str_) + +
Getters
+
- **get_point_count**(**self**, **channel**) +Retrieves the number of points sorted by channel that are generated by this measure. Sorting by channel allows to identify the original channel for every point. + - **Parameters:** + - `channel` (_int_) + +
Dunder methods
+
- **\__getitem__**(**self**, **pos**=int) +- **\__iter__**(**self**) +- **\__len__**(**self**) +- **\__setitem__**(**self**, **pos**=int, **detection**=[carla.SemanticLidarDetection](#carla.SemanticLidarDetection)) +- **\__str__**(**self**) + +--- + ## carla.Sensor
Inherited from _[carla.Actor](#carla.Actor)_

Sensors compound a specific family of actors quite diverse and unique. They are normally spawned as attachment/sons of a vehicle (take a look at [carla.World](#carla.World) to learn about actor spawning). Sensors are thoroughly designed to retrieve different types of data that they are listening to. The data they receive is shaped as different subclasses inherited from [carla.SensorData](#carla.SensorData) (depending on the sensor). @@ -1746,7 +1746,7 @@ Sets the log level. - [Gnss sensor](ref_sensors.md#gnss-sensor). - [IMU sensor](ref_sensors.md#imu-sensor). - [Lidar raycast](ref_sensors.md#lidar-raycast-sensor). - - [RawLidar raycast](ref_sensors.md#rawlidar-raycast-sensor). + - [SemanticLidar raycast](ref_sensors.md#semanticlidar-raycast-sensor). - [Radar](ref_sensors.md#radar-sensor). - [RGB camera](ref_sensors.md#rgb-camera). - [RSS sensor](ref_sensors.md#rss-sensor). @@ -1781,7 +1781,7 @@ Base class for all the objects containing data generated by a [carla.Sensor](#ca - IMU detector: [carla.IMUMeasurement](#carla.IMUMeasurement). - Lane invasion detector: [carla.LaneInvasionEvent](#carla.LaneInvasionEvent). - Lidar raycast: [carla.LidarMeasurement](#carla.LidarMeasurement). - - RawLidar raycast: [carla.LidarRawMeasurement](#carla.LidarRawMeasurement). + - SemanticLidar raycast: [carla.SemanticLidarMeasurement](#carla.SemanticLidarMeasurement). - Obstacle detector: [carla.ObstacleDetectionEvent](#carla.ObstacleDetectionEvent). - Radar detector: [carla.RadarMeasurement](#carla.RadarMeasurement). - RSS sensor: [carla.RssResponse](#carla.RssResponse). diff --git a/Docs/ref_sensors.md b/Docs/ref_sensors.md index 9bd7c8b0a..10e574e18 100644 --- a/Docs/ref_sensors.md +++ b/Docs/ref_sensors.md @@ -6,7 +6,7 @@ * [__IMU sensor__](#imu-sensor) * [__Lane invasion detector__](#lane-invasion-detector) * [__Lidar raycast sensor__](#lidar-raycast-sensor) - * [__RawLidar raycast sensor__](#rawlidar-raycast-sensor) + * [__SemanticLidar raycast sensor__](#semanticlidar-raycast-sensor) * [__Obstacle detector__](#obstacle-detector) * [__Radar sensor__](#radar-sensor) * [__RGB camera__](#rgb-camera) @@ -486,16 +486,16 @@ where a is the attenuation coefficient and d is the distance to the sensor. In order to increase the realism, we add the possibility of dropping cloud points. This is done in two different ways. In a general way, we can randomly drop points with a probability given by dropoff_general_rate. In this case, the drop off of points is done before tracing the ray cast so adjust this parameter can increase our performance. If that parameter is set to zero it will be ignored. The second way to regulate the drop off of points is in a rate proportional to the intensity. This drop off rate will be proportional to the intensity from zero at dropoff_intensity_limit to dropoff_zero_intensity at zero intensity. -This output contains a cloud of simulation points with its intensity and thus, can be iterated to retrieve a list of their [`carla.LidarDetection`](python_api.md#carla.LidarDetection): +This output contains a cloud of simulation points and thus, can be iterated to retrieve a list of their [`carla.Location`](python_api.md#carla.Location): ```py -for detection in lidar_measurement: - print(detection) +for location in lidar_measurement: + print(location) ``` -The rotation of the LIDAR can be tuned to cover a specific angle on every simulation step (using a [fixed time-step](adv_synchrony_timestep.md)). For example, to rotate once per step (full circle output, as in the picture below), the rotation frequency and the simulated FPS should be equal.
__1.__ Set the sensor's frequency `sensors_bp['lidar'][0].set_attribute('rotation_frequency','10')`.
__2.__ Run the simulation using `python config.py --fps=10`. +The rotation of the LIDAR can be tuned to cover a specific angle on every simulation step (using a [fixed time-step](adv_synchrony_timestep.md)). For example, to rotate once per step (full circle output, as in the picture below), the rotation frequency and the simulated FPS should be equal.
__1.__ Set the sensor's frequency `sensors_bp['lidar'][0].set_attribute('rotation_frequency','10')`.
__2.__ Run the simulation using `python config.py --fps=10`. -![LidarPointCloud](img/lidar_point_cloud.png) +![LidarPointCloud](img/lidar_point_cloud.gif) #### Lidar attributes @@ -613,10 +613,10 @@ The rotation of the LIDAR can be tuned to cover a specific angle on every simula --- -## RawLidar raycast sensor +## SemanticLidar raycast sensor -* __Blueprint:__ sensor.lidar.ray_cast_raw -* __Output:__ [carla.LidarRawMeasurement](python_api.md#carla.LidarRawMeasurement) per step (unless `sensor_tick` says otherwise). +* __Blueprint:__ sensor.lidar.ray_cast_semantic +* __Output:__ [carla.SemanticLidarMeasurement](python_api.md#carla.SemanticLidarMeasurement) per step (unless `sensor_tick` says otherwise). This sensor simulates a rotating Lidar implemented using ray-casting that exposes all the information about the raycast hit. Its behaviour is quite similar to the [Lidar raycast sensor](#lidar-raycast-sensor) but this sensor does not have any of the intensity, dropoff or noise features and its output is more complete. The points are computed by adding a laser for each channel distributed in the vertical FOV. The rotation is simulated computing the horizontal angle that the Lidar rotated in a frame. The point cloud is calculated by doing a ray-cast for each laser in every step: @@ -624,18 +624,18 @@ The points are computed by adding a laser for each channel distributed in the ve A Lidar measurement contains a packet with all the points generated during a `1/FPS` interval. During this interval the physics are not updated so all the points in a measurement reflect the same "static picture" of the scene. -This output contains a cloud of lidar raw detections and therefore, it can be iterated to retrieve a list of their [`carla.LidarRawDetection`](python_api.md#carla.LidarRawDetection): +This output contains a cloud of lidar semantic detections and therefore, it can be iterated to retrieve a list of their [`carla.SemanticLidarDetection`](python_api.md#carla.SemanticLidarDetection): ```py -for detection in lidar_raw_measurement: +for detection in semantic_lidar_measurement: print(detection) ``` The rotation of the LIDAR can be tuned to cover a specific angle on every simulation step (using a [fixed time-step](adv_synchrony_timestep.md)). For example, to rotate once per step (full circle output, as in the picture below), the rotation frequency and the simulated FPS should be equal.
__1.__ Set the sensor's frequency `sensors_bp['lidar'][0].set_attribute('rotation_frequency','10')`.
__2.__ Run the simulation using `python config.py --fps=10`. -![LidarPointCloud](img/rawlidar_point_cloud.png) +![LidarPointCloud](img/semantic_lidar_point_cloud.png) -#### Lidar attributes +#### SemanticLidar attributes @@ -720,10 +720,11 @@ The rotation of the LIDAR can be tuned to cover a specific angle on every simula - +
raw_data bytesArray that can be transform in raw detections, each of them have four 32-bits floats (XYZ of each point and consine of the incident angle) and two unsigned int (idx of the hitted actor and its semantic tag).Array that can be transform into semantic detections, each of them have four 32-bits floats (XYZ of each point and consine of the incident angle) and two unsigned int (idx of the hitted actor and its semantic tag).
+ --- ## Obstacle detector @@ -1232,21 +1233,21 @@ Since these effects are provided by UE, please make sure to check their document --- ## RSS sensor -* __Blueprint:__ sensor.other.rss -* __Output:__ [carla.RssResponse](python_api.md#carla.RssResponse) per step (unless `sensor_tick` says otherwise). +* __Blueprint:__ sensor.other.rss +* __Output:__ [carla.RssResponse](python_api.md#carla.RssResponse) per step (unless `sensor_tick` says otherwise). !!! Important - It is highly recommended to read the specific [rss documentation](adv_rss.md) before reading this. + It is highly recommended to read the specific [rss documentation](adv_rss.md) before reading this. -This sensor integrates the [C++ Library for Responsibility Sensitive Safety](https://github.com/intel/ad-rss-lib) in CARLA. It is disabled by default in CARLA, and it has to be explicitly built in order to be used. +This sensor integrates the [C++ Library for Responsibility Sensitive Safety](https://github.com/intel/ad-rss-lib) in CARLA. It is disabled by default in CARLA, and it has to be explicitly built in order to be used. -The RSS sensor calculates the RSS state of a vehicle and retrieves the current RSS Response as sensor data. The [carla.RssRestrictor](python_api.md#carla.RssRestrictor) will use this data to adapt a [carla.VehicleControl](python_api.md#carla.VehicleControl) before applying it to a vehicle. +The RSS sensor calculates the RSS state of a vehicle and retrieves the current RSS Response as sensor data. The [carla.RssRestrictor](python_api.md#carla.RssRestrictor) will use this data to adapt a [carla.VehicleControl](python_api.md#carla.VehicleControl) before applying it to a vehicle. -These controllers can be generated by an *Automated Driving* stack or user input. For instance, hereunder there is a fragment of code from `PythonAPI/examples/manual_control_rss.py`, where the user input is modified using RSS when necessary. +These controllers can be generated by an *Automated Driving* stack or user input. For instance, hereunder there is a fragment of code from `PythonAPI/examples/manual_control_rss.py`, where the user input is modified using RSS when necessary. -__1.__ Checks if the __RssSensor__ generates a valid response containing restrictions. -__2.__ Gathers the current dynamics of the vehicle and the vehicle physics. -__3.__ Applies restrictions to the vehicle control using the response from the RssSensor, and the current dynamics and physicis of the vehicle. +__1.__ Checks if the __RssSensor__ generates a valid response containing restrictions. +__2.__ Gathers the current dynamics of the vehicle and the vehicle physics. +__3.__ Applies restrictions to the vehicle control using the response from the RssSensor, and the current dynamics and physicis of the vehicle. ```py rss_restriction = self._world.rss_sensor.acceleration_restriction if self._world.rss_sensor and self._world.rss_sensor.response_valid else None @@ -1261,7 +1262,7 @@ if rss_restriction: #### The carla.RssSensor class -The blueprint for this sensor has no modifiable attributes. However, the [carla.RssSensor](python_api.md#carla.RssSensor) object that it instantiates has attributes and methods that are detailed in the Python API reference. Here is a summary of them. +The blueprint for this sensor has no modifiable attributes. However, the [carla.RssSensor](python_api.md#carla.RssSensor) object that it instantiates has attributes and methods that are detailed in the Python API reference. Here is a summary of them. @@ -1305,7 +1306,7 @@ def _on_rss_response(weak_self, response): !!! Warning This sensor works fully on the client side. There is no blueprint in the server. Changes on the attributes will have effect __after__ the *listen()* has been called. -The methods available in this class are related to the routing of the vehicle. RSS calculations are always based on a route of the ego vehicle through the road network. +The methods available in this class are related to the routing of the vehicle. RSS calculations are always based on a route of the ego vehicle through the road network. The sensor allows to control the considered route by providing some key points, which could be the [carla.Transform](python_api.md#carla.Transform) in a [carla.Waypoint](python_api.md#carla.Waypoint). These points are best selected after the intersections to force the route to take the desired turn. @@ -1412,11 +1413,11 @@ def _on_actor_constellation_request(self, actor_constellation_data): --- ## Semantic segmentation camera -* __Blueprint:__ sensor.camera.semantic_segmentation -* __Output:__ [carla.Image](python_api.md#carla.Image) per step (unless `sensor_tick` says otherwise). +* __Blueprint:__ sensor.camera.semantic_segmentation +* __Output:__ [carla.Image](python_api.md#carla.Image) per step (unless `sensor_tick` says otherwise). This camera classifies every object in sight by displaying it in a different color according to its tags (e.g., pedestrians in a different color than vehicles). -When the simulation starts, every element in scene is created with a tag. So it happens when an actor is spawned. The objects are classified by their relative file path in the project. For example, meshes stored in `Unreal/CarlaUE4/Content/Static/Pedestrians` are tagged as `Pedestrian`. +When the simulation starts, every element in scene is created with a tag. So it happens when an actor is spawned. The objects are classified by their relative file path in the project. For example, meshes stored in `Unreal/CarlaUE4/Content/Static/Pedestrians` are tagged as `Pedestrian`. The server provides an image with the tag information __encoded in the red channel__: A pixel with a red value of `x` belongs to an object with tag `x`. This raw [carla.Image](python_api.md#carla.Image) can be stored and converted it with the help of __CityScapesPalette__ in [carla.ColorConverter](python_api.md#carla.ColorConverter) to apply the tags information and show picture with the semantic segmentation. @@ -1612,8 +1613,8 @@ The following tags are currently available: --- ## DVS camera -* __Blueprint:__ sensor.camera.dvs -* __Output:__ [carla.DVSEventArray](python_api.md#carla.DVSEventArray) per step (unless `sensor_tick` says otherwise). +* __Blueprint:__ sensor.camera.dvs +* __Output:__ [carla.DVSEventArray](python_api.md#carla.DVSEventArray) per step (unless `sensor_tick` says otherwise). A Dynamic Vision Sensor (DVS) or Event camera is a sensor that works radically differently from a conventional camera. Instead of capturing diff --git a/PythonAPI/docs/sensor.yml b/PythonAPI/docs/sensor.yml index 21b93fba2..466685b31 100644 --- a/PythonAPI/docs/sensor.yml +++ b/PythonAPI/docs/sensor.yml @@ -14,7 +14,7 @@ - [Gnss sensor](ref_sensors.md#gnss-sensor). - [IMU sensor](ref_sensors.md#imu-sensor). - [Lidar raycast](ref_sensors.md#lidar-raycast-sensor). - - [RawLidar raycast](ref_sensors.md#rawlidar-raycast-sensor). + - [SemanticLidar raycast](ref_sensors.md#semanticlidar-raycast-sensor). - [Radar](ref_sensors.md#radar-sensor). - [RGB camera](ref_sensors.md#rgb-camera). - [RSS sensor](ref_sensors.md#rss-sensor). diff --git a/PythonAPI/docs/sensor_data.yml b/PythonAPI/docs/sensor_data.yml index 7aa07e153..470266aae 100644 --- a/PythonAPI/docs/sensor_data.yml +++ b/PythonAPI/docs/sensor_data.yml @@ -12,7 +12,7 @@ - IMU detector: carla.IMUMeasurement. - Lane invasion detector: carla.LaneInvasionEvent. - Lidar raycast: carla.LidarMeasurement. - - RawLidar raycast: carla.LidarRawMeasurement. + - SemanticLidar raycast: carla.SemanticLidarMeasurement. - Obstacle detector: carla.ObstacleDetectionEvent. - Radar detector: carla.RadarMeasurement. - RSS sensor: carla.RssResponse. @@ -188,11 +188,11 @@ - def_name: __str__ # -------------------------------------- - - class_name: LidarRawMeasurement + - class_name: SemanticLidarMeasurement parent: carla.SensorData # - DESCRIPTION ------------------------ doc: > - Class that defines the raw lidar data retrieved by a sensor.lidar.ray_cast_raw. This essentially simulates a rotating lidar using ray-casting. Learn more about this [here](ref_sensors.md#rawlidar-raycast-sensor). + Class that defines the semantic lidar data retrieved by a sensor.lidar.ray_cast_semantic. This essentially simulates a rotating lidar using ray-casting. Learn more about this [here](ref_sensors.md#semanticlidar-raycast-sensor). # - PROPERTIES ------------------------- instance_variables: - var_name: channels @@ -237,15 +237,15 @@ - param_name: pos type: int - param_name: detection - type: carla.LidarRawDetection + type: carla.SemanticLidarDetection # -------------------------------------- - def_name: __str__ # -------------------------------------- - - class_name: LidarRawDetection + - class_name: SemanticLidarDetection # - DESCRIPTION ------------------------ doc: > - Data contained inside a carla.LidarRawMeasurement. Each of these represents one of the points in the cloud with its location and its asociated intensity. + Data contained inside a carla.SemanticLidarMeasurement. Each of these represents one of the points in the cloud with its location and its asociated intensity. # - PROPERTIES ------------------------- instance_variables: - var_name: point diff --git a/PythonAPI/util/raycast_sensor_testing.py b/PythonAPI/util/raycast_sensor_testing.py index 5029edfff..4f185642f 100644 --- a/PythonAPI/util/raycast_sensor_testing.py +++ b/PythonAPI/util/raycast_sensor_testing.py @@ -22,8 +22,8 @@ to 1.5M per second. In this mode we do not render anything but processing of the data is done. For example for profiling one lidar: python raycast_sensor_testing.py -ln 1 --profiling -For example for profiling one raw lidar: - python raycast_sensor_testing.py -rln 1 --profiling +For example for profiling one semantic lidar: + python raycast_sensor_testing.py -sln 1 --profiling And for profiling one radar: python raycast_sensor_testing.py -rn 1 --profiling @@ -156,8 +156,8 @@ class SensorManager: lidar.listen(self.save_lidar_image) return lidar - elif sensor_type == 'RawLiDAR': - lidar_bp = self.world.get_blueprint_library().find('sensor.lidar.ray_cast_raw') + elif sensor_type == 'SemanticLiDAR': + lidar_bp = self.world.get_blueprint_library().find('sensor.lidar.ray_cast_semantic') lidar_bp.set_attribute('range', '100') for key in sensor_options: @@ -165,7 +165,7 @@ class SensorManager: lidar = self.world.spawn_actor(lidar_bp, transform, attach_to=attached) - lidar.listen(self.save_rawlidar_image) + lidar.listen(self.save_semanticlidar_image) return lidar elif sensor_type == "Radar": @@ -225,7 +225,7 @@ class SensorManager: self.time_processing += (t_end-t_start) self.tics_processing += 1 - def save_rawlidar_image(self, image): + def save_semanticlidar_image(self, image): t_start = self.timer.time() disp_size = self.display_man.get_display_size() @@ -334,17 +334,17 @@ def one_run(args, client): SensorManager(world, display_manager, 'LiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '200', 'points_per_second': lidar_points_per_second, 'rotation_frequency': '20'}, [1, 1]) - # If any, we instanciate the required rawlidars - rawlidar_points_per_second = args.rawlidar_points + # If any, we instanciate the required semantic lidars + semanticlidar_points_per_second = args.semanticlidar_points - if args.rawlidar_number >= 3: - SensorManager(world, display_manager, 'RawLiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '50', 'points_per_second': rawlidar_points_per_second, 'rotation_frequency': '20'}, [1, 0]) + if args.semanticlidar_number >= 3: + SensorManager(world, display_manager, 'SemanticLiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '50', 'points_per_second': semanticlidar_points_per_second, 'rotation_frequency': '20'}, [1, 0]) - if args.rawlidar_number >= 2: - SensorManager(world, display_manager, 'RawLiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '100', 'points_per_second': rawlidar_points_per_second, 'rotation_frequency': '20'}, [0, 1]) + if args.semanticlidar_number >= 2: + SensorManager(world, display_manager, 'SemanticLiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '100', 'points_per_second': semanticlidar_points_per_second, 'rotation_frequency': '20'}, [0, 1]) - if args.rawlidar_number >= 1: - SensorManager(world, display_manager, 'RawLiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '200', 'points_per_second': rawlidar_points_per_second, 'rotation_frequency': '20'}, [1, 1]) + if args.semanticlidar_number >= 1: + SensorManager(world, display_manager, 'SemanticLiDAR', carla.Transform(carla.Location(x=0, z=2.4)), vehicle, {'channels' : '64', 'range' : '200', 'points_per_second': semanticlidar_points_per_second, 'rotation_frequency': '20'}, [1, 1]) # If any, we instanciate the required radars @@ -415,7 +415,7 @@ def one_run(args, client): time_procc = 0 for sensor in display_manager.sensor_list: time_procc += sensor.time_processing - prof_str = "%-10s %-9s %-9s %-15s %-7.2f %-20.3f" % (args.lidar_number, args.rawlidar_number, args.radar_number, lidar_points_per_second, float(frame) / time_frames, time_procc/time_frames) + prof_str = "%-10s %-9s %-9s %-15s %-7.2f %-20.3f" % (args.lidar_number, args.semanticlidar_number, args.radar_number, lidar_points_per_second, float(frame) / time_frames, time_procc/time_frames) break if call_exit: @@ -484,17 +484,17 @@ def main(): choices=range(0, 4), help='Number of lidars to render (from zero to three)') argparser.add_argument( - '-rlp', '--rawlidar_points', - metavar='RLP', + '-slp', '--semanticlidar_points', + metavar='SLP', default='100000', - help='lidar points per second (default: "100000")') + help='semantic lidar points per second (default: "100000")') argparser.add_argument( - '-rln', '--rawlidar_number', - metavar='RLN', + '-sln', '--semanticlidar_number', + metavar='SLN', default=0, type=int, choices=range(0, 4), - help='Number of raw lidars to render (from zero to three)') + help='Number of semantic lidars to render (from zero to three)') argparser.add_argument( '-rp', '--radar_points', metavar='RP', @@ -538,7 +538,7 @@ def main(): if args.profiling: print("-------------------------------------------------------") - print("# Running profiling with %s lidars, %s raw lidars and %s radars." % (args.lidar_number, args.rawlidar_number, args.radar_number)) + print("# Running profiling with %s lidars, %s semantic lidars and %s radars." % (args.lidar_number, args.semanticlidar_number, args.radar_number)) args.render_cam = False args.render_window = False runs_output = [] @@ -548,7 +548,7 @@ def main(): '1100000', '1200000', '1300000', '1400000', '1500000'] for points in points_range: args.lidar_points = points - args.rawlidar_points = points + args.semanticlidar_points = points args.radar_points = points run_str = one_run(args, client) runs_output.append(run_str) @@ -562,7 +562,7 @@ def main(): print("#Hardware information not available, please install the " \ "multiprocessing module") - print("#NumLidars NumRawLids NumRadars PointsPerSecond FPS PercentageProcessing") + print("#NumLidars NumSemLids NumRadars PointsPerSecond FPS PercentageProcessing") for o in runs_output: print(o)