new draft

This commit is contained in:
sergi-e 2020-02-27 14:46:18 +01:00 committed by Marc Garcia Puig
parent 30ebbfe60f
commit 0fac729318
1 changed files with 5 additions and 5 deletions

View File

@ -67,7 +67,7 @@ sensor.listen(lambda data: do_something(data))
...
# This supposed collision sensor would print everytime a collision is detected.
# This collision sensor would print everytime a collision is detected.
def callback(event):
for actor_id in event:
vehicle = world_ref().get_actor(actor_id)
@ -95,14 +95,14 @@ Sensor data differs a lot between sensor types, but it is always tagged with:
<h4>Cameras</h4>
These sensors take a shot of the world from their point of view and then use the helper class to alter this image and provide different types of information.
These sensors take a shot of the world from their point of view and then use the helper class to alter this image and provide different types of information.
__Retrieve data:__ every simulation step.
| Sensor | Output | Overview |
| ---------- | ---------- | ---------- |
| Depth | [carla.Image](python_api.md#carla.Image) | Combines the photo with the distance of the elements on scene to provide with a gray-scale depth map. |
| Depth | [carla.Image](python_api.md#carla.Image) | Renders the depth of the elements in the field of view in a gray-scale depth map. |
| RGB | [carla.Image](python_api.md#carla.Image) | Provides clear vision of the surroundings. Looks like a normal photo of the scene. |
| Semantic segmentation | [carla.Image](python_api.md#carla.Image) | Uses the tags of the different actors in the photo to group the elements by color. |
| Semantic segmentation | [carla.Image](python_api.md#carla.Image) | Renders elements in the field of view with a specific color according to their tags. |
<h4>Detectors</h4>
@ -128,7 +128,7 @@ __Retrieve data:__ every simulation step.
| Radar | [carla.RadarMeasurement](python_api.md#carla.RadarMeasurement) | 2D point map that models elements in sight and their movement regarding the sensor. |
---------------
That is a wrap on sensors and how do these retrieve simulation data and thus, the introduction to CARLA is finished. However there is yet a lot to learn. Here are some different paths opened at the moment:
That is a wrap on sensors and how do these retrieve simulation data and thus, the introduction to CARLA is finished. However there is yet a lot to learn. Some of the different paths to follow now are listed here:
* __Gain some practise__: if diving alone in CARLA is still frightening, it may be a good idea to try some of the code recipes provided in this documentation and combine them with the example scripts or some ideas of your own.