-
 

robot_body_filter package from robot_body_filter repo

robot_body_filter

Package Summary

Tags No category tags.
Version 1.3.2
License BSD
Build type CATKIN
Use RECOMMENDED

Repository Summary

Checkout URI https://github.com/peci1/robot_body_filter.git
VCS Type git
VCS Version master
Last Updated 2024-02-03
Dev Status MAINTAINED
CI status
Released RELEASED
Tags No category tags.
Contributing Help Wanted (0)
Good First Issues (0)
Pull Requests to Review (0)

Package Description

Filters the robot's body out of laser scans or point clouds.

Additional Links

Maintainers

  • Martin Pecka

Authors

  • Eitan Marder-Eppstein
  • Tomas Petricek
  • Martin Pecka

robot_body_filter

Filters the robot’s body out of point clouds and laser scans.

Tutorial

Check out the webinar recording where a lot of options for this filter are explained and demonstrated! https://www.youtube.com/watch?v=j0ljV0uZy3Q

Changes vs PR2/robot_self_filter:

  • Now the package is a normal filters::FilterBase filter and not a standalone node.
  • Using both containment and ray-tracing tests.
  • Using all collision elements for each link instead of only the first one.
  • Enabling generic point type, removing PCL dependency and unnecessary params.
  • Using bodies.h and shapes.h from geometric_shapes.
  • As a by-product, the filter can compute robot’s bounding box or sphere.

Build Status

Development versions: Github Actions Dev melodic Dev noetic

Release jobs Melodic Melodic version: Bin melodic-amd64 Bin melodic-arm64 Bin melodic-armhf

Release jobs Noetic Noetic version: Bin noetic focal-amd64 Bin noetic focal-arm64 Bin noetic focal-armhf

Basic Operation

filters::FilterBase API

The basic workings of this filter are done via the filters::FilterBase API implemented for sensor_msgs::LaserScan and sensor_msgs::PointCloud2 types. This means you can load this filter into a FilterChain along other filters as usual. Different from the standard filters, this one can also publish several interesting topics and subscribes to TF.

General overview

This filter reads robot model and the filter config, subscribes to TF, waits for data (laserscans or point clouds) and then cleans them from various artifacts (this is called data filtering).

It can perform 3 kinds of data filters: clip the data based on the provided sensor limits (parameter filter/do_clipping), remove points that are inside or on the surface of the robot body (parameter filter/do_contains_test) and remove points that are seen through a part of the robot body (parameter filter/do_shadow_test). These kinds of tests are further referenced as “clipping”, “contains test” and “shadow test”.

If working with point clouds, the filter automatically recognizes whether it works with organized or non-organized clouds. In organized clouds, it marks the filtered-out points as NaN. In non-organized clouds, it removes the filtered-out points. In laserscans, removal is not an option, so the filtered-out points are marked with NaN (some guides suggest that max_range + 1 should be used for marking invalid points, but this filter uses NaN as a safer option).

Performance tips

In general, the filter will be computationally expensive (clipping is fast, contains test is medium CPU intensive and shadow test is the most expensive part, because it basically performs raytracing). You can limit the required CPU power by limiting the filter only to parts that matter. E.g. if the robot has a link that can never be seen by the sensor, put it in the list of ignored links. The less links are processed, the better performance. If you’re only interested in removing a few links, consider using the only_links parameter.

To speed up shadow filtering, you can set filter/max_shadow_distance, which limits the number of points considered for shadow tests just to points close to the sensor. Setting this to e.g. three times the diameter of the robot should remove all of the shadow points caused by refraction by a part of the robot body. But you have to test this with real data.

Performance also strongly depends on representation of the robot model. The filter reads <collision> tags from the robot URDF. You can use boxes, spheres and cylinders (which are fast to process), or you can use convex meshes (these are much worse performance-wise). If you pass a non-convex mesh, its convex hull will be used for the tests. Don’t forget that each link can have multiple <collision> tags. If you do not have time to convert your meshes to the basic shapes, try to at least reduce the number of triangles in your meshes. You can use your high-quality meshes in <visual> tags. To simplify your model to primitive shapes, you can either manually edit the URDF, or you can utilize ColliderGen.

Model inflation

You can utilize the builtin model inflation mechanism to slightly alter the size of the model. You will probably want to add a bit “margin” to the contains and shadow tests so that points that are millimeters outside the robot body will anyways get removed. You can set a default scale and padding which are used for all collisions. Different inflation can be used for contains tests and for shadow tests. Inflation can also be specified differently for each link. Look at the body_model/inflation/* parameters for details.

Scaling means multiplying the shape dimensions by the given factor (with its center staying in the same place). Padding means adding the specified metric distance to each “dimension” of the shape. Padding a sphere by p just adds p to its radius; padding a cylinder adds p to its radius and 2p to its length (you pad both the top and the bottom of the cylinder); padding a box adds 2p to all its extents (again, you pad both of the opposing sides); padding a mesh pads each vertex of its convex hull by p along the direction from the mesh center to the vertex (mesh center is the center specified in the mesh file, e.g. DAE). Have a good look at the effects of mesh padding, as the results can be non-intuitive.

Data acquisition modes

The filter supports data captured in two modes - all at once (e.g. RGBD cameras), or each point at a different time instant (mostly lidars). This is handled by sensor/point_by_point setting. Each mode supports both laser scans and point cloud input (although all-at-once laserscans aren’t that common). Point-by-point pointclouds are e.g. the output of 3D lidars like Ouster (where more points can have the same timestamp, but not all). If you want to use the point-by-point mode with pointclouds, make sure they contain not only the usual x, y and z fields, but also a float32 field stamps (with time difference from the time in the header) and float32 fields vp_x, vp_y and vp_z which contain the viewpoint (position of the sensor in filtering frame) from which the robot saw that point.

When filtering in the point-by-point mode, the robot posture has to be updated several times during processing a single scan (to reflect the motion the robot has performed during acquiring the scan). The frequency of these updates can also have a significant impact on performance. Use parameter filter/model_pose_update_interval to set the interval for which the robot is considered stationary. The positions of the robot at the beginning and at the end of the scan are queried from TF. The intermediate positions are linearly interpolated between these two positions.

TF frames

The filter recognizes four logical TF frames (some of which may be the same physical frames).

Fixed frame is a frame that doesn’t change within the duration of the processed scan. This means for all-at-once scans, this frame is not needed because the duration of the scan is zero. For point-by-point scans, it depends on the particular scenario. In static installations (like manipulators with sensors not attached to them), it can be the sensor frame. For stationary robots with the sensor attached to a movable part of their body, base_link will be a good choice. For completely mobile robots, you will need an external frame, e.g. odom or map (beware of cyclic dependencies - if the map is built from the filtered scans, you obviously cannot use map as the fixed frame for filtering the scans…).

Sensor frame is the frame in which the data were captured. Generally, it would be whatever is in header.frame_id field of the processed messages. You can use the filter for data from multiple sensors - in that case, you can leave the sensor frame unfilled and each message will be processed in the frame it has in its header.

Filtering frame is the frame in which the data filtering is done. For point-by-point scans, it has to be a fixed frame (possibly different from the fixed frame set in frames/fixed). For pointcloud scans, it should be the sensor frame (if all data are coming from a single sensor), or any other frame. It is also used as the frame in which all debugging outputs are published.

Output frame can only be used with pointcloud scans, and allows to transform the filtered pointcloud to a different frame before being published. It is just a convenience which can save you launching a transformation nodelet. By default, filtered pointclouds are output in the filtering frame.

Bounding shapes

As a byproduct, the filter can also compute various bounding shapes of the robot model. There are actually four robot models - one for contains test, one for shadow test, one for bounding sphere computation and one for bounding box (these models can differ by inflation and considered links). All bounding shapes are published in the filtering frame. For point-by-point scans, the bounding shapes correspond to the time instant specified in the header of the processed scan. The computation of bounding shapes is off by default, but enabling it is cheap (performance-wise).

The bounding sphere is easy - the smallest sphere that contains the whole collision model for bounding sphere computation (with the specified exclusions removed).

The bounding box is the smallest axis-aligned bounding box aligned to the filtering frame. It is built from the model for bounding box computation.

The local bounding box is the smallest axis-aligned bounding box aligned to the frame specified in local_bounding_box/frame_id. It is especially useful with mobile robots when the desired frame is base_link. It is built from the model for bounding box computation.

The oriented bounding box should be the smallest box containing the collision model. However, its computation is very bad conditioned, so the results can be very unsatisfactory. Currently, the oriented bounding box of each of the basic collision shapes is “tight”, but merging the boxes is not optimal. A good algorithm would probably require costly and advanced iterative methods. The current implementation uses FCL in the background and merges the boxes using fcl::OBB::operator+=() without any further optimizations. It is built from the model for bounding box computation.

The filter also supports publishing auxiliary pointclouds which “cut out” each of these bounding shapes. These are the input data converted to pointcloud in filtering frame from which all points belonging to the bounding shape are removed. Please note that the “base” used for cutting out is the input pointcloud, not the filtered one.

First setup/debugging

The filter offers plenty of debugging outputs to make sure it does exactly what you want it to do. All the options are described in the last part of this page. Generally, you should look at the pointclouds visualizing which points got filtered out and you should also check the robot models used for filtering.

Also, have a look in the [examples] folder to get some inspiration.

Usage

This is a standard filters::FilterBase<T>-based filter which implements the configure() and update(const T&, T&) methods. This means it can be loaded e.g. as a part of a filter chain via the laser_filters package or the relatively new sensor_filters. This means the input and output data are not supplied in the form of topics, but they are instead passed to the update() method via the C++ API.

This filter is a bit unusual - it subscribes and publishes several topics. Normally, filters only operate via the update() method and are not expected to publish anything. This filter is different, it requires a working ROS node handle, and can publish a lot of auxiliary or debug data.

Subscribed Topics

  • /tf, /tf_static

    Transforms

  • dynamic_robot_model_server/parameter_updates (dynamic_reconfigure/Config)

    Dynamic reconfigure topic on which updated robot model can be subscribed. The model is read from a field defined by parameter body_model/dynamic_robot_description/field_name.

Published Topics

  • scan_point_cloud_no_bsphere (sensor_msgs/PointCloud2)

    Point cloud with points inside body bounding sphere removed. Turned on by bounding_sphere/publish_cut_out_pointcloud parameter. Published in filtering frame.

  • scan_point_cloud_no_bbox (sensor_msgs/PointCloud2)

    Point cloud with points inside filtering frame axis-aligned bounding box removed. Turned on by bounding_box/publish_cut_out_pointcloud parameter. Published in filtering frame.

  • scan_point_cloud_no_oriented_bbox (sensor_msgs/PointCloud2)

    Point cloud with points inside oriented bounding box removed. Turned on by oriented_bounding_box/publish_cut_out_pointcloud parameter. Published in filtering frame. Please read the remarks above in the overview - the results can be non-satisfying.

  • scan_point_cloud_no_local_bbox (sensor_msgs/PointCloud2)

    Point cloud with points inside local_bounding_box/frame_id axis-aligned bounding box removed. Turned on by local_bounding_box/publish_cut_out_pointcloud parameter. Published in filtering frame.

  • robot_bounding_sphere (robot_body_filter/SphereStamped)

    Bounding sphere of the robot body. Turned on by bounding_sphere/compute parameter. Published in filtering frame.

  • robot_bounding_box (geometry_msgs/PolygonStamped)

    Axis-aligned bounding box of the robot body aligned to the filtering frame. First point is the minimal point, second one is the maximal point. Turned on by bounding_box/compute parameter. Published in filtering frame.

  • robot_oriented_bounding_box (robot_body_filter/OrientedBoundingBoxStamped)

    Oriented bounding box of the robot body. Turned on by oriented_bounding_box/compute parameter. Please read the remarks above in the overview - the results can be non-satisfying. Published in filtering frame.

  • robot_local_bounding_box (geometry_msgs/PolygonStamped)

    Axis-aligned bounding box of the robot body aligned to frame local_bounding_box/frame_id. First point is the minimal point, second one is the maximal point. Turned on by local_bounding_box/compute parameter. Published in frame local_bounding_box/frame_id.

Provided Services

  • ~reload_model (std_srvs/Trigger)

    Call this service to re-read the URDF model from parameter server.

Filter Parameters

Have a look in the examples folder to get inspiration for configuration of your filter.

  • sensor/point_by_point (bool, default: false for PointCloud2 version, true for LaserScan)

    If true, suppose that every point in the scan was captured at a different time instant. Otherwise, the scan is assumed to be taken at once. If this is true, the processing pipeline expects the pointcloud to have fields int32 index, float32 stamps, and float32 vp_x, vp_y and vp_z viewpoint positions. If one of these fields is missing, computeMask() throws runtime exception.

  • frames/fixed (string, default: "base_link")

    The fixed frame. Usually base_link for stationary robots (or sensor frame if both robot and sensor are stationary). For mobile robots, it can be e.g. odom or map. Only needed for point-by-point scans.

  • frames/sensor (string, default "")

    Frame of the sensor. If set to empty string, it will be read from header.frame_id of each incoming message. In LaserScan version, if nonempty, it has to match the frame_id of the incoming scans. In PointCloud2 version, the data can come in a different frame from frames/sensor.

  • frames/filtering (string, default: frames/fixed)

    Frame in which the filter is applied. For point-by-point scans, it has to be a fixed frame, otherwise, it can be the sensor frame or any other frame. Setting to sensor frame will save some computations, but this frame cannot be empty string, so sensor frame can only be used if all data are coming from a single sensor and you know the scan frame in advance.

  • frames/output (string, default: frames/filtering)

    Frame into which output data are transformed. Only applicable in PointCloud2 version.

  • sensor/min_distance (float, default: 0.0 m)

    The minimum distance of points from the laser to keep them.

  • sensor/max_distance (float, default: 0.0 m)

    The maximum distance of points from the laser to keep them. Set to zero to disable this limit.

  • filter/keep_clouds_organized (bool, default true)

    Whether to keep pointclouds organized (if they were). Organized cloud has height > 1.

  • filter/model_pose_update_interval (float, default 0.0 s)

    The interval between two consecutive model pose updates when processing a pointByPointScan. If set to zero, the model will be updated for each point separately (might be computationally exhaustive). If non-zero, it will only update the model once in this interval, which makes the masking algorithm a little bit less precise but more computationally affordable.

  • filter/do_clipping (bool, default true)

    If true, the filter will mark points outside sensor/(min|max)_distance as CLIP. If false, such points will be marked OUTSIDE.

  • filter/do_contains_test (bool, default true)

    If true, the filter will mark points inside robot body as INSIDE. If false, such points will be marked OUTSIDE.

  • filter/do_shadow_test (bool, default true)

    If true, the filter will mark points shadowed by robot body as SHADOW. If false, such points will be marked OUTSIDE.

  • filter/max_shadow_distance (float, default is the value of sensor/max_distance)

    If greater than zero, specifies the maximum distance of a point from the sensor frame within which the point can be considered for shadow testing. All further points are classified as OUTSIDE. Setting this parameter to a low value may greatly improve performance of the shadow filtering.

  • body_model/inflation/scale (float, default 1.0)

    A scale that is applied to the collision model for the purposes of collision checking.

  • body_model/inflation/padding (float, default 0.0 m)

    Padding to be added to the collision model for the purposes of collision checking.

  • body_model/inflation/contains_test/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for contains tests.

  • body_model/inflation/contains_test/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for contains tests.

  • body_model/inflation/shadow_test/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for shadow tests.

  • body_model/inflation/shadow_test/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for shadow tests.

  • body_model/inflation/bounding_sphere/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for bounding sphere computation.

  • body_model/inflation/bounding_sphere/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for bounding sphere computation.

  • body_model/inflation/bounding_box/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for bounding box computation.

  • body_model/inflation/bounding_box/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for bounding box computation.

  • body_model/inflation/per_link/scale (dict[str:float], default {})

    A scale that is applied to the specified links for the purposes of collision checking. Links not specified here will use the default scale set in body_model/inflation/contains_test/scale or body_model/inflation/shadow_test/scale. Keys are names, values are scale. Names can be either names of links (link), names of collisions ( *::my_collision), a combination of link name and zero-based collision index (link::1), or link name and collision name, e.g. link::my_collision. Any such name can have suffix ::contains, ::shadow, ::bounding_sphere or ::bounding_box, which will only change the scale for contains or shadow tests or for bounding sphere or box computation. If a collision is matched by multiple entries, they have priority corresponding to the order they were introduced here (the entries do not “add up”, but replace each other).

  • body_model/inflation/per_link/padding (dict[str:float], default {})

    Padding to be added to the specified links for the purposes of collision checking. Links not specified here will use the default padding set in body_model/inflation/contains_test/padding or body_model/inflation/shadow_test/padding. Keys are names, values are padding. Names can be either names of links (link), names of collisions ( *::my_collision), a combination of link name and zero-based collision index (link::1), or link name and collision name, e.g. link::my_collision. Any such name can have suffix ::contains, ::shadow, ::bounding_sphere or ::bounding_box, which will only change the padding for contains or shadow tests or for bounding sphere or box computation. If a collision is matched by multiple entries, they have priority corresponding to the order they were introduced here (the entries do not “add up”, but replace each other).

  • body_model/robot_description_param (string, default: "robot_description")

    Name of the parameter where the robot model can be found.

  • transforms/buffer_length (float, default 60.0 s)

    Duration for which transforms will be stored in TF buffer.

  • transforms/timeout/reachable (float, default 0.1 s)

    How long to wait while getting reachable TF (i.e. transform which was previously available). Please note that this timeout is computed not from the lookup start time, but from the scan timestamp - this allows you to tell how old scans you still want to process.

  • transforms/timeout/unreachable (float, default 0.2 s)

    How long to wait while getting unreachable TF.

  • transforms/require_all_reachable (bool, default false)

    If true, the filter won’t publish anything until all transforms are reachable.

  • ignored_links/bounding_sphere (list[string], default [])

    List of links to be ignored in bounding sphere computation. Can be either names of links (link), names of collisions (*::my_collision), a combination of link name and zero-based collision index (link::1), or link name and collision name, e.g. link::my_collision.

  • ignored_links/bounding_box (list[string], default [])

    List of links to be ignored in bounding box computation. Same naming rules as above.

  • ignored_links/contains_test (list[string], default [])

    List of links to be ignored when testing if a point is inside robot body. Same naming rules as above.

  • ignored_links/shadow_test (list[string], default ['laser'])

    List of links to be ignored when testing if a point is shadowed by robot body. Same naming rules as above. It is essential that this list contains the sensor link - otherwise all points would be shadowed by the sensor itself.

  • ignored_links/bounding_sphere (list[string], default [])

    List of links to be ignored when computing the bounding sphere. Same naming rules as above.

  • ignored_links/bounding_box (list[string], default [])

    List of links to be ignored when computing the bounding box. Same naming rules as above.

  • ignored_links/everywhere (list[string], default [])

    List of links to be completely ignored. Same naming rules as above.

  • only_links (list[string], default [])

    If non-empty, only the specified links will be processed. The above-mentioned ignored_links/* filters will still be applied.

  • bounding_sphere/compute (bool, default false)

    Whether to compute and publish bounding sphere.

  • bounding_box/compute (bool, default false)

    Whether to compute and publish axis-aligned bounding box aligned to filtering frame.

  • oriented_bounding_box/compute (bool, default false)

    Whether to compute and publish oriented bounding box (see the remarks in the overview above; the result might be pretty bad).

  • local_bounding_box/compute (bool, default false)

    Whether to compute and publish axis-aligned bounding box aligned to frame local_bounding_box/frame_id.

  • local_bounding_box/frame_id (str, default: frames/fixed)

    The frame to which local bounding box is aligned.

  • bounding_sphere/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the bounding sphere are removed. Will be published on scan_point_cloud_no_bsphere. Implies bounding_sphere/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • bounding_box/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the bounding box are removed. Will be published on scan_point_cloud_no_bbox. Implies bounding_box/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • oriented_bounding_box/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the oriented bounding box are removed. Will be published on scan_point_cloud_no_oriented_bbox. Implies oriented_bounding_box/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • local_bounding_box/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the local bounding box are removed. Will be published on scan_point_cloud_no_local_bbox. Implies local_bounding_box/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • body_model/dynamic_robot_description/field_name (string, default robot_model)

    If robot model is published by dynamic reconfigure, this is the name of the Config message field which holds the robot model.

  • cloud/point_channels (list[string], default ["vp_"])

    List of channels of the incoming pointcloud that should be transformed as positional data. The 3D positions channel given by fields x, y and z is always transformed. This list contains prefixes of fields that form another channel(s). E.g. to transform the channel given by fields vp_x, vp_y and vp_z, add item "vp_" to the list. If a channel is not present in the cloud, nothing happens. This parameter is only available in the PointCloud2 version of the filter.

  • cloud/direction_channels (list[string], default ["normal_"])

    List of channels of the incoming pointcloud that should be transformed as directional data. This list contains prefixes of fields that form channel(s). E.g. to transform the channel given by fields normal_x, normal_y and normal_z, add item "normal_" to the list. If a channel is not present in the cloud, nothing happens. This parameter is only available in the PointCloud2 version of the filter.

Debug Operation

These options are there to help correctly set up and debug the filter operation and should be turned off in production environments since they can degrade performance of the filter.

Published Topics

  • robot_bounding_sphere_marker (visualization_msgs/Marker)

    Marker of the bounding sphere of the robot body. Turned on by bounding_sphere/marker parameter. Published in filtering frame.

  • robot_bounding_box_marker (visualization_msgs/Marker)

    Marker of the bounding box of the robot body. Turned on by bounding_box/marker parameter. Published in filtering frame.

  • robot_oriented_bounding_box_marker (visualization_msgs/Marker)

    Marker of the oriented bounding box of the robot body. Turned on by oriented_bounding_box/marker parameter. Published in filtering frame.

  • robot_local_bounding_box_marker (visualization_msgs/Marker)

    Marker of the local bounding box of the robot body. Turned on by local bounding_box/marker parameter. Published in frame local_bounding_box/frame_id.

  • robot_bounding_sphere_debug (visualization_msgs/MarkerArray)

    Marker array containing the bounding sphere for each collision element. Turned on by bounding_sphere/debug parameter. Published in filtering frame.

  • robot_bounding_box_debug (visualization_msgs/MarkerArray)

    Marker array containing the bounding box for each collision element. Turned on by bounding_box/debug parameter. Published in filtering frame.

  • robot_oriented_bounding_box_debug (visualization_msgs/MarkerArray)

    Marker array containing the oriented bounding box for each collision element. Turned on by oriented_bounding_box/debug parameter. Published in filtering frame.

  • robot_local_bounding_box_debug (visualization_msgs/MarkerArray)

    Marker array containing the local bounding box for each collision element. Turned on by local_bounding_box/debug parameter. Published in frame local_bounding_box/frame_id.

  • robot_model_for_contains_test (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for contains tests. Turned on by debug/marker/contains parameter.

  • robot_model_for_shadow_test (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for shadow tests. Turned on by debug/marker/shadow parameter.

  • robot_model_for_bounding_sphere (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for computation of bounding sphere. Turned on by debug/marker/bounding_sphere parameter.

  • robot_model_for_bounding_box (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for computation of bounding box. Turned on by debug/marker/bounding_box parameter.

  • scan_point_cloud_inside (sensor_msgs/PointCloud2)

    Debugging pointcloud with points classified as INSIDE. Turned on by debug/pcl/inside parameter.

  • scan_point_cloud_clip (sensor_msgs/PointCloud2)

    Debugging pointcloud with points classified as CLIP. Turned on by debug/pcl/clip parameter.

  • scan_point_cloud_shadow (sensor_msgs/PointCloud2)

    Debugging pointcloud with points classified as SHADOW. Turned on by debug/pcl/shadow parameter.

Filter Parameters

  • bounding_sphere/marker (bool, default false)

    Whether to publish a marker representing the bounding sphere.

  • bounding_box/marker (bool, default false)

    Whether to publish a marker representing the axis-aligned bounding box aligned to the filtering frame.

  • oriented_bounding_box/marker (bool, default false)

    Whether to publish a marker representing the oriented bounding box.

  • local_bounding_box/marker (bool, default false)

    Whether to publish a marker representing the axis-aligned bounding box aligned to frame local_bounding_box/frame_id.

  • bounding_sphere/debug (bool, default false)

    Whether to compute and publish debug bounding spheres (marker array of spheres for each collision).

  • bounding_box/debug (bool, default false)

    Whether to compute and publish debug bounding boxes (marker array of axis-aligned boxes for each collision).

  • oriented_bounding_box/debug (bool, default false)

    Whether to compute and publish debug oriented bounding boxes (marker array of oriented boxes for each collision).

  • local_bounding_box/debug (bool, default false)

    Whether to compute and publish debug axis-aligned bounding boxes aligned to frame local_bounding_box/frame_id (marker array of axis-aligned boxes for each collision).

  • debug/pcl/inside (bool, default false)

    Whether to publish debugging pointcloud with points marked as INSIDE.

  • debug/pcl/clip (bool, default false)

    Whether to publish debugging pointcloud with points marked as CLIP.

  • debug/pcl/shadow (bool, default false)

    Whether to publish debugging pointcloud with points marked as SHADOW.

  • debug/marker/contains (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for containment test.

  • debug/marker/shadow (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for shadow test.

  • debug/marker/bounding_sphere (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for computing the bounding sphere.

  • debug/marker/bounding_box (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for computing the bounding box.

CHANGELOG

Changelog for package robot_body_filter

1.3.2 (2024-01-22)

  • Do not segfault when a mesh resource is not found.
  • Added full_example.
  • Contributors: Martin Pecka

1.3.1 (2023-07-14)

  • Explicitly specify minimum versions of required libraries.
  • Contributors: Martin Pecka

1.3.0 (2023-04-12)

  • Removed local implementation of oriented bounding boxes in favor of geometric_shapes/obb. This change breaks API and ABI, but I hope nobody explicitly used the OBB part of this library. This change requires geometric_shapes version 0.6.6+ (Melodic) or 0.7.5+ (Noetic) (released April 2023).
  • Add example launch & config for ease of use. Thanks Doan Nguyen for the contribution!
  • Changed xmlrpc_traits variables to constexpr static instead of inline static to decrease the required C++ language standard for this part. Changed stringType from std::string to const char*.
  • Improved xmlrpc_traits to recognize more types of valid structures.
  • Make filter_utils FilterBase::getParamVerbose() methods const. Allowed by https://github.com/ros/filters/pull/35 (released in Melodic filters 1.8.2 (October 2021) and Noetic filters 1.9.1 (September 2021)).
  • Contributors: Doan Nguyen, Martin Pecka

1.2.2 (2021-08-25)

  • Change ROS_WARN to ROS_INFO when loading a value of an undefined parameter
  • Add link to the youtube tutorial
  • Contributors: Martin Pecka

1.2.1 (2021-08-06)

  • Merge pull request #15 from universal-field-robots/master TFFramesWatchdog and RayCastingShapeMask need to be installed in the CMakeLists.txt
  • Added RayCastingShapeMask and TFFramesWatchdog to install targets in cmake
  • Contributors: Josh Owen, Martin Pecka

1.2.0 (2021-07-30)

  • Merge pull request #11 from peci1/correct-pointcloud-transforms Add possibility to specify pointcloud channels that should be transformed together with positional data.
  • Merge pull request #14 from peci1/per_link_scaling Add support for scaling/padding each link differently
  • Short-circuit classification of NaN points.
  • Warn about missing collision elements only for non-ignored links and only if they have at least one visual.
  • Fixed bounding shapes computation
  • Added filter/max_shadow_distance for great performance increase
  • Added possibility to scale/pad collisions separately for computation of bounding box and sphere.
  • Unique-ify test target name to allow building with geometric_shapes.
  • Use correct Eigen allocator.
  • Reflected the newly added support for non-uniformly scaled meshes.
  • Contributors: Martin Pecka

1.1.9 (2021-04-17)

  • Compatibility with Noetic.
  • Contributors: Martin Pecka

1.1.8 (2020-04-06)

  • Fixed typo.
  • Contributors: Martin Pecka

1.1.7 (2020-04-05)

  • When sensor frame is empty (autodetected), do not add it as initial monitored frame to watchdog.
  • Added configuration examples.
  • Make sure use_sim_time is set to false for tests.
  • Fixed computation of pointcloud without local bounding box.
  • Added tests for RobotBodyFilter.
  • No longer require the index pointcloud field in computeMask(), as it is not used anywhere.
  • Surprise! CropBox::setTransform() doesn't set transform of the cropbox, but instead a transform of the points. Fixed that.
  • Fixed copy-paste bug in bounding sphere publisher.
  • Fix incorrect usage of fixed frame at many places (substituted with filtering frame).
  • Make sure orientation.w = 1 is set in all published markers.
  • Correctly exclude the ignored bounding shapes from the published markers and shapes.
  • Make sure the published bounding shapes are computed as they appear at the time of the beginning of the scan.
  • Fix getShapeTransform() to correctly interpolate between before- and after-scan times. The logic was twisted, so with cacheLookupBetweenScansRation was zero, it returned the transform after scan.
  • Fix copy-paste bug in transform cache.
  • Make use of the newly optional scanFrame argument of computeMask()
  • Prevent division by zero in point-by-point scans with zero duration (which is weird, but well...)
  • Only compute sensor->filtering transform for all-at-once scans, because point-by-point scans have viewpoints instead.
  • Reworked how TFFramesWatchdog is initialized to allow easier reconfiguring of the filter.
  • Added RobotBodyFilter::failWithoutRobotDescription which aids in tests.
  • Clarified documentation of RobotBodyFilter regarding frames.
  • Add TFFramesWatchdog::isRunning().
  • Workaround PCL 1.8 ignoring CropBox::setKeepOrganized().
  • Added tests for filter utils.
  • Improved filter_utils documentation.
  • Improved getParamVerbose() with recursive resolution of slash-separated param names.
  • Make different performance test requirements for release and debug builds.
  • Added tests for RayCastingShapeMask and improved its documentation.
  • Added utility functions to compare transforms and generate random ones.
  • Implemented TF watchdog tests.
  • Prevented some more SIGABRTs from TFFramesWatchdog.
  • Added unit tests for utils.
  • Fixed a bug in OBB.contains(OBB) which compared against vertices of the wrong OBB.
  • Improved efficiency of Eigen expression by not using auto type.
  • Fix the reported transformation of default-constructed OBBs.
  • Fixed constructShapeFromBody() to correctly process meshes.
  • Fixed eigen and urdf includes.
  • Fixed printing of std containers in string_utils.
  • CREATE_FILTERED_CLOUD now prepends robot_body_filter stuff with a global namespace to be usable even from other namespaces.
  • Make sure sensor frame is monitored by the watchdog because we need to use it during the filtering. Fixes #6.
  • Added parameter transforms/require_all_reachable.
  • Added to_string(bool) specialization.
  • Contributors: Martin Pecka

1.1.6 (2019-11-11)

  • Avoid subtracting from ROS time. This may lead to exceptions in simulation, where current time is 0. ros::Time cannot be negative.
  • Contributors: Martin Pecka

1.1.5 (2019-11-08)

  • refactor(RobotBodyFilter): Made frames/sensor parameter optional (#2) The sensor frame will now be derived from the incoming sensor messages. This way, messages from different sensors can be processed and the filter can be configured without any knowledge of the sensor.
  • Improved readme.
  • Reorganized the readme.
  • Added support for non-uniform scaling of collision meshes.
  • Contributors: Martin Pecka, Rein Appeldoorn

1.1.4 (2019-08-07)

  • Fixed a bug in filtering unorganized point clouds.
  • Contributors: Martin Pecka

1.1.1 (2019-07-30)

  • Fixed dependencies
  • Contributors: Martin Pecka

1.1.0 (2019-07-18)

  • Increasing performance.
  • More descriptive frame configurations. Added the possibility to leave out clipping, contains or shadow tests.
  • Repurposed param ~transforms/timeout/reachable to be always used with remainingTime.
  • More efficient point transformation
  • Little improvements. Renamed the utils library to robot_body_filter_utils.
  • Removed compiler pragmas - they seem no longer needed.
  • Provided a fix for wrong box-ray intersections.
  • Fixed bodies.h header guard.
  • Fixed bug in cylinder shape creation.
  • Fixed a bug in constructMarkerFromBody() - references do not correctly support polymorphism.
  • Fixes for robot body filter. Added new bounding box types.
  • Fixed a few bugs while running under a nodelet.
  • Prepared remainingTime() to a situation where ROS time hasn't yet been initialized.
  • Added possibility to update only robot pose only with every n-th point in point_by_point scans. Added possibility to publish bounding sphere and box markers.
  • Renamed to robot_body_filter.
  • Reworked as a laser filter combining contains and shadow tests.
  • Make use of bodies.h and shapes.h from geometric_shapes.
  • Enabling generic point types, removed PCL dependency, removed unnecessary params.
  • Using all collision elements for each link instead of only the first one.
  • Testing all intersections instead of only the first one.
  • Merge branch 'master' into indigo-devel
  • Add robot_self_filter namespace before bodies and shapes namespace. geometric_shapes package also provides bodies and shapes namespace and same classes and functions. If a program is linked with geometric_shapes and robot_self_filter, it may cause strange behavior because of symbol confliction.
  • Contributors: Martin Pecka, Ryohei Ueda, Tomas Petricek

0.1.31 (2018-11-24)

  • update CHANGELOG
  • Merge pull request #16 from mikaelarguedas/tinyxml_dependency depends on tinyxml and link against it
  • Merge branch 'indigo-devel' into tinyxml_dependency
  • Merge pull request #18 from k-okada/add_travis update travis.yml
  • update travis.yml
  • depend on tinyxml and link against it
  • Merge pull request #14 from traclabs/indigo-devel Minor changes to indigo-devel CMake allow this to be used in kinetic and indigo
  • Changes for kinetic
  • Contributors: Devon Ash, Kei Okada, Mikael Arguedas, Patrick Beeson

0.1.30 (2017-01-20)

  • Update CHANGELOG.rst
  • Merge pull request #15 from PR2/fix-typo-cmakelists Fix typo in CMakeLists.txt: CATKIN-DEPENDS -> CATKIN_DEPENDS
  • Fix typo in CMakeLists.txt: CATKIN-DEPENDS -> CATKIN_DEPENDS
  • Merge pull request #12 from garaemon/max-queue-size Add ~max_queue_size parameter for subscription queue size
  • Add ~max_queue_size parameter for subscription queue size
  • Contributors: Devon Ash, Kentaro Wada, Ryohei Ueda

0.1.29 (2015-12-05)

  • Re-create changelog for robot_self_filter
  • Merge pull request #10 from garaemon/pr-4-indigo-devel Add robot_self_filter namespace before bodies and shapes namespace.
  • Add robot_self_filter namespace before bodies and shapes namespace. geometric_shapes package also provides bodies and shapes namespace and same classes and functions. If a program is linked with geometric_shapes and robot_self_filter, it may cause strange behavior because of symbol confliction.
  • Contributors: Ryohei Ueda

0.1.28 (2015-12-04)

  • Merge pull request #8 from wkentaro/indigo-devel-merge-master Merge master branch to indigo-devel
  • Merge remote-tracking branch 'origin/master' into indigo-devel
  • Added indigo devel
  • Merge pull request #7 from wkentaro/self_filter-timestamp Set correct timestamp for self filtered cloud
  • Set correct timestamp for self filtered cloud This is needed because pcl drops some value of timestamp. So pcl::fromROSMsg and pcl::toROSMsg does not work to get correct timestamp.
  • Merge pull request #5 from garaemon/use-protected-member Protected member variables in SelfMask for subclass of SelfMask
  • Protected member variables in SelfMask for subclass of SelfMask
  • Contributors: Devon Ash, Kentaro Wada, Ryohei Ueda, TheDash

0.1.27 (2015-12-01)

  • Merge pull request #1 from garaemon/robot-self-filter Porting robot_self_filter from pr2_navigation_self_filter
  • Porting robot_self_filter from pr2_navigation_self_filter
  • Initial commit
  • Contributors: Devon Ash, Ryohei Ueda

Wiki Tutorials

This package does not provide any links to tutorials in it's rosindex metadata. You can check on the ROS Wiki Tutorials page for the package.

Launch files

Services

No service files found

Plugins

Recent questions tagged robot_body_filter at Robotics Stack Exchange

robot_body_filter package from robot_body_filter repo

robot_body_filter

Package Summary

Tags No category tags.
Version 1.3.2
License BSD
Build type CATKIN
Use RECOMMENDED

Repository Summary

Checkout URI https://github.com/peci1/robot_body_filter.git
VCS Type git
VCS Version master
Last Updated 2024-02-03
Dev Status MAINTAINED
CI status
Released RELEASED
Tags No category tags.
Contributing Help Wanted (0)
Good First Issues (0)
Pull Requests to Review (0)

Package Description

Filters the robot's body out of laser scans or point clouds.

Additional Links

Maintainers

  • Martin Pecka

Authors

  • Eitan Marder-Eppstein
  • Tomas Petricek
  • Martin Pecka

robot_body_filter

Filters the robot’s body out of point clouds and laser scans.

Tutorial

Check out the webinar recording where a lot of options for this filter are explained and demonstrated! https://www.youtube.com/watch?v=j0ljV0uZy3Q

Changes vs PR2/robot_self_filter:

  • Now the package is a normal filters::FilterBase filter and not a standalone node.
  • Using both containment and ray-tracing tests.
  • Using all collision elements for each link instead of only the first one.
  • Enabling generic point type, removing PCL dependency and unnecessary params.
  • Using bodies.h and shapes.h from geometric_shapes.
  • As a by-product, the filter can compute robot’s bounding box or sphere.

Build Status

Development versions: Github Actions Dev melodic Dev noetic

Release jobs Melodic Melodic version: Bin melodic-amd64 Bin melodic-arm64 Bin melodic-armhf

Release jobs Noetic Noetic version: Bin noetic focal-amd64 Bin noetic focal-arm64 Bin noetic focal-armhf

Basic Operation

filters::FilterBase API

The basic workings of this filter are done via the filters::FilterBase API implemented for sensor_msgs::LaserScan and sensor_msgs::PointCloud2 types. This means you can load this filter into a FilterChain along other filters as usual. Different from the standard filters, this one can also publish several interesting topics and subscribes to TF.

General overview

This filter reads robot model and the filter config, subscribes to TF, waits for data (laserscans or point clouds) and then cleans them from various artifacts (this is called data filtering).

It can perform 3 kinds of data filters: clip the data based on the provided sensor limits (parameter filter/do_clipping), remove points that are inside or on the surface of the robot body (parameter filter/do_contains_test) and remove points that are seen through a part of the robot body (parameter filter/do_shadow_test). These kinds of tests are further referenced as “clipping”, “contains test” and “shadow test”.

If working with point clouds, the filter automatically recognizes whether it works with organized or non-organized clouds. In organized clouds, it marks the filtered-out points as NaN. In non-organized clouds, it removes the filtered-out points. In laserscans, removal is not an option, so the filtered-out points are marked with NaN (some guides suggest that max_range + 1 should be used for marking invalid points, but this filter uses NaN as a safer option).

Performance tips

In general, the filter will be computationally expensive (clipping is fast, contains test is medium CPU intensive and shadow test is the most expensive part, because it basically performs raytracing). You can limit the required CPU power by limiting the filter only to parts that matter. E.g. if the robot has a link that can never be seen by the sensor, put it in the list of ignored links. The less links are processed, the better performance. If you’re only interested in removing a few links, consider using the only_links parameter.

To speed up shadow filtering, you can set filter/max_shadow_distance, which limits the number of points considered for shadow tests just to points close to the sensor. Setting this to e.g. three times the diameter of the robot should remove all of the shadow points caused by refraction by a part of the robot body. But you have to test this with real data.

Performance also strongly depends on representation of the robot model. The filter reads <collision> tags from the robot URDF. You can use boxes, spheres and cylinders (which are fast to process), or you can use convex meshes (these are much worse performance-wise). If you pass a non-convex mesh, its convex hull will be used for the tests. Don’t forget that each link can have multiple <collision> tags. If you do not have time to convert your meshes to the basic shapes, try to at least reduce the number of triangles in your meshes. You can use your high-quality meshes in <visual> tags. To simplify your model to primitive shapes, you can either manually edit the URDF, or you can utilize ColliderGen.

Model inflation

You can utilize the builtin model inflation mechanism to slightly alter the size of the model. You will probably want to add a bit “margin” to the contains and shadow tests so that points that are millimeters outside the robot body will anyways get removed. You can set a default scale and padding which are used for all collisions. Different inflation can be used for contains tests and for shadow tests. Inflation can also be specified differently for each link. Look at the body_model/inflation/* parameters for details.

Scaling means multiplying the shape dimensions by the given factor (with its center staying in the same place). Padding means adding the specified metric distance to each “dimension” of the shape. Padding a sphere by p just adds p to its radius; padding a cylinder adds p to its radius and 2p to its length (you pad both the top and the bottom of the cylinder); padding a box adds 2p to all its extents (again, you pad both of the opposing sides); padding a mesh pads each vertex of its convex hull by p along the direction from the mesh center to the vertex (mesh center is the center specified in the mesh file, e.g. DAE). Have a good look at the effects of mesh padding, as the results can be non-intuitive.

Data acquisition modes

The filter supports data captured in two modes - all at once (e.g. RGBD cameras), or each point at a different time instant (mostly lidars). This is handled by sensor/point_by_point setting. Each mode supports both laser scans and point cloud input (although all-at-once laserscans aren’t that common). Point-by-point pointclouds are e.g. the output of 3D lidars like Ouster (where more points can have the same timestamp, but not all). If you want to use the point-by-point mode with pointclouds, make sure they contain not only the usual x, y and z fields, but also a float32 field stamps (with time difference from the time in the header) and float32 fields vp_x, vp_y and vp_z which contain the viewpoint (position of the sensor in filtering frame) from which the robot saw that point.

When filtering in the point-by-point mode, the robot posture has to be updated several times during processing a single scan (to reflect the motion the robot has performed during acquiring the scan). The frequency of these updates can also have a significant impact on performance. Use parameter filter/model_pose_update_interval to set the interval for which the robot is considered stationary. The positions of the robot at the beginning and at the end of the scan are queried from TF. The intermediate positions are linearly interpolated between these two positions.

TF frames

The filter recognizes four logical TF frames (some of which may be the same physical frames).

Fixed frame is a frame that doesn’t change within the duration of the processed scan. This means for all-at-once scans, this frame is not needed because the duration of the scan is zero. For point-by-point scans, it depends on the particular scenario. In static installations (like manipulators with sensors not attached to them), it can be the sensor frame. For stationary robots with the sensor attached to a movable part of their body, base_link will be a good choice. For completely mobile robots, you will need an external frame, e.g. odom or map (beware of cyclic dependencies - if the map is built from the filtered scans, you obviously cannot use map as the fixed frame for filtering the scans…).

Sensor frame is the frame in which the data were captured. Generally, it would be whatever is in header.frame_id field of the processed messages. You can use the filter for data from multiple sensors - in that case, you can leave the sensor frame unfilled and each message will be processed in the frame it has in its header.

Filtering frame is the frame in which the data filtering is done. For point-by-point scans, it has to be a fixed frame (possibly different from the fixed frame set in frames/fixed). For pointcloud scans, it should be the sensor frame (if all data are coming from a single sensor), or any other frame. It is also used as the frame in which all debugging outputs are published.

Output frame can only be used with pointcloud scans, and allows to transform the filtered pointcloud to a different frame before being published. It is just a convenience which can save you launching a transformation nodelet. By default, filtered pointclouds are output in the filtering frame.

Bounding shapes

As a byproduct, the filter can also compute various bounding shapes of the robot model. There are actually four robot models - one for contains test, one for shadow test, one for bounding sphere computation and one for bounding box (these models can differ by inflation and considered links). All bounding shapes are published in the filtering frame. For point-by-point scans, the bounding shapes correspond to the time instant specified in the header of the processed scan. The computation of bounding shapes is off by default, but enabling it is cheap (performance-wise).

The bounding sphere is easy - the smallest sphere that contains the whole collision model for bounding sphere computation (with the specified exclusions removed).

The bounding box is the smallest axis-aligned bounding box aligned to the filtering frame. It is built from the model for bounding box computation.

The local bounding box is the smallest axis-aligned bounding box aligned to the frame specified in local_bounding_box/frame_id. It is especially useful with mobile robots when the desired frame is base_link. It is built from the model for bounding box computation.

The oriented bounding box should be the smallest box containing the collision model. However, its computation is very bad conditioned, so the results can be very unsatisfactory. Currently, the oriented bounding box of each of the basic collision shapes is “tight”, but merging the boxes is not optimal. A good algorithm would probably require costly and advanced iterative methods. The current implementation uses FCL in the background and merges the boxes using fcl::OBB::operator+=() without any further optimizations. It is built from the model for bounding box computation.

The filter also supports publishing auxiliary pointclouds which “cut out” each of these bounding shapes. These are the input data converted to pointcloud in filtering frame from which all points belonging to the bounding shape are removed. Please note that the “base” used for cutting out is the input pointcloud, not the filtered one.

First setup/debugging

The filter offers plenty of debugging outputs to make sure it does exactly what you want it to do. All the options are described in the last part of this page. Generally, you should look at the pointclouds visualizing which points got filtered out and you should also check the robot models used for filtering.

Also, have a look in the [examples] folder to get some inspiration.

Usage

This is a standard filters::FilterBase<T>-based filter which implements the configure() and update(const T&, T&) methods. This means it can be loaded e.g. as a part of a filter chain via the laser_filters package or the relatively new sensor_filters. This means the input and output data are not supplied in the form of topics, but they are instead passed to the update() method via the C++ API.

This filter is a bit unusual - it subscribes and publishes several topics. Normally, filters only operate via the update() method and are not expected to publish anything. This filter is different, it requires a working ROS node handle, and can publish a lot of auxiliary or debug data.

Subscribed Topics

  • /tf, /tf_static

    Transforms

  • dynamic_robot_model_server/parameter_updates (dynamic_reconfigure/Config)

    Dynamic reconfigure topic on which updated robot model can be subscribed. The model is read from a field defined by parameter body_model/dynamic_robot_description/field_name.

Published Topics

  • scan_point_cloud_no_bsphere (sensor_msgs/PointCloud2)

    Point cloud with points inside body bounding sphere removed. Turned on by bounding_sphere/publish_cut_out_pointcloud parameter. Published in filtering frame.

  • scan_point_cloud_no_bbox (sensor_msgs/PointCloud2)

    Point cloud with points inside filtering frame axis-aligned bounding box removed. Turned on by bounding_box/publish_cut_out_pointcloud parameter. Published in filtering frame.

  • scan_point_cloud_no_oriented_bbox (sensor_msgs/PointCloud2)

    Point cloud with points inside oriented bounding box removed. Turned on by oriented_bounding_box/publish_cut_out_pointcloud parameter. Published in filtering frame. Please read the remarks above in the overview - the results can be non-satisfying.

  • scan_point_cloud_no_local_bbox (sensor_msgs/PointCloud2)

    Point cloud with points inside local_bounding_box/frame_id axis-aligned bounding box removed. Turned on by local_bounding_box/publish_cut_out_pointcloud parameter. Published in filtering frame.

  • robot_bounding_sphere (robot_body_filter/SphereStamped)

    Bounding sphere of the robot body. Turned on by bounding_sphere/compute parameter. Published in filtering frame.

  • robot_bounding_box (geometry_msgs/PolygonStamped)

    Axis-aligned bounding box of the robot body aligned to the filtering frame. First point is the minimal point, second one is the maximal point. Turned on by bounding_box/compute parameter. Published in filtering frame.

  • robot_oriented_bounding_box (robot_body_filter/OrientedBoundingBoxStamped)

    Oriented bounding box of the robot body. Turned on by oriented_bounding_box/compute parameter. Please read the remarks above in the overview - the results can be non-satisfying. Published in filtering frame.

  • robot_local_bounding_box (geometry_msgs/PolygonStamped)

    Axis-aligned bounding box of the robot body aligned to frame local_bounding_box/frame_id. First point is the minimal point, second one is the maximal point. Turned on by local_bounding_box/compute parameter. Published in frame local_bounding_box/frame_id.

Provided Services

  • ~reload_model (std_srvs/Trigger)

    Call this service to re-read the URDF model from parameter server.

Filter Parameters

Have a look in the examples folder to get inspiration for configuration of your filter.

  • sensor/point_by_point (bool, default: false for PointCloud2 version, true for LaserScan)

    If true, suppose that every point in the scan was captured at a different time instant. Otherwise, the scan is assumed to be taken at once. If this is true, the processing pipeline expects the pointcloud to have fields int32 index, float32 stamps, and float32 vp_x, vp_y and vp_z viewpoint positions. If one of these fields is missing, computeMask() throws runtime exception.

  • frames/fixed (string, default: "base_link")

    The fixed frame. Usually base_link for stationary robots (or sensor frame if both robot and sensor are stationary). For mobile robots, it can be e.g. odom or map. Only needed for point-by-point scans.

  • frames/sensor (string, default "")

    Frame of the sensor. If set to empty string, it will be read from header.frame_id of each incoming message. In LaserScan version, if nonempty, it has to match the frame_id of the incoming scans. In PointCloud2 version, the data can come in a different frame from frames/sensor.

  • frames/filtering (string, default: frames/fixed)

    Frame in which the filter is applied. For point-by-point scans, it has to be a fixed frame, otherwise, it can be the sensor frame or any other frame. Setting to sensor frame will save some computations, but this frame cannot be empty string, so sensor frame can only be used if all data are coming from a single sensor and you know the scan frame in advance.

  • frames/output (string, default: frames/filtering)

    Frame into which output data are transformed. Only applicable in PointCloud2 version.

  • sensor/min_distance (float, default: 0.0 m)

    The minimum distance of points from the laser to keep them.

  • sensor/max_distance (float, default: 0.0 m)

    The maximum distance of points from the laser to keep them. Set to zero to disable this limit.

  • filter/keep_clouds_organized (bool, default true)

    Whether to keep pointclouds organized (if they were). Organized cloud has height > 1.

  • filter/model_pose_update_interval (float, default 0.0 s)

    The interval between two consecutive model pose updates when processing a pointByPointScan. If set to zero, the model will be updated for each point separately (might be computationally exhaustive). If non-zero, it will only update the model once in this interval, which makes the masking algorithm a little bit less precise but more computationally affordable.

  • filter/do_clipping (bool, default true)

    If true, the filter will mark points outside sensor/(min|max)_distance as CLIP. If false, such points will be marked OUTSIDE.

  • filter/do_contains_test (bool, default true)

    If true, the filter will mark points inside robot body as INSIDE. If false, such points will be marked OUTSIDE.

  • filter/do_shadow_test (bool, default true)

    If true, the filter will mark points shadowed by robot body as SHADOW. If false, such points will be marked OUTSIDE.

  • filter/max_shadow_distance (float, default is the value of sensor/max_distance)

    If greater than zero, specifies the maximum distance of a point from the sensor frame within which the point can be considered for shadow testing. All further points are classified as OUTSIDE. Setting this parameter to a low value may greatly improve performance of the shadow filtering.

  • body_model/inflation/scale (float, default 1.0)

    A scale that is applied to the collision model for the purposes of collision checking.

  • body_model/inflation/padding (float, default 0.0 m)

    Padding to be added to the collision model for the purposes of collision checking.

  • body_model/inflation/contains_test/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for contains tests.

  • body_model/inflation/contains_test/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for contains tests.

  • body_model/inflation/shadow_test/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for shadow tests.

  • body_model/inflation/shadow_test/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for shadow tests.

  • body_model/inflation/bounding_sphere/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for bounding sphere computation.

  • body_model/inflation/bounding_sphere/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for bounding sphere computation.

  • body_model/inflation/bounding_box/scale (float, default body_model/inflation/scale)

    A scale that is applied to the collision model used for bounding box computation.

  • body_model/inflation/bounding_box/padding (float, default body_model/inflation/padding)

    Padding to be added to the collision model used for bounding box computation.

  • body_model/inflation/per_link/scale (dict[str:float], default {})

    A scale that is applied to the specified links for the purposes of collision checking. Links not specified here will use the default scale set in body_model/inflation/contains_test/scale or body_model/inflation/shadow_test/scale. Keys are names, values are scale. Names can be either names of links (link), names of collisions ( *::my_collision), a combination of link name and zero-based collision index (link::1), or link name and collision name, e.g. link::my_collision. Any such name can have suffix ::contains, ::shadow, ::bounding_sphere or ::bounding_box, which will only change the scale for contains or shadow tests or for bounding sphere or box computation. If a collision is matched by multiple entries, they have priority corresponding to the order they were introduced here (the entries do not “add up”, but replace each other).

  • body_model/inflation/per_link/padding (dict[str:float], default {})

    Padding to be added to the specified links for the purposes of collision checking. Links not specified here will use the default padding set in body_model/inflation/contains_test/padding or body_model/inflation/shadow_test/padding. Keys are names, values are padding. Names can be either names of links (link), names of collisions ( *::my_collision), a combination of link name and zero-based collision index (link::1), or link name and collision name, e.g. link::my_collision. Any such name can have suffix ::contains, ::shadow, ::bounding_sphere or ::bounding_box, which will only change the padding for contains or shadow tests or for bounding sphere or box computation. If a collision is matched by multiple entries, they have priority corresponding to the order they were introduced here (the entries do not “add up”, but replace each other).

  • body_model/robot_description_param (string, default: "robot_description")

    Name of the parameter where the robot model can be found.

  • transforms/buffer_length (float, default 60.0 s)

    Duration for which transforms will be stored in TF buffer.

  • transforms/timeout/reachable (float, default 0.1 s)

    How long to wait while getting reachable TF (i.e. transform which was previously available). Please note that this timeout is computed not from the lookup start time, but from the scan timestamp - this allows you to tell how old scans you still want to process.

  • transforms/timeout/unreachable (float, default 0.2 s)

    How long to wait while getting unreachable TF.

  • transforms/require_all_reachable (bool, default false)

    If true, the filter won’t publish anything until all transforms are reachable.

  • ignored_links/bounding_sphere (list[string], default [])

    List of links to be ignored in bounding sphere computation. Can be either names of links (link), names of collisions (*::my_collision), a combination of link name and zero-based collision index (link::1), or link name and collision name, e.g. link::my_collision.

  • ignored_links/bounding_box (list[string], default [])

    List of links to be ignored in bounding box computation. Same naming rules as above.

  • ignored_links/contains_test (list[string], default [])

    List of links to be ignored when testing if a point is inside robot body. Same naming rules as above.

  • ignored_links/shadow_test (list[string], default ['laser'])

    List of links to be ignored when testing if a point is shadowed by robot body. Same naming rules as above. It is essential that this list contains the sensor link - otherwise all points would be shadowed by the sensor itself.

  • ignored_links/bounding_sphere (list[string], default [])

    List of links to be ignored when computing the bounding sphere. Same naming rules as above.

  • ignored_links/bounding_box (list[string], default [])

    List of links to be ignored when computing the bounding box. Same naming rules as above.

  • ignored_links/everywhere (list[string], default [])

    List of links to be completely ignored. Same naming rules as above.

  • only_links (list[string], default [])

    If non-empty, only the specified links will be processed. The above-mentioned ignored_links/* filters will still be applied.

  • bounding_sphere/compute (bool, default false)

    Whether to compute and publish bounding sphere.

  • bounding_box/compute (bool, default false)

    Whether to compute and publish axis-aligned bounding box aligned to filtering frame.

  • oriented_bounding_box/compute (bool, default false)

    Whether to compute and publish oriented bounding box (see the remarks in the overview above; the result might be pretty bad).

  • local_bounding_box/compute (bool, default false)

    Whether to compute and publish axis-aligned bounding box aligned to frame local_bounding_box/frame_id.

  • local_bounding_box/frame_id (str, default: frames/fixed)

    The frame to which local bounding box is aligned.

  • bounding_sphere/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the bounding sphere are removed. Will be published on scan_point_cloud_no_bsphere. Implies bounding_sphere/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • bounding_box/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the bounding box are removed. Will be published on scan_point_cloud_no_bbox. Implies bounding_box/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • oriented_bounding_box/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the oriented bounding box are removed. Will be published on scan_point_cloud_no_oriented_bbox. Implies oriented_bounding_box/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • local_bounding_box/publish_cut_out_pointcloud (bool, default false)

    Whether to compute and publish pointcloud from which points in the local bounding box are removed. Will be published on scan_point_cloud_no_local_bbox. Implies local_bounding_box/compute. The “base” point cloud before cutting out are the input data, not the filtered data.

  • body_model/dynamic_robot_description/field_name (string, default robot_model)

    If robot model is published by dynamic reconfigure, this is the name of the Config message field which holds the robot model.

  • cloud/point_channels (list[string], default ["vp_"])

    List of channels of the incoming pointcloud that should be transformed as positional data. The 3D positions channel given by fields x, y and z is always transformed. This list contains prefixes of fields that form another channel(s). E.g. to transform the channel given by fields vp_x, vp_y and vp_z, add item "vp_" to the list. If a channel is not present in the cloud, nothing happens. This parameter is only available in the PointCloud2 version of the filter.

  • cloud/direction_channels (list[string], default ["normal_"])

    List of channels of the incoming pointcloud that should be transformed as directional data. This list contains prefixes of fields that form channel(s). E.g. to transform the channel given by fields normal_x, normal_y and normal_z, add item "normal_" to the list. If a channel is not present in the cloud, nothing happens. This parameter is only available in the PointCloud2 version of the filter.

Debug Operation

These options are there to help correctly set up and debug the filter operation and should be turned off in production environments since they can degrade performance of the filter.

Published Topics

  • robot_bounding_sphere_marker (visualization_msgs/Marker)

    Marker of the bounding sphere of the robot body. Turned on by bounding_sphere/marker parameter. Published in filtering frame.

  • robot_bounding_box_marker (visualization_msgs/Marker)

    Marker of the bounding box of the robot body. Turned on by bounding_box/marker parameter. Published in filtering frame.

  • robot_oriented_bounding_box_marker (visualization_msgs/Marker)

    Marker of the oriented bounding box of the robot body. Turned on by oriented_bounding_box/marker parameter. Published in filtering frame.

  • robot_local_bounding_box_marker (visualization_msgs/Marker)

    Marker of the local bounding box of the robot body. Turned on by local bounding_box/marker parameter. Published in frame local_bounding_box/frame_id.

  • robot_bounding_sphere_debug (visualization_msgs/MarkerArray)

    Marker array containing the bounding sphere for each collision element. Turned on by bounding_sphere/debug parameter. Published in filtering frame.

  • robot_bounding_box_debug (visualization_msgs/MarkerArray)

    Marker array containing the bounding box for each collision element. Turned on by bounding_box/debug parameter. Published in filtering frame.

  • robot_oriented_bounding_box_debug (visualization_msgs/MarkerArray)

    Marker array containing the oriented bounding box for each collision element. Turned on by oriented_bounding_box/debug parameter. Published in filtering frame.

  • robot_local_bounding_box_debug (visualization_msgs/MarkerArray)

    Marker array containing the local bounding box for each collision element. Turned on by local_bounding_box/debug parameter. Published in frame local_bounding_box/frame_id.

  • robot_model_for_contains_test (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for contains tests. Turned on by debug/marker/contains parameter.

  • robot_model_for_shadow_test (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for shadow tests. Turned on by debug/marker/shadow parameter.

  • robot_model_for_bounding_sphere (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for computation of bounding sphere. Turned on by debug/marker/bounding_sphere parameter.

  • robot_model_for_bounding_box (visualization_msgs/MarkerArray)

    Marker array containing the exact robot model used for computation of bounding box. Turned on by debug/marker/bounding_box parameter.

  • scan_point_cloud_inside (sensor_msgs/PointCloud2)

    Debugging pointcloud with points classified as INSIDE. Turned on by debug/pcl/inside parameter.

  • scan_point_cloud_clip (sensor_msgs/PointCloud2)

    Debugging pointcloud with points classified as CLIP. Turned on by debug/pcl/clip parameter.

  • scan_point_cloud_shadow (sensor_msgs/PointCloud2)

    Debugging pointcloud with points classified as SHADOW. Turned on by debug/pcl/shadow parameter.

Filter Parameters

  • bounding_sphere/marker (bool, default false)

    Whether to publish a marker representing the bounding sphere.

  • bounding_box/marker (bool, default false)

    Whether to publish a marker representing the axis-aligned bounding box aligned to the filtering frame.

  • oriented_bounding_box/marker (bool, default false)

    Whether to publish a marker representing the oriented bounding box.

  • local_bounding_box/marker (bool, default false)

    Whether to publish a marker representing the axis-aligned bounding box aligned to frame local_bounding_box/frame_id.

  • bounding_sphere/debug (bool, default false)

    Whether to compute and publish debug bounding spheres (marker array of spheres for each collision).

  • bounding_box/debug (bool, default false)

    Whether to compute and publish debug bounding boxes (marker array of axis-aligned boxes for each collision).

  • oriented_bounding_box/debug (bool, default false)

    Whether to compute and publish debug oriented bounding boxes (marker array of oriented boxes for each collision).

  • local_bounding_box/debug (bool, default false)

    Whether to compute and publish debug axis-aligned bounding boxes aligned to frame local_bounding_box/frame_id (marker array of axis-aligned boxes for each collision).

  • debug/pcl/inside (bool, default false)

    Whether to publish debugging pointcloud with points marked as INSIDE.

  • debug/pcl/clip (bool, default false)

    Whether to publish debugging pointcloud with points marked as CLIP.

  • debug/pcl/shadow (bool, default false)

    Whether to publish debugging pointcloud with points marked as SHADOW.

  • debug/marker/contains (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for containment test.

  • debug/marker/shadow (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for shadow test.

  • debug/marker/bounding_sphere (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for computing the bounding sphere.

  • debug/marker/bounding_box (bool, default false)

    Whether to publish debugging marker array containing the exact robot body model used for computing the bounding box.

CHANGELOG

Changelog for package robot_body_filter

1.3.2 (2024-01-22)

  • Do not segfault when a mesh resource is not found.
  • Added full_example.
  • Contributors: Martin Pecka

1.3.1 (2023-07-14)

  • Explicitly specify minimum versions of required libraries.
  • Contributors: Martin Pecka

1.3.0 (2023-04-12)

  • Removed local implementation of oriented bounding boxes in favor of geometric_shapes/obb. This change breaks API and ABI, but I hope nobody explicitly used the OBB part of this library. This change requires geometric_shapes version 0.6.6+ (Melodic) or 0.7.5+ (Noetic) (released April 2023).
  • Add example launch & config for ease of use. Thanks Doan Nguyen for the contribution!
  • Changed xmlrpc_traits variables to constexpr static instead of inline static to decrease the required C++ language standard for this part. Changed stringType from std::string to const char*.
  • Improved xmlrpc_traits to recognize more types of valid structures.
  • Make filter_utils FilterBase::getParamVerbose() methods const. Allowed by https://github.com/ros/filters/pull/35 (released in Melodic filters 1.8.2 (October 2021) and Noetic filters 1.9.1 (September 2021)).
  • Contributors: Doan Nguyen, Martin Pecka

1.2.2 (2021-08-25)

  • Change ROS_WARN to ROS_INFO when loading a value of an undefined parameter
  • Add link to the youtube tutorial
  • Contributors: Martin Pecka

1.2.1 (2021-08-06)

  • Merge pull request #15 from universal-field-robots/master TFFramesWatchdog and RayCastingShapeMask need to be installed in the CMakeLists.txt
  • Added RayCastingShapeMask and TFFramesWatchdog to install targets in cmake
  • Contributors: Josh Owen, Martin Pecka

1.2.0 (2021-07-30)

  • Merge pull request #11 from peci1/correct-pointcloud-transforms Add possibility to specify pointcloud channels that should be transformed together with positional data.
  • Merge pull request #14 from peci1/per_link_scaling Add support for scaling/padding each link differently
  • Short-circuit classification of NaN points.
  • Warn about missing collision elements only for non-ignored links and only if they have at least one visual.
  • Fixed bounding shapes computation
  • Added filter/max_shadow_distance for great performance increase
  • Added possibility to scale/pad collisions separately for computation of bounding box and sphere.
  • Unique-ify test target name to allow building with geometric_shapes.
  • Use correct Eigen allocator.
  • Reflected the newly added support for non-uniformly scaled meshes.
  • Contributors: Martin Pecka

1.1.9 (2021-04-17)

  • Compatibility with Noetic.
  • Contributors: Martin Pecka

1.1.8 (2020-04-06)

  • Fixed typo.
  • Contributors: Martin Pecka

1.1.7 (2020-04-05)

  • When sensor frame is empty (autodetected), do not add it as initial monitored frame to watchdog.
  • Added configuration examples.
  • Make sure use_sim_time is set to false for tests.
  • Fixed computation of pointcloud without local bounding box.
  • Added tests for RobotBodyFilter.
  • No longer require the index pointcloud field in computeMask(), as it is not used anywhere.
  • Surprise! CropBox::setTransform() doesn't set transform of the cropbox, but instead a transform of the points. Fixed that.
  • Fixed copy-paste bug in bounding sphere publisher.
  • Fix incorrect usage of fixed frame at many places (substituted with filtering frame).
  • Make sure orientation.w = 1 is set in all published markers.
  • Correctly exclude the ignored bounding shapes from the published markers and shapes.
  • Make sure the published bounding shapes are computed as they appear at the time of the beginning of the scan.
  • Fix getShapeTransform() to correctly interpolate between before- and after-scan times. The logic was twisted, so with cacheLookupBetweenScansRation was zero, it returned the transform after scan.
  • Fix copy-paste bug in transform cache.
  • Make use of the newly optional scanFrame argument of computeMask()
  • Prevent division by zero in point-by-point scans with zero duration (which is weird, but well...)
  • Only compute sensor->filtering transform for all-at-once scans, because point-by-point scans have viewpoints instead.
  • Reworked how TFFramesWatchdog is initialized to allow easier reconfiguring of the filter.
  • Added RobotBodyFilter::failWithoutRobotDescription which aids in tests.
  • Clarified documentation of RobotBodyFilter regarding frames.
  • Add TFFramesWatchdog::isRunning().
  • Workaround PCL 1.8 ignoring CropBox::setKeepOrganized().
  • Added tests for filter utils.
  • Improved filter_utils documentation.
  • Improved getParamVerbose() with recursive resolution of slash-separated param names.
  • Make different performance test requirements for release and debug builds.
  • Added tests for RayCastingShapeMask and improved its documentation.
  • Added utility functions to compare transforms and generate random ones.
  • Implemented TF watchdog tests.
  • Prevented some more SIGABRTs from TFFramesWatchdog.
  • Added unit tests for utils.
  • Fixed a bug in OBB.contains(OBB) which compared against vertices of the wrong OBB.
  • Improved efficiency of Eigen expression by not using auto type.
  • Fix the reported transformation of default-constructed OBBs.
  • Fixed constructShapeFromBody() to correctly process meshes.
  • Fixed eigen and urdf includes.
  • Fixed printing of std containers in string_utils.
  • CREATE_FILTERED_CLOUD now prepends robot_body_filter stuff with a global namespace to be usable even from other namespaces.
  • Make sure sensor frame is monitored by the watchdog because we need to use it during the filtering. Fixes #6.
  • Added parameter transforms/require_all_reachable.
  • Added to_string(bool) specialization.
  • Contributors: Martin Pecka

1.1.6 (2019-11-11)

  • Avoid subtracting from ROS time. This may lead to exceptions in simulation, where current time is 0. ros::Time cannot be negative.
  • Contributors: Martin Pecka

1.1.5 (2019-11-08)

  • refactor(RobotBodyFilter): Made frames/sensor parameter optional (#2) The sensor frame will now be derived from the incoming sensor messages. This way, messages from different sensors can be processed and the filter can be configured without any knowledge of the sensor.
  • Improved readme.
  • Reorganized the readme.
  • Added support for non-uniform scaling of collision meshes.
  • Contributors: Martin Pecka, Rein Appeldoorn

1.1.4 (2019-08-07)

  • Fixed a bug in filtering unorganized point clouds.
  • Contributors: Martin Pecka

1.1.1 (2019-07-30)

  • Fixed dependencies
  • Contributors: Martin Pecka

1.1.0 (2019-07-18)

  • Increasing performance.
  • More descriptive frame configurations. Added the possibility to leave out clipping, contains or shadow tests.
  • Repurposed param ~transforms/timeout/reachable to be always used with remainingTime.
  • More efficient point transformation
  • Little improvements. Renamed the utils library to robot_body_filter_utils.
  • Removed compiler pragmas - they seem no longer needed.
  • Provided a fix for wrong box-ray intersections.
  • Fixed bodies.h header guard.
  • Fixed bug in cylinder shape creation.
  • Fixed a bug in constructMarkerFromBody() - references do not correctly support polymorphism.
  • Fixes for robot body filter. Added new bounding box types.
  • Fixed a few bugs while running under a nodelet.
  • Prepared remainingTime() to a situation where ROS time hasn't yet been initialized.
  • Added possibility to update only robot pose only with every n-th point in point_by_point scans. Added possibility to publish bounding sphere and box markers.
  • Renamed to robot_body_filter.
  • Reworked as a laser filter combining contains and shadow tests.
  • Make use of bodies.h and shapes.h from geometric_shapes.
  • Enabling generic point types, removed PCL dependency, removed unnecessary params.
  • Using all collision elements for each link instead of only the first one.
  • Testing all intersections instead of only the first one.
  • Merge branch 'master' into indigo-devel
  • Add robot_self_filter namespace before bodies and shapes namespace. geometric_shapes package also provides bodies and shapes namespace and same classes and functions. If a program is linked with geometric_shapes and robot_self_filter, it may cause strange behavior because of symbol confliction.
  • Contributors: Martin Pecka, Ryohei Ueda, Tomas Petricek

0.1.31 (2018-11-24)

  • update CHANGELOG
  • Merge pull request #16 from mikaelarguedas/tinyxml_dependency depends on tinyxml and link against it
  • Merge branch 'indigo-devel' into tinyxml_dependency
  • Merge pull request #18 from k-okada/add_travis update travis.yml
  • update travis.yml
  • depend on tinyxml and link against it
  • Merge pull request #14 from traclabs/indigo-devel Minor changes to indigo-devel CMake allow this to be used in kinetic and indigo
  • Changes for kinetic
  • Contributors: Devon Ash, Kei Okada, Mikael Arguedas, Patrick Beeson

0.1.30 (2017-01-20)

  • Update CHANGELOG.rst
  • Merge pull request #15 from PR2/fix-typo-cmakelists Fix typo in CMakeLists.txt: CATKIN-DEPENDS -> CATKIN_DEPENDS
  • Fix typo in CMakeLists.txt: CATKIN-DEPENDS -> CATKIN_DEPENDS
  • Merge pull request #12 from garaemon/max-queue-size Add ~max_queue_size parameter for subscription queue size
  • Add ~max_queue_size parameter for subscription queue size
  • Contributors: Devon Ash, Kentaro Wada, Ryohei Ueda

0.1.29 (2015-12-05)

  • Re-create changelog for robot_self_filter
  • Merge pull request #10 from garaemon/pr-4-indigo-devel Add robot_self_filter namespace before bodies and shapes namespace.
  • Add robot_self_filter namespace before bodies and shapes namespace. geometric_shapes package also provides bodies and shapes namespace and same classes and functions. If a program is linked with geometric_shapes and robot_self_filter, it may cause strange behavior because of symbol confliction.
  • Contributors: Ryohei Ueda

0.1.28 (2015-12-04)

  • Merge pull request #8 from wkentaro/indigo-devel-merge-master Merge master branch to indigo-devel
  • Merge remote-tracking branch 'origin/master' into indigo-devel
  • Added indigo devel
  • Merge pull request #7 from wkentaro/self_filter-timestamp Set correct timestamp for self filtered cloud
  • Set correct timestamp for self filtered cloud This is needed because pcl drops some value of timestamp. So pcl::fromROSMsg and pcl::toROSMsg does not work to get correct timestamp.
  • Merge pull request #5 from garaemon/use-protected-member Protected member variables in SelfMask for subclass of SelfMask
  • Protected member variables in SelfMask for subclass of SelfMask
  • Contributors: Devon Ash, Kentaro Wada, Ryohei Ueda, TheDash

0.1.27 (2015-12-01)

  • Merge pull request #1 from garaemon/robot-self-filter Porting robot_self_filter from pr2_navigation_self_filter
  • Porting robot_self_filter from pr2_navigation_self_filter
  • Initial commit
  • Contributors: Devon Ash, Ryohei Ueda

Wiki Tutorials

This package does not provide any links to tutorials in it's rosindex metadata. You can check on the ROS Wiki Tutorials page for the package.

Launch files

Services

No service files found

Plugins

Recent questions tagged robot_body_filter at Robotics Stack Exchange