Skip to content

Toolkit for benchmarking and understanding robotic grasping

License

Notifications You must be signed in to change notification settings

mrudorfer/burg-toolkit

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

BURG toolkit

This is a Python toolkit for Benchmarking and Understanding Robotic Grasping, developed in the scope of BURG project. Features are:

  • core data structures for object types and instances, scenes, grippers, grasps and grasp sets
  • antipodal grasp sampling
  • physics simulation with pybullet
  • depth/point cloud rendering with pybullet and/or pyrender
  • metrics for evaluation of grasps and grasp sets
  • visualization of scenes and grasps using open3d
  • dataset creation

It is the backend to the BURG Setup Tool.

project structure

The project contains the following directories:

  • burg_toolkit - the core Python library
  • scripts - examples
  • tests - messy manual, test files
  • docs - configuration files to create documentation with sphinx
  • data - simple data samples to be used in some example scripts

installation

Developed for:

  • Python 3.6-3.8
  • Ubuntu 20.04 and Windows, but at some point we stopped testing on Windows

Recommended way is to install in a virtual environment:

git clone [email protected]:mrudorfer/burg-toolkit.git
cd burg-toolkit

# create virtual environment
python3 -m venv venv
source venv/bin/activate

python -m pip install --upgrade pip
pip install --upgrade setuptools wheel

# install burg_toolkit in editable mode, also see install options
pip install -e .

If you experience any problems, please open an issue.

Sometimes, especially if numpy has already been installed in the environment, there may be a mismatch of compiler versions which cause numpy-quaternion to fail. See moble/quaternion#72 for details.

install options

There are some dependencies not included in the default installation. These are mostly packages that are either not required for regular usage or not installing smoothly on all platforms. The following extras are available (general installation procedure as above, explanation in the subsequent sections):

pip install -e .['docs']  # includes packages required to build documentation
pip install -e .['collision']  # mesh-mesh collision checks
pip install -e .['openexr']  # read/write .exr image files
pip install -e .['full']  # install all optional dependencies

docs

Dependencies for the documentation should install smoothly on all platforms. You can build the documentation on the command line (with virtual environment activated) via:

sphinx-apidoc -f -o docs burg_toolkit
cd docs/
make html  # linux
./make.bat html  # windows

The docs can then be found in docs/_build/htmlfolder.

collision

Mesh-mesh collisions currently rely on python-fcl which is only available for linux. There are some efforts to bring python-fcl to Windows, e.g. there are packages like python-fcl-win32 and python-fcl-win32-nr, but there seem to be issues installing those, see: BerkeleyAutomation/python-fcl#17. So far I have not been able to install either of those packages with pip.

openexr

This is required for saving images to .exr file format specified by OpenEXR. It requires to install OpenEXR on your system. On Linux, you can sudo apt-get install libopenexr-dev upon which the package should install smoothly. On Windows, it is more complicated. See https://stackoverflow.com/a/68102521/1264582.

usage

Example:

import numpy as np
import burg_toolkit as burg

gs1 = burg.GraspSet.from_translations(np.random.random(50, 3))
gs2 = burg.GraspSet.from_translations(np.random.random(30, 3))
print('grasp coverage is:', burg.metrics.coverage(gs1, gs2))

You can download an exemplary object library and two scenes here: setuptool-examples.zip [10.4MB]. The objects are borrowed from the YCB dataset. The object library can be used with the setup_scenes_example.py, it also shows how to generate the required additional files for simulation. Scenes can be loaded and visualised with:

import burg_toolkit as burg

scene_filename = 'path/to/scene.yaml'
scene, object_library, printout = burg.Scene.from_yaml(scene_filename)

print(object_library)
print(scene)
burg.visualization.show_geometries([scene])

plans for the project

house keeping

  • replace messy manual test files with more reasonable unit testing
  • proper packaging/distribution of example files
  • resolve numpy-quaternion issues in installation (maybe switch to quaternionic?)
  • package-level logging
  • update to newer open3d version
  • properly load textures for meshes
  • documentation on readthedocs

features

  • grasp sampling
    • there is some unused code currently, needs better structure
    • more configurable AGS, e.g. with option to use random rotation offset for creating grasp orientations
    • more consistency in grasp representations, i.e. canonical forms, opening width, grasping depth, contact points etc.
    • computation of analytic success metrics
  • simulation-based grasp assessment using pybullet
    • determine simulation-based grasp success rate for grasp sets
    • use various grippers

Citation

If you use this toolkit in your research, please consider a citation:

@misc{rudorfer2022,
  author = {Martin Rudorfer},
  title = {BURG toolkit, a Python module for benchmarking and understanding robotic grasping},
  howpublished = {\url{https://github.com/mrudorfer/burg-toolkit}},
  year = {2022}
}

Acknowledgments

This work was conducted within the BURG research project for Benchmarking and Understanding Robotic Grasping. It is supported by CHIST-ERA and EPSRC grant no. EP/S032487/1.

References

research

  • BURG research project: https://burg.acin.tuwien.ac.at/
  • Clemens Eppner, Arsalan Mousavian and Dieter Fox: "A Billion Ways to Grasps - An Evaluation of Grasp Sampling Schemes on a Dense, Physics-based Grasp Data Set", ISRR 2019
  • Berk Calli, Aaron Walsman, Arjun Singh, Siddhartha Srinivasa, Pieter Abbeel, Aaron M. Dollar: "Benchmarking in Manipulation Research: Using the Yale-CMU-Berkeley Object and Model Set", IEEE Robotics & Automation Magazine, vol. 22, no. 3, pp. 36-52, Sept. 2015

software

About

Toolkit for benchmarking and understanding robotic grasping

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages