Renderers are used to visualize the simulation, and can be used either in on-screen mode or headless (off-screen) mode. Renderers are also responsible for generating image-based observations that are returned from a given environment, and compute virtual images of the environment based on the properties defined in the cameras.
Currently, the following ground-truth vision modalities are supported across the two renderers, MuJoCo, and NVISII:
RGB: Standard 3-channel color frames with values in range
[0, 255]. This is set during environment construction with the
Depth: 1-channel frame with normalized values in range
[0, 1]. This is set during environment construction with the
Segmentation: 1-channel frames with pixel values corresponding to integer IDs for various objects. Segmentation can occur by class, instance, or geom, and is set during environment construction with the
Additional modalities are supported by a subset of the renderers. In robosuite, the user has the following rendering options:
MuJoCo exposes users to an OpenGL context supported by mujoco. Based on OpenGL, our assets and environment definitions have been tuned to look good with this renderer. The rendered frames can be displayed in a window with OpenCV’s imshow.
NVISIIRenderer is a ray tracing-based renderer. It is primarily used for training perception models and visualizing results in high quality. Through NVISII, we can obtain different vision modalities, including depth, segmentations, surface normals, texture coordinates, and texture positioning.
Using the NVISII renderer#
Installing NVISII can be done using the command
pip install nvisii. Note that NVISII requires users’ drivers to be up to date. Please refer here for more information. You can try the NVISII renderer with the
demo_renderers.py script and learn about the APIs for obtaining vision modalities with
The following table shows the estimated frame rate of each renderer in frames per second (FPS). The profiling was conducted on a machine with Ubuntu 18.04, Intel Core i9-900K CPU@3.60GHz, and Nvidia RTX. The FPS numbers of each rendering option are reported below. These numbers are estimated on the Door environment with IIWA robot and Joint Velocity controller and 256x256 image size.
|Simulation + rendering
For the same environment setup, we profiled the renderer on a machine with Ubuntu 18.04, Intel Core i7-8700K CPU@3.70GHz and Nvidia GTX 1080ti.
|Simulation + rendering
In practice, mujoco renderer is best-suited for vision-based policy learning. NVISII is best suited for photorealistic rendering; however, the ray-tracing computation substantially slows down its rendering speed compared to the other two renderers. It is mainly intended for perception tasks and qualitative visualizations, rather than online policy training.