Platform FAQs
I do not see the simulation streaming after I start a session- In certain browsers, the streaming does not start automatically and we’re investigating this. In the meantime, you can try clicking on the simulation streaming pane (top right) to start the streaming manually.
- This is most likely to happen if you’ve already run some code, such as the first cell in the notebook. The initialization of the simulation client enables API control for the robot, thus disabling keyboard control. To regain keyboard control, run
airgen_car_0.client.enableApiControl(False)
in a notebook cell.
- When you click inside the simulation window, the mouse cursor is captured to allow for interaction within the simulation environment (like camera control). To release the cursor and make it visible again, simply press the
Esc
key.
- Currently, the LLM only has access to a limited set of API for the simulation/robot, so it will not be able to run every type of instruction. We are working on integrating more functionality into the scope of the LLM.
- The Platform does not currently allow for multiple simultaneous sessions. To start a new session while one is ongoing, navigate to Session History → Active Sessions and terminate the ongoing one.
- This is a known issue and we are working on fixing it. The code is still running in the background, so you can ignore the progress bar. Once the code finishes running, the rotating circle icon will disappear.
- Press
Backspace
to reset the robot to its default starting position.
- Click on the simulator, and then press M to enter manual camera mode. Pressing H within the simulator context also shows a list of keyboard shortcuts and functionality.
- You cannot access the terminal (LLM interaction window) when there is code running in the notebook.
- This feature has only been tested for Chrome and we are in the process of enabling it for other browsers.
- Unlike image data which are passed as numpy arrays, for videos, we pass the video path as a string to the model. Given each model has a different preprocessing pipeline, we request you to write your video data to disk and pass the path to the model. The model will read the video and process it accordingly. We will be working on enabling video data to be passed as numpy arrays directly to the model in the future.
Isaac Sim FAQs
How to enable the third person camera to follow my robot on Isaac so I don’t loose track of it while it is moving?- GRID provides the feature to always track the robot with Isaac ensuring that the user never looses track of the robot.
Adding the following configuration to the
custom.yaml
ensures that the third person camera follows the robot.
- The users can change the
livestream
variable to3
and theheadless
variable totrue
in thecustom.yaml
. This would allow the users to use the mouse scroll to zoom in and out along with all the other mouse shortcuts for Isaac Sim.
To view the simulation, users will now have to go to this link.
- If you’re getting a CUDA error, it is suggested to run the command
terminate
to stop the containers and restarting usinginit
.