The MiniCPM
class provides a wrapper for MiniCPM v2.6 that answers questions based on both images and videos.
If True, inference call is run on the local VM, else offloaded onto GRID-Cortex. Defaults to True.
This model is currently not available via Cortex.
If True, offloads inference to GRID-Cortex else runs locally on the session VM. Defaults to False.
If True, falls back to local inference if cloud inference fails. Defaults to True.
The input RGB image of shape (M,N,3). The question to answer about the media.
The response to the prompt.
from grid.model.perception.vlm.minicpm import MiniCPM
car = AirGenCar()
# We will be capturing an image from the AirGen simulator
# and run model inference on it.
img = car.getImage("front_center", "rgb").data
model = MiniCPM(use_local = True)
result = model.run(media=img, prompt=<prompt>)
print(result)
This code is licensed under the Apache 2.0 License. We have obtained official license from the company to offer this model on GRID.