RT-2: New model translates vision and language into action

RT-2: New model translates vision and language into action

Research Published 28 July 2023 Authors Yevgen Chebotar, Tianhe Yu Robotic Transformer 2 (RT-2) is a novel vision-language-action (VLA) model that learns from both web and robotics data, and translates this knowledge into generalised instructions for robotic control High-capacity vision-language models (VLMs) are trained on web-scale datasets, making these systems remarkably good at recognising visual…

Read More