T-araVLN: Translator for Agricultural Robotic Agents on Vision-and-Language Navigation
By: Xiaobei Zhao, Xingqi Lyu, Xiang Li
Potential Business Impact:
Robots follow farm directions better to harvest crops.
Agricultural robotic agents have been becoming powerful helpers in a wide range of agricultural tasks, however, still heavily rely on manual operation or fixed railways for movement. To address this limitation, the AgriVLN method and the A2A benchmark pioneeringly extend Vision-and-Language Navigation (VLN) to the agricultural domain, enabling agents to navigate to the target positions following the natural language instructions. AgriVLN effectively understands the simple instructions, but often misunderstands the complex ones. To bridge this gap, we propose the method of Translator for Agricultural Robotic Agents on Vision-and-Language Navigation (T-araVLN), in which the Instruction Translator module translates the original instruction to be more refined and precise. When evaluated on the A2A benchmark, our T-araVLN effectively improves Success Rate from 0.47 to 0.63 and reduces Navigation Error from 2.91m to 2.28m, demonstrating the state-of-the-art performance in the agricultural domain. Code: https://github.com/AlexTraveling/T-araVLN.
Similar Papers
T-araVLN: Translator for Agricultural Robotic Agents on Vision-and-Language Navigation
Robotics
Helps farm robots follow tricky directions better.
T-araVLN: Translator for Agricultural Robotic Agents on Vision-and-Language Navigation
Robotics
Helps farm robots follow complex spoken directions.
T-araVLN: Translator for Agricultural Robotic Agents on Vision-and-Language Navigation
Robotics
Helps farm robots follow complex spoken directions.