Here's the link to the youtube video I played in the meeting.
https://youtube.com/shorts/LRyp4u0X1vA
For this video I told Elsabot to go to the playroom TV and see what it was showing and then come back and tell me. It then used these LLM tool calls I have implemented for Elsabot:
1. Requested the list of known locations (maps location name to x,y,yaw). From that it determined the Nav2 map location of the TV.
2. Requested the robot move to the playroom TV location.
3. Once the move request finished, it requested a camera frame and VLM analysis of the frame.
4. Requested the robot move back to the original location.
5. Described what it saw after it arrived back.
In addition to describing the scene on the TV, it also noticed the hand-written note I had placed in view.
It is using Gemma 4 26B which is pretty amazing.
Scott