Been seeing a lot of people building robots that use the ChatGPT API to give them autonomy, but that’s like asking a writer to be a gymnast, so I’m building a software that makes better use of VLMs, Depth Estimation and World Models, to give autonomy to your robot. Building this in public. (skipped DAY 5 bc there was no much progress really) Today:
Tested out different visual odometry algorithms Turns out DA3 is also pretty good for pose estimation/odometry Was struggling for a bit generating a reasonable occupancy grid Reused some old code from my robotics research in college Turns out Bayesian Log-Odds Mapping yielded some kinda good results at least Pretty low definition voxels for now, but pretty good for SLAM that just uses a camera and no IMU or other odometry methods Working towards releasing this as an API alongside a Python SDK repo, for any builder to be able to add autonomy to their robot as long as it has a camera submitted by /u/L42ARO
Originally posted by u/L42ARO on r/ArtificialInteligence
