This project integrates Vision-Language Models (VLMs) with autonomous driving systems to enhance decision-making through scene understanding and reasoning. Techniques like YOLOv8, SAM2, and optical flow are utilized for robust object tracking and motion estimation.
This project integrates Vision-Language Models (VLMs) with autonomous driving systems to enhance decision-making through scene understanding and reasoning. Techniques like YOLOv8, SAM2, and optical flow are utilized for robust object tracking and motion estimation.