2023-03-08
What happens when we train the largest vision-language model and add in robot experiences? The result is PaLM-E 🌴🤖, a 562-billion parameter, general-purpose, embodied visual-language generalist - across robotics, vision, and language. Website: https://palm-e.github.io/ https://twitter.com/...
Ars Technica
Google and the Technical University of Berlin unveil PaLM-E, a visual language model with 562B parameters, integrating vision and language for robotic control
ChatGPT-style AI model adds vision to guide a robot without special training. — On Monday, a group of AI researchers from Google …