There’s been a lot of excitement around agentic AI as of late, and with Microsoft’s recently released Magma, the company believes its new foundation AI model will empower AI agents to execute multimodal tasks efficiently in both digital and real-world contexts, whether via software or physical robots.
Arising from a collaboration between researchers at Microsoft, KAIST, the University of Maryland, the University of Wisconsin-Madison, and the University of Washington, Magma expands on prior work in vision language models (VLMs), making it…