Nvidia unveiled Alpamayo at CES 2026, which includes a reasoning vision language action model that allows an autonomous ...
These are the LLMs that caught our attention in 2025—from autonomous coding assistants to vision models processing entire codebases.
Multimodal large language models have shown powerful abilities to understand and reason across text and images, but their ...
Cohere Labs unveils AfriAya, a vision-language dataset aimed at improving how AI models understand African languages and ...
Chinese AI startup Zhipu AI aka Z.ai has released its GLM-4.6V series, a new generation of open-source vision-language models (VLMs) optimized for multimodal reasoning, frontend automation, and ...
Milestone announced the traffic-focused VLM, powered by NVIDIA Cosmos Reason, supports automated video summarization in ...
A research team affiliated with UNIST has unveiled a novel AI system capable of grading and providing detailed feedback on ...
Our first deployments began with, contrary to model tuning, a store audit: camera inventory, network strength and in-store ...
You're currently following this author! Want to unfollow? Unsubscribe via the link in your email. Follow Lakshmi Varanasi Every time Lakshmi publishes a story, you’ll get an alert straight to your ...
Bridging communication gaps between hearing and hearing-impaired individuals is an important challenge in assistive ...