Deepseek VL-2 is a sophisticated vision-language model designed to address complex multimodal tasks with remarkable efficiency and precision. Built on a new mixture of experts (MoE) architecture, this ...
Hugging Face Inc. today open-sourced SmolVLM-256M, a new vision language model with the lowest parameter count in its category. The algorithm’s small footprint allows it to run on devices such as ...
A research team has developed a chest X-ray vision-language foundation model, MaCo, reducing the dependency on annotations while improving both clinical efficiency and diagnostic accuracy. The study ...
As I highlighted in my last article, two decades after the DARPA Grand Challenge, the autonomous vehicle (AV) industry is still waiting for breakthroughs—particularly in addressing the “long tail ...
Indian start-up Sarvam AI's sovereign Large Language Model (LLM) challenges larger global models in multilingual tasks, particularly in Indic languages.
Alibaba Cloud, the cloud services and storage division of the Chinese e-commerce giant, has announced the release of Qwen2-VL, its latest advanced vision-language model designed to enhance visual ...
BharatGen’s launch of its sovereign AI models was hailed as a decisive step towards technological self-reliance.
Sarvam is ideal for India-specific applications, where local language support is crucial.
Milestone Systems has released an advanced vision language model (VLM) specializing in traffic understanding, powered by NVIDIA Cosmos Reason, a framework designed to enable advanced reasoning across ...
Xiaomi is best known for smartphones, smart home gear, and the occasional electric vehicle update. Now it wants a place in robotics research too. The company has announced Xiaomi-Robotics-0, an ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results