News

The emerging field of neurosymbolic AI could solve challenges in the field, while also reducing the enormous amounts of data required to training LLMs.
For DOD, the future of large language models is smaller Everyone loves big AI, but “maybe there is a smaller-parameter model that could run on a laptop.” ...
Researchers found that vision-language models, widely used to analyze medical images, do not understand negation words like 'no' and 'not.' This could cause them to fail unexpectedly when asked to ...
A vision encoder is a necessary component for allowing many leading LLMs to be able to work with images uploaded by users.
The first presents Anthropic’s use of a technique called circuit tracing, which lets researchers track the decision-making processes inside a large language model step by step.
Cohere for AI, Cohere's nonprofit research lab, has released an 'open' multimodal AI model, Aya Vision, the lab claims is best-in-class.
Phi-4-multimodal is a 5.6 billion parameter model that uses the mixture-of-LoRAs technique to process speech, vision, and language simultaneously.
Microsoft Corp. today expanded its Phi line of open-source language models with two new algorithms optimized for multimodal processing and hardware efficiency. The first addition is the text-only ...