Pavlo Molchanov – NVIDIA Technical Blog News and tutorials for developers, data scientists, and IT admins 2025-01-07T04:01:29Z http://www.open-lab.net/blog/feed/ Pavlo Molchanov <![CDATA[Hymba Hybrid-Head Architecture Boosts Small Language Model Performance]]> http://www.open-lab.net/blog/?p=92595 2024-12-12T19:38:36Z 2024-11-22T17:31:14Z Transformers, with their attention-based architecture, have become the dominant choice for language models (LMs) due to their strong performance,...]]>

Transformers, with their attention-based architecture, have become the dominant choice for language models (LMs) due to their strong performance, parallelization capabilities, and long-term recall through key-value (KV) caches. However, their quadratic computational cost and high memory demands pose efficiency challenges. In contrast, state space models (SSMs) like Mamba and Mamba-2 offer constant…

Source

]]>
Pavlo Molchanov <![CDATA[Visual Language Models on NVIDIA Hardware with VILA]]> http://www.open-lab.net/blog/?p=81571 2025-01-07T04:01:29Z 2024-05-03T15:00:00Z Note: As of January 6, 2025 VILA is now part of the new Cosmos Nemotron vision language models. Visual language models have evolved significantly recently....]]>

Note: As of January 6, 2025 VILA is now part of the new Cosmos Nemotron vision language models. Visual language models have evolved significantly recently. However, the existing technology typically only supports one single image. They cannot reason among multiple images, support in context learning or understand videos. Also, they don’t optimize for inference speed. We developed VILA…

Source

]]>
1
���˳���97caoporen����