Meta open-sourced Byte Latent Transformer (BLT), an LLM architecture that uses a learned dynamic scheme for processing patches of bytes instead of a tokenizer. This allows BLT models to match the ...
First set out in a scientific paper last September, Pathway’s post-transformer architecture, BDH (Dragon hatchling), gives LLMs native reasoning powers with intrinsic memory mechanisms that support ...
Since the groundbreaking 2017 publication of “Attention Is All You Need,” the transformer architecture has fundamentally reshaped artificial intelligence research and development. This innovation laid ...
ETH Zurich and EPFL’s open-weight LLM offers a transparent alternative to black-box AI built on green compute and set for public release. Large language models (LLMs), which are neural networks that ...
In a new paper, researchers from clinical stage artificial intelligence (AI)-driven drug discovery company Insilico Medicine ("Insilico"), in collaboration with NVIDIA, present a new large language ...
Researchers at the Tokyo-based startup Sakana AI have developed a new technique that enables language models to use memory more efficiently, helping enterprises cut the costs of building applications ...
Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now Improving the capabilities of large ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results