Meta AI Unveils AU-Net: A Scalable Byte-Level Model Surpassing Transformers in Language Tasks
Meta AI researchers have developed AU-Net, a scalable byte-level autoregressive U-Net model that outperforms traditional token-based transformers across multiple language modeling benchmarks, offering faster and more efficient text generation.
The Evolution of Language Modeling
Language modeling is essential in natural language processing (NLP), enabling machines to predict and generate text resembling human language. Traditionally, models progressed from statistical methods to neural architectures and now to large transformer-based systems. These models power applications like chatbots, translation, and text completion by interpreting and generating sequences of words or bytes. Effectiveness depends on architecture and data representation. As demand grows for efficient, scalable models, combining convolutional architectures with autoregressive prediction has become a promising direction.
Limitations of Token-Based Transformers
Most current language models rely on token-based transformers, which are computationally expensive and inefficient at the byte level or across multiple languages. Tokenization methods like Byte Pair Encoding (BPE) manage sequence lengths but cause inconsistencies across languages and domains. Transformers have quadratic complexity, limiting scalability. Sparse attention methods trade simplicity or performance for efficiency. Byte-level modeling with flat transformers has had limited success, highlighting the need for new architectures that process raw bytes without tokenization while maintaining high performance.
Introducing AU-Net: Autoregressive U-Net for Byte-Level Modeling
Researchers from Meta AI (FAIR), TAU, INRIA, and other institutions introduced AU-Net, a novel autoregressive U-Net model that works directly on bytes without tokenization. It integrates convolutional U-Net designs with autoregressive decoding, enabling parallel and efficient text generation. AU-Net uses hierarchical down-sampling and up-sampling convolutional stages to encode and restore sequence length. Its splitting mechanism allows prediction over subsegments concurrently, resulting in linear complexity with respect to sequence length rather than quadratic.
AU-Net Architecture and Training
AU-Net employs multiple scale stages with stride convolutions to reduce and reconstruct input sequences. During training, input segments are predicted with masking to maintain autoregressive properties. A learned splitting function divides sequences into non-overlapping groups for concurrent prediction, which are then combined into a full output. The model supports configurations ranging from 3% to 75% of training compute compared to baselines. For instance, an 8-billion parameter model trained on 200 billion tokens achieved highly competitive results, while a 1-billion parameter model trained on 60 billion tokens attained a 35.7 BLEU score in translation tasks, outperforming baselines trained on the same data. AU-Net also offers faster generation due to its parallel decoding, benefiting latency-sensitive applications.
Benchmark Performance and Multilingual Capabilities
AU-Net demonstrated strong results on various benchmarks. On Enwik8 (byte-level compression), it achieved 1.01 bits per byte (bpb), better than the transformer baseline at 1.02 bpb. On PG-19 (long-context modeling), AU-Net scored 2.61 bpb versus transformers’ 2.75 bpb. In FLORES-200 multilingual translation, the 8B model scored 43.3 BLEU, outperforming token-based transformers, especially in low-resource languages. It showed improved cross-lingual generalization and robustness to noise, with BLEU scores up to 33.0 across language families. AU-Net matched or exceeded transformer performance at equal compute budgets and improved generation speeds by 20–30%.
Key Contributions and Advantages
- Eliminates tokenization by operating on raw byte inputs.
- Outperforms transformer baselines on byte-level and multilingual benchmarks.
- Supports scalable training with linear complexity relative to sequence length.
- Delivers faster, parallel generation suited for real-time applications.
- Demonstrates robustness in multilingual and noisy data scenarios.
- Efficient compute usage, matching or surpassing transformers with less training cost.
Future Potential and Impact
AU-Net aligns with known scaling laws, improving steadily with increased model size and data. It scales effectively up to 8 billion parameters and performs well on downstream NLP tasks like generation and translation. The model is easier to train and more robust than token-based systems, positioning it as a promising alternative for large-scale, multilingual, and byte-level language modeling. This research challenges the dominance of token-based transformers by offering a more efficient, scalable, and inclusive architecture for future NLP developments.
For more details, check the original paper and GitHub repository maintained by the research teams.
Сменить язык
Читать эту статью на русском