BestHub
Discover
Artificial IntelligenceBackend DevelopmentMobile DevelopmentProduct ManagementCloud NativeFrontend DevelopmentFundamentalsBig DataCloud ComputingGame DevelopmentR&D ManagementOperationsDatabasesInformation SecurityBlockchainUser Experience DesignInterview ExperienceIndustry Insights
View all →
TopicsTagsTrendsRanking
Sign in
Discover
Artificial Intelligence Backend Development Mobile Development Product Management Cloud Native Frontend Development Fundamentals Big Data Cloud Computing Game Development R&D Management Operations Databases Information Security Blockchain User Experience Design Interview Experience Industry Insights View all →
TopicsTagsTrendsRanking
Sign in
  1. Home
  2. / Tags
  3. / Linear Complexity
Network Intelligence Research Center (NIRC)
Network Intelligence Research Center (NIRC)
Aug 22, 2023 · Artificial Intelligence

LONGNET: Extending Transformers to Over 1 Billion Tokens

LONGNET introduces dilated attention to enable Transformers to process sequences exceeding one billion tokens with linear computational cost, preserving performance on shorter inputs and demonstrating strong results on long‑sequence modeling and standard language tasks.

Dilated AttentionLONGNETLanguage Modeling
0 likes · 6 min read
LONGNET: Extending Transformers to Over 1 Billion Tokens
BestHub

Editorial precision for engineers who prefer signal over noise. Deep reads, careful curation, and sharper frontiers in software.

Best Hub for Dev. Power Your Build.
Navigation
Discover Tags Topics System Status Privacy Terms Rss Feed