Radio Measurements of the Depth of Air-Shower Maximum at the Pierre Auger Observatory
Physical review D/Physical review D(2024)
Univ Adelaide | UL | INFN | UNCuyo | UNSAM | Gran Sasso Sci Inst | Univ Santiago de Compostela | CUNY Herbert H Lehman Coll | Univ Nova Gorica | Rhein Westfal TH Aachen | Univ Michoacana | Observ Pierre Auger | Czech Acad Sci | Colorado Sch Mines | Univ Grenoble Alpes | Max Planck Inst Radioastron | UNR | Radboud Univ Nijmegen | Univ Paris Saclay | Karlsruhe Inst Technol KIT | Univ Estadual Campinas | Colorado State Univ | Univ Granada | Vrije Univ Brussel | Univ Autonoma Chiapas | Univ Sao Paulo | Palacky Univ | Univ Delaware | Univ Fed Fluminense | Case Western Reserve Univ | Univ Siegen | Univ Nacl La Plata | Univ Buenos Aires | Univ Fed Rio de Janeiro | Inst Fed Educ Ciencia & Tecnol Rio de Janeiro IFR | Horia Hulubei Natl Inst Phys & Nucl Engn | Univ Nacl Autonoma Mexico | Ctr Brasileiro Pesquisas Fis | Univ Fed Parana | Univ Chicago | NYU | Fermilab Natl Accelerator Lab | Michigan Technol Univ | Berg Univ Wuppertal | Univ Estadual Feira de Santana | Inst Space Sci | Ctr Fed Educ Tecnol Celso Suckow da Fonseca | Univ Fed ABC | Univ Paris | Univ Wisconsin | Univ Libre Bruxelles ULB | Benemerita Univ Autonoma Puebla | Louisiana State Univ | Consejo Nacl Invest Cient & Tecn | Charles Univ Prague | Univ Ind Santander | IPN | Univ Fed Campina Grande | Univ Hamburg | UBA | Univ Lodz | Univ Medellin | Univ Fed Rio de Janeiro UFRJ | Univ Amsterdam | Univ Leeds
- Pretraining has recently greatly promoted the development of natural language processing (NLP)
- We show that M6 outperforms the baselines in multimodal downstream tasks, and the large M6 with 10 parameters can reach a better performance
- We propose a method called M6 that is able to process information of multiple modalities and perform both single-modal and cross-modal understanding and generation
- The model is scaled to large model with 10 billion parameters with sophisticated deployment, and the 10 -parameter M6-large is the largest pretrained model in Chinese
- Experimental results show that our proposed M6 outperforms the baseline in a number of downstream tasks concerning both single modality and multiple modalities We will continue the pretraining of extremely large models by increasing data to explore the limit of its performance

被引用126 | 浏览
被引用166 | 浏览
被引用227 | 浏览
被引用83 | 浏览
被引用136 | 浏览
被引用132 | 浏览
被引用35 | 浏览
被引用73 | 浏览
被引用41 | 浏览
被引用69 | 浏览
被引用46 | 浏览
被引用6 | 浏览
被引用64 | 浏览