1 code implementation • Findings (NAACL) 2022 • Huan Lin, Baosong Yang, Liang Yao, Dayiheng Liu, Haibo Zhang, Jun Xie, Min Zhang, Jinsong Su
Diverse NMT aims at generating multiple diverse yet faithful translations given a source sentence.
4 code implementations • 19 Dec 2024 • Qwen, :, An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, TianHao Li, Tianyi Tang, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, Zihan Qiu
In addition, for hosted solutions, the proprietary models currently include two mixture-of-experts (MoE) variants: Qwen2. 5-Turbo and Qwen2. 5-Plus, both available from Alibaba Cloud Model Studio.
Ranked #6 on
on GPQA
1 code implementation • 29 Jul 2024 • Xin Zhang, Yanzhao Zhang, Dingkun Long, Wen Xie, Ziqi Dai, Jialong Tang, Huan Lin, Baosong Yang, Pengjun Xie, Fei Huang, Meishan Zhang, Wenjie Li, Min Zhang
We first introduce a text encoder (base size) enhanced with RoPE and unpadding, pre-trained in a native 8192-token context (longer than 512 of previous multilingual encoders).
5 code implementations • 15 Jul 2024 • An Yang, Baosong Yang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Zhou, Chengpeng Li, Chengyuan Li, Dayiheng Liu, Fei Huang, Guanting Dong, Haoran Wei, Huan Lin, Jialong Tang, Jialin Wang, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Ma, Jianxin Yang, Jin Xu, Jingren Zhou, Jinze Bai, Jinzheng He, Junyang Lin, Kai Dang, Keming Lu, Keqin Chen, Kexin Yang, Mei Li, Mingfeng Xue, Na Ni, Pei Zhang, Peng Wang, Ru Peng, Rui Men, Ruize Gao, Runji Lin, Shijie Wang, Shuai Bai, Sinan Tan, Tianhang Zhu, TianHao Li, Tianyu Liu, Wenbin Ge, Xiaodong Deng, Xiaohuan Zhou, Xingzhang Ren, Xinyu Zhang, Xipin Wei, Xuancheng Ren, Xuejing Liu, Yang Fan, Yang Yao, Yichang Zhang, Yu Wan, Yunfei Chu, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, Zhifang Guo, Zhihao Fan
This report introduces the Qwen2 series, the latest addition to our large language models and large multimodal models.
Ranked #1 on
Arithmetic Reasoning
on GSM8K
(using extra training data)
1 code implementation • 12 Jul 2023 • Xiangpeng Wei, Haoran Wei, Huan Lin, TianHao Li, Pei Zhang, Xingzhang Ren, Mei Li, Yu Wan, Zhiwei Cao, Binbin Xie, Tianxiang Hu, Shangjie Li, Binyuan Hui, Bowen Yu, Dayiheng Liu, Baosong Yang, Fei Huang, Jun Xie
Large language models (LLMs) demonstrate remarkable ability to comprehend, reason, and generate following nature language instructions.
1 code implementation • 26 May 2023 • Zhiwei Cao, Baosong Yang, Huan Lin, Suhang Wu, Xiangpeng Wei, Dayiheng Liu, Jun Xie, Min Zhang, Jinsong Su
$k$-Nearest neighbor machine translation ($k$NN-MT) has attracted increasing attention due to its ability to non-parametrically adapt to new translation domains.
no code implementations • 4 May 2023 • Binbin Xie, Jia Song, Liangying Shao, Suhang Wu, Xiangpeng Wei, Baosong Yang, Huan Lin, Jun Xie, Jinsong Su
In this paper, we comprehensively summarize representative studies from the perspectives of dominant models, datasets and evaluation metrics.
1 code implementation • 13 Nov 2022 • Binbin Xie, Xiangpeng Wei, Baosong Yang, Huan Lin, Jun Xie, Xiaoli Wang, Min Zhang, Jinsong Su
Keyphrase generation aims to automatically generate short phrases summarizing an input document.
no code implementations • 15 Jul 2022 • Jianwei Lin, Jiatai Lin, Cheng Lu, Hao Chen, Huan Lin, Bingchao Zhao, Zhenwei Shi, Bingjiang Qiu, Xipeng Pan, Zeyan Xu, Biao Huang, Changhong Liang, Guoqiang Han, Zaiyi Liu, Chu Han
To bridge the gap between Transformer and CNN features, we propose a Trans&CNN Feature Calibration block (TCFC) in the decoder.
no code implementations • 13 Apr 2022 • Chu Han, Xipeng Pan, Lixu Yan, Huan Lin, Bingbing Li, Su Yao, Shanshan Lv, Zhenwei Shi, Jinhai Mai, Jiatai Lin, Bingchao Zhao, Zeyan Xu, Zhizhen Wang, Yumeng Wang, Yuan Zhang, Huihui Wang, Chao Zhu, Chunhui Lin, Lijian Mao, Min Wu, Luwen Duan, Jingsong Zhu, Dong Hu, Zijie Fang, Yang Chen, Yongbing Zhang, Yi Li, Yiwen Zou, Yiduo Yu, Xiaomeng Li, Haiming Li, Yanfen Cui, Guoqiang Han, Yan Xu, Jun Xu, Huihua Yang, Chunming Li, Zhenbing Liu, Cheng Lu, Xin Chen, Changhong Liang, Qingling Zhang, Zaiyi Liu
According to the technical reports of the top-tier teams, CAM is still the most popular approach in WSSS.
Data Augmentation
Weakly supervised Semantic Segmentation
+1
1 code implementation • 14 Oct 2021 • Chu Han, Jiatai Lin, Jinhai Mai, Yi Wang, Qingling Zhang, Bingchao Zhao, Xin Chen, Xipeng Pan, Zhenwei Shi, Xiaowei Xu, Su Yao, Lixu Yan, Huan Lin, Zeyan Xu, Xiaomei Huang, Guoqiang Han, Changhong Liang, Zaiyi Liu
In the segmentation phase, we achieved tissue semantic segmentation by our proposed Multi-Layer Pseudo-Supervision.
1 code implementation • ACL 2021 • Huan Lin, Liang Yao, Baosong Yang, Dayiheng Liu, Haibo Zhang, Weihua Luo, Degen Huang, Jinsong Su
Furthermore, we contribute the first Chinese-English parallel corpus annotated with user behavior called UDT-Corpus.
1 code implementation • 4 Sep 2020 • Huan Lin, Fandong Meng, Jinsong Su, Yongjing Yin, Zhengyuan Yang, Yubin Ge, Jie zhou, Jiebo Luo
Particularly, we represent the input image with global and regional visual features, we introduce two parallel DCCNs to model multimodal context vectors with visual features at different granularities.
Ranked #3 on
Multimodal Machine Translation
on Multi30K
no code implementations • 25 Feb 2020 • DES Collaboration, Tim Abbott, Michel Aguena, Alex Alarcon, Sahar Allam, Steve Allen, James Annis, Santiago Avila, David Bacon, Alberto Bermeo, Gary Bernstein, Emmanuel Bertin, Sunayana Bhargava, Sebastian Bocquet, David Brooks, Dillon Brout, Elizabeth Buckley-Geer, David Burke, Aurelio Carnero Rosell, Matias Carrasco Kind, Jorge Carretero, Francisco Javier Castander, Ross Cawthon, Chihway Chang, Xinyi Chen, Ami Choi, Matteo Costanzi, Martin Crocce, Luiz da Costa, Tamara Davis, Juan De Vicente, Joseph DeRose, Shantanu Desai, H. Thomas Diehl, Jörg Dietrich, Scott Dodelson, Peter Doel, Alex Drlica-Wagner, Kathleen Eckert, Tim Eifler, Jack Elvin-Poole, Juan Estrada, Spencer Everett, August Evrard, Arya Farahi, Ismael Ferrero, Brenna Flaugher, Pablo Fosalba, Josh Frieman, Juan Garcia-Bellido, Marco Gatti, Enrique Gaztanaga, David Gerdes, Tommaso Giannantonio, Paul Giles, Sebastian Grandis, Daniel Gruen, Robert Gruendl, Julia Gschwend, Gaston Gutierrez, Will Hartley, Samuel Hinton, Devon L. Hollowood, Klaus Honscheid, Ben Hoyle, Dragan Huterer, David James, Mike Jarvis, Tesla Jeltema, Margaret Johnson, Stephen Kent, Elisabeth Krause, Richard Kron, Kyler Kuehn, Nikolay Kuropatkin, Ofer Lahav, Ting Li, Christopher Lidman, Marcos Lima, Huan Lin, Niall MacCrann, Marcio Maia, Adam Mantz, Jennifer Marshall, Paul Martini, Julian Mayers, Peter Melchior, Juan Mena, Felipe Menanteau, Ramon Miquel, Joe Mohr, Robert Nichol, Brian Nord, Ricardo Ogando, Antonella Palmese, Francisco Paz-Chinchon, Andrés Plazas Malagón, Judit Prat, Markus Michael Rau, Kathy Romer, Aaron Roodman, Philip Rooney, Eduardo Rozo, Eli Rykoff, Masao Sako, Simon Samuroff, Carles Sanchez, Alexandro Saro, Vic Scarpine, Michael Schubnell, Daniel Scolnic, Santiago Serrano, Ignacio Sevilla, Erin Sheldon, J. Allyn Smith, Eric Suchyta, Molly Swanson, Gregory Tarle, Daniel Thomas, Chun-Hao To, Michael A. Troxel, Douglas Tucker, Tamas Norbert Varga, Anja von der Linden, Alistair Walker, Risa Wechsler, Jochen Weller, Reese Wilkinson, Hao-Yi Wu, Brian Yanny, Zhuowen Zhang, Joe Zuntz
We perform a joint analysis of the counts and weak lensing signal of redMaPPer clusters selected from the Dark Energy Survey (DES) Year 1 dataset.
Cosmology and Nongalactic Astrophysics
no code implementations • 11 Sep 2018 • Boris Beranger, Huan Lin, Scott A. Sisson
We assume that, as with a standard statistical analysis, inference is required at the level of individual-level data.
1 code implementation • 6 Nov 2007 • Hiroaki Oyaizu, Marcos Lima, Carlos E. Cunha, Huan Lin, Joshua Frieman
Photometric redshift (photo-z) estimates are playing an increasingly important role in extragalactic astronomy and cosmology.