default search action
Xiaonan Nie
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [i14]Bin Xiao, Chunan Shi, Xiaonan Nie, Fan Yang, Xiangwei Deng, Lei Su, Weipeng Chen, Bin Cui:
Clover: Regressive Lightweight Speculative Decoding with Sequential Knowledge. CoRR abs/2405.00263 (2024) - [i13]Pinxue Zhao, Hailin Zhang, Fangcheng Fu, Xiaonan Nie, Qibin Liu, Fang Yang, Yuanbo Peng, Dian Jiao, Shuaipeng Li, Jinbao Xue, Yangyu Tao, Bin Cui:
Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs. CoRR abs/2407.12117 (2024) - [i12]Hailin Zhang, Xiaodong Ji, Yilin Chen, Fangcheng Fu, Xupeng Miao, Xiaonan Nie, Weipeng Chen, Bin Cui:
PQCache: Product Quantization-based KVCache for Long Context LLM Inference. CoRR abs/2407.12820 (2024) - [i11]Guosheng Dong, Da Pan, Yiding Sun, Shusen Zhang, Zheng Liang, Xin Wu, Yanjun Shen, Fan Yang, Haoze Sun, Tianpeng Li, Mingan Lin, Jianhua Xu, Yufan Zhang, Xiaonan Nie, Lei Su, Bingning Wang, Wentao Zhang, Jiaxin Mao, Zenan Zhou, Weipeng Chen:
BaichuanSEED: Sharing the Potential of ExtensivE Data Collection and Deduplication by Introducing a Competitive Large Language Model Baseline. CoRR abs/2408.15079 (2024) - [i10]Keer Lu, Zheng Liang, Xiaonan Nie, Da Pan, Shusen Zhang, Keshi Zhao, Weipeng Chen, Zenan Zhou, Guosheng Dong, Wentao Zhang, Bin Cui:
DataSculpt: Crafting Data Landscapes for LLM Post-Training through Multi-objective Partitioning. CoRR abs/2409.00997 (2024) - [i9]Ran Yan, Youhe Jiang, Wangcheng Tao, Xiaonan Nie, Bin Cui, Binhang Yuan:
FlashFlex: Accommodating Large Language Model Training over Heterogeneous Environment. CoRR abs/2409.01143 (2024) - 2023
- [j5]Xupeng Miao, Xiaonan Nie, Hailin Zhang, Tong Zhao, Bin Cui:
Hetu: a highly efficient automatic parallel distributed deep learning system. Sci. China Inf. Sci. 66(1) (2023) - [j4]Xiaonan Nie, Xupeng Miao, Zilong Wang, Zichao Yang, Jilong Xue, Lingxiao Ma, Gang Cao, Bin Cui:
FlexMoE: Scaling Large-scale Sparse Pre-trained Model Training via Dynamic Device Placement. Proc. ACM Manag. Data 1(1): 110:1-110:19 (2023) - [j3]Xiaonan Nie, Yi Liu, Fangcheng Fu, Jinbao Xue, Dian Jiao, Xupeng Miao, Yangyu Tao, Bin Cui:
Angel-PTM: A Scalable and Economical Large-scale Pre-training System in Tencent. Proc. VLDB Endow. 16(12): 3781-3794 (2023) - [c4]Youhe Jiang, Fangcheng Fu, Xupeng Miao, Xiaonan Nie, Bin Cui:
OSDP: Optimal Sharded Data Parallel for Distributed Deep Learning. IJCAI 2023: 2142-2150 - [i8]Xiaonan Nie, Yi Liu, Fangcheng Fu, Jinbao Xue, Dian Jiao, Xupeng Miao, Yangyu Tao, Bin Cui:
Angel-PTM: A Scalable and Economical Large-scale Pre-training System in Tencent. CoRR abs/2303.02868 (2023) - [i7]Xiaonan Nie, Xupeng Miao, Zilong Wang, Zichao Yang, Jilong Xue, Lingxiao Ma, Gang Cao, Bin Cui:
FlexMoE: Scaling Large-scale Sparse Pre-trained Model Training via Dynamic Device Placement. CoRR abs/2304.03946 (2023) - [i6]Yujie Wang, Youhe Jiang, Xupeng Miao, Fangcheng Fu, Xiaonan Nie, Bin Cui:
Improving Automatic Parallel Training via Balanced Memory Workload Optimization. CoRR abs/2307.02031 (2023) - 2022
- [j2]Xupeng Miao, Yujie Wang, Youhe Jiang, Chunan Shi, Xiaonan Nie, Hailin Zhang, Bin Cui:
Galvatron: Efficient Transformer Training over Multiple GPUs Using Automatic Parallelism. Proc. VLDB Endow. 16(3): 470-479 (2022) - [c3]Xiaonan Nie, Xupeng Miao, Zhi Yang, Bin Cui:
TSPLIT: Fine-grained GPU Memory Management for Efficient DNN Training via Tensor Splitting. ICDE 2022: 2615-2628 - [c2]Xupeng Miao, Yining Shi, Hailin Zhang, Xin Zhang, Xiaonan Nie, Zhi Yang, Bin Cui:
HET-GMP: A Graph-based System Approach to Scaling Large Embedding Model Training. SIGMOD Conference 2022: 470-480 - [i5]Xiaonan Nie, Pinxue Zhao, Xupeng Miao, Tong Zhao, Bin Cui:
HetuMoE: An Efficient Trillion-scale Mixture-of-Expert Distributed Training System. CoRR abs/2203.14685 (2022) - [i4]Youhe Jiang, Xupeng Miao, Xiaonan Nie, Bin Cui:
OSDP: Optimal Sharded Data Parallel for Distributed Deep Learning. CoRR abs/2209.13258 (2022) - [i3]Xupeng Miao, Yujie Wang, Youhe Jiang, Chunan Shi, Xiaonan Nie, Hailin Zhang, Bin Cui:
Galvatron: Efficient Transformer Training over Multiple GPUs Using Automatic Parallelism. CoRR abs/2211.13878 (2022) - 2021
- [j1]Xupeng Miao, Hailin Zhang, Yining Shi, Xiaonan Nie, Zhi Yang, Yangyu Tao, Bin Cui:
HET: Scaling out Huge Embedding Model Training via Cache-enabled Distributed Framework. Proc. VLDB Endow. 15(2): 312-320 (2021) - [c1]Xupeng Miao, Xiaonan Nie, Yingxia Shao, Zhi Yang, Jiawei Jiang, Lingxiao Ma, Bin Cui:
Heterogeneity-Aware Distributed Machine Learning Training via Partial Reduce. SIGMOD Conference 2021: 2262-2270 - [i2]Xupeng Miao, Hailin Zhang, Yining Shi, Xiaonan Nie, Zhi Yang, Yangyu Tao, Bin Cui:
HET: Scaling out Huge Embedding Model Training via Cache-enabled Distributed Framework. CoRR abs/2112.07221 (2021) - [i1]Xiaonan Nie, Shijie Cao, Xupeng Miao, Lingxiao Ma, Jilong Xue, Youshan Miao, Zichao Yang, Zhi Yang, Bin Cui:
Dense-to-Sparse Gate for Mixture-of-Experts. CoRR abs/2112.14397 (2021)
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2024-10-07 02:34 CEST by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint