


default search action
USENIX Annual Technical Conference 2025: Boston, MA, USA
- Deniz Altinbüken, Ryan Stutsman:
Proceedings of the 2025 USENIX Annual Technical Conference, USENIX ATC 2025, Boston, MA, USA, July 7-9, 2025. USENIX Association 2025, ISBN 978-1-939133-48-9
Cloud Computing: Speed, Scale, and Serverless
- Sushant Kumar Gupta, Anil Raghunath Iyer, Chang Yu, Neel Bagora, Olivier Pomerleau, Vivek Kumar, Prunthaban Kanthakumar:
Fast ACS: Low-Latency File-Based Ordered Message Delivery at Scale. 1-17 - Zihao Chang, Jiaqi Zhu, Haifeng Sun, Yunlong Xie, Kan Shi, Ninghui Sun, Yungang Bao, Sa Wang:
Poby: SmartNIC-accelerated Image Provisioning for Coldstart in Clouds. 19-37 - Daniel Barcelona Pons, Aitor Arjona, Pedro García López, Enrique Molina-Giménez, Stepan Klymonchuk:
Burst Computing: Quick, Sudden, Massively Parallel Processing on Serverless Resources. 39-56 - Junhao Hu, Jiang Xu, Zhixia Liu, Yulong He, Yuetao Chen, Hao Xu, Jiang Liu, Jie Meng, Baoquan Zhang, Shining Wan, Gengyuan Dan, Zhiyu Dong, Zhihao Ren, Changhong Liu, Tao Xie, Dayun Lin, Qin Zhang, Yue Yu, Hao Feng, Xusheng Chen, Yizhou Shan:
DEEPSERVE: Serverless Large Language Model Serving at Scale. 57-72 - Yuan Yao, Chuan He, Chinedum Emmanuel Okwudire, Harsha V. Madhyastha:
Cosmic: Cost-Effective Support for Cloud-Assisted 3D Printing. 73-88
Accelerating ML Training: Parallelism, Tuning, and Modalities
- Yuke Wang, Boyuan Feng, Zheng Wang, Guyue Huang, Tony Tong Geng, Ang Li, Yufei Ding:
GMI-DRL: Empowering Multi-GPU DRL with Adaptive-Grained Parallelism. 89-103 - Kezhao Huang, Siqi Zhu, Mingshu Zhai, Liyan Zheng, Kinman Lei, Jiaao He, Yuyang Jin, Jidong Zhai:
mTuner: Accelerating Parameter-Efficient Fine-Tuning on Multi-GPU Servers with Elastic Tensor. 105-121 - Tuowei Wang, Xingyu Chen, Kun Li, Ting Cao, Ju Ren, Yaoxue Zhang:
JENGA: Enhancing LLM Long-Context Fine-tuning with Contextual Token Sparsity. 123-141 - Hairui Zhao, Qi Tian, Hongliang Li, Zizhong Chen:
FlexPipe: Maximizing Training Efficiency for Transformer-based Models with Variable-Length Inputs. 143-159 - Weiqi Feng, Yangrui Chen, Shaoyu Wang, Yanghua Peng, Haibin Lin, Minlan Yu:
Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation. 161-177
Networking: From Cloud to In-Network Intelligence
- Xudong Liao, Han Tian, Xinchen Wan, Chaoliang Zeng, Hao Wang, Junxue Zhang, Mengyu Ma, Guyue Liu, Kai Chen:
Towards Optimal Rack-scale μs-level CPU Scheduling through In-Network Workload Shaping. 179-198 - Yifan Yang, Lin He, Jiasheng Zhou, Xiaoyi Shi, Yichi Xu, Shicheng Wang, Jinlong E, Ying Liu, Junwei Zhang, Zhuang Yuan, Hengyang Xu:
TGW: Operating an Efficient and Resilient Cloud Gateway at Scale. 199-215 - Yuankang Zhao, Furong Yang, Gerui Lv, Qinghua Wu, Yanmei Liu, Jiuhai Zhang, Yutang Peng, Feng Peng, Hongyu Guo, Ying Chen, Zhenyu Li, Gaogang Xie:
MARC: Motion-Aware Rate Control for Mobile E-commerce Cloud Rendering. 217-232 - Zhaoyi Li, Jiawei Huang, Yijun Li, Jingling Liu, Junxue Zhang, Hui Li, Xiaojun Zhu, Shengwen Zhou, Jing Shao, Xiaojuan Lu, Qichen Su, Jianxin Wang, Chee Wei Tan, Yong Cui, Kai Chen:
Accelerating Distributed Graph Learning by Using Collaborative In-Network Multicast and Aggregation. 233-247 - Shinichi Awamoto, Michio Honda:
Opening Up Kernel-Bypass TCP Stacks. 249-262
Operating Systems: Scheduling, Security, and Extensibility
- Ruwen Fan, Tingxu Ren, Minhui Xie, Shiwei Gao, Jiwu Shu, Youyou Lu:
GPREEMPT: GPU Preemptive Scheduling Made General and Efficient. 263-272 - Le Chen, Yiyang Wu, Jinyu Gu, Yubin Xia, Haibo Chen:
μEFI: A Microkernel-Style UEFI with Isolation and Transparency. 273-289 - Anil Yelam, Kan Wu, Zhiyuan Guo, Suli Yang, Rajath Shashidhara, Wei Xu, Stanko Novakovic, Alex C. Snoeren, Kimberly Keeton:
PageFlex: Flexible and Efficient User-space Delegation of Linux Paging Policies with eBPF. 291-306 - Yuke Peng, Hongliang Tian, Junyang Zhang, Ruihan Li, Chengjun Chen, Jianfeng Jiang, Jinyi Xian, Xiaolin Wang, Chenren Xu, Diyu Zhou, Yingwei Luo, Shoumeng Yan, Yinqian Zhang:
ASTERINAS: A Linux ABI-Compatible, Rust-Based Framekernel OS with a Small and Sound TCB. 307-323 - Jinghao Jia, Ruowen Qin, Milo Craun, Egor Lukiyanov, Ayush Bansal, Minh Phan, Michael V. Le, Hubertus Franke, Hani Jamjoom, Tianyin Xu, Dan Williams:
Rex: Closing the language-verifier gap with safe and usable kernel extensions. 325-342
The Programmable Data Plane: SmartNICs and Beyond
- Yajuan Peng, Haoran Wei, Xiaolong Zhong, Junkai Huang, Haohan Xu, Zicheng Wang, Yang Bai, Zhuo Jiang, Jianxi Ye, Xiaoliang Wang, Xiaoming Fu, Huichen Dai:
Barre: Empowering Simplified and Versatile Programmable Congestion Control in High-Speed AI Clusters. 343-363 - Jinbin Hu, Wenxue Li, Xiangzhou Liu, Junfeng Wang, Bowen Liu, Ping Yin, Jianxin Wang, Jiawei Huang, Kai Chen:
FLB: Fine-grained Load Balancing for Lossless Datacenter Networks. 365-380 - Qiaoyin Gan, Heng Pan, Luyang Li, Kai Lv, Hongtao Guan, Zhaohua Wang, Zhenyu Li, Gaogang Xie:
SNARY: A High-Performance and Generic SmartNIC-accelerated Retrieval System. 381-398 - Zihao Wang, Qing Li, Guorui Xie, Dan Zhao, Kejun Li, Zhuochen Fan, Lianbo Ma, Yong Jiang:
Minos : A Lightweight and Dynamic Defense against Traffic Analysis in Programmable Data Planes. 399-415
Performance: Benchmarking, Caching, and Workload Characterization
- Yaoyu Wang, Xiao Guo, Junmin Xiao, De Chen, Guangming Tan:
GeneralSparse: Bridging the Gap in SpMM for Pruned Large Language Model Inference on GPUs. 417-432 - Keshav Vinayak Jha, Shweta Pandey, Murali Annavaram, Arkaprava Basu:
HyCache: Hybrid Caching for Accelerating DNN Input Preprocessing Pipelines. 433-448 - Evangelos Lamprou, Ethan Williams, Georgios Kaoukis, Zhuoxuan Zhang, Michael Greenberg, Konstantinos Kallas, Lukas Lazarek, Nikos Vasilakis:
The Koala Benchmarks for the Shell: Characterization and Implications. 449-464 - Jiahao Wang, Jinbo Han, Xingda Wei, Sijie Shen, Dingyan Zhang, Chenguang Fang, Rong Chen, Wenyuan Yu, Haibo Chen:
KVCache Cache in the Wild: Characterizing and Optimizing KVCache Cache at a Large Cloud Provider. 465-482
Storage Innovations: Logs, Tiers, and Modern Flash
- Junyu Wei, Guangyan Zhang, Junchao Chen, Qi Zhou:
LogCrisp: Fast Aggregated Analysis on Large-scale Compressed Logs by Enabling Two-Phase Pattern Extraction and Vectorized Queries. 483-496 - Jiansheng Qiu, Fangzhou Yuan, Mingyu Gao, Huanchen Zhang:
HotRAP: Hot Record Retention and Promotion for LSM-trees with Tiered Storage. 497-511 - Qingyang Zhang, Yongkun Li, Yubiao Pan, Haoting Tang, Yinlong Xu:
Mitigating Resource Usage Dependency in Sorting-based KV Stores on Hybrid Storage Devices via Operation Decoupling. 513-529 - Riwei Pan, Yu Liang, Lei Li, Hongchao Du, Tei-Wei Kuo, Chun Jason Xue:
SolFS: An Operation-Log Versioning File System for Hash-free Efficient Mobile Cloud Backup. 531-545 - Inhwi Hwang, Sangjin Lee, Sunggon Kim, Hyeonsang Eom, Yongseok Son:
Z-LFS: A Zoned Namespace-tailored Log-structured File System for Commodity Small-zone ZNS SSDs. 547-562
Serving Intelligence: Efficient LLM Inference at the Edge and Cloud
- Chunlin Tian, Xinpeng Qin, Kahou Tam, Li Li, Zijian Wang, Yuanzhe Zhao, Minglei Zhang, Chengzhong Xu:
CLONE: Customizing LLMs for Efficient Latency-Aware Inference at the Edge. 563-585 - Shiwei Gao, Qing Wang, Shaoxun Zeng, Youyou Lu, Jiwu Shu:
Weaver: Efficient Multi-LLM Serving with Attention Offloading. 587-595 - Minchen Yu, Ao Wang, Dong Chen, Haoxuan Yu, Xiaonan Luo, Zhuohao Li, Wei Wang, Ruichuan Chen, Dapeng Nie, Haoran Yang, Yu Ding:
Torpor: GPU-Enabled Serverless Computing for Low-Latency, Resource-Efficient Inference. 597-612 - Suyi Li, Hanfeng Lu, Tianyuan Wu, Minchen Yu, Qizhen Weng, Xusheng Chen, Yizhou Shan, Binhang Yuan, Wei Wang:
Toppings: CPU-Assisted, Rank-Aware Adapter Serving for LLM Inference. 613-629 - Qihao Zhang, Mingshu Zhai, Rui Sun, Jidong Zhai:
QFactory: Accelerating Quantized Large Language Model Serving with Qtile Graphs. 631-646
Optimizing ML Execution: Compilers, Pipelines, and Runtimes
- Ruofan Wu, Zhen Zheng, Feng Zhang, Chuanjie Liu, Zaifeng Pan, Jidong Zhai, Xiaoyong Du:
PluS: Highly Efficient and Expandable ML Compiler with Pluggable Graph Schedules. 647-663 - Yuzhou Huang, Yapeng Jiang, Zicong Hong, Wuhui Chen, Bin Wang, Weixi Zhu, Yue Yu, Zibin Zheng:
Obscura: Concealing Recomputation Overhead in Training of Large Language Models with Bubble-filling Pipeline Transformation. 665-678 - Z. Jonny Kong, Qiang Xu, Y. Charlie Hu:
PPipe: Efficient Video Analytics Serving on Heterogeneous GPU Clusters via Pool-Based Pipeline Parallelism. 679-698 - Yaqi Xia, Weihu Wang, Donglin Yang, Xiaobo Zhou, Dazhao Cheng:
Voltrix: Sparse Matrix-Matrix Multiplication on Tensor Cores with Asynchronous and Balanced Kernel Optimization. 699-714
Resilient Systems: Failure Detection, Consistency, and Scalability
- Zhaoyang Wan, Rongxin Han, Haifeng Sun, Qi Qi, Zirui Zhuang, Bo He, Liang Zhang, Jianxin Liao, Jingyu Wang:
NetKeeper: Enhancing Network Resilience with Autonomous Network Configuration Update on Traffic Patterns and Anomalies. 715-730 - Tianyuan Wu, Wei Wang, Yinghao Yu, Siran Yang, Wenchao Wu, Qinkai Duan, Guodong Yang, Jiamang Wang, Lin Qu, Liping Zhang:
GREYHOUND: Hunting Fail-Slows in Hybrid-Parallel Training at Scale. 731-747 - Shaohua Duan, Youmin Chen:
Crash Consistency in Block-Level Caching Systems: An Open CAS Case Study. 749-764 - Davide Rovelli, Pavel Chuprikov, Philipp Berdesinski, Ali Pahlevan, Patrick Jahnke, Patrick Eugster:
FiDe: Reliable and Fast Crash Failure Detection to Boost Datacenter Coordination. 765-788
Network Performance & Protocols: From Space to VR
- Suvam Basak, Amitangshu Pal, Debopam Bhattacherjee:
LEOCraft: Towards Designing Performant LEO Networks. 789-813 - Liying Wang, Qing Li, Yuhan Zhou, Zhaofeng Luo, Donghao Zhang, Shangguang Wang, Xuanzhe Liu, Chenren Xu:
Emulating Space Computing Networks with RHONE. 815-831 - Yifei Xu, Xumiao Zhang, Yuning Chen, Pan Hu, Xuan Zeng, Zhilong Zheng, Xianshang Lin, Yanmei Liu, Songwu Lu, Z. Morley Mao, Wan Du, Dennis Cai, Ennan Zhai, Yunfei Ma:
Roaming Free in the VR World with MP2. 833-850 - Liekun Hu, Changlong Li:
STORM: a Multipath QUIC Scheduler for Quick Streaming Media Transport under Unstable Mobile Networks. 851-866 - Gina Yuan, Thea Rossman, Keith Winstein:
Internet Connection Splitting: What's Old is New Again. 867-887
Distributed Systems: Communication, Consensus, and Data Structures
- Jiajian Zhang, Fangyu Wu, Hai Jiang, Qiufeng Wang, Genlang Chen, Chaoyi Pang:
WIC: Hiding Producer-Consumer Synchronization Delays with Warp-Level Interrupt-based GPU Communications. 889-904 - Jixi Shan, Xiuqi Huang, Yang Guo, Hongyue Mao, Ho-Pang Hsu, Hang Cheng, Can Wang, Jun Song, Rui Shi, Xiaofeng Gao, Jingwei Xu, Shiru Ren, Jiaxiao Zheng, Hua Huang, Lele Yu, Peng Xu, Guihai Chen:
Primus: Unified Training System for Large-Scale Deep Learning Recommendation Models. 905-922 - Rongji Huang, Xiangzhe Wang, Xiaofeng Yan, Lei Fan, Guangtao Xue, Shengyun Liu:
Chitu: Avoiding Unnecessary Fallback in Byzantine Consensus. 923-942 - Haodi Lu, Haikun Liu, Yujian Zhang, Zhuohui Duan, Xiaofei Liao, Hai Jin, Yu Zhang:
Fast Distributed Transactions for RDMA-based Disaggregated Memory. 943-958 - Stewart Grant, Alex C. Snoeren:
Cuckoo for Clients: Disaggregated Cuckoo Hashing. 959-972
Virtualization and Isolation: Security, Sharing, and Performance
- Kaesi Manakkal, Nathan Daughety, Marcus Pendleton, Hui Lu:
LITESHIELD: Secure Containers via Lightweight, Composable Userspace μKernel Services. 973-985 - Ori Ben Zur, Jakob Krebs, Shai Aviram Bergman, Mark Silberstein:
Accelerating Nested Virtualization with HyperTurtle. 987-1002 - Shulai Zhang, Ao Xu, Quan Chen, Han Zhao, Weihao Cui, Zhen Wang, Yan Li, Limin Xiao, Minyi Guo:
Efficient Performance-Aware GPU Sharing with Compatibility and Isolation through Kernel Space Interception. 1003-1019 - Tong Meng, Wei Zhang, Dong Chen, Zhen Wang, Quanqing Li, Changqing Yan, Wei Yang, Chao Yuan, Le Zhang, Jianxin Kuang, Jianlin Xu:
AnchorNet: Bridging Live and Collaborative Streaming with a Unified Architecture. 1021-1036
Scaling Complex Models: Distribution, Heterogeneity, and Efficiency
- Suyi Li, Lingyun Yang, Xiaoxiao Jiang, Hanfeng Lu, Dakai An, Zhipeng Di, Weiyi Lu, Jiawei Chen, Kan Liu, Yinghao Yu, Tao Lan, Guodong Yang, Lin Qu, Liping Zhang, Wei Wang:
Katz: Efficient Workflow Serving for Diffusion Models with Many Adapters. 1037-1052 - Junyi Zhang, Chuanhu Ma, Xiong Wang, Yuntao Nie, Yuqing Li, Yuedong Xu, Xiaofei Liao, Bo Li, Hai Jin:
PopFetcher: Towards Accelerated Mixture-of-Experts Training Via Popularity Based Expert-Wise Prefetch. 1053-1069 - Jiaao He, Shengqi Chen, Kezhao Huang, Jidong Zhai:
HypeReca: Distributed Heterogeneous In-Memory Embedding Database for Training Recommender Models. 1071-1087 - Tiancheng Chen, Ales Kubicek, Langwen Huang, Torsten Hoefler:
CrossPipe: Towards Optimal Pipeline Schedules for Cross-Datacenter Training. 1089-1108
Hunting Elusive Bugs: Verification and Analysis from Compilers to Hardware
- Jiangchang Wu, Yibiao Yang, Maolin Sun, Yuming Zhou:
Unveiling Compiler Faults via Attribute-Guided Compilation Space Exploration. 1109-1125 - Gen Dong, Yu Hua, Yongle Zhang, Zhangyu Chen, Menglei Chen:
Understanding and Detecting Fail-Slow Hardware Failure Bugs in Cloud Systems. 1127-1142 - Ruize Tang, Minghua Wang, Xudong Sun, Lin Huang, Yu Huang, Xiaoxing Ma:
Converos: Practical Model Checking for Verifying Rust OS Kernel Concurrency. 1143-1159 - Zao Yang, Stefan Nagy:
Bin2Wrong: a Unified Fuzzing Framework for Uncovering Semantic Errors in Binary-to-C Decompilers. 1161-1179 - Jiaqi Yin, Zhan Song, Nicolas Bohm Agostini, Antonino Tumeo, Cunxi Yu:
HEC: Equivalence Verification Checking for Code Transformation via Equality Saturation. 1181-1196
Software-Hardware Synergy: Accelerators, Memory, and Interconnects
- Houxiang Ji, Minho Kim, Seonmu Oh, Daehoon Kim, Nam Sung Kim:
Para-ksm: Parallelized Memory Deduplication with Data Streaming Accelerator. 1197-1212 - Ruili Liu, Teng Ma, Mingxing Zhang, Jialiang Huang, Yingdi Shan, Zheng Liu, Lingfeng Xiang, Zhen Lin, Hui Lu, Jia Rao, Kang Chen, Yongwei Wu:
DSA-2LM: A CPU-Free Tiered Memory Architecture with Intel DSA. 1213-1222 - Puqing Wu, Minhui Xie, Enrui Zhao, Dafang Zhang, Jing Wang, Xiao Liang, Kai Ren, Yunpeng Chai:
Turbocharge ANNS on Real Processing-in-Memory by Enabling Fine-Grained Per-PIM-Core Scheduling. 1223-1241 - Hongjing Huang, Jie Zhang, Xuzheng Chen, Ziyu Song, Jiajun Qin, Zeke Wang:
SwCC: Software-Programmable and Per-Packet Congestion Control in RDMA Engine. 1243-1260 - Xu Zhang, Ke Liu, Yuan Hui, Xiaolong Zheng, Yisong Chang, Yizhou Shan, Guanghui Zhang, Ke Zhang, Yungang Bao, Mingyu Chen, Chenxi Wang:
DRack: A CXL-Disaggregated Rack Architecture to Boost Inter-Rack Communication. 1261-1279
Securing the Stack: Attestation, Memory Protection, and Privacy
- Jingyuan Yang, Jun Wu, Ruilin Wu, Jingwei Li, Patrick P. C. Lee, Xiong Li, Xiaosong Zhang:
ShieldReduce: Fine-Grained Shielded Data Reduction. 1281-1296 - Chenke Luo, Jiang Ming, Dongpeng Xu, Guojun Peng, Jianming Fu:
MemoryTrap: Booby Trapping Memory to Counter Memory Disclosure Attacks with Hardware Support. 1297-1318 - Carsten Weinhold, Muhammad Usama Sardar, Ionut Mihalcea, Yogesh Deshpande, Hannes Tschofenig, Yaron Sheffer, Thomas Fossati, Michael Roitzsch:
Separate but Together: Integrating Remote Attestation into TLS. 1319-1326 - Zhiyong Wu, Jie Liang, Jingzhou Fu, Wenqian Deng, Yu Jiang:
DDLumos: Understanding and Detecting Atomic DDL Bugs in DBMSs. 1327-1341
Hardware-Specific Optimizations: Space, Accelerators, and AI Chips
- Jiacheng Liu, Xiaozhi Zhu, Tongqiao Xu, Xiaofeng Hou, Chao Li:
SpaceExit: Enabling Efficient Adaptive Computing in Space with Early Exits. 1343-1358 - Neel Patel, Mohammad Alian:
XRT: An Accelerator-Aware Runtime for Accelerated Chip Multiprocessors. 1359-1369 - Chen Ding, Sicen Li, Kai Lu, Ting Yao, Daohui Wang, Huatao Wu, Jiguang Wan, Zhihu Tan, Changsheng Xie:
DShuffle: DPU-Optimized Shuffle Framework for Large-scale Data Processing. 1371-1386 - Yuhang Zhou, Zibo Wang, Zhibin Wang, Ruyi Zhang, Chen Tian, Xiaoliang Wang, Wanchun Dou, Guihai Chen, Bingqiang Wang, Yonghong Tian, Yan Zhang, Hui Wang, Fuchun Wei, Boquan Sun, Jingyi Zhang, Bin She, Teng Su, Yifan Yao, Chunsheng Li, Ziyang Zhang, Yaoyuan Wang, Bin Zhou, Guyue Liu:
Accelerating Model Training on Ascend Chips: An Industrial System for Profiling, Analysis and Optimization. 1387-1408
Finding Faults: Concurrency, Numerics, and Kernel Fuzzing
- Yuanliang Chen, Fuchen Ma, Yuanhang Zhou, Zhen Yan, Yu Jiang:
CAFault: Enhance Fault Injection Technique in Practical Distributed Systems via Abundant Fault-Dependent Configurations. 1409-1424 - Peichen Xie, Yanjie Gao, Yang Wang, Jilong Xue:
Revealing Floating-Point Accumulation Orders in Software/Hardware Implementations. 1425-1440 - Yunmo Zhang, Junqiao Qiu, Hong Xu, Chun Jason Xue:
Inferring Likely Counting-related Atomicity Program Properties for Persistent Memory. 1441-1450 - Hui Guo, Hao Sun, Shan Huang, Ting Su, Geguang Pu, Shaohua Li:
Optimizing Input Minimization in Kernel Fuzzing. 1451-1465 - Tobias Landsberg, Johannes Grunenberg, Christian Dietrich, Daniel Lohmann:
IRHash: Efficient Multi-Language Compiler Caching by IR-Level Hashing. 1467-1479
Advanced Distributed Systems: ML, Data, and Performance Tuning
- Giovanni Bartolomeo, Navidreza Asadi, Wolfgang Kellerer, Jörg Ott, Nitinder Mohan:
On-Demand Container Partitioning for Distributed ML. 1481-1500 - Sukjin Kim, Seongyeon Park, Si Ung Noh, Junguk Hong, Taehee Kwon, Hunseong Lim, Jinho Lee:
PathWeaver: A High-Throughput Multi-GPU System for Graph-Based Approximate Nearest Neighbor Search. 1501-1517 - Xinyu Lian, Sam Ade Jacobs, Lev Kurilenko, Masahiro Tanaka, Stas Bekman, Olatunji Ruwase, Minjia Zhang:
Universal Checkpointing: A Flexible and Efficient Distributed Checkpointing System for Large-Scale DNN Training with Reconfigurable Parallelism. 1519-1534 - Jianjun Zhao, Haikun Liu, Shuhao Zhang, Haodi Lu, Yancan Mao, Zhuohui Duan, Xiaofei Liao, Hai Jin:
Towards High-Performance Transactional Stateful Serverless Workflows with Affinity-Aware Leasing. 1535-1551 - Chao Chen, Shixin Huang, Xuehai Qian, Zhibin Yu:
Swift: Fast Performance Tuning with GAN-Generated Configurations. 1553-1568
OS, Mobile, and Reliability Challenges
- Wentong Li, Li-Pin Chang, Yu Mao, Liang Shi:
PMR: Fast Application Response via Parallel Memory Reclaim on Mobile Devices. 1569-1584 - Wenxin Zheng, Bin Xu, Jinyu Gu, Haibo Chen:
SAVE: Software-Implemented Fault Tolerance for Model Inference against GPU Memory Bit Flips. 1585-1604 - Yidong Gong, Arnab Kanti Tarafder, Saima Afrin, Pradeep Kumar:
Identifying and Analyzing Pitfalls in GNN Systems. 1605-1624 - Dominik-Philip Schreiber, Manuel Leithner, Jovan Zivanovic, Dimitris E. Simos:
Bluetooth Low Energy Security Testing with Combinatorial Methods. 1625-1638
Intelligent Resource Management: Federation, Colocation, and ML for Systems
- Yongjun He, Haofeng Yang, Yao Lu, Ana Klimovic, Gustavo Alonso:
Resource Multiplexing in Tuning and Serving Large Language Models. 1639-1655 - Jiali Wang, Yankui Wang, Mingcong Han, Rong Chen:
Colocating ML Inference and Training with Fast GPU Memory Handover. 1657-1675 - Shichen Zhan, Li Li, Chengzhong Xu:
AssyLLM: Efficient Federated Fine-tuning of LLMs via Assembling Pre-trained Blocks. 1677-1691 - Guanglin Duan, Yucheng Huang, Zhengxin Zhang, Qing Li, Dan Zhao, Zili Meng, Dirk Kutscher, Ruoyu Li, Yong Jiang, Mingwei Xu:
Learning-Enhanced High-Throughput Pattern Matching Based on Programmable Data Plane. 1693-1712

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.