Full News List

  • [2025.01]  🎉🎉 Our paper “Hot-pluggable Federated Learning: Briding General and Personalized FL via Dynamic Selection” is selected at ICLR 2025. This paper proposes a selective federated learning approach to integate personalized modules into general federated learning. (Paper)
  • [2025.01]  🎉🎉 Our paper “STBLLM: Breaking the 1-Bit Barrier with Structured Binary LLMs” is selected at ICLR 2025. This paper introduces STBLLM, a novel approach that breaks the 1-bit barrier in language models by leveraging Structured Binary LLMs. (Paper)
  • [2025.01]  🎉🎉 Our paper “The Lottery LLM Hypothesis, Rethinking What Abilities Should LLM Compression Preserve?” is selected as ICLR Blogpost 2025. This blog proposes a lottery LLM hypothesis suggesting that for a given LLM and task, there exists a smaller lottery LLM capable of producing the same performance with the original LLM with the assistances of multi-step reasoning and external tools. (Paper)
  • [2025.01]  🎉🎉 Our paper “Can LLM Simulations Truly Reflect Humanity? A Deep Dive.” is selected as ICLR Blogpost 2025. This blog rethinks LLM-based simulations by emphasizing both their limitations and the necessities for advancing LLM simulations. It offer actionable insights and strategies for enhancing the applicability of LLM simulations in human society in the future. (Paper)
  • [2025.01]  🎉🎉 Our paper “ParZC: Parametric Zero-Cost Proxies for Efficient NAS. “ is selected at AAAI 2025 for Oral Presentation!
  • [2024.12]  🎉🎉 Our paper “ParZC: Parametric Zero-Cost Proxies for Efficient NAS. “ is accepted at AAAI 2025! Parametric Zero-Cost Proxies (ParZC) method improves zero-shot Neural Architecture Search by addressing unequal node importance and using novel techniques for uncertainty estimation and architecture ranking. (paper and codes will come soon…)
  • [2024.11]  🎉🎉 I’m selected as the Top Reviewer of NeurIPS 2024 for both main and D&B tracks (Link).
  • [2024.10]  🎉🎉 Our paper “Hot Pluggable Federated Learning.” has been selected by the FL@FM-NeurIPS’24 workshop to receive the Outstanding Student Paper Award!. Congratulations to all co-authors!
  • [2024.10]  🎉🎉 Our paper “FSMoE: A Flexible and Scalable Training System for Sparse Mixture-of-Experts Models.” is accepted In ASPLOS 2025! In this paper, we design and implement a new training system modularizes various operators in the entire MoE model, providing more fine-grained computation and communication scheduling, and achieving better computation communication overlap through appropriate gradient segmentation.. (paper and codes will come soon…)
  • [2024.09]  🎉🎉 Our paper “Hot Pluggable Federated Learning.” is accepted at Workshop Federated Foundation Models@NeurIPS 2024 as an Oral paper!. In this paper, we propose a new method to regard model heads as pluggable modules appended after the model backbone. (paper and codes will come soon…)
  • [2024.09]  🎉🎉 Our paper “FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion.” is accepted at NeurIPS 2024 as a Spotlight paper ! This work identifies the cause of low performance of one-shot FL, and proposes FuseFL to progressively train and fuses DNN model following a bottom-up manner, reducing communication costs to an extremely low degree. (paper and codes will come soon…)
  • [2024.09]  🎉🎉 Our paper “Discovering Sparsity Allocation for Layer-wise Pruning of Large Language Models.” is accepted at NeurIPS 2024</ strong>. In this paper, we present a new method for optimizing layerwise sparsity allocation in large language models. (paper and codes will come soon…)
  • [2024.09]  🎉🎉 Our paper “Should We Really Edit Language Models? On the Evaluation of Edited Language Models.” is accepted at NeurIPS 2024. In this paper, we benchmark the methods of editing LLMs and see how they influence LLM performance. (paper and codes will come soon…)
  • [2024.09]  🎉🎉 Our paper “LPZero: Language Model Zero-cost Proxy Search from Zero.” is accepted at EMNLP 2024 (Findings). In this paper, we propose LPZero, which can automatically design zero-cost proxies for NLP tasks. It uses genetic programming to find optimal symbolic equations, outperforming human-designed proxies in ranking consistency. (paper and codes will come soon…)
  • [2024.06]  🎉🎉 Our paper “Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning.” is accepted at ICPP 2024. This work finds that the overlapness between indexes of compressed client model parameters demonstrates important information that can be utilized to adjust averging weights. (paper)
  • [2024.05]  🎉🎉 Our paper “Pruner-Zero: Evolving Symbolic Pruning Metric From Scratch for Large Language Models.” is accepted at ICML 2024. This work finds new pruning metric to prune LLMs to achieve SOTA performance under the same sparsity ratio. (paper)
  • [2024.03]  🎉🎉 VMRNN is available. This work proposes the VMRNN cell, a new recurrent unit that integrates the strengths of Vision Mamba blocks with LSTM. We construct a network centered on VMRNN cells to tackle spatiotemporal prediction tasks effectively. (paper, code)
  • [2024.01]  🎉🎉 Our paper “FedImpro: Measuring and Improving Client Update in Federated Learning” is accepted at ICLR 2024. (paper,code). This work trains the high-level neural network on reconstructed feature distributions, to accelerate FL training and enhance the model performance.
  • [2024.01]  🎉🎉 Our paper “Towards Efficient and Reliable LLM Serving: A Real-World Workload Study” is available. (paper), code. This paper introduces the first real-world trace dataset of LLM serving workloads, detailing user, system, and LLM behaviors. Many new insights of GPT services are found in this work.
  • [2023.10]  🎉🎉 Our paper “Reliable and Efficient In-Memory Fault Tolerance of Large Language Model Pretraining” is available. (paper). This work desings an in-memory fault-tolerance framework for large-scale LLM pretraining.
  • [2023.09]  🎉🎉 Our paper “FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs” is available. (paper). This paper envisions a decentralized system unlocking the potential vast untapped consumer-level GPUs in pre-training, inference and fine-tuning of LLMs with privacy protection.
  • [2023.02]  🎉🎉 Our paper “FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training” is available. (paper). This work aims to democratize simulating large-scale FL experiments. BTW, our open-source FL framework FedML has reached 2.6k stars at github.
  • [2022.12]  🎉🎉 Our paper “GossipFL: A Decentralized Federated Learning Framework with Sparsified and Adaptive Communication” has been accepted by IEEE TPDS. (paper).
  • [2022.10]  🎉🎉 Our paper “NAS-LID: Efficient Neural Architecture Search with Local Intrinsic Dimension” is accepted at AAAI 2023</ strong>. (paper).
  • [2022.05]  🎉🎉 Our paper “Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning” is accepted at ICML 2022. (paper, [code] (https://github.com/wizard1203/VHL)).
  • [2022.03]  🎉🎉 Our paper “FedCV: A Federated Learning Framework for Diverse Computer Vision Tasks” is accepted to FL-AAAI workshop '22 as Oral Presentation. paper.
  • [2021.07]  🎉🎉 Our paper “Data Resampling for Federated Learning with Non-IID Labels.” is accepted to FTL-IJCAI workshop ‘21 paper.
  • [2020.10]  🎉🎉 Our survey paper “A Quantitative Survey of Communication Optimizations in Distributed Deep Learning” [code, paper] has been accepted by IEEE Network Magazine.