Stars
Speed up Stable Diffusion with this one simple trick!
Code for CVPR 2022 paper: Coarse-to-Fine Feature Mining for Video Semantic Segmentation
🎉 PILOT: A Pre-trained Model-Based Continual Learning Toolbox
[ECCV 2024] SparseRefine: Sparse Refinement for Efficient High-Resolution Semantic Segmentation
[ICLR 2024] The Need for Speed: Pruning Transformers with One Recipe
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions
[CVPR'24 Oral] Official repository of Point Transformer V3 (PTv3)
A paper list of some recent Transformer-based CV works.
[ICLR 2025] Code for Progressive Token Length Scaling in Transformer Encoders for Efficient Universal Segmentation
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites
Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]
Anytime Dense Prediction with Confidence Adaptivity (ICLR 2022)
Code for "Fast yet Safe: Early-Exiting with Risk Control" paper
Refign: Align and Refine for Adaptation of Semantic Segmentation to Adverse Conditions [WACV 2023]
[Arxiv 2024] Official code for Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time Adaptation
[CVPR 2024] Official implement of <Stronger, Fewer, & Superior: Harnessing Vision Foundation Models for Domain Generalized Semantic Segmentation>
Stable Neighbor Denoising for Source-free Domain Adaptive Segmentation (CVPR-2024)
[ACL 2024] Not All Experts are Equal: Efficient Expert Pruning and Skipping for Mixture-of-Experts Large Language Models
[ICLR 2025] MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts
Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024
Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts
The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".