-
21:24
(UTC +11:00) - https://orcid.org/0000-0002-9948-1400
Lists (1)
Sort Name ascending (A-Z)
Stars
Official Implementation of "Towards Open-Vocabulary Semantic Segmentation without Semantic Labels" (NeurIPS 2024)
Easily fine-tune, evaluate and deploy gpt-oss, Qwen3, DeepSeek-R1, or any open source LLM / VLM!
Circular visualization in Python (Circos Plot, Chord Diagram, Radar Chart)
[NeurIPS 2024] Code for Dual Prototype Evolving for Test-Time Generalization of Vision-Language Models
A curated list of awesome prompt/adapter learning methods for vision-language models like CLIP.
AIHawk aims to easy job hunt process by automating the job application process. Utilizing artificial intelligence, it enables users to apply for multiple jobs in a tailored way.
[CVPR 2024] Offical implemention of the paper "DePT: Decoupled Prompt Tuning"
The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".
[IJCV 2026] Offical implemention of the paper "A Closer Look at Conditional Prompt Tuning for Vision-Language Models"
Official implementation of PrevMatch: Revisiting and Maximizing Temporal Knowledge in Semi-supervised Semantic Segmentation
A curated list of awesome papers on dataset distillation and related applications.
Official code for ICLR 2024 paper, "A Hard-to-Beat Baseline for Training-free CLIP-based Adaptation"
[CVPR 2023] Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners
This is the official code release for our work, Denoising Vision Transformers.
VMamba: Visual State Space Models,code is based on mamba
[ICML 2024] Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
Code and Pretrained Models for ICLR 2023 Paper "Contrastive Audio-Visual Masked Autoencoder".
[NeurIPS'23] DropPos: Pre-Training Vision Transformers by Reconstructing Dropped Positions
[CVPR'23 & TPAMI'25] Hard Patches Mining for Masked Image Modeling & Bootstrap Masked Visual Modeling via Hard Patch Mining
VISSL is FAIR's library of extensible, modular and scalable components for SOTA Self-Supervised Learning with images.
Existing Literature about Machine Unlearning
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions


