Follow
Taiqiang Wu
Taiqiang Wu
University of Hong Kong | Tsinghua University
Verified email at connect.hku.hk - Homepage
Title
Cited by
Cited by
Year
Riformer: Keep your vision backbone effective but removing token mixer
J Wang, S Zhang, Y Liu, T Wu, Y Yang, X Liu, K Chen, P Luo, D Lin
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023
332023
Syngen: A syntactic plug-and-play module for generative aspect-based sentiment analysis
C Yu, T Wu, J Li, X Bai, Y Yang
ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023
202023
Tencentpretrain: A scalable and flexible toolkit for pre-training models of different modalities
Z Zhao, Y Li, C Hou, J Zhao, R Tian, W Liu, Y Chen, N Sun, H Liu, W Mao, ...
ACL 2023 system paper, 2022
202022
Mixture-of-Subspaces in Low-Rank Adaptation
T Wu, J Wang, Z Zhao, N Wong
EMNLP 2024, 2024
14*2024
Rethinking kullback-leibler divergence in knowledge distillation for large language models
T Wu, C Tao, J Wang, R Yang, Z Zhao, N Wong
COLING 2025, 2024
142024
Modeling fine-grained information via knowledge-aware hierarchical graph for zero-shot entity retrieval
T Wu, X Bai, W Guo, W Liu, S Li, Y Yang
Proceedings of the Sixteenth ACM International Conference on Web Search and …, 2023
132023
Unchosen Experts Can Contribute Too: Unleashing MoE Models' Power by Self-Contrast
C Shi, C Yang, X Zhu, J Wang, T Wu, S Li, D Cai, Y Yang, Y Meng
NIPS 2024, 2024
92024
Edge-free but structure-aware: Prototype-guided knowledge distillation from gnns to mlps
T Wu, Z Zhao, J Wang, X Bai, L Wang, N Wong, Y Yang
COLING 2025, 2023
82023
Weight-inherited distillation for task-agnostic bert compression
T Wu, C Hou, S Lao, J Li, N Wong, Z Zhao, Y Yang
NAACL 2024, 2023
72023
A survey on the honesty of large language models
S Li, C Yang, T Wu, C Shi, Y Zhang, X Zhu, Z Cheng, D Cai, M Yu, L Liu, ...
arXiv preprint arXiv:2409.18786, 2024
52024
Adapting LLaMA Decoder to Vision Transformer
J Wang, W Shao, M Chen, C Wu, Y Liu, T Wu, K Zhang, S Zhang, K Chen, ...
arXiv preprint arXiv:2404.06773, 2024
32024
Prompt-based Model for Acronym Disambiguation via Negative Sampling
T Wu, X Bai, Y Yang
AAAI 2022 workshop SDU@2022, 2022
32022
Llm-neo: Parameter efficient knowledge distillation for large language models
R Yang, T Wu, J Wang, P Hu, N Wong, Y Yang
arXiv preprint arXiv:2411.06839, 2024
12024
Autoregressive Models in Vision: A Survey
J Xiong, G Liu, L Huang, C Wu, T Wu, Y Mu, Y Yao, H Shen, Z Wan, ...
arXiv preprint arXiv:2411.05902, 2024
12024
Multi-stage Distillation Framework for Cross-Lingual Semantic Similarity Matching
K Ding, W Liu, Y Fang, Z Zhao, Q Ju, X Yang, R Tian, Z Tao, H Liu, H Guo, ...
NAACL 2022 Findings, 2022
12022
LiT: Delving into a Simplified Linear Diffusion Transformer for Image Generation
J Wang, N Kang, L Yao, M Chen, C Wu, S Zhang, S Xue, Y Liu, T Wu, ...
arXiv preprint arXiv:2501.12976, 2025
2025
A Unified View for Attention and MoE
T Wu, N Wong
Report, 2025
2025
MCUBERT: Memory-Efficient BERT Inference on Commodity Microcontrollers
Z Yang, R Chen, T Wu, N Wong, Y Liang, R Wang, R Huang, M Li
ICCAD 2024, 2024
2024
LoCa: Logit Calibration for Knowledge Distillation
R Yang, T Wu, Y Yang
ECAI 2024, 2024
2024
Recouple Event Field via Probabilistic Bias for Event Extraction
X Bai, T Wu, H Guo, Z Zhao, X Yang, J Li, W Liu, Q Ju, W Guo, Y Yang
ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023
2023
The system can't perform the operation now. Try again later.
Articles 1–20