Jianguo Li


2023

pdf bib
ZeroAE: Pre-trained Language Model based Autoencoder for Transductive Zero-shot Text Classification
Kaihao Guo | Hang Yu | Cong Liao | Jianguo Li | Haipeng Zhang
Findings of the Association for Computational Linguistics: ACL 2023

Many text classification tasks require handling unseen domains with plenty of unlabeled data, thus giving rise to the self-adaption or the so-called transductive zero-shot learning (TZSL) problem. However, current methods based solely on encoders or decoders overlook the possibility that these two modules may promote each other. As a first effort to bridge this gap, we propose an autoencoder named ZeroAE. Specifically, the text is encoded with two separate BERT-based encoders into two disentangled spaces, i.e., label-relevant (for classification) and label-irrelevant respectively. The two latent spaces are then decoded by prompting GPT-2 to recover the text as well as to further generate text with labels in the unseen domains to train the encoder in turn. To better exploit the unlabeled data, a novel indirect uncertainty-aware sampling (IUAS) approach is proposed to train ZeroAE. Extensive experiments show that ZeroAE largely surpasses the SOTA methods by 15.93% and 8.70% on average respectively in the label-partially-unseen and label-fully-unseen scenario. Notably, the label-fully-unseen ZeroAE even possesses superior performance to the label-partially-unseen SOTA methods.

2008

pdf bib
Which Are the Best Features for Automatic Verb Classification
Jianguo Li | Chris Brew
Proceedings of ACL-08: HLT

2006

pdf bib
Parsing and Subcategorization Data
Jianguo Li | Chris Brew
Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions

pdf bib
Parsing and Subcategorization Data
Jianguo Li
Proceedings of the COLING/ACL 2006 Student Research Workshop

2005

pdf bib
Robust Extraction of Subcategorization Data from Spoken Language
Jianguo Li | Chris Brew | Eric Fosler-Lussier
Proceedings of the Ninth International Workshop on Parsing Technology