Paper: Adapting Text instead of the Model: An Open Domain Approach

ACL ID W11-0327
Title Adapting Text instead of the Model: An Open Domain Approach
Venue International Conference on Computational Natural Language Learning
Session Main Conference
Year 2011
Authors

Natural language systems trained on labeled data from one domain do not perform well on other domains. Most adaptation algorithms proposed in the literature train a new model for the new domain using unlabeled data. How- ever, it is time consuming to retrain big mod- els or pipeline systems. Moreover, the domain of a new target sentence may not be known, and one may not have significant amount of unlabeled data for every new domain. To pursue the goal of an Open Domain NLP (train once, test anywhere), we propose ADUT (ADaptation Using label-preserving Transfor- mation), an approach that avoids the need for retraining and does not require knowledge of the new domain, or any data from it. Our ap- proach applies simple label-preserving trans- formations to the target text so that the trans- f...