skip to main content
10.1145/3292500.3330841acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

A Minimax Game for Instance based Selective Transfer Learning

Published:25 July 2019Publication History

ABSTRACT

Deep neural network based transfer learning has been widely used to leverage information from the domain with rich data to help domain with insufficient data. When the source data distribution is different from the target data, transferring knowledge between these domains may lead to negative transfer. To mitigate this problem, a typical way is to select useful source domain data for transferring. However, limited studies focus on selecting high-quality source data to help neural network based transfer learning. To bridge this gap, we propose a general Minimax Game based model for selective Transfer Learning (MGTL). More specifically, we build a selector, a discriminator and a TL module in the proposed method. The discriminator aims to maximize the differences between selected source data and target data, while the selector acts as an attacker to selected source data that are close to the target to minimize the differences. The TL module trains on the selected data and provides rewards to guide the selector. Those three modules play a minimax game to help select useful source data for transferring. Our method is also shown to speed up the training process of the learning task in the target domain than traditional TL methods. To the best of our knowledge, this is the first to build a minimax game based model for selective transfer learning. To examine the generality of our method, we evaluate it on two different tasks: item recommendation and text retrieval. Extensive experiments over both public and real-world datasets demonstrate that our model outperforms the competing methods by a large margin. Meanwhile, the quantitative evaluation shows our model can select data which are close to target data. Our model is also deployed in a real-world system and significant improvement over the baselines is observed.

References

  1. Andreas Argyriou, Theodoros Evgeniou, and Massimiliano Pontil. 2007. Multitask feature learning. In NIPS.Google ScholarGoogle Scholar
  2. John Blitzer, Mark Dredze, and Fernando Pereira. 2007. Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. ACL (2007).Google ScholarGoogle Scholar
  3. Zhangjie Cao, Mingsheng Long, Jianmin Wang, and Michael I. Jordan. 2017. Partial Transfer Learning with Selective Adversarial Networks. CoRR (2017).Google ScholarGoogle Scholar
  4. Minmin Chen, Kilian Q. Weinberger, and John C. Blitzer. 2011. Co-training for Domain Adaptation. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Heng-Tze Cheng, Levent Koc, Jeremiah Harmsen, Tal Shaked, Tushar Chandra, Hrishi Aradhye, Glen Anderson, Greg Corrado, Wei Chai, Mustafa Ispir, Rohan Anil, Zakaria Haque, Lichan Hong, Vihan Jain, Xiaobing Liu, and Hemal Shah. 2016. Wide & Deep Learning for Recommender Systems. CoRR abs/1606.07792 (2016).Google ScholarGoogle Scholar
  6. Paul Covington, Jay Adams, and Emre Sargin. 2016. Deep Neural Networks for YouTube Recommendations. In RecSys '16. 191--198. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Wenyuan Dai, Qiang Yang, Gui-Rong Xue, and Yong Yu. 2007. Boosting for Transfer Learning. In ICML. 193--200. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Hal Daume III. 2007. Frustratingly Easy Domain Adaptation. In ACL.Google ScholarGoogle Scholar
  9. Yang Fan, Fei Tian, Tao Qin, Jiang Bian, and Tie-Yan Liu. 2017. Learning What Data to Learn. CoRR (2017).Google ScholarGoogle Scholar
  10. Meng Fang, Yuan Li, and Trevor Cohn. 2017. Learning how to Active Learn: A Deep Reinforcement Learning Approach. In EMNLP.Google ScholarGoogle Scholar
  11. Jun Feng, Minlie Huang, Li Zhao, Yang Yang, and Xiaoyan Zhu. 2018. Reinforcement Learning for Relation Classification From Noisy Data. In AAAI.Google ScholarGoogle Scholar
  12. Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky. 2016. Domain-adversarial Training of Neural Networks. J. Mach. Learn. Res. 17, 1 (Jan. 2016), 2096--2030. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial nets. In NIPS. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Huifeng Guo, Ruiming Tang, Yunming Ye, Zhenguo Li, and Xiuqiang He. 2017. DeepFM: A Factorization-Machine based Neural Network for CTR Prediction. CoRR abs/1703.04247 (2017).Google ScholarGoogle Scholar
  15. Xiangnan He and Tat-Seng Chua. 2017. Neural Factorization Machines for Sparse Predictive Analytics. In Proceedings of SIGIR. 355--364. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Xiangnan He, Zhankui He, Xiaoyu Du, and Tat-Seng Chua. 2018. Adversarial personalized ranking for recommendation. In The 41st International ACM SIGIR Conference on Research & Development in Information Retrieval. 355--364. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Jiayuan Huang, Alexander J. Smola, Arthur Gretton, Karsten M. Borgwardt, and Bernhard. Scholkopf. 2006. Correcting Sample Selection Bias by Unlabeled Data. In NIPS. (2006). Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Ferenc Huszar. 2015. How (not) to Train your Generative Model: Scheduled Sampling, Likelihood, Adversary? arXiv:1511.05101. (2015).Google ScholarGoogle Scholar
  19. Tushar Khot, Ashish Sabharwal, and Peter Clark. 2018. SciTail: A Textual Entailment Dataset from Science Question Answering. In AAAI.Google ScholarGoogle Scholar
  20. Pengfei Liu, Xipeng Qiu, and Xuanjing Huang. 2017. Adversarial Multi-task Learning for Text Classification. In Proceedings of ACL. (2017).Google ScholarGoogle ScholarCross RefCross Ref
  21. Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin A. Riedmiller, Andreas Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. 2015. Human-level control through deep reinforcement learning. Nature (2015).Google ScholarGoogle Scholar
  22. Lili Mou, Rui Men, Ge Li, Yan Xu, Lu Zhang, Rui Yan, and Zhi Jin. 2016. Natural Language Inference by Tree-Based Convolution and Heuristic Matching. In ACL.Google ScholarGoogle Scholar
  23. Lili Mou, Zhao Meng, Rui Yan, Ge Li, Yan Xu, Lu Zhang, and Zhi Jin. 2016. How Transferable are Neural Networks in NLP Applications?. In EMNLP.Google ScholarGoogle Scholar
  24. Sinno Jialin Pan and Qiang Yang. 2010. A survey on transfer learning. IEEE Transactions on knowledge and data engineering (2010), 1345--1359. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Ankur P. Parikh, Oscar Täckström, Dipanjan Das, and Jakob Uszkoreit. 2016. A Decomposable Attention Model for Natural Language Inference. In EMNLP.Google ScholarGoogle Scholar
  26. Yash Patel, Kashyap Chitta, and Bhavan Jasani. {n. d.}. Learning Sampling Policies for Domain Adaptation. CoRR, abs/1805.07641, 2018. ({n. d.}).Google ScholarGoogle Scholar
  27. Chen Qu, Feng Ji, Minghui Qiu, Liu Yang, Zhiyu Min, Haiqing Chen, Jun Huang, and W. Bruce Croft. 2019. Learning to Selectively Transfer: Reinforced Transfer Learning for Deep Text Matching. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining (WSDM '19). Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Michael T. Rosenstein, Zvika Marx, Leslie Pack Kael-bling, and Thomas G.Dietterich. 2005. To Transfer or Not To Transfer. NIPS Workshop on Inductive Transfer (2005).Google ScholarGoogle Scholar
  29. Sebastian Ruder and Barbara Plank. 2017. Learning to select data for transfer learning with Bayesian Optimization. In EMNLP. (2017).Google ScholarGoogle ScholarCross RefCross Ref
  30. Gavin A. Rummery and Mahesan Niranjan. 1994. OnLine Q-Learning Using Connectionist Systems. Technical Report. University of Cambridge.Google ScholarGoogle Scholar
  31. Tobias Schnabel and Hinrich SchuÌtze. 2014. FLORS: Fast and Simple Domain Adaptation for Part-of-Speech Tagging. TACL, 2:15-26. (2014).Google ScholarGoogle Scholar
  32. Jian Shen, Yanru Qu, Weinan Zhang, and Yong Yu. 2018. Wasserstein Distance Guided Representation Learning for Domain Adaptation. In AAAI. AAAI Press.Google ScholarGoogle Scholar
  33. Richard S. Sutton and Andrew G. Barto. 1998. Reinforcement Learning - An Introduction. MIT Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  34. Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. 2017. Adversarial Discriminative Domain Adaptation. CoRR abs/1702.05464 (2017).Google ScholarGoogle Scholar
  35. ChangWang and Sridhar Mahadevan. 2008. Manifold alignment using procrustes analysis. In ICML.Google ScholarGoogle Scholar
  36. Jun Wang, Lantao Yu, Weinan Zhang, Yu Gong, Yinghui Xu, Benyou Wang, Peng Zhang, and Dell Zhang. 2017. Irgan: A minimax game for unifying generative and discriminative information retrieval models. In Proceedings of SIGIR. ACM, 515--524. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. RuoxiWang, Bin Fu, Gang Fu, and MingliangWang. 2017. Deep & Cross Network for Ad Click Predictions. In Proceedings of the ADKDD'17. 12:1--12:7. Google ScholarGoogle ScholarDigital LibraryDigital Library
  38. TianyangWang, Jun Huan, and Michelle Zhu. 2018. Instance-based Deep Transfer Learning. In WACV.Google ScholarGoogle Scholar
  39. Junfeng Wen, Chun-Nam Yu, and Russell Greiner. 2014. Robust Learning Under Uncertain Test Distributions: Relating Covariate Shift to Model Misspecification (ICML'14). JMLR.org, II--631--II--639. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Adina Williams, Nikita Nangia, and Samuel Bowman. 2018. A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference. In NAACL.Google ScholarGoogle Scholar
  41. Ronald J Williams. 1992. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning (1992) (1992). Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Jiawei Wu, Lei Li, and William Yang Wang. 2018. Reinforced Co-Training. In NAACL.Google ScholarGoogle Scholar
  43. Zhilin Yang, Ruslan Salakhutdinov, and William W. Cohen. 2017. Transfer Learning for Sequence Tagging with Hierarchical Recurrent Networks. In ICLR (2017).Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. Wenpeng Yin, Hinrich Schütze, Bing Xiang, and Bowen Zhou. 2016. ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs. TACL (2016).Google ScholarGoogle Scholar
  45. Lantao Yu, Weinan Zhang, Jun Wang, and Yong Yu. 2017. SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient. In AAAI. (2017). Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Fuzhen Zhuang, Lang Huang, Jia He, Jixin Ma, and Qing He. 2017. Transfer Learning with Manifold Regularized Convolutional Neural Network, Gang Li, Yong Ge, Zili Zhang, Zhi Jin, and Michael Blumenstein (Eds.). Springer International Publishing, Cham, 483--494.Google ScholarGoogle Scholar

Index Terms

  1. A Minimax Game for Instance based Selective Transfer Learning

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        KDD '19: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
        July 2019
        3305 pages
        ISBN:9781450362016
        DOI:10.1145/3292500

        Copyright © 2019 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 25 July 2019

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        KDD '19 Paper Acceptance Rate110of1,200submissions,9%Overall Acceptance Rate1,133of8,635submissions,13%

        Upcoming Conference

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader