首页 | 本学科首页   官方微博 | 高级检索  
   检索      


Analyzing Machine-Learned Representations: A Natural Language Case Study
Authors:Ishita Dasgupta  Demi Guo  Samuel J Gershman  Noah D Goodman
Institution:1. Departments of Psychology and Computer Science, Princeton University;2. Department of Computer Science, Harvard University;3. Department of Psychology and Center for Brain Science, Harvard University;4. Departments of Psychology and Computer Science, Stanford University
Abstract:As modern deep networks become more complex, and get closer to human-like capabilities in certain domains, the question arises as to how the representations and decision rules they learn compare to the ones in humans. In this work, we study representations of sentences in one such artificial system for natural language processing. We first present a diagnostic test dataset to examine the degree of abstract composable structure represented. Analyzing performance on these diagnostic tests indicates a lack of systematicity in representations and decision rules, and reveals a set of heuristic strategies. We then investigate the effect of training distribution on learning these heuristic strategies, and we study changes in these representations with various augmentations to the training set. Our results reveal parallels to the analogous representations in people. We find that these systems can learn abstract rules and generalize them to new contexts under certain circumstances—similar to human zero-shot reasoning. However, we also note some shortcomings in this generalization behavior—similar to human judgment errors like belief bias. Studying these parallels suggests new ways to understand psychological phenomena in humans as well as informs best strategies for building artificial intelligence with human-like language understanding.
Keywords:Representation learning  Natural language inference  Compositionality  Heuristic  Strategies  Sentence embeddings  Generalization  Test datasets
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号