語系:
繁體中文
English
說明(常見問題)
回圖書館首頁
手機版館藏查詢
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Training Models to Ignore Dataset Bias.
~
Clark, Christopher.
FindBook
Google Book
Amazon
博客來
Training Models to Ignore Dataset Bias.
紀錄類型:
書目-電子資源 : Monograph/item
正題名/作者:
Training Models to Ignore Dataset Bias./
作者:
Clark, Christopher.
出版者:
Ann Arbor : ProQuest Dissertations & Theses, : 2020,
面頁冊數:
106 p.
附註:
Source: Dissertations Abstracts International, Volume: 82-05, Section: B.
Contained By:
Dissertations Abstracts International82-05B.
標題:
Artificial intelligence. -
電子資源:
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=28092295
ISBN:
9798684669552
Training Models to Ignore Dataset Bias.
Clark, Christopher.
Training Models to Ignore Dataset Bias.
- Ann Arbor : ProQuest Dissertations & Theses, 2020 - 106 p.
Source: Dissertations Abstracts International, Volume: 82-05, Section: B.
Thesis (Ph.D.)--University of Washington, 2020.
This item must not be sold to any third party vendors.
Modern machine learning algorithms have been able to achieve impressive results on complex tasks such as language comprehension or image understanding. However, recent work has cautioned that this success is often partially due to exploiting incidental correlations that were introduced during dataset creation, and are not fundamental to the target task. For example, sentence entailment datasets can have spurious word-class correlations if nearly all contradiction sentences contain the word ``not'', and image recognition datasets can have tell-tale object-background correlations if dogs are always indoors. Models that exploit these incidental correlation, which we call dataset bias, can be brittle and perform poorly on out-of-domain examples. In this thesis, we present several methods of solving this issue by preventing models from using dataset bias.A key challenge for this task is determining which predictive patterns in the training data are bias. This thesis proposes several solutions, ranging from methods that exploit domain expertise when such knowledge is available, to more broadly applicable domain-general solutions. Solving this task also requires preventing complex neural models from exploiting these biased patterns, even though they are often easy to learn and effective on the training data. We present ensembling and data augmentation based methods to handle this difficulty. In all cases, we evaluate our models by showing improved performance on out-of-domain datasets that were built to penalize biased models.Our first focus is on question answering, motivated by the observation biases can lead to poor performance when applying a model to multiple paragraphs. To solve this task, we propose a modified training scheme that exposes the model to additional paragraphs that do not answer the question. We then consider the case where expert knowledge of the bias can be used to construct a \extit{bias-only} model that captures biased methods. In this case, we can build an unbiased model by ensembling it with the bias-only model in order to disincentive it from learning bias. Finally, we generalize this approach by proposing a method to automatically construct the bias-only model when no such expert knowledge is available. Overall, this thesis shows that it is possible to train unbiased models on biased datasets, and proposes some fundamental answers to question about how bias can be detected and avoided.
ISBN: 9798684669552Subjects--Topical Terms:
516317
Artificial intelligence.
Subjects--Index Terms:
Computer vision
Training Models to Ignore Dataset Bias.
LDR
:03553nmm a2200361 4500
001
2277371
005
20210521101709.5
008
220723s2020 ||||||||||||||||| ||eng d
020
$a
9798684669552
035
$a
(MiAaPQ)AAI28092295
035
$a
AAI28092295
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Clark, Christopher.
$3
3224412
245
1 0
$a
Training Models to Ignore Dataset Bias.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2020
300
$a
106 p.
500
$a
Source: Dissertations Abstracts International, Volume: 82-05, Section: B.
500
$a
Advisor: Zettlemoyer, Luke.
502
$a
Thesis (Ph.D.)--University of Washington, 2020.
506
$a
This item must not be sold to any third party vendors.
520
$a
Modern machine learning algorithms have been able to achieve impressive results on complex tasks such as language comprehension or image understanding. However, recent work has cautioned that this success is often partially due to exploiting incidental correlations that were introduced during dataset creation, and are not fundamental to the target task. For example, sentence entailment datasets can have spurious word-class correlations if nearly all contradiction sentences contain the word ``not'', and image recognition datasets can have tell-tale object-background correlations if dogs are always indoors. Models that exploit these incidental correlation, which we call dataset bias, can be brittle and perform poorly on out-of-domain examples. In this thesis, we present several methods of solving this issue by preventing models from using dataset bias.A key challenge for this task is determining which predictive patterns in the training data are bias. This thesis proposes several solutions, ranging from methods that exploit domain expertise when such knowledge is available, to more broadly applicable domain-general solutions. Solving this task also requires preventing complex neural models from exploiting these biased patterns, even though they are often easy to learn and effective on the training data. We present ensembling and data augmentation based methods to handle this difficulty. In all cases, we evaluate our models by showing improved performance on out-of-domain datasets that were built to penalize biased models.Our first focus is on question answering, motivated by the observation biases can lead to poor performance when applying a model to multiple paragraphs. To solve this task, we propose a modified training scheme that exposes the model to additional paragraphs that do not answer the question. We then consider the case where expert knowledge of the bias can be used to construct a \extit{bias-only} model that captures biased methods. In this case, we can build an unbiased model by ensembling it with the bias-only model in order to disincentive it from learning bias. Finally, we generalize this approach by proposing a method to automatically construct the bias-only model when no such expert knowledge is available. Overall, this thesis shows that it is possible to train unbiased models on biased datasets, and proposes some fundamental answers to question about how bias can be detected and avoided.
590
$a
School code: 0250.
650
4
$a
Artificial intelligence.
$3
516317
650
4
$a
Information science.
$3
554358
650
4
$a
Information technology.
$3
532993
653
$a
Computer vision
653
$a
Dataset bias
653
$a
Machine learning
653
$a
Natural language processing
690
$a
0800
690
$a
0723
690
$a
0489
710
2
$a
University of Washington.
$b
Computer Science and Engineering.
$3
2097608
773
0
$t
Dissertations Abstracts International
$g
82-05B.
790
$a
0250
791
$a
Ph.D.
792
$a
2020
793
$a
English
856
4 0
$u
https://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=28092295
筆 0 讀者評論
館藏地:
全部
電子資源
出版年:
卷號:
館藏
1 筆 • 頁數 1 •
1
條碼號
典藏地名稱
館藏流通類別
資料類型
索書號
使用類型
借閱狀態
預約狀態
備註欄
附件
W9429105
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
在架
0
1 筆 • 頁數 1 •
1
多媒體
評論
新增評論
分享你的心得
Export
取書館
處理中
...
變更密碼
登入