Language:
English
繁體中文
Help
回圖書館首頁
手機版館藏查詢
Login
Back
Switch To:
Labeled
|
MARC Mode
|
ISBD
Using Reinforcement Learning to Pers...
~
Humphrey, Kyle.
Linked to FindBook
Google Book
Amazon
博客來
Using Reinforcement Learning to Personalize Dosing Strategies in a Simulated Cancer Trial With High Dimensional Data.
Record Type:
Electronic resources : Monograph/item
Title/Author:
Using Reinforcement Learning to Personalize Dosing Strategies in a Simulated Cancer Trial With High Dimensional Data./
Author:
Humphrey, Kyle.
Published:
Ann Arbor : ProQuest Dissertations & Theses, : 2017,
Description:
49 p.
Notes:
Source: Masters Abstracts International, Volume: 56-04.
Contained By:
Masters Abstracts International56-04(E).
Subject:
Biostatistics. -
Online resource:
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10282783
ISBN:
9781369827217
Using Reinforcement Learning to Personalize Dosing Strategies in a Simulated Cancer Trial With High Dimensional Data.
Humphrey, Kyle.
Using Reinforcement Learning to Personalize Dosing Strategies in a Simulated Cancer Trial With High Dimensional Data.
- Ann Arbor : ProQuest Dissertations & Theses, 2017 - 49 p.
Source: Masters Abstracts International, Volume: 56-04.
Thesis (M.S.)--The University of Arizona, 2017.
In a simulation of an advanced generic cancer trial, I use Q-learning, a reinforcement learning algorithm, to develop dynamic treatment regimes for a continuous treatment, the dose of a single drug. Selected dynamic treatment regimes are tailored to time-varying patient characteristics and to patient subgroups with differential treatment effects. This approach allows estimation of optimal dynamic treatment regimes without a model of the disease process or a priori hypotheses about subgroup membership. Using observed patient characteristics and outcomes from the simulated trial, I estimate Q-functions based on 1) a single regression tree grown by the Classification And Regression Trees (CART) method, 2) random forests, and 3) a slightly modified version of Multivariate Adaptive Regression Splines (MARS). I then compare the survival times of an independent group of simulated patients under treatment regimes estimated using Q-learning with each of the three methods, 10 constant dose regimes, and the best possible treatment regime chosen using a brute force search over all possible treatment regimes with complete knowledge of disease processes and their effects on survival. I also make these comparisons in scenarios with and without spurious high dimensional covariates and with and without patient subgroups with differential treatment effects. Treatment regimes estimated using Q-learning with MARS and random forests greatly increased survival times when compared to the constant dose regimes, but were still considerably lower than the best possible dose regime. Q-learning with a single regression tree did not outperform the constant dose regimes. These results hold across high dimensional and subgroup scenarios. While the MARS method employed produces much more interpretable models than random forests, and therefore has more promise for patient subgroup identification, I show that it is also more sensitive to variations in training data.
ISBN: 9781369827217Subjects--Topical Terms:
1002712
Biostatistics.
Using Reinforcement Learning to Personalize Dosing Strategies in a Simulated Cancer Trial With High Dimensional Data.
LDR
:02901nmm a2200289 4500
001
2164111
005
20181030085011.5
008
190424s2017 ||||||||||||||||| ||eng d
020
$a
9781369827217
035
$a
(MiAaPQ)AAI10282783
035
$a
(MiAaPQ)arizona:15579
035
$a
AAI10282783
040
$a
MiAaPQ
$c
MiAaPQ
100
1
$a
Humphrey, Kyle.
$3
3352146
245
1 0
$a
Using Reinforcement Learning to Personalize Dosing Strategies in a Simulated Cancer Trial With High Dimensional Data.
260
1
$a
Ann Arbor :
$b
ProQuest Dissertations & Theses,
$c
2017
300
$a
49 p.
500
$a
Source: Masters Abstracts International, Volume: 56-04.
500
$a
Adviser: Jin Zhou.
502
$a
Thesis (M.S.)--The University of Arizona, 2017.
520
$a
In a simulation of an advanced generic cancer trial, I use Q-learning, a reinforcement learning algorithm, to develop dynamic treatment regimes for a continuous treatment, the dose of a single drug. Selected dynamic treatment regimes are tailored to time-varying patient characteristics and to patient subgroups with differential treatment effects. This approach allows estimation of optimal dynamic treatment regimes without a model of the disease process or a priori hypotheses about subgroup membership. Using observed patient characteristics and outcomes from the simulated trial, I estimate Q-functions based on 1) a single regression tree grown by the Classification And Regression Trees (CART) method, 2) random forests, and 3) a slightly modified version of Multivariate Adaptive Regression Splines (MARS). I then compare the survival times of an independent group of simulated patients under treatment regimes estimated using Q-learning with each of the three methods, 10 constant dose regimes, and the best possible treatment regime chosen using a brute force search over all possible treatment regimes with complete knowledge of disease processes and their effects on survival. I also make these comparisons in scenarios with and without spurious high dimensional covariates and with and without patient subgroups with differential treatment effects. Treatment regimes estimated using Q-learning with MARS and random forests greatly increased survival times when compared to the constant dose regimes, but were still considerably lower than the best possible dose regime. Q-learning with a single regression tree did not outperform the constant dose regimes. These results hold across high dimensional and subgroup scenarios. While the MARS method employed produces much more interpretable models than random forests, and therefore has more promise for patient subgroup identification, I show that it is also more sensitive to variations in training data.
590
$a
School code: 0009.
650
4
$a
Biostatistics.
$3
1002712
690
$a
0308
710
2
$a
The University of Arizona.
$b
Biostatistics.
$3
3352147
773
0
$t
Masters Abstracts International
$g
56-04(E).
790
$a
0009
791
$a
M.S.
792
$a
2017
793
$a
English
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10282783
based on 0 review(s)
Location:
ALL
電子資源
Year:
Volume Number:
Items
1 records • Pages 1 •
1
Inventory Number
Location Name
Item Class
Material type
Call number
Usage Class
Loan Status
No. of reservations
Opac note
Attachments
W9363658
電子資源
11.線上閱覽_V
電子書
EB
一般使用(Normal)
On shelf
0
1 records • Pages 1 •
1
Multimedia
Reviews
Add a review
and share your thoughts with other readers
Export
pickup library
Processing
...
Change password
Login