Author Biographies |
|
xi | |
|
|
xiii | |
|
|
xvii | |
Preface |
|
xix | |
|
Part I Human-robot Interaction Control |
|
|
1 | (96) |
|
|
3 | (14) |
|
1.1 Human-Robot Interaction Control |
|
|
3 | (3) |
|
1.2 Reinforcement Learning for Control |
|
|
6 | (1) |
|
1.3 Structure of the Book |
|
|
7 | (3) |
|
|
10 | (7) |
|
2 Environment Model of Human-Robot Interaction |
|
|
17 | (16) |
|
2.1 Impedance and Admittance |
|
|
17 | (4) |
|
2.2 Impedance Model for Human-Robot Interaction |
|
|
21 | (3) |
|
2.3 Identification of Human-Robot Interaction Model |
|
|
24 | (6) |
|
|
30 | (1) |
|
|
30 | (3) |
|
3 Model Based Human-Robot Interaction Control |
|
|
33 | (12) |
|
3.1 Task Space Impedance/Admittance Control |
|
|
33 | (3) |
|
3.2 Joint Space Impedance Control |
|
|
36 | (1) |
|
3.3 Accuracy and Robustness |
|
|
37 | (2) |
|
|
39 | (3) |
|
|
42 | (2) |
|
|
44 | (1) |
|
4 Model Free Human-Robot Interaction Control |
|
|
45 | (28) |
|
4.1 Task-Space Control Using Joint-Space Dynamics |
|
|
45 | (7) |
|
4.2 Task-Space Control Using Task-Space Dynamics |
|
|
52 | (1) |
|
|
53 | (1) |
|
|
54 | (1) |
|
|
55 | (13) |
|
|
68 | (3) |
|
|
71 | (2) |
|
5 Human-in-the-loop Control Using Euler Angles |
|
|
73 | (24) |
|
|
73 | (1) |
|
|
74 | (5) |
|
|
79 | (4) |
|
|
83 | (9) |
|
|
92 | (2) |
|
|
94 | (3) |
|
Part II Reinforcement Learning for Robot Interaction Control |
|
|
97 | (138) |
|
6 Reinforcement Learning for Robot Position/Force Control |
|
|
99 | (20) |
|
|
99 | (1) |
|
6.2 Position/Force Control Using an Impedance Model |
|
|
100 | (3) |
|
6.3 Reinforcement Learning Based Position/Force Control |
|
|
103 | (7) |
|
6.4 Simulations and Experiments |
|
|
110 | (7) |
|
|
117 | (1) |
|
|
117 | (2) |
|
7 Continuous-Time Reinforcement Learning for Force Control |
|
|
119 | (20) |
|
|
119 | (1) |
|
7.2 K-means Clustering for Reinforcement Learning |
|
|
120 | (4) |
|
7.3 Position/Force Control Using Reinforcement Learning |
|
|
124 | (6) |
|
|
130 | (6) |
|
|
136 | (1) |
|
|
136 | (3) |
|
8 Robot Control in Worst-Case Uncertainty Using Reinforcement Learning |
|
|
139 | (34) |
|
|
139 | (2) |
|
8.2 Robust Control Using Discrete-Time Reinforcement Learning |
|
|
141 | (3) |
|
8.3 Double Q-Learning with k-Nearest Neighbors |
|
|
144 | (6) |
|
8.4 Robust Control Using Continuous-Time Reinforcement Learning |
|
|
150 | (4) |
|
8.5 Simulations and Experiments: Discrete-Time Case |
|
|
154 | (7) |
|
8.6 Simulations and Experiments: Continuous-Time Case |
|
|
161 | (9) |
|
|
170 | (1) |
|
|
170 | (3) |
|
9 Redundant Robots Control Using Multi-Agent Reinforcement Learning |
|
|
173 | (20) |
|
|
173 | (2) |
|
9.2 Redundant Robot Control |
|
|
175 | (4) |
|
9.3 Multi-Agent Reinforcement Learning for Redundant Robot Control |
|
|
179 | (4) |
|
9.4 Simulations and experiments |
|
|
183 | (4) |
|
|
187 | (2) |
|
|
189 | (4) |
|
10 Robot H2 Neural Control Using Reinforcement Learning |
|
|
193 | (40) |
|
|
193 | (1) |
|
10.2 H2 Neural Control Using Discrete-Time Reinforcement Learning |
|
|
194 | (13) |
|
10.3 H2 Neural Control in Continuous Time |
|
|
207 | (12) |
|
|
219 | (10) |
|
|
229 | (1) |
|
|
229 | (4) |
|
|
233 | (2) |
|
A Robot Kinematics and Dynamics |
|
|
235 | (12) |
|
|
235 | (2) |
|
|
237 | (3) |
|
|
240 | (6) |
|
|
246 | (1) |
|
B Reinforcement Learning for Control |
|
|
247 | (12) |
|
B.1 Markov decision processes |
|
|
247 | (1) |
|
|
248 | (2) |
|
|
250 | (1) |
|
|
251 | (7) |
|
|
258 | (1) |
Index |
|
259 | |