Web-Books
im Austria-Forum
Austria-Forum
Web-Books
Informatik
Joint Austrian Computer Vision and Robotics Workshop 2020
Seite - 45 -
  • Benutzer
  • Version
    • Vollversion
    • Textversion
  • Sprache
    • Deutsch
    • English - Englisch

Seite - 45 - in Joint Austrian Computer Vision and Robotics Workshop 2020

Bild der Seite - 45 -

Bild der Seite - 45 - in Joint Austrian Computer Vision and Robotics Workshop 2020

Text der Seite - 45 -

Webcam Hand Tracking Xtion DemonstrationsDemonstrationsDemonstrations Robot State Robot Control Hand Poses End-Effector Poses RGB Images RGB-D Images RGB-D Images Desired End-Effector Poses End-Effector Poses (120, 160, 3) 7x7 Conv 64 /2 Concatenation Current Position 7x7 Conv 16 /2 (120, 160, 1) Final Position Auxiliary Predictions Recent Positions (5 ⨉ 3) (1 ⨉ 3) Figure 2. Overview of the system. The dashed lines show the procedure to collect demonstrations for training. The continuous lines show the information flow duringpolicy execution. the changing lighting conditions in the test environ- ment. Additionally, we added dropout of the recent end-effector positions to avoid the robot following the same trajectory during most executions and not taking theobjectposition intoaccount. Theoverall loss isdefinedas L(θ)=λl1Ll1+λl2Ll2+λcLc+λsLs+λauxΣaL(a)aux. (2) The first two terms are the l1and l2 losses.Lc is the cosine lossandLauxare the l2 lossesof theauxiliary predictions. Compared to [27],weadded the loss Ls=exp(−||πθ(ot)||2) (3) that penalizes very slow speeds. The weights were chosen as λl1=1.0, λl2=0.01, λc=0.05, λs=0.1, andλaux=0.01. 4. Experiments Thissectionpresents theexperimental results. We first describe the setup and procedure for collecting demonstration data. We analyze the performance of ourmethodwith respect to thenetworkdesign. 4.1.ExperimentalSetup All experiments are conducted with a KUKA LWR IV+ [3] robotic arm using the provided con- trol unit. The arm has 7 degrees of freedom and is controlled with position commands for the joints. The arm is mounted on the ceiling with a small ta- ble standing underneath it on which the target object (box) rests. The goal region is marked with tape. An ASUS Xtion RGB-D camera is mounted to the ceil- ing to capture the scene from above. For hand track- ing,aseparatewebcamisusedandfacestheoperator. The algorithms for the hand tracking and the task execution run on a remote PC connected to the KUKA control unit via Ethernet. The communica- tion between the remote PC and the control unit is enabled through thekuka-lwr-rospackage1 using the fast research interface (FRI) [23]. For data collection, the teleoperator directly faces the robot and the webcam. For each demonstration, the box is positioned randomly on the table. The teleoperator moves the box to the goal position us- ing our control scheme. We collected 98 demon- strations with an average length of 42.8s with a rate of 10Hz for our evaluation. That is signifi- cantlyabovetheaveragedemonstration timeper task of [27], which is between 3.7s and 11.6s and neces- sitates our changes to the architecture to deal with these imperfectdemonstrations. 4.2. Results For the evaluation, the workspace of the robot on the table is divided into a grid of 9 different posi- tions with 20cm intervals. Per position, the learned policy is executed for4different rotationsof thebox (−45◦,0◦,45◦,90◦). We measure both if the box is pushed towards the goal (started push) as well as if at least part of it is pushed into the goal (success). If the robot starts to push the box, but loses it, we restart the policy manually and keep the box in the same position when the end-effector stops or leaves the workspace. This could be automated with a sim- pleheuristic. If the taskcanbeachieved inaconsec- utive trial, westill count it as a success. As shown in Table 1, our learned policy started to push the box in the right direction in 86.1% of the cases and reached the goal in 58.3% of the overall attempts. A reason for most failure cases is the grid natureofourworkspaceseparation,whichinherently tests the roboton theedgesof itsworkspacewhere it ismuchmoredifficult toperformthe task. We conducted an ablation study to evaluate our changes to the original architecture of [27]. We re- 1https://github.com/epfl-lasa/kuka-lwr-ros 45
zurück zum  Buch Joint Austrian Computer Vision and Robotics Workshop 2020"
Joint Austrian Computer Vision and Robotics Workshop 2020
Titel
Joint Austrian Computer Vision and Robotics Workshop 2020
Herausgeber
Graz University of Technology
Ort
Graz
Datum
2020
Sprache
englisch
Lizenz
CC BY 4.0
ISBN
978-3-85125-752-6
Abmessungen
21.0 x 29.7 cm
Seiten
188
Kategorien
Informatik
Technik
Web-Books
Bibliothek
Datenschutz
Impressum
Austria-Forum
Austria-Forum
Web-Books
Joint Austrian Computer Vision and Robotics Workshop 2020