Difference between revisions of "CISC849 S2018"
(→Schedule) |
(→Schedule) |
||
(37 intermediate revisions by the same user not shown) | |||
Line 51: | Line 51: | ||
* [https://arxiv.org/pdf/1509.06825.pdf Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours], Pinto and Gupta, ICRA 2016 | * [https://arxiv.org/pdf/1509.06825.pdf Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours], Pinto and Gupta, ICRA 2016 | ||
* [http://www.cs.ucf.edu/~rrahmati/manipulationLearning.html Learning real manipulation tasks from virtual demonstrations using LSTM], Rahmatizadeh ''et al.'', AAAI 2018 | * [http://www.cs.ucf.edu/~rrahmati/manipulationLearning.html Learning real manipulation tasks from virtual demonstrations using LSTM], Rahmatizadeh ''et al.'', AAAI 2018 | ||
− | |||
* [https://arxiv.org/abs/1704.05588 Learning to Fly by Crashing], Gandhi, Pinto, and Gupta, IROS 2017 | * [https://arxiv.org/abs/1704.05588 Learning to Fly by Crashing], Gandhi, Pinto, and Gupta, IROS 2017 | ||
− | |||
<!--* "Collaborative mapping of an earthquake-damaged building via ground and aerial robots", N. Michael ''et al.'', JFR 2012. ''UAV, UGV, disaster, mapping''--> | <!--* "Collaborative mapping of an earthquake-damaged building via ground and aerial robots", N. Michael ''et al.'', JFR 2012. ''UAV, UGV, disaster, mapping''--> | ||
* [http://ieeexplore.ieee.org/document/7353481/ VoxNet: A 3D Convolutional Neural Network for real-time object recognition], Maturana and Scherer, IROS 2015 | * [http://ieeexplore.ieee.org/document/7353481/ VoxNet: A 3D Convolutional Neural Network for real-time object recognition], Maturana and Scherer, IROS 2015 | ||
Line 189: | Line 187: | ||
|<!--Localization-->Finish detection & segmentation background, introduction to TensorFlow library | |<!--Localization-->Finish detection & segmentation background, introduction to TensorFlow library | ||
|[https://arxiv.org/abs/1612.08242 YOLOv2, Redmon and Farhadi] (CVPR 2017) | |[https://arxiv.org/abs/1612.08242 YOLOv2, Redmon and Farhadi] (CVPR 2017) | ||
− | |<!--ETH localization lectures: [http://nameless.cis.udel. | + | |<!--ETH localization lectures: [http://nameless.cis.udel.edAishwarya Dangu/class_data/849_s2018/ETH_Localization_I_2017.pdf 1] [http://nameless.cis.udel.edu/class_data/849_s2018/ETH_Localization_II_2017.pdf 2]--><!--[https://docs.google.com/presentation/d/1CTGsRG7yvJVfZyBfaC1ZpBoUJ3ALbLaYnMeYZDM1wRA/edit?usp=sharing slides]-->Stanford CS231n [http://cs231n.stanford.edu/slides/2017/cs231n_2017_lecture11.pdf Detection and Segmentation slides] (44-49, 53-85), [http://cs231n.stanford.edu/slides/2017/cs231n_2017_lecture8.pdf deep learning software slides] (21-33, 40-59)<br>[https://drive.google.com/file/d/104GFBl4J6bseaTNBcqW2ub8hyXtL99d1/view?usp=sharing TensorFlow code sample from Stanford lecture]<br>[[CISC849_S2018_HW2|HW #2]] |
|- | |- | ||
|12 | |12 | ||
Line 224: | Line 222: | ||
|style="background:rgb(102, 204, 255)"|15 | |style="background:rgb(102, 204, 255)"|15 | ||
|Apr. 3 | |Apr. 3 | ||
− | |Imitation | + | |Imitation learning |
− | |[https://www.youtube.com/watch?v=NJU9ULQUwng Dave-2 driving a Lincoln], [https://www.youtube.com/watch?v=umRdt3zGgpU Quadcopter Navigation in the Forest...], [https://www.youtube.com/watch?v=hNsP6-K3Hn4 DAgger for drones] | + | | |
− | |[http://rll.berkeley.edu/deeprlcourse/ Berkeley Levine course] [http://rll.berkeley.edu/deeprlcourse/f17docs/lecture_2_behavior_cloning.pdf IL slides]<br>''Paper presentation choice due'' | + | * [http://rpg.ifi.uzh.ch/docs/RAL16_Giusti.pdf A Machine Learning Approach to Visual Perception of Forest Trails for Mobile Robots], Giusti ''et al.'', ''IEEE Robotics and Automation Letters'', 2016 |
+ | * [https://arxiv.org/pdf/1604.07316.pdf End to End Learning for Self-Driving cars], Bojarski ''et al.'', 2016 | ||
+ | * [https://www.youtube.com/watch?v=NJU9ULQUwng Dave-2 driving a Lincoln], [https://www.youtube.com/watch?v=umRdt3zGgpU Quadcopter Navigation in the Forest...], [https://www.youtube.com/watch?v=hNsP6-K3Hn4 DAgger for drones] | ||
+ | * [https://www.youtube.com/watch?v=ZMhO1FO_j0o&t=566s Hal Daume DAgger explanation] (9:26-12:43) | ||
+ | |[http://rll.berkeley.edu/deeprlcourse/ Berkeley Levine course] [http://rll.berkeley.edu/deeprlcourse/f17docs/lecture_2_behavior_cloning.pdf IL slides] (DAgger defined on slide 14); [https://katefvision.github.io/katefSlides/immitation_learning_I_katef.pdf CMU Fragkiadaki IL slides] (1-29)<br>''Paper presentation choice due'' | ||
|- | |- | ||
|16 | |16 | ||
|Apr. 5 | |Apr. 5 | ||
− | | | + | |Reinforcement learning |
| | | | ||
− | |[http://www0.cs.ucl.ac.uk/staff/d.silver/web/Teaching.html UCL Silver course]: [http://nameless.cis.udel.edu/class_data/849_s2018/intro_RL.pdf RL slides], [https://www.youtube.com/watch?v=2pWv7GOvuf0 RL video] (contains movies missing from PDF) | + | |[http://www0.cs.ucl.ac.uk/staff/d.silver/web/Teaching.html UCL Silver course]: [http://nameless.cis.udel.edu/class_data/849_s2018/intro_RL.pdf RL slides] (7-42), [https://www.youtube.com/watch?v=2pWv7GOvuf0 RL video] (15:40-21:20, 1:02:13-1:04:40 contains movies missing from PDF); [http://cs231n.stanford.edu/slides/2017/cs231n_2017_lecture14.pdf Stanford deep-Q lecture slides] (23-62); [https://www.nature.com/articles/nature14236.pdf original deep-Q paper] |
|- | |- | ||
|style="background:rgb(102, 204, 255)"|17 | |style="background:rgb(102, 204, 255)"|17 | ||
|Apr. 10<br> | |Apr. 10<br> | ||
''Withdraw deadline Apr. 9'' | ''Withdraw deadline Apr. 9'' | ||
− | |Student paper | + | |Student paper presentation |
| | | | ||
+ | * Peng Su, [https://www.cv-foundation.org/openaccess/content_cvpr_2016/papers/Yeung_End-To-End_Learning_of_CVPR_2016_paper.pdf "End-to-end Learning of Action Detection from Frame Glimpses in Videos"] | ||
| | | | ||
|- | |- | ||
|18 | |18 | ||
|Apr. 12 | |Apr. 12 | ||
− | |Student paper | + | |Student paper presentation |
| | | | ||
+ | * Patrick Geneva, [http://openaccess.thecvf.com/content_cvpr_2017/papers/Dong_Visual-Inertial-Semantic_Scene_Representation_CVPR_2017_paper.pdf "Visual-Inertial-Semantic Scene Representation for 3D Object Detection"] | ||
|''Project proposal due Sunday, April 15'' | |''Project proposal due Sunday, April 15'' | ||
|- | |- | ||
Line 251: | Line 255: | ||
|Student paper presentations | |Student paper presentations | ||
| | | | ||
+ | * Zengxiang Lu, "Image Generation from Scene Graphs" | ||
+ | * Aishwarya Dang, [http://ieeexplore.ieee.org/document/7353481/ "VoxNet: A 3D Convolutional Neural Network for real-time object recognition"] | ||
| | | | ||
|- | |- | ||
|20 | |20 | ||
|Apr. 19 | |Apr. 19 | ||
− | | | + | |Student paper presentations |
| | | | ||
+ | * Sumeet Gupta, [https://arxiv.org/abs/1704.05588 "Learning to Fly by Crashing"] | ||
+ | * Yulin Yang, [http://rpg.ifi.uzh.ch/docs/RAL18_Loquercio.pdf "DroNet: Learning to Fly by Driving"] | ||
| | | | ||
|- | |- | ||
|style="background:rgb(102, 204, 255)"|21 | |style="background:rgb(102, 204, 255)"|21 | ||
|Apr. 24 | |Apr. 24 | ||
− | |Student paper | + | |Student paper presentation |
| | | | ||
+ | Matt Schmittle, [https://arxiv.org/pdf/1509.06825.pdf Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours] | ||
| | | | ||
|- | |- | ||
Line 269: | Line 278: | ||
|Student paper presentations | |Student paper presentations | ||
| | | | ||
+ | * Apoorva Patil, [http://www.cs.ubc.ca/~van/papers/2017-TOG-deepLoco/ DeepLoco: Dynamic Locomotion Skills Using Hierarchical Deep Reinforcement Learning] | ||
+ | * Leighanne Hsu, [http://www2.informatik.uni-freiburg.de/~hornunga/pub/maier12humanoids.pdf Real-Time Navigation in 3D Environments Based on Depth Camera Data] | ||
|<!--[https://docs.google.com/presentation/d/19Wm-H2ttW_m5pan_lCXDzmk468gLoud6Xug_2_i0mWI/edit?usp=sharing&authkey=CKmYmJYI slides]--> | |<!--[https://docs.google.com/presentation/d/19Wm-H2ttW_m5pan_lCXDzmk468gLoud6Xug_2_i0mWI/edit?usp=sharing&authkey=CKmYmJYI slides]--> | ||
|- | |- | ||
|style="background:rgb(102, 204, 255)"|23 | |style="background:rgb(102, 204, 255)"|23 | ||
|May 1 | |May 1 | ||
− | | | + | |Student paper presentation |
− | | | + | |Yiqun Jia, [https://arxiv.org/pdf/1603.03833.pdf From Virtual Demonstration to Real-World Manipulation Using LSTM and MDN] |
| | | | ||
|- | |- | ||
|24 | |24 | ||
|May 3 | |May 3 | ||
− | | | + | |Instructor paper presentation |
− | | | + | |Rasmussen, [http://www.ri.cmu.edu/publication_view.html?pub_id=5590 "Real-Time SLAM with Octree Evidence Grids for Exploration in Underwater Tunnels"], N. Fairfield, G. Kantor, and D. Wettergreen, JFR 2007 |
− | | | + | |[https://docs.google.com/presentation/d/1CTGsRG7yvJVfZyBfaC1ZpBoUJ3ALbLaYnMeYZDM1wRA/edit#slide=id.i31 DepthX slides] |
|- | |- | ||
|style="background:rgb(102, 204, 255)"|25 | |style="background:rgb(102, 204, 255)"|25 | ||
|May 8 | |May 8 | ||
− | | | + | |Final project review |
| | | | ||
| | | | ||
|- | |- | ||
− | | | + | | |
|May 10 | |May 10 | ||
− | | | + | |style="background:rgb(255, 102, 0)"|NO CLASS<br>''Work on projects...'' |
| | | | ||
| | | | ||
|- | |- | ||
− | |style="background:rgb(102, 204, 255)"| | + | |style="background:rgb(102, 204, 255)"|26 |
|May 15 | |May 15 | ||
− | |Final project presentations | + | |Final project presentations part I |
+ | | | ||
+ | | | ||
+ | |- | ||
+ | |27 | ||
+ | |May 16 | ||
+ | |Final project presentations part II | ||
+ | | | ||
+ | | | ||
+ | |- | ||
+ | |28 | ||
+ | |May 21 | ||
+ | |Final project presentations part III | ||
+ | | | ||
| | | | ||
− | |||
|} | |} |
Latest revision as of 09:53, 10 May 2018
Contents
Course information
Title | CISC849 Robot Vision and Learning |
Shortened URL | https://goo.gl/ektJij |
Description | Survey of image-based 2-D and 3-D sensing algorithms for mobile robot navigation and interaction, including motion estimation, obstacle segmentation, terrain modeling, and object recognition, with a particular focus on deep learning techniques to dramatically improve performance. |
When | Tuesdays and Thursdays, 11 am-12:15 pm |
Where | Smith 102A |
Instructor | Christopher Rasmussen, 446 Smith Hall, cer@cis.udel.edu |
Office hours | Wednesdays, 10 am -- 12 pm |
Grading |
|
Academic policies | Programming projects are due by midnight of the deadline day (with a grace period of a few hours afterward...after sunrise is definitely late). A late homework is a 0 without a valid prior excuse. To give you a little flexibility, you have 6 "late days" to use on homeworks to extend the deadline by one day each without penalty. No more than three late days may be used per assignment. Late days will automatically be subtracted, but as a courtesy please notify the instructor in an e-mail of your intention to use late days before the deadline. See submission instructions below.
Students can discuss problems with one another in general terms, but must work independently on programming assignments. This also applies to online and printed resources: you may consult them as references (as long as you cite them), but the words and source code you turn in must be yours alone. The University's policies on academic dishonesty are set forth in the student code of conduct here. |
Homeworks | Assignment submissions should consist of a directory containing all code (your .cpp files, makefile, etc.), any output data generated (e.g., images, movies, etc.), and an explanation of your approach, what worked and didn't work, etc. contained in a separate text or HTML file. Do not submit executables or .o files, please! The directory you submit for each assignment should be packaged by tar'ing and gzip'ing it or just zip'ing it. The resulting file should be submitted through Canvas.
You may develop your C/C++ code in any fashion that is convenient--that is, with any compiler and operating system that you want. However, we will be grading your homework on a Linux system with a makefile, and so you must avoid OS- and hardware-specific functions and provide a makefile for us that will work (like one of the templates above). |
Possible Papers to Present (not a complete list)
- Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours, Pinto and Gupta, ICRA 2016
- Learning real manipulation tasks from virtual demonstrations using LSTM, Rahmatizadeh et al., AAAI 2018
- Learning to Fly by Crashing, Gandhi, Pinto, and Gupta, IROS 2017
- VoxNet: A 3D Convolutional Neural Network for real-time object recognition, Maturana and Scherer, IROS 2015
- PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation, Qi, Su, Mo, and Guibas, CVPR 2017
- Real-Time Navigation in 3D Environments Based on Depth Camera Data, D. Maier, A. Hornung, and M. Bennewitz, Humanoids 2012
- Robotic Grasping of Novel Objects using Vision, A. Saxena, J. Driemeyer, A. Ng, IJRR 2008
- DeepLoco: Dynamic Locomotion Skills Using Hierarchical Deep Reinforcement Learning, X. Peng, G. Berseth, K. Yin, and M. van de Panne, SIGGRAPH 2017
- High fidelity day/night stereo mapping with vegetation and negative obstacle detection for vision-in-the-loop walking, M. Bajracharya et al., IROS 2013
Instructions for Homeworks
Software |
|
Note |
|
Schedule
Note: The blue squares in the "#" column below indicate Tuesdays.
# | Date | Topic | Links/Readings/videos | Assignments/slides |
---|---|---|---|---|
1 | Feb. 6 | Background | slides | |
2 | Feb. 8 | Finish background | slides | |
3 | Feb. 13 | Introduction to DARPA Urban Challenge (UC), Robotics Challenge (DRC) and algorithm components |
|
slides |
4 | Feb. 15 | Introduction to ROS; PCL tutorial |
ARGOS challenge overview (8:00) |
slides |
5 | Feb. 20 Register/add deadline Feb. 19 |
Plane/obstacle/object segmentation (3-D) | RANSAC background | HW #1 plane_fit.cpp (No ROS required) |
6 | Feb. 22 | Finish plane segmentation, clustering, ICP | slides | |
7 | Feb. 27 | Image classification background | Stanford CS231n Image classification slides (6-61), | |
8 | Mar. 1 | NO CLASS Instructor away |
HW #1 due | |
9 | Mar. 6 | More classification background | Loss/optimization slides (8-10, 37-77, finish with web demo), Intro to NN (83-88) | |
10 | Mar. 8 | Introduction to convolutional neural networks | Stanford CS231n Convolutional Neural Networks slides (15-63, 71-78), CNN architectures (8-36), Detection and Segmentation slides (15-27) | |
11 | Mar. 13 | Finish detection & segmentation background, introduction to TensorFlow library | YOLOv2, Redmon and Farhadi (CVPR 2017) | Stanford CS231n Detection and Segmentation slides (44-49, 53-85), deep learning software slides (21-33, 40-59) TensorFlow code sample from Stanford lecture HW #2 |
12 | Mar. 15 | More on TensorFlow programming | ||
13 | Mar. 20 | Perception for stepping | "Learning Locomotion over Rough Terrain using Terrain Templates", M. Kalakrishnan, J. Buchli, P. Pastor, and S. Schaal, IROS 2009 | slides DeepLoco highlights video (6:23) |
14 | Mar. 22 | NO CLASS | HW #2 due Friday, March 23 | |
Mar. 27 | NO CLASS Spring break |
|||
Mar. 29 | NO CLASS Spring break |
|||
15 | Apr. 3 | Imitation learning |
|
Berkeley Levine course IL slides (DAgger defined on slide 14); CMU Fragkiadaki IL slides (1-29) Paper presentation choice due |
16 | Apr. 5 | Reinforcement learning | UCL Silver course: RL slides (7-42), RL video (15:40-21:20, 1:02:13-1:04:40 contains movies missing from PDF); Stanford deep-Q lecture slides (23-62); original deep-Q paper | |
17 | Apr. 10 Withdraw deadline Apr. 9 |
Student paper presentation | ||
18 | Apr. 12 | Student paper presentation | Project proposal due Sunday, April 15 | |
19 | Apr. 17 |
Student paper presentations |
|
|
20 | Apr. 19 | Student paper presentations |
|
|
21 | Apr. 24 | Student paper presentation |
Matt Schmittle, Supersizing Self-supervision: Learning to Grasp from 50K Tries and 700 Robot Hours |
|
22 | Apr. 26 | Student paper presentations | ||
23 | May 1 | Student paper presentation | Yiqun Jia, From Virtual Demonstration to Real-World Manipulation Using LSTM and MDN | |
24 | May 3 | Instructor paper presentation | Rasmussen, "Real-Time SLAM with Octree Evidence Grids for Exploration in Underwater Tunnels", N. Fairfield, G. Kantor, and D. Wettergreen, JFR 2007 | DepthX slides |
25 | May 8 | Final project review | ||
May 10 | NO CLASS Work on projects... |
|||
26 | May 15 | Final project presentations part I | ||
27 | May 16 | Final project presentations part II | ||
28 | May 21 | Final project presentations part III |