Skip to Main Content

Course Materials - Fall 2024

ROB801: Advanced Robotic Motion Planning

 TB = Textbook or Required reading                     REF = Reference or supplemental reading

Type          Title

eBook

Print  

Call Number
A. Silberschatz, G. Gagne, and P. B. Galvin, Operating System Concepts, 10th ed. Hoboken, NJ, USA: Wiley, 2018. NA On Order  

B. Recht, C. Re, S. Wright, and F. Niu, Hogwild!: A lock-free approach to parallelizing stochastic gradient descent, Advances in neural information processing systems 24, 2011.

ACM NA  

Q. Ho, J. Cipar, H. Cui, S. Lee, J. Kim, P. Gibbons, G. Gibson, G. Ganger, and E. Xing, More effective distributed ml via a stale synchronous parallel parameter server, Advances in neural information processing systems 26, 2013.

ADM NA    

J. Kim, Jin Kyu, Q. Ho, S. Lee, X. Zheng, W. Dai, G. Gibson, and E. Xing, Strads: A distributed framework for scheduled model parallel machine learning, In Proceedings of the Eleventh European Conference on Computer Systems, pp. 1-16 ,2016.

ADM NA  

H, Zhang, Z. Zheng, S. Xu, W. Dai, Q. Ho, X. Liang, Z. Hu, J. Wei, P. Xie, and E. Xing, Poseidon: Anefficient communication architecture for distributed deep learning on GPU clusters, In 2017 USENIX Annual Technical Conference USENIX ATC 17,pp. 181-193, 2017.

ADM NA  

P. Xie, J. Kim, Q. Ho, Y. Yu, and E. Xing, Orpheus: Efficient distributed machine learning via system and algorithm codesign, In Proceedings of the ACM Symposium on Cloud Computing, pp. 1-13, 2018.

ACM NA  

H. Wang, S. Sievert, S. Liu, Z. Charles, D. Papailiopoulos, and S. Wright, Atomo: Communication-efficient learning via atomic sparsification,  Advances in neural information processing systems 31, 2018.

ACM NA    

M. Shoeybi, M. Patwary, R. Puri, P. LeGresley, J. Casper, and B. Catanzaro, Megatron-lm: Training multi-billionparameter language models using model parallelism, arXiv preprint, 2019.

arXiv NA  

A. Qiao, S.  Choe, S. Subramanya, W. Neiswanger, Q. Ho, H. Zhang, G. Ganger, and E. Xing, Pollux: Co-adaptive cluster scheduling for goodput-optimized deep learning, In 15th USENIX Symposium on Operating Systems Design and Implementation OSDI 21, 2021.

Open Access NA  

L. Zheng, Z. Li, H. Zhang, Y. Zhuang, Z. Chen, Y. Huang, Y. Wang, Alpa: Automating inter-and Intra- Operator parallelism for distributed deep learning, In 16th USENIX Symposium on Operating Systems Design and Implementation OSDI 22, 2022.

Open Access NA  

T. Dao, D. Fu, S. Ermon, A. Rudra, and C. Ré, Flashattention: Fast and memory-efficient exact attention with io-awareness, Advances in Neural Information Processing Systems 35 ,2022.

ACM NA  

Y. Zhuang, H. Zhao, L. Zheng, Z. Li, E. Xing, Q0 Ho, J. Gonzalez, I. Stoica, and H. Zhang, On optimizing the communication of model parallelism, Proceedings of Machine Learning and Systems 5 ,2023.

ADM NA