ROB801: Advanced Robotic Motion Planning
TB = Textbook or Required reading
REF = Reference or supplemental reading
Type | Title |
eBook |
|
Call Number | ||
![]() |
![]() |
A. Silberschatz, G. Gagne, and P. B. Galvin, Operating System Concepts, 10th ed. Hoboken, NJ, USA: Wiley, 2018. | NA | On Order | ||
![]() |
![]() |
B. Recht, C. Re, S. Wright, and F. Niu, Hogwild!: A lock-free approach to parallelizing stochastic gradient descent, Advances in neural information processing systems 24, 2011. |
ACM | NA | ||
![]() |
![]() |
Q. Ho, J. Cipar, H. Cui, S. Lee, J. Kim, P. Gibbons, G. Gibson, G. Ganger, and E. Xing, More effective distributed ml via a stale synchronous parallel parameter server, Advances in neural information processing systems 26, 2013. |
ADM | NA | ||
![]() |
![]() |
J. Kim, Jin Kyu, Q. Ho, S. Lee, X. Zheng, W. Dai, G. Gibson, and E. Xing, Strads: A distributed framework for scheduled model parallel machine learning, In Proceedings of the Eleventh European Conference on Computer Systems, pp. 1-16 ,2016. |
ADM | NA | ||
![]() |
![]() |
H, Zhang, Z. Zheng, S. Xu, W. Dai, Q. Ho, X. Liang, Z. Hu, J. Wei, P. Xie, and E. Xing, Poseidon: Anefficient communication architecture for distributed deep learning on GPU clusters, In 2017 USENIX Annual Technical Conference USENIX ATC 17,pp. 181-193, 2017. |
ADM | NA | ||
![]() |
![]() |
P. Xie, J. Kim, Q. Ho, Y. Yu, and E. Xing, Orpheus: Efficient distributed machine learning via system and algorithm codesign, In Proceedings of the ACM Symposium on Cloud Computing, pp. 1-13, 2018. |
ACM | NA | ||
![]() |
![]() |
H. Wang, S. Sievert, S. Liu, Z. Charles, D. Papailiopoulos, and S. Wright, Atomo: Communication-efficient learning via atomic sparsification, Advances in neural information processing systems 31, 2018. |
ACM | NA | ||
![]() |
![]() |
M. Shoeybi, M. Patwary, R. Puri, P. LeGresley, J. Casper, and B. Catanzaro, Megatron-lm: Training multi-billionparameter language models using model parallelism, arXiv preprint, 2019. |
arXiv | NA | ||
![]() |
![]() |
A. Qiao, S. Choe, S. Subramanya, W. Neiswanger, Q. Ho, H. Zhang, G. Ganger, and E. Xing, Pollux: Co-adaptive cluster scheduling for goodput-optimized deep learning, In 15th USENIX Symposium on Operating Systems Design and Implementation OSDI 21, 2021. |
Open Access | NA | ||
![]() |
![]() |
L. Zheng, Z. Li, H. Zhang, Y. Zhuang, Z. Chen, Y. Huang, Y. Wang, Alpa: Automating inter-and Intra- Operator parallelism for distributed deep learning, In 16th USENIX Symposium on Operating Systems Design and Implementation OSDI 22, 2022. |
Open Access | NA | ||
![]() |
![]() |
T. Dao, D. Fu, S. Ermon, A. Rudra, and C. Ré, Flashattention: Fast and memory-efficient exact attention with io-awareness, Advances in Neural Information Processing Systems 35 ,2022. |
ACM | NA | ||
![]() |
![]() |
Y. Zhuang, H. Zhao, L. Zheng, Z. Li, E. Xing, Q0 Ho, J. Gonzalez, I. Stoica, and H. Zhang, On optimizing the communication of model parallelism, Proceedings of Machine Learning and Systems 5 ,2023. |
ADM | NA |