专栏名称: 机器学习研究会
机器学习研究会是北京大学大数据与机器学习创新中心旗下的学生组织,旨在构建一个机器学习从事者交流的平台。除了及时分享领域资讯外,协会还会举办各种业界巨头/学术神牛讲座、学术大牛沙龙分享会、real data 创新竞赛等活动。
目录
相关文章推荐
爱可可-爱生活  ·  【[134星]LangFair:一个Pyth ... ·  昨天  
爱可可-爱生活  ·  晚安~ #晚安# -20250125230440 ·  昨天  
爱可可-爱生活  ·  《爱可可微博热门分享(1.24)》 ... ·  2 天前  
新智元  ·  OpenAI首个智能体Operator大测评 ... ·  3 天前  
机器学习研究组订阅  ·  ICLR ... ·  4 天前  
51好读  ›  专栏  ›  机器学习研究会

【学习】华盛顿大学深度学习系统课程

机器学习研究会  · 公众号  · AI  · 2017-04-07 18:50

正文



点击上方“机器学习研究会”可以订阅哦
摘要
 

转自:陈天奇怪

这学期在UW上深度学习系统课程,介绍深度学习系统的原理。有兴趣的同学可以做一做作业。

课程简介:

Over the past few years, deep learning has become an important technique to successfully solve problems in many different fields, such as vision, NLP, robotics. An important ingredient that is driving this success is the development of deep learning systems that efficiently support the task of learning and inference of complicated models using many devices and possibly using distributed resources. The study of how to build and optimize these deep learning systems is now an active area of research and commercialization, and yet there isn’t a course that covers this topic.


This course is designed to fill this gap. We will be covering various aspects of deep learning systems, including: basics of deep learning, programming models for expressing machine learning models, automatic differentiation, memory optimization, scheduling, distributed learning, hardware acceleration, domain specific languages, and model serving. Many of these topics intersect with existing research directions in databases, systems and networking, architecture and programming languages. The goal is to offer a comprehensive picture on how deep learning systems works, discuss and execute on possible research opportunities, and build open-source software that will have broad appeal.


We will have two classes per week. Each week will have one lecture. Another class will either be lab/discuss session or guest lectures. Each lecture will study a specific aspect of deep learning systems. The lab/discussion session will contain tutorials to implement that specific aspect and will include case studies of existing systems, such as Tensorflow, Caffe, Mxnet, PyTorch, and others.


作业一:

Assignment 1: Reverse-mode Automatic Differentiation

In this assignment, we would implement reverse-mode auto-diff.

Our code should be able to construct simple expressions, e.g. y=x1*x2+x1, and evaluate their outputs as well as their gradients (or adjoints), e.g. y, dy/dx1 and dy/dx2.

There are many ways to implement auto-diff, as explained in the slides for Lecture 4. For this assignment, we use the approach of a computation graph and an explicit construction of gradient (adjoint) nodes, similar to what MXNet and Tensorflow do.

Key concepts and data structures that we would need to implement are

  • Computation graph and Node

  • Operator, e.g. Add, MatMul, Placeholder, Oneslike

  • Construction of gradient nodes given forward graph

  • Executor


课程主页:

http://dlsys.cs.washington.edu/


作业链接:

https://github.com/dlsys-course/assignment1


原文链接:

http://weibo.com/2397265244/EDlCWpCeQ?type=comment

“完整内容”请点击【阅读原文】
↓↓↓