Date | Time | Activities | Lecturers |
Dec. 2, 2016 | 08:30-09:00 | Opening Ceremony, Group Photo | |
09:00-12:00 | Lecture 1: 语言、理解与认知:从瞎子摸象说起 | Guodong Zhou, Professor, Soochow University | |
14:00-17:00 | Lecture 2: Recent Progress on Deep Learning for Natural Language Processing | Zhengdong Lu, Founder and CTO of DeeplyCourious.Ai | |
Dec. 3, 2016 | 09:00-12:00 | Lecture 3: Efficient Training and Deployment of Large Scale Deep Learning Systems for NLP | Jacob Devlin, Principal NLP Scientist, Microsoft Research |
14:00-16:00 | Lecture 4: Panel: Natural Language Processing and Artificial Intelligence: Challenges and Opportunities | Panelist: Jacob Devlin, Zhengdong Lu, Guodong Zhou, Chengqing Zong, Lei Li In this panel, we are going to discuss potential and opportunities and challenges for the research of natural language processing in the artificial intelligence era. In particular, the panelist members will share their thoughts on theory, algorithms, and killer applications underlying the fast development of natural language processing technologies, and those benefit from the development of NLP. |
|
Dec. 4, 2016 | 09:00-10:00 | Lecture 5: NLP Technologies in Question Answerers and Voice Assistants | Dekang Lin, ACL Fellow, co-founder and co-CEO of naturali.io. Jidian Jizhi |
10:15-12:00 | Lecture 6: Deep Learning for Answering Questions | Lei Li, Toutiao | |
14:00-17:00 | Lecture 7: Fundamentals and Challenges in Mining Social Media | Huan Liu, Computer Science and Engineering at Arizona State University | |
17:00-17:20 | Closing Ceremony |
◇ Lecture 1: 语言、理解与认知:从瞎子摸象说起
◇ Guodong Zhou, Soochow University
◇ ABSTRACT: 虽然自然语言一直在社会、经济和国家安全等领域中扮演着重要角色,但是一直以来计算机的自然语言理解与认知能力远逊于人类。近几年,随着移动互联网的不断普及,以及云计算、大数据、GPU、深度学习等相关平台和技术的快速发展,我们越来越感到自然语言处理方面的突破就在眼前。
本报告将从语言、理解、认知等多方位系统探讨如何提高自然语言理解与认知能力。具体包括:语言的本质特点、经典语言学理论简介(结构主义语言学、形式语言学)、现代语言学理论简介(功能语言学、认知语言学)、自然语言理解基础(任务体系、篇章分析)、语言与认知基础(认知科学、认知语言学)。
◇ Short Bio: I received my B.Sc., M.Sc. and Ph.D. from XI'AN Jiaotong Univ. in 1989, Shanghai Jiaotong Univ. in 1992 and the National Univ. of Singapore in 1999 (Ph.D. thesis submitted in Dec 1997 and Ph.D. degree conferred in June 1999), respectively. Besides, I worked as a lecturer at Shanghai Jiaotong Univ. from April 1992 to March 1995 and as a post-doc research fellow at the National Univ. of Singapore from Jan 1998 to March 1999. I joined the Institute for Infocomm Research, Singapore as an associate scientist in April 1999 and was promoted to scientist and associate lead scientist in April 2002 and April 2005, respectively. In the meantime, I had been a Ph.D. supervisor at the institute (jointly with School of Computing, the National Univ. of Singapore) since April 2002. I joined Soochow Univ. as a full-time distinguished professor and Ph.D. supervisor in Aug 2006. Currently, I am an associate editor of ACM Transaction on Asian Language Information Processing(2010.07-2016.06), an editorial member of Journal of Software (Chinese)(2012.01-2014.12) and a vice chair of Technical Committees on Chinese Information/China Computer Federation(2010.12-2016.12), Computational Linguistics/Chinese Information Processing Society of China and Natural Language Understanding/Artificial Intelligence Society of China. Besides, I had been a member of the Editorial Board of Computational Linguistics (2010.01-2012.12), and acted as a panel project appraisal expert of Information Division, NSFC during 2009-2014. In addition, I have severed on the program committees/reviewer boards of several prestigious international journals and conferences, including Bioinformatics (Journal), BMC Bioinformatics (Journal), Information Sciences (Journal), Information Systems(Journal), Information Processing and Management (Journal), Journal of Computer Science and Technology (Journal), IEEE TASLP(Journal), ACM Transactions on Asian Language Information Processing (Journal), Natural Language Engineering (Journal), Computer Speech and Language (Journal), ACL (Annual Meeting of the Association for Computational Linguistics), EMNLP(Empirical Methods on Natural Language Processing), COLING (International Conference on Computational Linguistics ), SIGIR, AAAI, IJCAI, CIKM.
◇ Lecture 2: Recent Progress on Deep Learning for NLP
◇ Zhengdong Lu: Senior Researcher, Founder and CTO of DeeplyCourious.Ai
◇ ABSTRACT: Dr. Lu will give a talk on the recent progress of deep learning for NLP (DL4NLP) as a particular direction of Deep Learning. More specifically he will talk briefly about the basic concepts of DL4NLP, and then he will elaborate on the recent progress on differentiable data structure and learning paradigms.
◇ Short Bio: Zhengdong Lu is a founder and CTO of DeeplyCourious.Ai. Before that he was a senior researcher at Noah’s Ark Lab, Huawei. Before joining Huawei, Dr. Lu worked as an associate researcher in MSRA and postdoctoral researcher at University of Texas at Austin. He got his bachelor’s degree from Xi’an Institute of Post and Telecommunication, and Ph.D degree from Oregon Health & Science University. His research interest includes deep learning,natural language processing, reasoning and general machine learning. Dr. Lu has authored over 40 papers on NIPS, ICML, ACL, and JMLR etc.
◇ Lecture 3: Efficient Training and Deployment of Large Scale Deep Learning Systems for NLP
◇ Jacob Devlin, Principal NLP Scientist, Microsoft Research
◇ ABSTRACT: Large-scale, complex neural networks such as attentional sequence-to-sequence models for machine translation are often slow to train and very memory intensive, even on powerful GPUs. Additionally, it is often difficult to deploy the trained model running on the CPU (or GPU) with acceptable latency for a real time service.
In this tutorial, I will detail several techniques for drastically improving model training time and memory usage on a single GPU (or a small number of GPUs). These techniques can improve the training time for various recurrent neural network architectures architecture by 5x or more. They also allow much wider and deeper networks to be trained on a single GPU.
I will also describe a number of tricks improving the speed of the networks on the CPU at inference time, with a specific focus on attentional sequence-to-sequence models.
◇ Short Bio: Jacob Devlin is a Principal NLP Scientist in the Machine Translation Group at Microsoft Research in Redmond, Washington. His primary research focus is on developing fast, scalable deep learning systems for machine translation and other applications. Prior to joining Microsoft, Mr. Devlin worked as a Research Scientist at BBN Technologies in Cambridge, Massachussets. His work on machine translation received the ACL best paper award in 2014 and the NAACL best paper award in 2012. He received his B.A./M.S. in Computer Science from the University of Maryland, College Park in 2009.
◇ Lecture 5: NLP Technologies in Question Answerers and Voice Assistants
◇ Dekang Lin (co-founder and co-CEO of naturali.io.), Jidian Jizhi
◇ ABSTRACT: Question answerers and voice assistants are two of the most important NLP applications. The former aims at providing precise, direct answers answers to users questions. The latter goes a step further by not only figuring out the semantic interpretation of a request, but also execute a sequence of actions to fulfil the request on the user's behalf. Both applications require their NLP components to be precise, robust, as well as efficient. I will survey and discuss the NLP technologies used in these two applicaitons.
◇ Short Bio: Dekang Lin is co-founder and co-CEO of naturali.io, a Beijing-based startup for voice assistant. He was a Senior Staff Research Scientist at Google from 2004 to 2016, where he led a team of researchers and engineers to build a question-answerer in Google search.
Before joining Google, Dekang Lin was a full professor of Computer Science at University of Alberta. He authored 90 papers with over 12000 citations. He was elected as Fellow of Association for Computational Linguistics (ACL) in 2012 and served as the program co-chair and general chair for ACL-2002 and ACL-2011 respectively.
◇ Lecture 6: Deep Learning for Answering Questions
◇ Lei Li, Toutiao
◇ ABSTRACT: Human language is still elusive for machines to comprehend. In recent years, there are much effort in pushing the performance of machines in complex tasks such as dialog and question answering. In this talk, we will review recent progress of deep recurrent neural networks in answering factoid questions regarding large knowledge bases. We will give a deep dive to our CFO system. It builds upon techniques essential to success of retrieving candidate answers from a Knowledge Base given a question: including word meaning in vector space, entity mentions in an utterance, parsing user intents in queries, and linking entities and relations in a large-scale knowledge base. It can retrieve answers for questions such as “who is the creator of Harry Potter” from tens of millions of facts.
◇ Short Bio: Lei Li is a research scientist at Toutiao and director of Toutiao Lab. Lei received his B.S. in Computer Science and Engineering from Shanghai Jiao Tong University (ACM class) and Ph.D. in Computer Science from Carnegie Mellon University, respectively. His dissertation work on fast algorithms for mining co-evolving time series was awarded ACM KDD best dissertation (runner up). Before Toutiao, he worked at Baidu's Institute of Deep Learning in Silicon Valley as a Principal Research Scientist (Young Leading Scholar). Previously, he worked briefly at Microsoft Research (Asia and Redmond), Google (Mountain View), and IBM (TJ Watson Reserch Center). Before joining Baidu, he was working in EECS department of UC Berkeley as a Post-Doctoral Researcher. His research interest lies in the intersection of deep learning, statistical inference, natural language understanding, and time series analysis.
◇ Lecture 7: Fundamentals and Challenges in Mining Social Media
◇ Huan Liu, Computer Science and Engineering at Arizona State University
◇ ABSTRACT: This ADL lecture consists of three parts: (1) Fundamentals in social media mining, (2) some recent algorithmic developments, and (3) data and evaluation challenges. We start with the basics of social media data and mining and show how it is different from traditional data mining. We will provide a brief review based on the Cambridge University Press textbook on Social Media Mining. Participants can access the text at the speaker’s website. In the second part of this lecture, we introduce some interesting and ingenious ideas and algorithms and demonstrate that research opportunities are where challenges are. In the last part of this lecture, we discuss some current challenges we are facing: where we can find data to suit our needs, how we can know whether data is sufficient if we crawl data ourselves, how we can separate wheat from chaff and sieve misinformation in search of knowledge, and how we can evaluate our findings where ground truth is not readily available. Questions are most welcome.
◇ Short Bio: Dr. Huan Liu is a professor of Computer Science and Engineering at Arizona State University. He obtained his Ph.D. in Computer Science at University of Southern California and B.Eng. in EECS at Shanghai JiaoTong University. He was recognized for excellence in teaching and research in Computer Science and Engineering at Arizona State University. His research interests are in data mining, machine learning, social computing, and artificial intelligence, investigating problems that arise in real-world applications with high-dimensional data of disparate forms. His well-cited publications include books, book chapters, encyclopedia entries as well as conference and journal papers. He serves on journal editorial/advisory boards and numerous conference program committees. He is a Fellow of IEEE and a member of several professional societies. His website is at http://www.public.asu.edu/~huanliu.