点击蓝字

关注我们

AI TIME欢迎每一位AI爱好者的加入!

9月28日 15:00~20:30

AI TIME特别邀请了多位PhD,带来ICML-6!

哔哩哔哩直播通道

扫码关注AITIME哔哩哔哩官方账号

观看直播

链接:https://live.bilibili.com/21813994

15:00-17:00

★ 嘉宾介绍 ★

贾彬彬

从2017年9月开始在东南大学计算机科学与工程学院攻读博士学位,主要研究方向为多维分类(Multi-Dimensional Classification),目前发表CCF A类期刊1篇、会议3篇,CCF B类期刊3篇,CCF C类会议1篇。

报告题目:

基于稀疏标记编码的多维分类方法

内容简介:

In multi-dimensional classification (MDC), there are multiple class variables in the output space with each of them corresponding to one heterogeneous class space. Due to the heterogeneity of class spaces, it is quite challenging to consider the dependencies among class variables when learning from MDC examples. In this paper, we propose a novel MDC approach named SLEM which learns the predictive model in an encoded label space instead of the original heterogeneous one. Specifically, SLEM works in an encoding-training-decoding framework. In the encoding phase, each class vector is mapped into a real-valued one via three cascaded operations including pairwise grouping, one-hot conversion and sparse linear encoding. In the training phase, a multi-output regression model is learned within the encoded label space. In the decoding phase, the predicted class vector is obtained by adapting orthogonal matching pursuit over outputs of the learned multi-output regression model. Experimental results clearly validate the superiority of SLEM against state-of-the-art MDC approaches.

吴齐天

上海交大博士生,导师为严骏驰教授,研究方向为机器学习与数据挖掘。在ICML/NeurIPS/KDD发表多篇一作论文,获评2021年百度AI新星。

报告题目:

基于图结构学习的归纳式协同过滤

内容简介:

Recommendation models can effectively estimate underlying user interests and predict one's future behaviors by factorizing an observed user-item rating matrix into products of two sets of latent factors. However, the user-specific embedding factors can only be learned in a transductive way, making it difficult to handle new users on-the-fly. In this paper, we propose an inductive collaborative filtering framework that contains two representation models. The first model follows conventional matrix factorization which factorizes a group of key users' rating matrix to obtain meta latents. The second model resorts to attention-based structure learning that estimates hidden relations from query to key users and learns to leverage meta latents to inductively compute embeddings for query users via neural message passing. Our model enables inductive representation learning for users and meanwhile guarantees equivalent representation capacity as matrix factorization. Experiments demonstrate that our model achieves promising results for recommendation on few-shot users with limited training ratings and new unseen users which are commonly encountered in open-world recommender systems.

唐瀚霖

从2017年9月就读于University of Rochester大学CS,于2021年7月获得CS博士学位。主要研究方向是大规模训练下的通信加速的优化器设计以及广义优化器研究。导师是刘霁老师。在ICML和NIPS上发表过相关论文。

报告题目:

一种使用一比特通信的Adam优化器

内容简介:

Scalable training of large models (like BERT and GPT-3) requires careful optimization rooted in model design, architecture, and system capabilities. From a system standpoint, communication has become a major bottleneck, especially on commodity systems with standard TCP interconnects that offer limited network bandwidth. Communication compression is an important technique to reduce training time on such systems. One of the most effective methods is error-compensated compression, which offers robust convergence speed even under 1-bit compression. However, state-of-the-art error compensation techniques only work with basic optimizers like SGD and momentum SGD, which are linearly dependent on the gradients. They do not work with non-linear gradient-based optimizers like Adam, which offer state-of-the-art convergence efficiency and accuracy for models like BERT. In this paper, we propose 1-bit Adam that reduces the communication volume by up to 5×, offers much better scalability, and provides the same convergence speed as uncompressed Adam. Our key finding is that Adam's variance (non-linear term) becomes stable (after a warmup phase) and can be used as a fixed precondition for the rest of the training (compression phase). Experiments on up to 256 GPUs show that 1-bit Adam enables up to 3.3× higher throughput for BERT-Large pre-training and up to 2.9× higher throughput for SQuAD fine-tuning. In addition, we provide theoretical analysis for our proposed work.

牛帅程

自2018年9月开始在华南理工大学软件学院攻读博士学位。导师为谭明奎教授,以及腾讯AI Lab的吴家祥和赵沛霖研究员。主要研究方向为神经网络结构搜索和迁移学习,并在相关领域会议和期刊发表论文多篇,包括ICML, CVPR, IJCAI, TIP, TKDE等。

报告题目:

针对动态变化数据的模型结构自动调整

内容简介:

In real-world applications, data often come in a growing manner, where the data volume and the number of classes may increase dynamically. This will bring a critical challenge for learning: given the increasing data volume or the number of classes, one has to instantaneously adjust the neural model capacity to obtain promising performance. Existing methods either ignore the growing nature of data or seek to independently search an optimal architecture for a given dataset, and thus are incapable of promptly adjusting the architectures for the changed data.  To address this, we present a neural architecture adaptation method, namely Adaptation eXpert (AdaXpert), to efficiently adjust previous architectures on the growing data. Specifically, we introduce an architecture adjuster to generate a suitable architecture for each data snapshot, based on the previous architecture and the different extent between current and previous data distributions. Furthermore, we propose an adaptation condition to determine the necessity of adjustment, thereby avoiding unnecessary and time-consuming adjustments. Extensive experiments on two growth scenarios (increasing data volume and number of classes) demonstrate the effectiveness of the proposed method.

19:30-20:30

钟方威

现为北京大学人工智能研究院博士后,获博士后创新人才计划资助。在此之前于2021年获得北京大学博士学位。他的研究兴趣是融合计算机视觉、机器人学习、多智能体、虚拟现实和认知科学等多个领域知识,实现高效自主的机器人。他已在人工智能领域顶级学术期刊和会议发表论文多篇,包括了IEEE TPAMI、ICML、ICLR、NeurIPS、CVPR、AAAI等。他多次受邀担任 NeurIPS、ICML、ICLR、CVPR、ICCV、AAAI等人工智能领域顶级国际会议程序委员/审稿人。

报告题目:

向抗视觉混淆的主动目标跟踪迈进

内容简介:

In active visual tracking, it is notoriously difficult when distracting objects appear, as distractors often mislead the tracker by occluding the target or bringing a confusing appearance. To address this issue, we propose a mixed cooperative-competitive multi-agent game, where a target and multiple distractors form a collaborative team to play against a tracker and make it fail to follow. Through learning in our game, diverse distracting behaviors of the distractors naturally emerge, thereby exposing the tracker's weakness, which helps enhance the distraction-robustness of the tracker.For effective learning, we then present a bunch of practical methods, including a reward function for distractors, a cross-modal teacher-student learning strategy, and a recurrent attention mechanism for the tracker. The experimental results show that our tracker performs desired distraction-robust active visual tracking and can be well generalized to unseen environments. We also show that the multi-agent game can be used to adversarially test the robustness of trackers.

孙胜扬

加拿大多伦多大学的五年级在读博士生,主要研究方向是基于贝叶斯理论的不确定性估计,侧重于高斯过程,核方法,以及贝叶斯深度学习。曾在ICML,ICLR,NeurIPS,AISTATS等会议上发表文章。

报告题目:

基于谐波核分解的变分高斯过程

内容简介:

We introduce a new scalable variational Gaussian process approximation which provides a high fidelity approximation while retaining general applicability. We propose the harmonic kernel decomposition (HKD), which uses Fourier series to decompose a kernel as a sum of orthogonal kernels. Our variational approximation exploits this orthogonality to enable a large number of inducing points at a low computational cost. We demonstrate that, on a range of regression and classification problems, our approach can exploit input space symmetries such as translations and reflections, and it significantly outperforms standard variational methods in scalability and accuracy. Notably, our approach achieves state-of-the-art results on CIFAR-10 among pure GP models.

# 今日视推荐 #

直播结束后我们会邀请讲者在微信群中与大家答疑交流,请添加“AI TIME小助手(微信号:AITIME_HY)”,回复“icml”,将拉您进“AI TIME ICML 会议交流群”!

AI TIME微信小助手

主       办:AI TIME

合作媒体:学术头条、AI 数据派

合作伙伴:智谱·AI、中国工程院知领直播、学堂在线、学术头条、biendata、 Ever链动

AI TIME欢迎AI领域学者投稿,期待大家剖析学科历史发展和前沿技术。针对热门话题,我们将邀请专家一起论道。同时,我们也长期招募优质的撰稿人,顶级的平台需要顶级的你,

请将简历等信息发至yun.he@aminer.cn!

微信联系:AITIME_HY

AI TIME是清华大学计算机系一群关注人工智能发展,并有思想情怀的青年学者们创办的圈子,旨在发扬科学思辨精神,邀请各界人士对人工智能理论、算法、场景、应用的本质问题进行探索,加强思想碰撞,打造一个知识分享的聚集地。

更多资讯请扫码关注

我知道你“在看”哟~

点击 阅读原文 预约直播

直播预告|ICML专场最后一场啦!来蹲守直播间呀相关推荐

  1. 直播预告| ICML专场四~

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 9月16日 15:00~21:00 AI TIME特别邀请了多位PhD,带来ICML-4! 哔哩哔哩直播通道 扫码关注AITIME哔哩哔 ...

  2. 直播预告 | NeurIPS 专场一 青年科学家专场

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 01 预告一:NeurIPS 2021专场一 2月17日下午2点,本期AI TIME NeurIPS邀请了来自宾夕法尼亚州立大学.伊利诺 ...

  3. 直播预告 | NeurIPS 专场七 青年科学家专场

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 01 预告一:NeurIPS 2021专场七 3月10日下午2点,本期AI TIME NeurIPS邀请了来自蒙特利尔学习算法研究所.杜 ...

  4. 直播预告 | NeurIPS 专场六 青年科学家专场

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 01 预告一:NeurIPS 2021专场六 3月9日下午2点,本期AI TIME NeurIPS邀请了来自加州大学尔湾分校.加州大学伯 ...

  5. 直播预告 | NeurIPS 专场二 青年科学家专场

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 01 预告一:NeurIPS 2021专场二 2月23日下午2点,本期AI TIME NeurIPS邀请了来自多伦多大学.北京大学.斯坦 ...

  6. 直播预告 | NeurIPS 专场八

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 3月16日下午2点,本期AI TIME NeurIPS邀请了来自加州大学旧金山分校.哈佛大学.亚利桑那州立大学.西安交通大学.埃默里大学 ...

  7. 直播预告 | ICLR专场二

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 『今日视频推荐』 5月26日晚7:30-9:00 AI TIME特别邀请了三位优秀的讲者跟大家共同开启ICLR专场二! 哔哩哔哩直播通道 ...

  8. 直播预告 | NeurIPS 专场三 青年科学家专场

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 01 预告一:NeurIPS 2021专场三 2月24日下午2点,本期AI TIME NeurIPS邀请了来自莱斯大学.清华大学.佐治亚 ...

  9. 直播预告 | ICLR专场一

    点击蓝字 关注我们 AI TIME欢迎每一位AI爱好者的加入! 5月21日晚7:30-9:00 AI TIME特别邀请了三位优秀的讲者跟大家共同开启ICLR专场一! 哔哩哔哩直播通道 扫码关注AITI ...

最新文章

  1. 方法 注释_注释模板导入操作方法
  2. [翻译完成] Pi+Raspbian+CGMiner+TightVNC+WIFI五合一矿机超级教程
  3. 高逼格的画图:VIM原来可以这样玩
  4. C#——《C#语言程序设计》实验报告——数据库编程——基于ADO.NET技术的数据库操作DEMO
  5. Lync Server 2010的部署系列_第三章 证书、架构、DNS规划
  6. .NET Core log4net 使用
  7. java 修饰_Java 修饰符
  8. 大数据每日哔哔、新征程
  9. vscode beautify sass 格式化_最强编辑器 VSCode 系列之插件推荐【不定期更新】
  10. node定时任务php 卡,在NodeJs中如何实现定时任务(详细教程)
  11. Linux源码安装Python3.7出现的各种坑
  12. 游戏设计情境探秘之动画
  13. linux下环境变量设置的问题
  14. php 创建任务程序命令行,PHP Yii 命令行程序以及定时任务详解
  15. 如何强制使用某一大小的包去ping某个IP地址?
  16. 计算机网络四个阶段的主要特征,1计算机的四个发展阶段的主要特征
  17. Linux epoll 实现封装
  18. 【STM32H7】第30章 ThreadX GUIX炫酷实用的时钟表盘设计,结合硬件RTC实时时钟
  19. linux csgo 是英文,linux csgo服务器
  20. 优秀WinRT用户界面开发工具推荐

热门文章

  1. Redis 的高并发实战:抢购系统 --浅奕
  2. 2022年深圳市科技型中小微企业贷款贴息资助标准及申报条件,补贴100万
  3. Linux网络包接收过程
  4. 读懂DeFi四大金融原语的演变:流动性、杠杆、风险和套利
  5. Power Apps函数笔记
  6. 龙芯粉的“芯”动日记
  7. 佐治亚理工学计算机硕士,佐治亚理工学院电子与计算机工程硕士专业
  8. 【渝粤题库】陕西师范大学201511先秦历史散文研究 作业(专升本)
  9. HTML网页内嵌入网页
  10. 汽车VIN码与二手车市场