AI

Exploring Frontiers of AI at ICLR 2018

Share this post:

IBM Research is a pioneer across many aspects of AI. At the 6th International Conference on Learning Representations (ICLR 2018), our team will share recent discoveries in learning data representations, techniques that are key to the success of machine learning algorithms. These techniques enable machine learning systems to automatically discover how to represent raw data for subsequent analysis. Learning data representations is an important learning task that powers computer vision, speech recognition, natural language processing, drug design, and other advances in AI. At ICLR 2018, IBM Research will present technical papers on adversarial learning, self-organizing networks for multi-task learning, open-domain question answering, disentanglement of representations, reinforcement learning, and deep learning for graphical data. Below are details on the papers IBM Research will be presenting at ICLR 2018.

In one paper, IBM Research will be introducing the concept of a “routing network,” a kind of self-organizing neural network consisting of two components: a router and a set of function blocks. Given an input, the router makes a routing decision, choosing a function block to apply and passing the output back to the router recursively. This allows similar tasks to share layers, while enabling differing tasks to choose distinct layers. Routing networks perform much better than alternative approaches on a set of multi-task learning benchmarks.

Another IBM AI project featured at ICLR 2018 involves a new metric called CLEVER that can be used to evaluate the robustness of a neural network against attack. The CLEVER score calculates the minimum strength that an adversarial attack would need in order to succeed at deceiving the neural network. Higher scores indicate more robust neural networks. The scores are attack-agnostic, meaning they can be applied to existing and unseen attacks, and are computationally feasible for large neural networks. CLEVER scores can be used to compare the robustness of different network designs and training procedures to help researchers build more reliable AI systems.

IBM Research is a gold sponsor of ICLR 18. We will be at the conference expo at booth #500, where you can try your hand at attacking imaginary banks’ AI check image processing systems by distorting check digits and using CLEVER scores to guide your decisions and maximize your profits.
Big Check demo

Accepted papers at ICLR 2018

Evaluating the Robustness of Neural Networks: An Extreme Value Theory Approach
Tsui-Wei Weng, Huan Zhang, Pin-Yu Chen, Jinfeng Yi, Dong Su, Yupeng Gao, Cho-Jui Hsieh, Luca Daniel

Sobolev GAN
Youssef Mroueh, Chun-Liang Li, Tom Sercu, Anant Raj, Yu Cheng

Routing Networks: Adaptive Selection of Non-linear Functions for Multi-Task Learning
Clemens Rosenbaum, Tim Klinger, Matthew Riemer

Evidence Aggregation for Answer Re-Ranking in Open-Domain Question Answering
Shuohang Wang, Mo Yu, Jing Jiang, Wei Zhang, Xiaoxiao Guo, Shiyu Chang, Zhiguo Wang, Tim Klinger, Gerald Tesauro, Murray Campbell

Variational Inference of Disentangled Latent Concepts from Unlabeled Observations
Abhishek Kumar, Prasanna Sattigeri, Avinash Balakrishnan

Eigenoption Discovery through the Deep Successor Representation
Marlos C. Machado, Clemens Rosenbaum, Xiaoxiao Guo, Miao Liu, Gerald Tesauro, Murray Campbell

FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling
Jie Chen, Tengfei Ma, Cao Xiao

More AI stories

SpotTune: Transfer Learning through Adaptive Fine-Tuning

New techniques make fine tuning an AI model more efficient when doing transfer learning

Continue reading

IBM Research AI at CVPR 2019

The annual conference on Computer Vision and Pattern Recognition (CVPR 2019) takes place June 16–20 in Long Beach, CA. IBM Research AI will present technical papers describing our latest results in our quest to give AI systems sight. Our research explores a variety of areas within computer vision and multimedia, with a focus on multimodal […]

Continue reading

Beyond Backprop: Online Alternating Minimization with Auxiliary Variables

Since its introduction in the 1970s, the backpropagation algorithm (backprop) has been the workhorse for training neural networks has and contributed to impressive successes in deep learning for a wide range of applications. Backprop plays an important role in enabling neural networks to track the errors they make, learn from those mistakes and improve over […]

Continue reading