IBM researchers published the first major release of the Adversarial Robustness 360 Toolbox (ART). Initially released in April 2018, ART is an open-source library for adversarial machine learning that provides researchers and developers with state-of-the-art tools to defend and verify AI models against adversarial attacks. ART addresses growing concerns about people’s trust in AI, specifically the security of AI in mission-critical applications.
Convex optimization problems, which involve the minimization of a convex function over a convex set, can be approximated in theory to any fixed precision in polynomial time. However, practical algorithms are known only for special cases. An important question is whether it is possible to develop algorithms for a broader subset of convex optimization problems that are efficient in both theory and practice.
IBM researchers have partnered with scientists from MIT, Northeastern University, Boston University and University of Minnesota to publish two papers on novel attacks and defenses for graph neural networks and on a new robust training algorithm called hierarchical random switching at IJCAI 2019.
We report new research results relevant to AI planning in our paper, "Depth-First Memory-Limited AND/OR Search and Unsolvability in Cyclic Search Spaces," presented at the International Joint Conference on Artificial Intelligence, IJCAI-19.
At IJCAI'19, IBM researchers present new results on causal knowledge extraction from large amounts of text for applications in enterprise risk management.
At ACL 2019, IBM researchers released a paper detailing the model they trained to answer complex questions using Neural Program Induction, which allows an AI model can be taught to procedurally decompose a complex task into a program.
There is a growing number of adversarial attacks and nefarious behaviors aimed at AI systems. To combat this, IBM Research AI will present multiple papers that yield new scientific discoveries and recommendations related to adversarial learning at KDD 2019.