| CARVIEW |
| CVPR-19 Workshop on Explainable AI |
Deep neural networks (DNNs) have no doubt brought great successes to a wide range of applications in computer vision, computational linguistics and AI. However, foundational principles underlying the DNNs’ success and their resilience to adversarial attacks are still largely missing. Interpreting and theorizing the internal mechanisms of DNNs becomes a compelling yet controversial topic. The statistical methods and rule-based methods for network interpretation have much to offer in semantically disentangling inference patterns inside DNNs and quantitatively explaining the decisions made by DNNs. Rethinking DNNs explicitly toward building explainable systems from scratch is another interesting topic, including new neural architectures, new parameter estimation methods, new training protocols, and new interpretability-sensitive loss functions.
This workshop aims to bring together researchers, engineers as well as industrial practitioners, who concern about interpretability, safety, and reliability of artificial intelligence. Joint force efforts along this direction are expected to open the black box of DNNs and, ultimately, to bridge the gap between connectionism and symbolism of AI research. The main theme of the workshop is therefore to build up consensus on a variety of topics including motivations, typical methodologies, prospective innovations of transparent and trustworthy AI. Research outcomes are also expected to have profound influences on critical industrial applications such as medical diagnosis, finance, and autonomous driving.
Schedule
June 16, 2019
Hyatt, Beacon, A
| 08:40 - 08:45 | Welcome |
|   |   |
| 08:45 - 09:15 | Invited talk: Dr. Song-Chun Zhu, University of California, Los Angeles |
|   |   |
| 09:15 - 09:45 | Invited talk: Dr. Klaus-Robert Muller, TU Berlin |
|   |   |
| 09:45 - 10:15 | Invited talk: Dr. Kate Saenko, Boston University (PPT) |
|   |   |
| 10:15 - 10:30 | Coffee break |
|   |   |
| 10:30 - 11:00 | Invited talk: Dr. Devi Parikh & Dr. Dhruv Batra, Georgia Tech (PPT) |
|   |   |
| 11:00 - 11:30 | Invited talk: Dr. Been Kim, Google Brain (PPT) |
|   |   |
| 11:30 - 13:00 | Poster session |
Topics
Topics of interests include, but are not limited to, following fields
- Theories of interpretable AI models.
- Visualizing feature representations in deep neural networks.
- Deep coupling of neural networks and grammars or graphical models
- Deep coupling of AI models and the theory of mind
- Qualitative and quantitative diagnosis and analysis of the decision-making process of deep models.
- Probabilistic logic interpretation of deep learning.
- Causality reasoning and learning
- Safety and fairness of artificial intelligence models.
- Industrial applications of trustworthy AI, e.g. in medical diagnosis, autonomous driving, and finance.
- Evaluation of interpretable AI systems.
All above topics are core issues in the development of explainable AI and have received an increasing attention in recent years. We believe these topics will receive broad interests in fields of computer vision and machine learning.
Calling for papers
This workshop is a half-day event, which will include invited talks, spotlight and poster presentations of accepted papers. We are calling for extended abstracts with 2–4 pages. Papers accepted by this workshop can be re-submitted to other conferences or journals.Please submit your papers to cvpr19xai@gmail.com
- Submission deadline: March, 30th, 2019
- Notification date: April, 11st, 2019
- Camera-ready deadline: April 18th, 2019
Organizers
Please contact Quanshi Zhang if you have question.





