Thirty-Fifth Conference on Artificial Intelligence
February 3, 2021
What Is the Tutorial Forum?
The Tutorial Forum provides an opportunity for researchers and practitioners to explore exciting advances in disciplines outside their normal focus. We believe this type of forum is essential for the cross fertilization, cohesiveness, and vitality of the AI field. We all have a lot to learn from each other; the Tutorial Forum promotes the continuing education of each member of AAAI.
MQ1: Explainable AI for Societal Event Predictions: Foundations, Methods, and Applications
Songgaojun Deng, Yue Ning and Huzefa Rangwala
The goal of this tutorial is to bring together data scientists, AI researchers, and social scientists to discuss research problems and challenges in computational event modeling. Traditional machine learning models for societal event forecasting focus on predictive performance using structured data. We present new directions of interpretable AI/ML models for predictive analysis in dynamic, heterogeneous, and multi-source event data. The tutorial will cover technical material related to event detection, forecasting, and precursor identification with assumed preliminary knowledge in supervised learning, deep learning, and temporal event modeling. First, we will introduce formal definitions of event prediction and event precursors, along with a brief discussion on the methods applied in this field. Next, we will present recently developed deep learning methods including dynamic graph convolutional networks in event forecasting and actor (participant) inference. Last, we will discuss real-world applications based on temporal event modeling.
Songgaojun Deng
Stevens Institute of Technology
Songgaojun Deng is a Ph.D. candidate in the Department of Computer Science at Stevens Institute of Technology. Her research interests are machine learning and deep learning in social informatics and health informatics. Her current research focuses on developing graph neural networks to capture dynamic and interpretable graph-based patterns.
Yue Ning
Stevens Institute of Technology
Yue Ning is an Assistant Professor in the Department of Computer Science at Stevens Institute of Technology. She received her Ph.D. degree in Computer Science from Virginia Tech. Her research interests are machine learning, data analytics, and social media analysis.
Huzefa Rangwala
George Mason University
Huzefa Rangwala is the Lawrence Cranberg Faculty Fellow and Professor in the Department of Computer Science at George Mason University. His research interests are data mining and machine learning applications in the areas of biological sciences, learning sciences, and cyber-physical sciences involving fundamental contributions of multitask learning, hierarchical classification, and recommender systems.
MQ2: An Extensible Toolkit for Explainable Recommender Systems
Ludovik Coba, Roberto Confalonieri and Markus Zanker
In this tutorial, we will provide a review of eXplainable Recommender Systems (XRSs) and a practical hands-on session using a recently developed software toolkit called RecoXplainer. RecoXplainer is a unified, extendable, and easy-to-use Python toolkit that includes several explainability techniques that are useful for various groups of stakeholders.
The tutorial will be organized in two parts. The first part will provide the necessary background knowledge about Explainable AI and explanations techniques in recommender systems, such as model-based, post-hoc explanations, offline and online evaluation of explanations. The second part will consist of a hands-on session where attendees will be guided to the use of RecoXplainer, and to generate several examples of explainable recommendations.
The target audience is assumed to have a certain acquaintance with recommender systems, Matrix Factorisation, and information retrieval. For the hands-on session a basic understanding of Python is required. A more advanced programming knowledge is required for those who would like to experiment with the toolkit, e.g., for inspecting the code to find advanced features, or to learn how to extend the current framework.
Ludovik Coba
Free University of Bozen-Bolzano
Ludovik Coba is a post-doctoral researcher at the Free University of Bozen-Bolzano. He currently performs research in explainability and innovative algorithms for recommender systems. He publishes his results in venues like RecSys or IUI and journals like IT and Tourism, UMUAI or DKE.
Roberto Confalonieri
Free University of Bozen-Bolzano
Roberto Confalonieri is Assistant Professor at the Free University of Bozen-Bolzano. His current research focuses on Explainable AI, in particular, on how the integration of symbolic and non-symbolic reasoning approaches can convey human-understandable explanations of black-box models.
Markus Zanker
Free University of Bozen-Bolzano
Markus Zanker is a professor at the Faculty of Computer Science of the Free University of Bozen-Bolzano, where he also served as vice dean for studies and director for study programmes. Before moving to Bolzano he was an associate professor at the Alpen-Adria-Universitaet Klagenfurt, Austria. His research focuses on knowledge-based information systems supporting decision making processes such as personalized information filtering and retrieval and product recommendation. Until 2013 Markus Zanker was also a co-founder and director of a recommendation service company for more than 10 years. He is an associate editor of Information Technology & Tourism (Springer). Besides organizing numerous workshops in the field, he was a program chair of the 4th ACM Conference on Recommender Systems in 2010 and he co-chaired the 9th ACM Conference on Recommender Systems in 2015.
MQ3: Deep Randomized Neural Networks
Claudio Gallicchio
Deep Neural Networks (DNNs) are a fundamental tool in the modern development of Machine Learning. Beyond the merits of the training algorithms, a great part of DNNs success is due to the inherent properties of their layered architectures, i.e., to the introduced architectural biases. In this tutorial, we explore recent classes of DNN models wherein the majority of connections are randomized or more generally fixed according to some specific heuristic. Limiting the training algorithms to operate on a reduced set of weights implies intriguing features. Among them, the extreme efficiency of the learning processes is undoubtedly a striking advantage with respect to fully trained architectures. Besides, despite the involved simplifications, randomized neural systems possess remarkable properties both in practice, achieving state-of-the-art results in multiple domains, and theoretically, allowing to analyze intrinsic properties of neural architectures.
This tutorial covers all the major aspects regarding Deep Randomized Neural Networks, from feed-forward and convolutional neural networks, to dynamically recurrent deep neural systems for structures. The tutorial is targeted to both researchers and practitioners, from academia or industry, who are interested in developing DNNs that can be trained efficiently, and possibly embedded into low-powerful devices.
Requirements: basics on Machine Learning and DNNs.
Claudio Gallicchio
University of Pisa
Claudio Gallicchio is Assistant Professor at the Department of Computer Science, University of Pisa. His research interests include Machine Learning, Deep Learning, Randomized Neural Networks, Reservoir Computing, Recurrent and Recursive Neural Networks, Graph Neural Networks. He is founder and chair of the IEEE CIS Task Force on Reservoir Computing.
MQ4: Efficient Neural Architecture Search
Tejaswini Pedapati and Martin Wistuba
With the advances and success of deep learning technology in solving complex AI problems, the natural step forward lies in building systems that automate the decisions required for setting up a deep learning pipeline. Tackling this automation is not only crucial for speeding up the deployment of deep learning models, it also helps in expanding the capabilities of models to other challenging scenarios like modelling with limited data or modelling under resource constraints. This tutorial seeks to provide a comprehensive overview of the approaches used in this regard by means of neural architecture search. It is also the first tutorial that strongly focuses on transfer and meta-learning, going beyond classic neural architecture search.
The tutorial is geared toward graduate students, AI researchers, and practitioners, who are interested in automating parts of their deep learning pipelines, want to learn about principles of automated machine learning and deep learning, and apply those principles to make their own work more effective and less arduous.
The prerequisite knowledge assumed of the audience includes basic understanding of deep learning, optimization, and machine learning concepts. Familiarity with some state-of-the-art convolutional neural network architecture can facilitate the understanding but is not required.
Tejaswini Pedapati
IBM Research
Tejaswini Pedapati works at IBM Research. Her research is focused on interpretability and automating deep learning. To that end, she was involved in developing tools and algorithms to provide these capabilities for IBM products. She has a masters’ degree from Columbia University.
Martin Wistuba
IBM Research
Martin Wistuba is a researcher at IBM Research, where he develops tools to automate deep learning. Previously, he received his Ph.D. in Machine Learning from the University of Hildesheim. His research interest includes AutoML, in particular the idea of meta-knowledge transfer to speed up Bayesian optimization and Neural Architecture Search.
MH1: Artificial Intelligence for Drug Discovery
Jian Tang, Fei Wang and Feixiong Cheng
Drug discovery is a long and costly process, taking on average 10 years and 2.5 billion dollars to develop a new drug. Artificial intelligence has the potential to significantly accelerate the process of drug discovery by analyzing a large amount of data generated in the biomedical domain. In this tutorial, we will provide a detailed introduction to key problems in drug discovery such as molecular property prediction, de novo molecular design and molecular optimization, retrosynthesis reaction and prediction, and drug repurposing and combination, and also key technique advancements with artificial intelligence for these problems. This tutorial can be served as introduction materials for both computer scientists interested in drug discovery as well as drug discovery practitioners for learning the latest AI techniques along this direction.
The intended audience for the tutorial is participants that want to get up-to-speed with recent advancements in AI for drug discovery. The tutorial is targeted towards an entry-level audience, with knowledge of the fundamentals of machine learning and deep learning; e.g., empirical risk minimization and familiarity with LSTMS or CNNs. However, we expect that experts in the area will also benefit from the tutorial’s synthesis of disparate techniques and problems.
Jian Tang
Mila-Quebec AI Institute
Jian Tang is currently an assistant professor at Mila-Quebec AI Institute, led by Turing Award Winner Yoshua Bengio. His research focuses on graph representation learning, graph neural networks, drug discovery, and knowledge graphs. He is a recipient of the CIFAR AI Research Chair. He received the best paper award of ICML’14 and was nominated for the best paper of WWW’16. He published quite a few representative works on graph representation learning (including LINE, LargeVis, RotatE) and recently has been very actively working on graph representation learning for drug discovery.
Fei Wang
Cornell University
Fei Wang is currently an Associate Professor of Health Informatics in the Department of Population Health Sciences, Weill Cornell Medicine, Cornell University. His major research interest is data mining and its applications in health data science. He has published more than 250 papers in AI and medicine and has won 8 best paper awards at top international conferences on data mining and medical informatics. Dr. Wang is the recipient of the NSF CAREER Award in 2018, the inaugural research leadership award in IEEE International Conference on Health Informatics (ICHI) 2019.
Feixiong Cheng
Cleveland Clinic
Feixiong Cheng, PhD, is a principal investigator with Cleveland Clinic’s Genomic Medicine Institute. Dr. Cheng is working to develop computational and experimental network medicine technologies for advancing the characterization of disease heterogeneity, thereby approaching the goal of coordinated, patient-centered strategies to innovative diagnostics and therapeutics development. Dr. Cheng has received several awards, including NIH Pathway to Independence Award (K99/R00), SCI highly cited papers reward, and Vanderbilt Postdoc of the Year Honorable mention.
MH2: Commonsense Knowledge Acquisition and Representation
Filip Ilievski, Antoine Bosselut, Simon Razniewski and Mayank Kejriwal
The tutorial on Commonsense Knowledge Acquisition and Representation will consist of four main components, each covered by one of the presenters, followed by a discussion session. We will start by introducing our work on axiomatizating commonsense knowledge and comparing it to prior theories. The participants would then get familiar with the Commonsense Knowledge Graph (CSKG) (Ilievski et al., 2020), a recent resource that integrates seven diverse commonsense sources. Thirdly, we will discuss how commonsense knowledge can be automatically extracted, as well as qualified with DICE (Chalier et al., 2020) metrics like typicality and salience. We will use COMET (Bosselut et al., 2019) to improve the completeness of commonsense knowledge graphs based on language models, like BERT and GPT-2, and discuss how it can be leveraged on downstream question answering (QA) tasks. We will conclude the tutorial with a discussion of the way forward, and propose to combine language models (LMs), knowledge graphs (KGs), and axiomatization in the next-generation commonsense reasoning techniques. Prior knowledge expected from participants will be minimal. Some knowledge of machine learning and language modeling will be helpful, but not compulsory: we will introduce relevant machine learning concepts so that everyone has an opportunity to follow along.
Filip Ilievski
USC Viterbi School of Engineering
Filip Ilievski is a Computer Scientist within the USC Information Sciences Institute and Lecturer at the USC Viterbi School of Engineering. He obtained his Ph.D. at Vrije Universiteit Amsterdam. His research focuses on representing and consolidating commonsense knowledge and applying this knowledge to understand gaps in human communication.
Antoine Bosselut
EPFL
Antoine Bosselut is a Postdoctoral Researcher at Stanford University and will join EPFL as an Assistant Professor starting in Fall 2021. He completed his PhD at the University of Washington. His research interests are in building systems that can mix knowledge and language representations to solve problems in NLP, specializing in commonsense representation and reasoning.
Simon Razniewski
Max Planck Institute for Informatics
Simon Razniewski is senior researcher at the Max Planck Institute for Informatics in Saarbrücken, where he heads the knowledge base construction and quality area. He was previously assistant professor at the Free University of Bozen Bolzano. His research interests include commonsense knowledge base construction, information extraction, and data quality.
Mayank Kejriwal
USC Information Sciences Institute
Mayank Kejriwal is a research assistant professor in the department of Industrial and Systems Engineering at USC, and a research lead in the USC Information Sciences Institute. His research focuses on knowledge graphs and their applications, especially in social contexts such as fighting human trafficking and crisis response.
MH3: Dealing with Bias and Fairness in Building AI Systems: A Practical Hands-on Tutorial
Rayid Ghani, Pedro Saleiro and Kit Rodolfa
There is a need for training AI researchers and practitioners on how to deal with bias and fairness in practice through the entire lifecycle of an AI system, from the early stages of a project up to maintaining the system in production. This tutorial will help participants learn how to think about overall fairness of an AI system, from project scoping (through informed discussions with stakeholders) to thinking about the right metrics (based on the use case context), to performing bias audits and exploring different bias reduction strategies, to continuous monitoring to assess the need for retraining.
The discussions, breakout sessions, and hands-on coding sessions will involve real-world case studies based on our work with over 100 government agencies and non-profits over the last several years and will include working through Python notebooks, taking the participants through each step of the process.
The tutorial will have the following prerequisites: 1) prior programming experience in Python, and 2) prior experience in machine learning, with an understanding of different types of ML models as well as experience building those models.
Rayid Ghani
CMU
Rayid Ghani is a Distinguished Career Professor at CMU, working on the use of AI in solving public policy and social challenges in a fair and equitable manner. Rayid started the Data Science for Social Good Fellowship that trains people from around the world to work on AI problems with social impact.
Pedro Saleiro
Feedzai
Pedro Saleiro is a Senior Research Manager at Feedzai where he leads the Fairness, Accountability, Transparency, and Ethics research group. He is responsible for several initiatives related to bias auditing and algorithmic fairness, model explainability and ML governance. Previously, Pedro worked with Rayid Ghani as postdoc at UChicago.
Kit Rodolfa
CMU
Kit Rodolfa is a Senior Research Scientist at CMU working with Rayid Ghani. He previously led the data science efforts at Devoted Health, served as Chief Data Scientist at Hillary for America, and as Director of Digital Analytics for the Office of Digital Strategy in the Obama White House.
MH4: Designing Agents’ Preferences, Beliefs, and Identities
Vincent Conitzer
We often assume that each agent has a well-defined identity, well-defined preferences over outcomes, and well-defined beliefs about the world. However, when designing agents, we in fact need to specify where the boundaries between one agent and another in the system lie, what objective functions these agents aim to maximize, and to some extent even what belief formation processes they use. What is the right way to do so? As more and more AI systems are deployed in the world, this question becomes increasingly important. In this tutorial, I will show how it can be approached from the perspectives of decision theory, game theory, social choice theory, and the algorithmic and computational aspects of these fields. (No previous background required.)
AAAI’19 blue sky writeup on a subset of these ideas can be found here: https://users.cs.duke.edu/~conitzer/designingAAAI19.pdf
Vincent Conitzer
Duke University
Vincent Conitzer is the Kimberly J. Jenkins University Professor of New Technologies and Professor of Computer Science, Professor of Economics, and Professor of Philosophy at Duke University. He received the IJCAI Computers and Thought Award in 2011, and is a Fellow of AAAI and ACM.
MH5: Learning with Small Data
Huaxiu Yao, Fei Wang and Zhenhui Jessie Li
In the era of big data, data-driven methods have become increasingly popular. The superior performance of these data-driven approaches relies on large-scale labeled training data. But many real-world applications often face the “small (labeled) data” challenge. For example, precision medicine aims to provide tailored treatment or management plans to individual patients, but we only have data from a limited number of individuals (patients). For another example, smart city technologies aim at using data to make better decisions in applications such as transportation, security, and energy. But a city is a highly complex, interconnected, and dynamic system and we often do not have enough data from different sources to model a holistic view at the city scale.
In this tutorial, we will review the trending machine learning techniques for learning with small (labeled) data. These techniques are organized from three aspects: (1) providing a comprehensive review of recent studies about knowledge generalization, transfer, and sharing; (2) introducing the cutting-edge techniques which focus on incorporating domain knowledge into machine learning models; (3) discussing the small data challenges and solutions in real-world applications on precision medicine and smart city.
Huaxiu Yao
Pennsylvania State University
Huaxiu Yao is currently a Ph.D. candidate of the College of Information Sciences and Technology at the Pennsylvania State University. He obtained his B.Eng. degree from the University of Electronic Science and Technology of China. He also spent time in Amazon A9, Salesforce Research, Alibaba DAMO Academy, Tencent AI Lab and Didi AI Labs. His current research goal is to enable agents to learn quickly and efficiently via knowledge transfer and structure exploration. He is also passionate about applying these methods for solving real-world problems, especially in smart city and E-commerce. He has published over 15 papers on top conferences and journals such as ICML, ICLR, NeurIPS, KDD, AAAI, WWW and WSDM. He has served as a program committee member in major machine learning and data mining conferences such as ICML, ICLR, NeurIPS, KDD, AAAI, IJCAI.
Fei Wang
Cornell University
Fei Wang is an Associate Professor at Department of Population Health Sciences, Weill Cornell Medicine, Cornell University. His major research interest is data mining, machine learning and their applications in health data science. He has published on the top venues of related areas such as ICML, KDD, NeurIPS, AAAI, JAMA Internal Medicine, Annals of Internal Medicine, etc. His papers have received over 13,400 citations so far with an H-index 57. His (or his students’) papers have won 7 best paper (or nomination) awards at international academic conferences. His team won the championship of the NIPS/Kaggle Challenge on Classification of Clinically Actionable Genetic Mutations in 2017 and Parkinson’s Progression Markers Initiative data challenge organized by Michael J. Fox Foundation in 2016. Dr. Wang is the recipient of the NSF CAREER Award in 2018, as well as the inaugural research leadership award in IEEE International Conference on Health Informatics (ICHI) 2019. Dr. Wang is the chair of the Knowledge Discovery and Data Mining working group in American Medical Informatics Association (AMIA). Dr. Wang’s research has been supported by funding agencies including NSF, NIH, ONR, NMRC, MJFF, AHA , PCORI and industries including Amazon, Google, Boehringer Ingelheim and MITRE.
Zhenhui Li
UPenn
Zhenhui Li is an associate professor of Information Sciences and Technology at the Pennsylvania State University. Before joining Penn State, she received her Ph.D. degree in Computer Science from University of Illinois Urbana-Champaign in 2012. Her research has been focused on mining spatial-temporal data with applications in transportation, ecology, environment, social science, and urban computing. She is a passionate interdisciplinary researcher and has been actively collaborating with cross-domain researchers. She has served as organizing committee or senior program committee of many conferences including KDD, ICDM, SDM, CIKM, and SIGSPATIAL. She has been regularly offering classes on data organizing and data mining since 2012. Her classes have constantly received high student ratings. She has received NSF CAREER award, junior faculty excellence in research, and George J. McMurtry junior faculty excellence in teaching and learning award.
MH6: Meta Learning
Iddo Drori and Joaquin Vanschoren
Meta-learning allows machines to learn to learn new algorithms. It is an emerging and fast developing research area within machine learning with implications for all AI research. Recent successes include automatic model discovery, few-shot learning, multi-task learning, meta-reinforcement learning, as well as teaching machines to read, learn and reason. Just as humans do not learn new tasks from scratch, but rather draw on what they learn before, meta-learning is key to efficient and robust learning. This tutorial will cover important mathematical foundations of the field and its applications, including key methods underlying current state of the art in this fast-paced field that is increasingly relevant for a broad range of AAAI attendees.
The intended audience comprises both AI researchers and practitioners. For AI researchers, this tutorial presents a comprehensive overview of existing methods and how they are combined in novel ways. We explore the interactions between the sub-fields of meta-learning, automated machine learning, transfer learning, adaptation, multi-task learning, and continual learning. For practitioners and attendees from industry, we explain how meta-learning enables novel applications, such as building AI systems that learn models from very small amounts of data, that learn across tasks to get increasingly better, and enable disruptive applications.
Iddo Drori
MIT
Iddo Drori is Lecturer at MIT EECS. He holds a Ph.D in Computer Science and did his postdoc at Stanford in Statistics. He was Associate Professor at Colman; Lecturer at Tel Aviv University; visiting and adjunct Associate Professor at NYU, Columbia University, and Cornell University, working on machine learning.
Joaquin Vanschoren
Eindhoven University of Technology
MH7: Recent Advances in Integrating Machine Learning and Combinatorial Optimization
Elias Khalil, Andrea Lodi, Bistra Dilkina, Didier Chételat, Maxime Gasse, Antoine Prouvost, Giulia Zarpellon and Laurent Charlin
This tutorial will provide an overview of the recent impact machine learning is having on combinatorial optimization, particularly under the Mixed Integer Programming (MIP) framework. Topics covered will include ML and reinforcement learning for predicting feasible solutions, improving exact solvers with ML, a software framework for learning in exact MIP solvers, and the emerging paradigm of decision-focused learning.
The tutorial targets both junior and senior researchers in two prominent areas of interest to the AAAI community:
- Machine learning researchers looking for a challenging application domain, namely combinatorial optimization;
- Optimization practitioners and researchers who may benefit from learning about recent advances in ML methods for improving combinatorial optimization algorithms.
Basic prerequisites of the tutorial include:
- Combinatorial optimization: basic understanding of optimization modeling, algorithm design, computational complexity.
- Machine learning: basic knowledge of paradigms such as supervised and reinforcement learning; common techniques such as neural networks.
Elias B. Khalil
University of Toronto
Elias B. Khalil is an Assistant Professor of Industrial Engineering at the University of Toronto since July 2020. Prior to that, he was the IVADO Postdoctoral Scholar at Polytechnique Montréal. Elias obtained his Ph.D. from the College of Computing at Georgia Tech, where he was an IBM Ph.D. Fellow (2016).
Andrea Lodi
Polytechnique Montréal
As Canada Excellence Research Chair in Data Science for Real-Time Decision-Making at Polytechnique Montréal, Dr. Andrea Lodi holds Canada’s main chair in operations research. He has been Herman Goldstine Fellow at IBM in 2005–2006. He is author of more than 80 publications in the top journals of Mathematical Optimization.
Bistra Dilkina
USC
Bistra Dilkina is a co-Director of the Center for AI in Society (CAIS) at the University of Southern California (USC), and an Associate Professor of Computer Science at the Viterbi School of Engineering. Her work spans discrete optimization, network design, and machine learning with a strong focus on sustainability applications.
Didier Chételat
Polytechnique Montréal
Didier Chételat is a machine learning researcher at the Canada Excellence Research Chair in Data Science for Real Time Decision Making, under the leadership of Andrea Lodi. He obtained his Ph.D. in Statistics at Cornell University in 2015. His main research interests center around deep learning for combinatorial optimization.
Maxime Gasse
Polytechnique Montréal
Maxime Gasse is a machine learning researcher within the Canada Excellence Research Chair in Data Science for Real-Time Decision-Making at Polytechnique Montréal, and also part of the MILA research institute on artificial intelligence. His research interests include probabilistic graphical models, structured prediction problems, questions regarding causality, and combinatorial optimization problems.
Antoine Prouvost
Polytechnique Montréal
Antoine Prouvost is a Ph.D. candidate in Polytechnique Montréal under supervision of Prof. Andrea Lodi at CERC DS4DM and Prof. Yoshua Bengio at Mila. I work at the interplay of deep learning and operations research, building combinatorial optimization algorithm that leverage machine learning to adapt to different problem structure.
Giulia Zarpellon
Polytechnique Montréal
Giulia Zarpellon obtained her Ph.D. in Applied Mathematics at Polytechnique Montréal, under the supervision of Prof. Andrea Lodi, and a master’s degree in Mathematics at University of Padova. Her main research interests are in the integration of mathematical optimization and machine learning.
Laurent Charlin
HEC Montréal
Laurent Charlin is an assistant professor of artificial intelligence at HEC Montréal. He earned a master’s and PhD respectively from the universities of Waterloo and Toronto and was a postdoc at Columbia, Princeton and McGill. He develops machine learning models to analyze large collections of data and help in decision-making.
MH8: Responsible AI in Industry: Practical Challenges and Lessons Learned
Krishnaram Kenthapadi, Ben Packer, Mehrnoosh Sameki and Nashlie Sephus
Artificial Intelligence is increasingly being used in decisions and processes that are critical for individuals, businesses, and society, especially in areas such as hiring, lending, criminal justice, healthcare, and education. Recent ethical challenges and undesirable outcomes associated with AI systems have highlighted the need for regulations, best practices, and practical tools to help data scientists and ML developers build AI systems that are secure, privacy-preserving, transparent, explainable, fair, and accountable – to avoid unintended and potentially harmful consequences and compliance challenges.
In this tutorial, we will present an overview of responsible AI, highlighting model explainability, fairness, and privacy in AI, key regulations/laws, and techniques/tools for providing understanding around AI/ML systems. Then, we will focus on the application of explainability, fairness assessment/unfairness mitigation, and privacy techniques in industry, wherein we present practical challenges/guidelines for using such techniques effectively and lessons learned from deploying models for several web-scale machine learning and data mining applications. We will present case studies across different companies, spanning many industries and application domains. Finally, based on our experiences in industry, we will identify open problems and research directions for the AI community.
Krishnaram Kenthapadi
Amazon AWS AI
Krishnaram Kenthapadi is a Principal Scientist at Amazon AWS AI, where he leads the fairness, explainability, and privacy initiatives in Amazon AI platform. Previously, he held positions at LinkedIn AI and Microsoft Research. Krishnaram has published 40+ papers (2500+ citations; awards at conferences) and filed 140+ patents (30+ granted). He received his Ph.D. in Computer Science from Stanford University in 2006.
Ben Packer
Google AI
Ben Packer is a Software Engineer in Research at Google AI, working on Fairness and Robustness at the intersection of research and product development, and has been involved in Google’s Machine Learning Fairness Education effort. Ben received his Ph.D. in Computer Science from the AI lab at Stanford University.
Mehrnoosh Sameki
Microsoft
Mehrnoosh Sameki is a senior technical program manager at Microsoft, responsible for leading the product efforts on machine learning interpretability and fairness within the Azure Machine Learning platform. She earned her PhD degree in computer science at Boston University, where she currently serves as an adjunct assistant professor and lecturer, offering courses in responsible AI. Previously, she was a data scientist in the retail space, incorporating data science and machine learning to enhance customers’ personalized shopping experiences.
Nashlie Sephus
Amazon AWS AI
Nashlie Sephus is the Applied Science manager for Amazon Artificial Intelligence focusing on fairness and identifying biases in AWS AI technologies. She received her Ph.D. from the School of Electrical and Computer Engineering at Georgia Tech in 2014 and her B.S. from Mississippi State University in 2007.
MH9: A Visual Tour of Bias Mitigation Techniques for Word Representations
Sunipa Dev, Jeff M. Phillips, Archit Rathore, Vivek Srikumar and Bei Wang
Word vector embeddings have been shown to contain and amplify biases in data they are extracted from. Consequently, many techniques have been proposed to identify, mitigate, and attenuate these biases in word representations. In this tutorial, we will review a collection of state-of-the-art debiasing techniques. To aid this, we provide an open source web-based visualization tool and offer hands-on experience in exploring the effects of these debiasing techniques on the geometry of high-dimensional word vectors. To help understand how various debiasing techniques change the underlying geometry, we decompose each technique into interpretable sequences of primitive operations and study their effect on the word vectors using dimensionality reduction and interactive visual exploration.
Prerequisite Knowledge The attendees are expected to understand the basics of linear algebra and dimensionality reduction. Familiarity with basic NLP would be helpful but is not required. The audience is not assumed to have knowledge about bias in NLP; however, the tutorial will still be applicable to those who do.
Sunipa Dev
University of Utah
Sunipa Dev received her PhD at the School of Computing at the University of Utah in Fall 2020 and is a CI Postdoctoral Fellow at UCLA. Her research focuses on understanding the structure of language representations and leveraging that to isolate and decouple associations and concept subspaces within.
Jeff M. Phillips
University of Utah
Jeff M. Phillips is an Associate Professor at the School of Computing, and Director of the Utah Center for Data Science, at the University of Utah. He is an expert in the geometry of data, and actively publishes in top venues in machine learning & data mining, algorithms & geometry, and databases.
Archit Rathore
University of Utah
Archit Rathore is a 4th year Ph.D. student at the School of Computing at the University of Utah. His current research focuses on probing machine learning models through visualization techniques to improve interpretability.
Vivek Srikumar
University of Utah
Vivek Srikumar is an Associate Professor at the School of Computing at the University of Utah. His research focuses on machine learning in the context of natural learning processing and has primarily been driven by questions arising from the need to reason about textual data with limited explicit supervision and to scale NLP to large problems.
Bei Wang
University of Utah
Bei Wang is an Assistant Professor at the School of Computing, a faculty member in the Scientific Computing and Imaging (SCI) Institute, University of Utah. Her research interests include data visualization, topological data analysis, computational topology, computational geometry, machine learning, and data mining.
AQ1: Explaining Machine Learning Predictions: State-of-the-Art, Challenges, and Opportunities
Himabindu Lakkaraju, Julius Adebayo and Sameer Singh
As machine learning is deployed in all aspects of society, it has become increasingly important to ensure stakeholders understand and trust these models. Decision-makers must have a clear understanding of the model behavior so they can diagnose errors and potential biases in these models and decide when and how to employ them. However, most accurate models that are deployed in practice are not interpretable, making it difficult for users to understand where the predictions are coming from, and thus, difficult to trust. Recent work on explanation techniques in machine learning offers an attractive solution: they provide intuitive explanations for “any” machine learning model by approximating complex machine learning models with simpler ones.
In this tutorial, we discuss several post hoc explanation methods, and focus on their advantages and shortcomings. We cover three families of techniques: (a) single instance gradient-based attribution methods (saliency maps), (b) model agnostic explanations via perturbations, such as LIME and SHAP, and (c) surrogate modeling for global interpretability, such as MUSE. For each of these approaches, we provide their problem setup, prominent methods, example applications, and finally, discuss their vulnerabilities and shortcomings. We hope to provide a practical and insightful introduction to explainability in machine learning.
Hima Lakkaraju
Harvard University
Hima Lakkaraju is an Assistant Professor at Harvard University working on explaining complex machine learning models to decision-makers (e.g., judges, doctors). Hima was recently named in the 35 innovators under 35 by MIT Tech Review, and has given invited workshop talks at ICML, NeurIPS, AAAI, and CVPR. Website: https://himalakkaraju.github.io/
Julius Adebayo
MIT
Julius Adebayo is a Ph.D. student at MIT working on developing and understanding approaches that seek to make machine learning-based systems reliable when deployed. More broadly, he is interested in rigorous approaches to help develop models that are robust to spurious associations, distribution shifts, and align with ‘human’ values. Website: https://juliusadebayo.com/
Sameer Singh
UC Irvine
Sameer Singh is an Assistant Professor at UC Irvine working on robustness and interpretability of machine learning. Sameer has presented tutorials and invited workshop talks at EMNLP, NeurIPS, NAACL, WSDM, ICLR, ACL, and AAAI, and received paper awards at KDD 2016, ACL 2018, EMNLP 2019, AKBC 2020, and ACL 2020. Website: http://sameersingh.org/
AQ2: Topological Data Analysis in Document Classification
Wlodek Zadrozny and Shafie Gholizadeh
Topological Data Analysis (TDA) introduces methods that capture the underlying structure of shapes in data. Despite the old history in applied mathematics, utilization of topology in data science is relatively a new phenomenon. Within the last decade, TDA has been mostly examined in unsupervised machine learning tasks. As an alternative to the conventional algorithms, TDA has been often considered due to its capability to deal with high-dimensional data in different tasks including but not limited to clustering, dimensionality reduction or descriptive modeling. This tutorial will focus on applications of topological data analysis to text, and in particular to text classification. After introducing the fundamentals, we will show three ways in which topological information can be added to improve the accuracy of classification. More specifically, we explain three different methods of extracting topological features from textual documents, using as the underlying representations of text the two most popular methods, namely term frequency vectors and word embeddings, and also without using any conventional features. In addition, we show how even the simplest out of the box topological methods can be used to provide similarity judgments, e.g. topological plots of classical novels.
Wlodek Zadrozny
UNC Charlotte
Wlodek Zadrozny is a Professor of Computer Science at UNC Charlotte, lifetime member of AAAI, and recipient of the 2013 AAAI Feigenbaum Prize. His research focuses on natural language processing, including mathematical methods in NLP. Prior UNCC, he was a researcher and manager at the IBM T.J. Watson Research Center.
Shafie Gholizadeh
Wells Fargo & Company
Shafie Gholizadeh is a quantitative associate at Wells Fargo & Company. He received his Ph.D. from University of North Carolina in Charlotte. His thesis examined the feasibility of topological data analysis in text mining. A key focus of his dissertation is on novel methods to extract homological features from text.
AH1: AI for COVID-19: Battling the Pandemic with Computational Intelligence
Fei Wang
The new coronavirus disease 2019 (COVID-19) has become a global pandemic. Since the initial outbreak from January 2020 in Wuhan, China, COVID-19 has demonstrated a high transmission rate and a diverse set of clinical characteristics (e.g., high rate of hospital and intensive care unit admission rates, multi-organ dysfunction for critically ill patients due to hyperinflammation, thrombosis, etc.). To understand the disease mechanism of COVID-19 and develop effective control, treatment and prevention strategies, researchers from related disciplines are making tremendous efforts on different aspects including epidemiology and public health, biology and genomic medicine, as well as clinical care and patient management. This tutorial will summarize the efforts on battling the pandemic with artificial intelligence. In particular, I will focus on how AI methods can derive insights on the prediction, treatment and prevention of COVID-19 on the following 4 tasks: 1) disease spread forecasting and policy impact assessment from news reports, government documents, census, meteorology and climate data, etc.; 2) contact tracing and individual behavior modeling from mobile and sensor data; 3) drug research and development from biomedical literature and multi-omics data; 4) patient care and management from longitudinal clinical records. I will also summarize the implications and envision how AI can advance human healthcare in the post-pandemic era.
Fei Wang
Cornell University
Fei Wang is an Associate Professor at Department of Population Health Sciences, Weill Cornell Medicine, Cornell University. His major research interest is data mining, machine learning and their applications in health data science. He has published on the top venues of related areas such as ICML, KDD, NeurIPS, AAAI, JAMA Internal Medicine, Annals of Internal Medicine, etc. His papers have received over 13,400 citations so far with an H-index 57. His (or his students’) papers have won 7 best paper (or nomination) awards at international academic conferences. His team won the championship of the NIPS/Kaggle Challenge on Classification of Clinically Actionable Genetic Mutations in 2017 and Parkinson’s Progression Markers Initiative data challenge organized by Michael J. Fox Foundation in 2016. Dr. Wang is the recipient of the NSF CAREER Award in 2018, as well as the inaugural research leadership award in IEEE International Conference on Health Informatics (ICHI) 2019. Dr. Wang is the chair of the Knowledge Discovery and Data Mining working group in American Medical Informatics Association (AMIA). Dr. Wang’s research has been supported by funding agencies including NSF, NIH, ONR, NMRC, MJFF, AHA , PCORI and industries including Amazon, Google, Boehringer Ingelheim and MITRE.
AH2: Event-Centric Natural Language Understanding
Muhao Chen, Hongming Zhang, Qiang Ning, Manling Li, Heng Ji and Dan Roth
This tutorial targets researchers and practitioners who are interested in AI technologies that help machines understand natural language text, particularly real-world events described in the text. These include methods to extract the internal structures of an event regarding its protagonist(s), participant(s) and properties, as well as external structures concerning memberships, temporal and causal relations of multiple events. This tutorial will provide audience with a systematic introduction of (i) knowledge representations of events, (ii) various methods for automated extraction, conceptualization and prediction of events and their relations, (iii) induction of event processes and properties, and (iv) a wide range of NLU and commonsense understanding tasks that benefit from aforementioned techniques. We will conclude the tutorial by outlining emerging research problems in this area.
Muhao Chen
USC
Muhao Chen is a researcher at Information Sciences Institute, USC. He was a postdoctoral fellow in CIS, UPenn, and received his Ph.D. from UCLA Department of Computer Science in 2019. Muhao’s research focuses on data-driven machine learning approaches for processing structured and unstructured data, and extending their applications to NLU, KBC, computational biology and medicine. Additional information is available at http://muhaochen.github.io.
Hongming Zhang
UPenn
Hongming Zhang is a third-year Ph.D. student at HKUST and a visiting scholar at UPenn. His research focuses on commonsense reasoning and open domain event understanding. Additional information is available at https://www.cse.ust.hk/~hzhangal/
Qiang Ning
Amazon
Qiang Ning is an applied scientist on the Alexa Search team at Amazon. He was a research scientist on the AllenNLP team at AI2. He received his Ph.D. in 2019 from the Department of Electrical and Computer Engineering at University of Illinois at UrbanaChampaign (UIUC). His research interests include natural language understanding and learning from indirect supervision. Additional information is available at http://qiangning.info/.
Manlin Li
University of Illinois at Urbana-Champaign
Manlin Li is a third-year Ph.D. student at the Computer Science Department of the University of Illinois at Urbana-Champaign (UIUC). Her research interests lie in the general area of Natural Language Processing, with a special focus on multimedia. Additional information is available at https://limanling.github.io/.
Heng Ji
University of Illinois at Urbana-Champaign
Heng Ji is a professor at Computer Science Department, and an affiliated faculty member at Electrical and Computer Engineering Department of University of Illinois at Urbana-Champaign, she is also an Amazon Scholar. Her research interests focus on Natural Language Processing, especially on Multimedia Multilingual Information Extraction, Knowledge Base Population and Knowledge-driven Generation. Additional Information is available at http://blender.cs.illinois.edu/hengji.html
Dan Roth
UPenn
Dan Roth is the Eduardo D. Glandt Distinguished Professor at CIS, UPenn, and a Fellow of the AAAS, ACM, AAAI, and the ACL. Roth was recognized for major conceptual and theoretical advances in the modeling of natural language understanding, machine learning, and reasoning. Additional information is available at https://www.cis.upenn.edu/~danroth/.
AH3: From Explainability to Model Quality and Back Again
Anupam Datta, Matt Fredrikson, Klas Leino, Kaiji Lu, Shayak Sen and Zifan Wang
The goal of this tutorial is to provide a systematic view of the current knowledge relating to explainability to several key outstanding concerns regarding the quality of Machine Learning (ML) models; in particular, robustness, privacy, and bias. We will discuss the ways in which explainability can inform questions about these aspects of model quality, and how methods for improving them that are emerging from recent research in the AI, Security, Privacy, and Fairness communities can in turn lead to better outcomes for explainability. Through presentation and live demonstrations, we will discuss recent research results from these communities, as well as several open questions of interest to the AI community.
We aim to make these findings accessible to a general audience, including not only researchers who want to further engage with this area, but also practitioners who stand to benefit from the results, and policymakers who want to deepen their technical understanding of these important issues. Audience members should be familiar with supervised learning and have a working knowledge of gradient-based optimization techniques. Familiarity with Python and at least one deep learning framework is recommended but not required.
Anupam Datta
Carnegie Mellon University
Anupam Datta is a Professor of Electrical & Computer Engineering and Computer Science at Carnegie Mellon University, Co-founder and Chief Scientist of Truera, Director of the Accountable Systems Lab. He received his Ph.D. of Computer Science from Stanford University. His research focuses on enabling real-world complex systems to be accountable for their behavior, especially as they pertain to privacy, fairness, and security.
Matt Fredrikson
Carnegie Mellon University
Matt Fredrikson is an Assistant Professor of Computer Science at Carnegie Mellon University, where his research aims to make machine learning systems more accountable and reliable by addressing fundamental problems of security, privacy, and fairness that emerge in real-world settings.
Klas Leino
Carnegie Mellon University
Klas Leino is a PhD candidate in the Accountable Systems Lab at Carnegie Mellon University, advised by Matt Fredrikson. His research primarily concentrates on demystifying deep learning and understanding its weaknesses and vulnerabilities in order to improve the security, privacy, and transparency of deep neural networks.
Shayak Sen
Truera
Shayak Sen is Co-founder and Chief Technology Officer of Truera, a startup providing enterprise-class platform that delivers explainability for Machine Learning models. Shayak obtained his PhD in Computer Science from Carnegie Mellon University where his research aims to make machine learning and big data systems more explainable, privacy compliant, and fair.
AH4: From Statistical Relational to Neural Symbolic Artificial Intelligence
Luc De Raedt, Sebastijan Dumančić, Robin Manhaeve and Giuseppe Marra
The integration of learning and reasoning is one of the key challenges in artificial intelligence and machine learning today. The area neural symbolic computation tackles this challenge by integrating symbolic reasoning with neural networks. This tutorial will provide an introduction to Neural Symbolic Artificial Intelligence (NeSy) by drawing several parallels with another field that has a rich tradition in integrating learning and reasoning, namely Statistical Relational Artificial Intelligence (StarAI). The tutorial will discuss seven dimensions for introducing and categorizing several StarAI and NeSy approaches. By positioning a wide variety of systems along these dimensions, the tutorial makes it possible to underline common patterns and similarities between them. Interesting opportunities for further research will be described, by looking at areas across the dimensions that have not yet been covered. The tutorial is intended for artificial intelligence researchers and practitioners, as well as domain experts interested in the integration of symbolic reasoning and neural computation. Basic knowledge of logic and/or graphical models and/or neural computation at the level of an introductory course in AI will be helpful but is not a prerequisite.
Luc De Raedt
KU Leuven Institute for AI
Luc De Raedt is full professor at the Department of Computer Science, KU Leuven, and director of Leuven.AI, the newly founded KU Leuven Institute for AI. He is a guest professor at Örebro University in the Wallenberg AI, Autonomous Systems and Software Program. He received his PhD in Computer Science from KU Leuven (1991), and was full professor (C4) and Chair of Machine Learning at the Albert-Ludwigs-University Freiburg, Germany (1999-2006). His research interests are in Artificial Intelligence, Machine Learning and Data Mining, as well as their applications. He is well known for his contributions in the areas of learning and reasoning, in particular, for his work on probabilistic and inductive programming. He co-chaired important conferences such as ECMLPKDD 2001 and ICML 2005 (the European and International Conferences on Machine Learning), ECAI 2012 and will chair IJCAI in 2022 (the European and international AI conferences). He is on the editorial board of Artificial Intelligence, Machine Learning and the Journal of Machine Learning Research. He is a EurAI and AAAI fellow, and received an ERC Advanced Grant in 2015.
Sebastijan Dumančić
KU Leuven
Sebastijan Dumančić is post-doctoral researcher in the DTAI research group at KU Leuven. His post-doctoral research is funded by the FWO’s junior post-doctoral fellowship. He obtained his PhD summa cum laude in the same group in 2018 under the supervision of Prof. Dr. Hendrik Blockeel. His PhD thesis has been awarded an honourable mention for the EurAI Distinguished Dissertation Award, organised by EurAI. His research focuses on program learning and the integration of symbolic and sub-symbolic approaches to AI.
Robin Manhaeve
KU Leuven
Robin Manhaeve is a PhD student in the Declarative Languages and Artificial Intelligence (DTAI) research group at KU Leuven. He obtained both a Master’s degree in Computer Science at KU Leuven. He has obtained funding from the FWO to perform research in the field of neural symbolic computation, and with this helped develop the DeepProbLog framework.
Giuseppe Marra
KU Leuven
Giuseppe Marra is a post-doctoral researcher in the Declarative Languages and Artificial Intelligence (DTAI) research group at KU Leuven. He obtained his PhD at the University of Florence, Italy, defending a thesis titled “Merging Symbolic and Sub-symbolic Reasoning with MiniMax Entropy Models”. His research interests focus on the integration between neural computation and relational reasoning, with a particular focus on logical and probabilistic reasoning.
AH5: Graph Neural Networks: Models and Applications
Yao Ma, Wei Jin, Yiqi Wang, Tyler Derr and Jiliang Tang
Graph structured data such as social networks and molecular graphs are ubiquitous in the real world. It is of great research importance to design advanced algorithms for representation learning on graph structured data so that downstream tasks can be facilitated. Graph Neural Networks (GNNs), which generalize the deep neural network models to graph structured data, pave a new way to effectively learn representations for graph structured data either from the node level or the graph level. Thanks to their strong representation learning capability, GNNs have gained practical significance in various applications ranging from recommendation, natural language processing to healthcare. It has become a hot research topic and attracted increasing attention from the machine learning and data mining community recently. This tutorial of GNNs is timely for AAAI 2021 and covers relevant and interesting topics, including representation learning on graph structured data using GNNs, the robustness of GNNs, the scalability of GNNs and applications based on GNNs.
Yao Ma
Michigan State University
Yao Ma is a Ph.D. student of Computer Science and Engineering at Michigan State University. His research interests include network embedding and graph neural networks for representation learning on graph-structured data. He has published innovative works in top-tier conferences such as WSDM, ASONAM, ICDM, SDM, WWW, KDD and IJCAI.
Wei Jin
Michigan State University
Wei Jin is a CSE Ph.D. student at Michigan State University. He works on the area of graph neural networks, including the theory foundations, model robustness and applications. He has published his research in top conference proceedings (e.g., KDD and WWW). He also delivered tutorials at AAAI’20 and KDD’20.
Yiqi Wang
Michigan State University
Yiqi Wang is a Ph.D. student at Michigan State University. She is working on graph neural networks, including fundamental algorithms and applications. She has published her research in top conference proceedings (e.g., WSDM and WWW). She also delivered tutorials at KDD’20.
Tyler Derr
Vanderbilt University
Tyler Derr is an Assistant Professor at Vanderbilt University. His research includes social network analysis, deep learning on graphs, and data science for social good. He gave a KDD’20 tutorial and BigData’19 workshop on Deep Graph Learning, received Best Reviewer Award at ICWSM’19, and Best Student Poster Award at SDM’19.
Jiliang Tang
Michigan State University
Jiliang Tang is an assistant professor at MSU. His research interests include social computing, data mining and machine learning and their applications in social media and education. He was the recipient of SIGKDD Rising Star Award and NSF Career Award. His research received more than 12,000 citations with h-index 55.
AH6: New Frontiers of Automated Mechanism Design for Pricing and Auctions
Maria-Florina Balcan, Tuomas Sandholm and Ellen Vitercik
Mechanism design is a field of game theory with significant real-world impact, encompassing areas such as pricing and auction design. Mechanisms are used in sales settings ranging from large-scale internet marketplaces to the US government’s radio spectrum reallocation efforts. A powerful and prominent approach in this field is automated mechanism design, which uses optimization and machine learning to design mechanisms based on data. This automated approach helps overcome challenges faced by traditional, manual approaches to mechanism design, which have been stuck for decades due to inherent computational complexity challenges: the revenue-maximizing mechanism is not known even for just two items for sale! This workshop is focused on the rapidly growing area of automated mechanism design for revenue maximization. This encompasses both the foundations of batch and online learning (including statistical guarantees and optimization procedures), as well as real-world success stories.
Maria-Florina Balcan
Carnegie Mellon University
Maria-Florina Balcan is the Cadence Design Systems Professor of Computer Science at Carnegie Mellon University, working in machine learning, algorithmic game theory, and algorithms. She has been Program Committee Co-chair for COLT 2014, ICML 2016, and NeurIPS 2020. She will be General Chair for ICML 2021.
Tuomas Sandholm
Carnegie Mellon University
Tuomas Sandholm is Angel Jordan University Professor of Computer Science at Carnegie Mellon University and Co-Director of CMU AI. He is Founder and Director of the Electronic Marketplaces Laboratory. He is a successful serial entrepreneur. He has fielded over 800 combinatorial auctions, worth over $60 billion. He is Founder and CEO of Optimized Markets, Strategic Machine, and Strategy Robot.
Ellen Vitercik
Carnegie Mellon University
Ellen Vitercik is a final-year PhD student at Carnegie Mellon University. Her research interests include artificial intelligence, machine learning theory, and the interface between economics and computation. She has received the IBM PhD Fellowship, a fellowship from CMU’s Center for Machine Learning and Health, and the NSF Graduate Research Fellowship.
AH7: On Explainable AI: From Theory to Motivation, Industrial Applications and Coding Practices
Freddy Lecue, Pasquale Minervini, Fosca Giannotti and Riccardo Guidotti
The future of AI lies in enabling people to collaborate with machines to solve complex problems. Like any efficient collaboration, this requires good communication, trust, clarity and understanding. XAI (eXplainable AI) aims at addressing such challenges by combining the best of symbolic AI and traditional Machine Learning. Such topic has been studied for years by all different communities of AI, with different definitions, evaluation metrics, motivations and results.
This tutorial is a snapshot on the work of XAI to date, and surveys the work achieved by the AI community with a focus on machine learning and symbolic AI-related approaches (given the half-day format). We will motivate the needs of XAI in real-world and large-scale application, while presenting state-of-the-art techniques, with best XAI coding practices. In the first part of the tutorial, we give an introduction to the different aspects of explanations in AI. We then focus the tutorial on two specific approaches: (i) XAI using machine learning and (ii) XAI using a combination of graph-based knowledge representation and machine learning. For both we get into the specifics of the approach, the state of the art and the research challenges for the next steps. The final part of the tutorial gives an overview of real-world applications of XAI as well as best XAI coding practices.
Freddy Lecue
Accenture Technology Labs, Dublin
Freddy Lecue (PhD 2008, Habilitation 2015) is a principal scientist and research manager in Artificial Intelligent systems, systems combining learning and reasoning capabilities, in Accenture Technology Labs, Dublin – Ireland. He is also a research associate at INRIA, in WIMMICS, Sophia Antipolis – France. Before joining Accenture Labs, he was a Research Scientist at IBM Research, Smarter Cities Technology Center (SCTC) in Dublin, Ireland, and lead investigator of the Knowledge Representation and Reasoning group. His main research interests are Explainable AI systems. The application domain of his current research is Smarter Cities, with a focus on Smart Transportation and Building. In particular, he is interested in exploiting and advancing Knowledge Representation and Reasoning methods for representing and inferring actionable insight from large, noisy, heterogeneous and big data. He has over 40 publications in refereed journals and conferences related to Artificial Intelligence (AAAI, ECAI, IJCAI, IUI) and Semantic Web (ESWC, ISWC), all describing new system to handle expressive semantic representation and reasoning. He co-organized the first three workshops on semantic cities (AAAI 2012, 2014, 2015, IJCAI 2013), and the first two tutorials on smart cities at AAAI 2015 and IJCAI 2016. Prior to joining IBM, Freddy Lecue was a Research Fellow (2008-2011) with the Centre for Service Research at The University of Manchester, UK. He has been awarded by a second prize for his Ph.D. thesis by the French Association for the Advancement of Artificial Intelligence in 2009, and has been recipient of the Best Research Paper Award at the ACM/IEEE Web Intelligence conference in 2008.
Pasquale Minervini
University College London
Pasquale is a Senior Research Fellow at University College London (UCL). He received a PhD in Computer Science from University of Bari, Italy, with a thesis on relational learning. After his PhD, he worked as a postdoc researcher at the University of Bari, and at the INSIGHT Centre for Data Analytics (INSIGHT), where he worked in a group composed of researchers and engineers from INSIGHT and Fujitsu Ireland Research and Innovation. Pasquale published peer-reviewed papers in top-tier AI conferences, receiving two best paper awards, participated to the organisation of tutorials on Explainable AI and relational learning (three for AAAI, one for ECML, and others), and was a guest lecturer at UCL and at the Summer School on Statistical Relational Artificial Intelligence. He is the main inventor of a patent application assigned to Fujitsu Ltd, and recently he was awarded a seven-figures H2020 research grant involving applications of relational learning to cancer research. His website is neuralnoise.com.
Fosca Giannotti
National Research Council, Pisa, Italy
Fosca Giannotti is Director of Research at the Information Science and Technology Institute “A. Faedo” of the National Research Council, Pisa, Italy. Fosca is a scientist in Data mining and Machine Learning and Big Data Analytics. She leads the Pisa KDD Lab – Knowledge Discovery and Data Mining Laboratory http://kdd.isti.cnr.it, a joint research initiative of the University of Pisa and ISTI-CNR, founded in 1994 as one of the earliest research labs centered on data mining. Fosca’s research focus is on social mining from big data: human dynamics, social networks, diffusion of innovation, privacy enhancing technology and explainable AI. She has coordinated tens of research projects and industrial collaborations. Fosca is now the coordinator of SoBigData, the European research infrastructure on Big Data Analytics and Social Mining, an ecosystem of ten cutting edge European research centers providing an open platform for interdisciplinary data science and data-driven innovation http://www.sobigdata.eu. In 2012-2015 Fosca has been general chair of the Steering board of ECML-PKDD (European conference on Machine Learning) and is currently member of the steering committee EuADS (European Association on Data Science) and of the AIIS: Italian Lab. of Artificial Intelligence and Autonomous Systems
Riccardo Guidotti
University of Pisa, Italy
Riccardo Guidotti is currently a post-doc researcher at the Department of Computer Science, University of Pisa, Italy and a member of the Knowledge Discovery and Data Mining Laboratory (KDDLab), a joint research group with the Information Science and Technology Institute of the National Research Council in Pisa. Riccardo Guidotti was born in 1988 in Pitigliano (GR) Italy. In 2013 and 2010 he graduated cum laude in Computer Science (MS and BS) at University of Pisa. He received the PhD in Computer Science with a thesis on Personal Data Analytics in the same institution. He won the IBM fellowship program and has been an intern in IBM Research Dublin, Ireland in 2015. His research interests are in personal data mining, clustering, explainable models, analysis of transactional data related to recipes and to migration flows.
AH8: Recent Advances in Multiple Facets of Preference Learning
Arun Rajkumar
Analyzing preferences using pairwise/partially ordered data has a long history and has been core to several research areas including AI, statistics, operations research, game theory, social choice and theoretical computer science with a range of applications including ad-placement, voting, sports ranking, multi-criteria decision making among several others. Research in each of these areas has developed independently, typically, without a lot of interaction among them. The broad aim of this tutorial is to overview the exciting recent advances in preference learning with the aim of introducing tools and techniques across disciplines which will aid in fostering interdisciplinary AI research.
The target audience includes typical practitioners of AI who might have a high-level idea about preference learning but are not typically aware of the various challenging facets of the problem. The novelty of the tutorial lies in translating the different paradigms across communities into the language of AI so that it will benefit the ML/AI community. The tutorial will be self-contained and does not expect any prerequisites. Audience with basic knowledge of AI/ML would be able to follow most of the material.
Arun Rajkumar
Indian Institute of Technology, Madras
Arun Rajkumar is currently an assistant professor at the Computer Science and Engineering department of Indian Institute of Technology, Madras. His research interests are broadly in the areas of statistical learning theory, machine learning with a specific focus on ranking algorithms and online learning.
AH9: Simplifying and Automating Parallel Machine Learning via a Programmable and Composable Parallel ML System
Hao Zhang, Aurick Qiao, Qirong Ho and Eric Xing
ML scale-up is frequently underestimated. What does it really take to train an ML model, originally implemented for a single CPU/GPU, on multiple machines? A few pain points are: (1) Many new lines of code need to be written to convert the code to a distributed version; (2) Need to heavily tune the code to the satisfying system/statistical performance, which is an added process to model development; (3) Decide which/how many hardware resources to use to train and deploy the model; (4) From an organization’s perspective, automate resource sharing between many users and jobs to satisfy user needs while maximizing resource utilization and minimizing cost.
In this tutorial, we will present improved techniques to automate distributed ML infrastructure. We plan the tutorial to cover three areas critical to ML parallelization: (1) Cataloging and standardizing parallel ML building blocks; (2) Representations and software frameworks for ML parallelism; (3) Algorithms and systems to automate ML parallelization and the resource allocation of ML jobs on shared clusters. By exposing unique characteristics of ML programs, and by dissecting successful cases to reveal how they can be harnessed, we present opportunities for ML researchers and practitioners to further shape and grow the area of SysML.
The audience should be familiar with ML and DL basics. Knowledge of TensorFlow, PyTorch, and distributed ML techniques is also helpful but not required.
Hao Zhang
University of California, Berkeley
Hao Zhang is currently a postdoc scholar at RISE lab, the University of California, Berkeley. Hao’s general research interest is in scalable machine learning. Hao completed his Ph.D. at CMU. Hao’s past works including AutoDist, Poseidon, Cavs are now being commercialized at the Pittsburgh-based startup Petuum Inc.
Aurick Qiao
Carnegie Mellon University
Aurick Qiao is a Ph.D. candidate at CMU. His research interest is in resource management for distributed ML in shared-resource computing environments. Several of Aurick’s work including Litz and Bosen are parts of the Petuum project. Aurick is also an Engineering Lead at Petuum, building scalable and easy-to-use systems for ML which “just works.”
Qirong Ho
Carnegie Mellon University
Qirong Ho is Co-Founder and CTO at Petuum, Inc., a technology startup from the Petuum distributed ML team at Carnegie Mellon University. He holds a Ph.D. from CMU, and his research interests are in distributed ML systems with a view towards correctness, performance guarantees, robustness, programmability and usability.
Eric Xing
Petuum Inc.
Eric Xing is a Professor at CMU, and the Founder, CEO, and Chief Scientist of Petuum Inc. His main research interests are the development of ML and statistical methodology, and large-scale computational system and architectures. He is an AAAI Fellow and an IEEE Fellow.
AH10: Towards Ubiquitous Recommender Systems: Data, Approaches, and Applications
Shoujin Wang, Liang Hu, Yan Wang, Longbing Cao, Michael Sheng and Mehmet Orgun
This tutorial presents representative and state-of-the-art theories and approaches to build recommender systems (RS) and the ubiquitous applications of RS in nearly every aspect of our daily life, including eating, dressing, housing and travelling. To be specific, we will first present the background and foundations of RS, followed by the illustration of representative and advanced machine learning approaches that can be used for building RS, including latent factor models, deep learning models, graph learning models and knowledge graph models. Finally, we will demonstrate the ubiquity of RS by introducing the typical real-world applications of RS in both traditional domains, such as e-commerce, social network, and emerging domains including fashion industry, financial industry, and healthcare industry. Specifically, we will illustrate the domain characteristics and the data characteristics in each domain, and then introduce the appropriate RS approaches to generate recommendations for each specific domain based on the unique characteristics of the domain. We will conclude the tutorial by outlining future research directions in this area. No specific prerequisite knowledge is required, but a rudimentary knowledge of RS and some machine learning basic (e.g., factorization machine, deep learning) will be helpful.
Shoujin Wang
Macquarie University
Shoujin Wang is a research fellow at the Department of Computing, Macquarie University. His research interests include data mining, machine learning, and their general applications in recommender systems. He has delivered tutorials on recommender systems at IJCAI 2019, IJCAI 2020 and organized a workshop at ICDM 2020.
Liang Hu
University of Technology Sydney
Liang Hu is a research associate with Advanced Analytics Institute, University of Technology Sydney. His research interests include recommender systems, data mining, machine learning and general artificial intelligence.
Yan Wang
Macquarie University
Yan Wang is a Professor at the Department of Computing, Macquarie University. His research interests include recommender systems, trust management/computing, social computing and services computing.
Longbing Cao
University of Technology Sydney
Longbing Cao is a Professor and an ARC future fellow at Advanced Analytics Institute (AAI), University of Technology Sydney. He received the 2019 Eureka Prizes for Excellence in Data Science. His research interests include data science, analytics and machine learning, and behavior informatics and their enterprise applications.
Michael Sheng
Macquarie University
Michael Sheng is a full Professor and Head of Department of Computing at Macquarie University. His research interests include web of things, internet of things, big data analytics, web science, service-oriented computing, pervasive computing, sensor networks.
Mehmet A. Orgun
Macquarie University
Mehmet A. Orgun is a Professor at the Department of Computing, Macquarie University. His current research interests lie in the areas of computational intelligence, multi-agent systems, trust and security, temporal reasoning, formal methods.