Graphs are ubiquitous in many applications, such as molecular biochemistry, neural science, Internet, computer vision, NLP, and crowdsourcing.
Machine learning on graphs, especially with neural networks, has demonstrated prediction accuracy. However, accuracy is not the only desideratum,
and humans and society can still be negatively impacted by the models if care is not taken. For example, a model can lack of transparency so that it is hard to understand why a model make a prediction;
a model's accuracy may drop due to slight perturbations;
or an accurate model can treat different demographical groups or individuals in an unfair way.
We aim to make the models more responsible by investigating explainability, fairness, and robustness beyond accuracy of the models.
(i)
On large graphs, power-law degree distributions are common and can lead to fairness issues in the graphical models and affect end-users.
We propose a linear system to certificate if multiple desired fairness criteria can be fulfilled simultaneously, and if not,
a multi-objective optimization algorithm to find Pareto fronts for efficient trade-offs among the criteria
[
CIKM2021]).
To reduce optimization cost, the team proposes continuous Pareto front exploration
by exploiting the smoothness of the set of Pareto optima.
As graphs can contain hidden factors to complicate fairness issues,
we simultaneously learn a fair models and identify such hidden factors to mitigate the issues
[
KDD2023]).
(ii)
Graphical models can be hard to understand by human users due to multiplexed information propagations over many edges.
The team published a series of works addressing challenges in making graphical models more interpretable,
such as
large discrete search space
[
ICDM2019]),
axiomatic attribution
[
CIKM2020]),
multi-objective explanations
[
ICDM2021a]),
and differential geometric for interpretating nonlinear graph evolution
[
ICLR2023]).
(iii)
Robustness can be interpreted broadly as maintaining any desired properties under reasonably slight perturbations.
We provide robust explanations through self-supervision and constrained optimization
[
ICDM2021b]),
and robust optimization, statistical theory, and optimization convergence analysis
[
ICML2023])