Dr. Xiaoge Zhang delivered a talk on “Enhancing the Performance of Neural Networks Through Causal Discovery and Integration of Domain Knowledge” at Sichuan University, China

In this talk, I will present a generic methodology to encode hierarchical causal structure among observed variables into a neural network to improve its prediction performance. The proposed causality-informed neural network (CINN) leverages three coherent steps to systematically map the structural causal knowledge into the layer-to-layer design of neural network while strictly preserving the orientation of every causal relationship. In the first step, CINN discovers causal relationships from observational data via directed acyclic graph (DAG) learning, where causal discovery is recast as a continuous optimization problem to avoid the combinatorial nature. In the second step, the discovered hierarchical causal structure among observed variables is encoded into neural network through a dedicated architecture and customized loss function. By categorizing variables as root, intermediate, and leaf nodes, the hierarchical causal DAG is translated into CINN with a one-to-one correspondence between nodes in the DAG and units in the CINN while maintaining the relative order among these nodes. Regarding the loss function, both intermediate and leaf nodes in the DAG are treated as target outputs during CINN training to drive co-learning of causal relationships among different types of nodes. In the final step, as multiple loss components emerge in CINN, we leverage the projection of conflicting gradients to mitigate gradient interference among the multiple learning tasks. Computational experiments across a broad spectrum of UCI datasets demonstrate substantial advantages of CINN in prediction performance over other state-of-the-art methods. In addition, we conduct an ablation study by incrementally injecting structural and quantitative causal knowledge into neural network to demonstrate their role in enhancing neural network’s prediction performance.

Dr. Xiaoge Zhang delivered a talk on “Safety assessment and risk analysis of complex systems under uncertainty” at Nanjing University, China

This talk showcases two different strategies to assess and analyze the safety of air transportation system. In the first place, considering the rich information in the historical aviation accident events, we analyzed the accidents reported in the National Transporation Safety Board (NTSB) over the past two decades, and developed a large-scale Bayesian network to model the causal relationships among a variety of factors contributing to the occurrence of aviation accidents. The construction of Bayesian network greatly facilitates the root cause diagnosis and outcome analysis of aviation accident. Next, we analyze how to leverage deep learning to forecast flight trajectory. Using Bayesian neural network, we fully characterize the effect of exogenous variables on the flight trajectory. The predicted trajectory is then expanded to multiple flights, and used to assess safety based on horizontal and vertical separation distance between two flights, thus enabling real-time monitoring of in-flight safety.

Dr. Xiaoge Zhang delivered a talk on “A Review on Uncertainty Quantification of Neural Network and Its Application for Reliable Detection of Steel Wire Rope Defects” at Hunan University, China

This talk provides a holistic lens on emerging uncertainty quantification (UQ) methods for ML models with a particular focus on neural networks and gives a tutorial-style description of several state-of-the-art UQ methods: Gaussian process regression, Bayesian neural network, neural network ensemble, and deterministic UQ methods focusing on spectral-normalized neural Gaussian process (SNGP). Established upon the mathematical formulations, we subsequently examine the soundness of these UQ methods quantitatively and qualitatively (by a toy regression example) to examine their strengths and shortcomings from different dimensions. Based on the findings of the comparison, we exploit the advantages of SNGP in UQ and develop an uncertainty-aware deep neural network to detect the defects of steel wire rope. Computational experiments and comparisons with state-of-the-art models suggest that the principled uncertainty quantified by SNGP not only substantially enhances the prediction performance, but also provides an essential layer of protection for neural network against out-of-distribution data.

Dr. Xiaoge Zhang delivered an online talk on “A Tutorial on Uncertainty Quantification of Neural Network and Its Application for Reliable Detection of Steel Wire Rope Defects” at University of Tennessee, Knoxville (UTK)

This talk provides a holistic lens on emerging uncertainty quantification (UQ) methods for ML models with a particular focus on neural networks and gives a tutorial-style description of several state-of-the-art UQ methods: Gaussian process regression, Bayesian neural network, neural network ensemble, and deterministic UQ methods focusing on spectral-normalized neural Gaussian process (SNGP). Established upon the mathematical formulations, we subsequently examine the soundness of these UQ methods quantitatively and qualitatively (by a toy regression example) to examine their strengths and shortcomings from different dimensions. Based on the findings of the comparison, we exploit the advantages of SNGP in UQ and develop an uncertainty-aware deep neural network to detect the defects of steel wire rope. Computational experiments and comparisons with state-of-the-art models suggest that the principled uncertainty quantified by SNGP not only substantially enhances the prediction performance, but also provides an essential layer of protection for neural network against out-of-distribution data.

Dr. Xiaoge Zhang delivered a talk on “Causality-Informed Neural Network (CINN)” at Hong Kong Institute for Advanced Study (HKIAS)

Despite the impressive performance of deep learning in solving long-standing problems (e.g., machine translation, image classification), it has been oftentimes criticized for learning spurious correlations, vulnerable robustness, and poor generalizability. These deficiencies significantly hinder the adoptions of deep learning in high-stakes decision settings, such as healthcare. Although the state-of-the-art literature has made some advances in addressing some issues, the existing attempts to improve robustness, interpretability, and generalizability are unfortunately ad-hoc and unprincipled. In this talk, we develop a generic framework to inject causal knowledge in either structural or qualitative (or quantitative) form (or both) into neural network, where the orientation of each causal relationship is strictly preserved. The proposed causality-informed neural network (CINN) provides a one-stop solution with a great potential to fundamentally solve the existing issues in neural network. More importantly, the incorporation of causal knowledge greatly unlocks the power of neural network in scenarios more than prediction, such as causal effect estimation, interventional query. Several examples are used to demonstrate the appealing features of CINN.