Publications
publications by categories in reversed chronological order. generated by jekyll-scholar.
2023
- AIRA survey of Bayesian Network structure learningArtificial Intelligence Review, 2023
Bayesian Networks (BNs) have become increasingly popular over the last few decades as a tool for reasoning under uncertainty in fields as diverse as medicine, biology, epidemiology, economics and the social sciences. This is especially true in real-world areas where we seek to answer complex questions based on hypothetical evidence to determine actions for intervention. However, determining the graphical structure of a BN remains a major challenge, especially when modelling a problem under causal assumptions. Solutions to this problem include the automated discovery of BN graphs from data, constructing them based on expert knowledge, or a combination of the two. This paper provides a comprehensive review of combinatoric algorithms proposed for learning BN structure from data, describing 74 algorithms including prototypical, well-established and state-of-the-art approaches. The basic approach of each algorithm is described in consistent terms, and the similarities and differences between them highlighted. Methods of evaluating algorithms and their comparative performance are discussed including the consistency of claims made in the literature. Approaches for dealing with data noise in real-world datasets and incorporating expert knowledge into the learning process are also covered.
@article{kitson2021survey, url = {https://link.springer.com/article/10.1007/s10462-022-10351-w}, title = {A survey of Bayesian Network structure learning}, author = {Kitson, Neville K and Constantinou, Anthony C and Guo, Zhigao and Liu, Yang and Chobtham, Kiattikun}, journal = {Artificial Intelligence Review}, page = {1-94}, year = {2023}, doi = {10.1007/s10462-022-10351-w}, dimensions = {true}, }
- ICLRImproving the imputation of missing data with Markov Blanket discoveryYang Liu and Anthony C ConstantinouIn The Eleventh International Conference on Learning Representations, 2023
The process of imputation of missing data typically relies on generative and re- gression models. These approaches often operate on the unrealistic assumption that all of the data features are directly related with one another, and use all of the available features to impute missing values. In this paper, we propose a novel Markov Blanket discovery approach to determine the optimal feature set for a given variable by considering both observed variables and missingness of partially observed variables to account for systematic missingness. We then incorporate this method to the learning process of the state-of-the-art MissForest imputation algorithm, such that it informs MissForest which features to consider to impute missing values, depending on the variable the missing value belongs to. Experi- ments across different case studies and multiple imputation algorithms show that the proposed solution improves imputation accuracy, both under random and sys- tematic missingness.
@inproceedings{liu2023improving, title = {Improving the imputation of missing data with Markov Blanket discovery}, author = {Liu, Yang and Constantinou, Anthony C}, booktitle = {The Eleventh International Conference on Learning Representations}, url = {https://openreview.net/pdf?id=GrpU6dxFmMN}, year = {2023}, }
- ESAOpen problems in causal structure learning: A case study of COVID-19 in the UKExpert Systems with Applications, 2023
Causal machine learning (ML) algorithms recover graphical structures that tell us something about cause-and-effect relationships. The causal representation provided by these algorithms enables transparency and explainability, which is necessary for decision making in critical real-world problems. Yet, causal ML has had limited impact in practice compared to associational ML. This paper investigates the challenges of causal ML with application to COVID-19 UK pandemic data. We collate data from various public sources and investigate what the various structure learning algorithms learn from these data. We explore the impact of different data formats on algorithms spanning different classes of learning, and assess the results produced by each algorithm, and groups of algorithms, in terms of graphical structure, model dimensionality, sensitivity analysis, confounding variables, predictive and interventional inference. We use these results to highlight open problems in causal structure learning and directions for future research. To facilitate future work, we make all graphs, models, data sets, and source code publicly available online.
@article{constantinou2023open, title = {Open problems in causal structure learning: A case study of COVID-19 in the UK}, author = {Constantinou, Anthony C and Kitson, Neville K and Liu, Yang and Chobtham, Kiattikun and Amirkhizi, Arian Hashemzadeh and Nanavati, Praharsh A and Mbuvha, Rendani and Petrungaro, Bruno}, journal = {Expert Systems with Applications}, pages = {121069}, url = {https://pdf.sciencedirectassets.com/271506/1-s2.0-S0957417423X00207/1-s2.0-S0957417423015713/main.pdf?X-Amz-Security-Token=IQoJb3JpZ2luX2VjEJr%2F%2F%2F%2F%2F%2F%2F%2F%2F%2FwEaCXVzLWVhc3QtMSJHMEUCIQCUg1wv8wNN7ME1pq0f5PKI4eQWMMP20Q8ySzxfEfDvzAIgPxQU0MjwoDV7xkqYPXe%2BM3mdWKmNvXFJSDI5C3nKS%2BkquwUI0v%2F%2F%2F%2F%2F%2F%2F%2F%2F%2FARAFGgwwNTkwMDM1NDY4NjUiDNWEnpxjMSK5vgKvFiqPBfbiK4EwKe6DTwO8ZULx2PYvEa%2Bi%2Be1bvR7EtaebvOiRB285j5ws5QptnTBwq9qiYazuJ62ibErbE1k%2BoLALtRQrHdI%2FZMhayHsBeWG2vR0h8Y8macrRaqHKgGed6mHw1ASBq1VtkncPIV%2BtyDtBknTPDuGBT%2FRIanbEYJaQc278MMp1UD%2Fo3%2FJHb7k8yDxlSuFUpNh7fQ2gRIcbW0bcrn6G9LYZCur0Q4t%2FncylcdMick%2BxDLWZ9QgIEzPDOYAhEKnIb5YjKFFHdeS3Uus%2F1HKSWJWZ8vQy8ovcaDpJ%2B65WsWSSH4i7ZO82VVUbtF07p9%2B0PLDQ5AEDLgGhavehLhP4I0xOZv9jE9i2jOpVTD9Eg1VAbDw2twNki7G1%2Bl%2FBRH7RrzbtIskRhtm7tWkBwNM3GlWOZRnp8ugvaqS5RxvUxN96cZfEFRZ1%2B1btBk1i1rFo67gnMRbv%2FQQegPN%2BXn4AeYGi5BV8Q81oda8m7s5O8yWdknd4whYXxefEORoZJLg6OZkYjb7O3UI2fTrkWD92NMKWNoreHnHIpr6%2FCUw6B9S%2BgLCJyK9iMWEWUpspbHEWTAsyULuGc9KX3m3nLXyijoW479kQcMUbh2KerIBGX%2FAF%2BrOptwf22QbCgeDZSIb3vQdm2KBmJMdHAmmjhAzQme8UaFGh2pcwHxGXwbJDcmZ3op3z7zSeOwp9nMara7909a2aBU%2B%2FfFeOtXpioSReFaa4KshgwGgE29RpxVw5PA31wBe%2BFWVXJ4W3mB0g6EqHs6Fdtr6VnoOVXyNp865AotgtGLBCBQjqsmmmU2EExeAUSjU%2BeuPrlFAzyvquziLURIObomssqc2FKu%2BlV%2BAAaIftTLA7PyqtYELvE58wtMThxAY6sQGgpgP%2Fx2ewQGwhlPxBjNpOKDDU4ycbcSdreCERjPVh6aglA4tQadad%2FMSIPgWpRCciTqhXZqJMsDIhikFpTOuk6wop4YcTIHeJ7KI%2FDh5z%2BKqKGFG1DevvNA2%2FLvZY7mmHy2ForK0LbVn2o0xSWZ0afXIV3czsGXVy%2Fc5EMoxM7GGklG0P4SlGTBMyXGuC0UnkUAHyOS%2B3FTADzLpd3xn9JNe9OdbYP2eeS8tiuO9yy0o%3D&X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Date=20250810T102407Z&X-Amz-SignedHeaders=host&X-Amz-Expires=300&X-Amz-Credential=ASIAQ3PHCVTY77ZGFOCW%2F20250810%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Signature=a3154e02b2e2966f4406bd34207808ebf4df12ec79455185da94f84eba5250bb&hash=1ea76254f3908a3a84020b66eb328bae8f5211478ec390d5f7a83570241e8118&host=68042c943591013ac2b2430a89b270f6af2c76d8dfd086a07176afe7c76c2c61&pii=S0957417423015713&tid=spdf-0dd36597-bd9d-46ca-a656-0b720d1c19c8&sid=6004d8f495df6847538a5f2331816ba842e0gxrqa&type=client&tsoh=d3d3LnNjaWVuY2VkaXJlY3QuY29t&rh=d3d3LnNjaWVuY2VkaXJlY3QuY29t&ua=120e5e59010100585e0602&rr=96cec81ae825094e&cc=hk}, year = {2023}, publisher = {Elsevier}, }
2022
- MLJGreedy structure learning from data that contain systematic missing valuesYang Liu and Anthony C ConstantinouMachine Learning, 2022
Learning from data that contain missing values represents a common phenomenon in many domains. Relatively few Bayesian Network structure learning algorithms account for missing data, and those that do tend to rely on standard approaches that assume missing data are missing at random, such as the Expectation-Maximisation algorithm. Because missing data are often systematic, there is a need for more pragmatic methods that can effectively deal with data sets containing missing values not missing at random. The absence of approaches that deal with systematic missing data impedes the application of BN structure learning methods to real-world problems where missingness are not random. This paper describes three variants of greedy search structure learning that utilise pairwise deletion and inverse probability weighting to maximally leverage the observed data and to limit potential bias caused by missing values. The first two of the variants can be viewed as sub-versions of the third and best performing variant, but are important in their own in illustrating the successive improvements in learning accuracy. The empirical investigations show that the proposed approach outperforms the commonly used and state-of-the-art Structural EM algorithm, both in terms of learning accuracy and efficiency, as well as both when data are missing at random and not at random.
@article{liu2022greedy, url = {https://link.springer.com/article/10.1007/s10994-022-06195-8}, title = {Greedy structure learning from data that contain systematic missing values}, author = {Liu, Yang and Constantinou, Anthony C}, journal = {Machine Learning}, volume = {111}, number = {10}, pages = {3867--3896}, year = {2022}, publisher = {Springer US}, doi = {10.1007/s10994-022-06195-8}, dimensions = {true}, }
- IJAREffective and efficient structure learning with pruning and model averaging strategiesInternational Journal of Approximate Reasoning, 2022
Learning the structure of a Bayesian Network (BN) with score-based solutions involves exploring the search space of possible graphs and moving towards the graph that maximises a given objective function. Some algorithms offer exact solutions that guarantee to return the graph with the highest objective score, while others offer approximate solutions in exchange for reduced computational complexity. This paper describes an approximate BN structure learning algorithm, which we call Model Averaging Hill-Climbing (MAHC), that combines two novel strategies with hill-climbing search. The algorithm starts by pruning the search space of graphs, where the pruning strategy can be viewed as an aggressive version of the pruning strategies that are typically applied to combinatorial optimisation structure learning problems. It then performs model averaging in the hill-climbing search process and moves to the neighbouring graph that maximises the objective function, on average, for that neighbouring graph and over all its valid neighbouring graphs. Comparisons with other algorithms spanning different classes of learning suggest that the combination of aggressive pruning with model averaging is both effective and efficient, particularly in the presence of data noise.
@article{constantinou2022effective, url = {https://www.sciencedirect.com/science/article/pii/S0888613X22001591}, title = {Effective and efficient structure learning with pruning and model averaging strategies}, author = {Constantinou, Anthony C and Liu, Yang and Kitson, Neville K and Chobtham, Kiattikun and Guo, Zhigao}, journal = {International Journal of Approximate Reasoning}, volume = {151}, pages = {292--321}, year = {2022}, publisher = {Elsevier}, doi = {10.1016/j.ijar.2022.09.016}, dimensions = {true}, }
- JMLRImproving Bayesian network structure learning in the presence of measurement errorYang Liu, Anthony C Constantinou, and Zhigao GuoJournal of Machine Learning Research, 2022
Structure learning algorithms that learn the graph of a Bayesian network from observational data often do so by assuming the data correctly reflect the true distribution of the variables. However, this assumption does not hold in the presence of measurement error, which can lead to spurious edges. This is one of the reasons why the synthetic performance of these algorithms often overestimates real-world performance. This paper describes a heuristic algorithm that can be added as an additional learning phase at the end of any structure learning algorithm, and serves as a correction learning phase that removes potential false positive edges. The results show that the proposed correction algorithm successfully improves the graphical score of five well-established structure learning algorithms spanning different classes of learning in the presence of measurement error.
@article{liu2022improving, title = {Improving Bayesian network structure learning in the presence of measurement error}, author = {Liu, Yang and Constantinou, Anthony C and Guo, Zhigao}, journal = {Journal of Machine Learning Research}, volume = {23}, number = {324}, pages = {1--28}, year = {2022}, url = {https://www.jmlr.org/papers/volume23/20-1319/20-1319.pdf}, }
- OSNEMSWSR: A Chinese dataset and lexicon for online sexism detectionOnline Social Networks and Media,, 2022
Online sexism has become an increasing concern in social media platforms as it has affected the healthy development of the Internet and can have negative effects in society. While research in the sexism detection domain is growing, most of this research focuses on English as the language and on Twitter as the platform. Our objective here is to broaden the scope of this research by considering the Chinese language on Sina Weibo. We propose the first Chinese sexism dataset – Sina Weibo Sexism Review (SWSR) dataset –, as well as a large Chinese lexicon SexHateLex made of abusive and gender-related terms. We introduce our data collection and annotation process, and provide an exploratory analysis of the dataset characteristics to validate its quality and to show how sexism is manifested in Chinese. The SWSR dataset provides labels at different levels of granularity including (i) sexism or non-sexism, (ii) sexism category and (iii) target type, which can be exploited, among others, for building computational methods to identify and investigate finer-grained gender-related abusive language. We conduct experiments for the three sexism classification tasks making use of state-of-the-art machine learning models. Our results show competitive performance, providing a benchmark for sexism detection in the Chinese language, as well as an error analysis highlighting open challenges needing more research in Chinese NLP. The SWSR dataset and SexHateLex lexicon are publicly available.
@article{jiang2022swsr, title = {SWSR: A Chinese dataset and lexicon for online sexism detection}, author = {Jiang, Aiqi and Yang, Xiaohan and Liu, Yang and Zubiaga, Arkaitz}, journal = {Online Social Networks and Media,}, volume = {27}, pages = {100182}, year = {2022}, publisher = {Elsevier,}, url = {https://arxiv.org/pdf/2108.03070.pdf}, doi = {10.1016/j.osnem.2021.100182}, dimensions = {true}, }
2021
- IJARLarge-scale empirical validation of Bayesian Network structure learning algorithms with noisy dataInternational Journal of Approximate Reasoning, 2021
Numerous Bayesian Network (BN) structure learning algorithms have been proposed in the literature over the past few decades. Each publication makes an empirical or theoretical case for the algorithm proposed in that publication and results across studies are often inconsistent in their claims about which algorithm is ‘best’. This is partly because there is no agreed evaluation approach to determine their effectiveness. Moreover, each algorithm is based on a set of assumptions, such as complete data and causal sufficiency, and tend to be evaluated with data that conforms to these assumptions, however unrealistic these assumptions may be in the real world. As a result, it is widely accepted that synthetic performance overestimates real performance, although to what degree this may happen remains unknown. This paper investigates the performance of 15 state-of-the-art, well-established, or recent promising structure learning algorithms. We propose a methodology that applies the algorithms to data that incorporates synthetic noise, in an effort to better understand the performance of structure learning algorithms when applied to real data. Each algorithm is tested over multiple case studies, sample sizes, types of noise, and assessed with multiple evaluation criteria. This work involved learning approximately 10,000 graphs with a total structure learning runtime of seven months. In investigating the impact of data noise, we provide the first large scale empirical comparison of BN structure learning algorithms under different assumptions of data noise. The results suggest that traditional synthetic performance may overestimate real-world performance by anywhere between 10% and more than 50%. They also show that while score-based learning is generally superior to constraint-based learning, a higher fitting score does not necessarily imply a more accurate causal graph. The comparisons extend to other outcomes of interest, such as runtime, reliability, and resilience to noise, assessed over both small and large networks, and with both limited and big data. To facilitate comparisons with future studies, we have made all data, raw results, graphs and BN models freely available online.
@article{constantinou2021large, url = {https://www.sciencedirect.com/science/article/pii/S0888613X21000025}, title = {Large-scale empirical validation of Bayesian Network structure learning algorithms with noisy data}, author = {Constantinou, Anthony C and Liu, Yang and Chobtham, Kiattikun and Guo, Zhigao and Kitson, Neville K}, journal = {International Journal of Approximate Reasoning}, volume = {131}, pages = {151--188}, year = {2021}, publisher = {Elsevier}, doi = {10.1016/j.ijar.2021.01.001}, dimensions = {true}, }
2020
- The Bayesys data and Bayesian network repositoryQueen Mary University of London: London, UK, 2020
@article{constantinou2020bayesys, url = {http://constantinou.info/downloads/bayesys/bayesys_repository.pdf}, title = {The Bayesys data and Bayesian network repository}, author = {Constantinou, Anthony C and Liu, Yang and Chobtham, Kiattikun and Guo, Zhigao and Kitson, Neville K}, journal = {Queen Mary University of London: London, UK}, year = {2020}, }