Hierarchical Independence Thresholding for learning Bayesian network classifiers


Liu Y., Wang L., MƏMMƏDOV M., Chen S., Wang G., Qi S., ...daha çox

Knowledge-Based Systems, vol.212, 2021 (SCI-Expanded, Scopus) identifier

  • Nəşrin Növü: Article / Article
  • Cild: 212
  • Nəşr tarixi: 2021
  • Doi nömrəsi: 10.1016/j.knosys.2020.106627
  • jurnalın adı: Knowledge-Based Systems
  • Jurnalın baxıldığı indekslər: Science Citation Index Expanded (SCI-EXPANDED), Scopus, Academic Search Premier, Applied Science & Technology Source, Computer & Applied Sciences, INSPEC, Library and Information Science Abstracts, Library, Information Science & Technology Abstracts (LISTA)
  • Açar sözlər: Adaptive thresholding, Bayesian network, Hierarchical independence thresholding, Informational independence, Probabilistic independence
  • Açıq Arxiv Kolleksiyası: Məqalə
  • Adres: Bəli

Qısa məlumat

Bayesian networks are powerful tools for knowledge representation and inference under conditions of uncertainty. However, learning an optimal Bayesian network classifier (BNC) is an NP-hard problem since its topology complexity increases exponentially with the number of attributes. Researchers proposed to apply information-theoretic criteria to measure conditional dependence, and independence assumptions are introduced implicitly or explicitly to simplify the network topology of BNC. In this paper, we clarify the mapping relationship between conditional mutual information and local topology, and then illustrate that informational independence does not correspond to probabilistic independence, the criterion of probabilistic independence does not necessarily hold for the independence topology. A novel framework of semi-naive Bayesian operation, called Hierarchical Independence Thresholding (HIT), is presented to efficiently identify informational conditional independence and probabilistic conditional independence by applying an adaptive thresholding method, redundant edges will be filtered out and the learned topology will fit the data better. Extensive experimental evaluation on 58 publicly available datasets reveals that when HIT is applied to BNCs (such as tree augmented Naive Bayes or k-dependence Bayesian classifier), the final BNCs achieve competitive classification performance compared to state-of-the-art learners such as Random Forest and Logistic regression.