The following preprints are provided here to allow for a deeper view of our research work, as well as to
promote the rapid dissemination of research results. Please consider, on the other hand, that these preprints
can differ from their published version in ways that may not be entirely negligible, and for this reason we
recommend you to refer to the published version whenever they have to be used or cited.
top2020
Akata, Z., Ballier, D., de Rijke, M., Dignum, F., Dignum, V., Eiben, G., Fokkens, A., Grossi, D., Hindriks, K., Hoos, H., Hung, H., Jonker, C., Monz, C., Neerincx, M., Oliehoek, F., Prakken, H., Schlobach, S., van der Gaag, L.C., van Harmelen, F., van Hoof, H., van Riemsdijk, B., van Wynsberghe, A., Verbrugge, R., Berheij, B., Vossen, P., Welling, M. (2020). A research agenda for hybrid intelligence: Augmenting human intellect with collaborative, adaptive, responsible and explainable artificial intelligence. IEEE Computer 53, pp. 18-28.
A research agenda for hybrid intelligence: Augmenting human intellect with collaborative, adaptive, responsible and explainable artificial intelligence
Authors: Akata, Z. and Ballier, D. and de Rijke, M. and Dignum, F. and Dignum, V. and Eiben, G. and Fokkens, A. and Grossi, D. and Hindriks, K. and Hoos, H. and Hung, H. and Jonker, C. and Monz, C. and Neerincx, M. and Oliehoek, F. and Prakken, H. and Schlobach, S. and van der Gaag, L.C. and van Harmelen, F. and van Hoof, H. and van Riemsdijk, B. and van Wynsberghe, A. and Verbrugge, R. and Berheij, B. and Vossen, P. and Welling, M.
Year: 2020
Abstract: We define Hybrid Intelligence (HI) as the combination of human and machine intelligence, augmenting human intellect and capabilities instead of replacing them, and achieve goals that were unreachable by either humans or machines alone. We identify Hybrid Intelligence as an important new research focus for the field of Artificial Intelligence, and we set a research agenda for Hybrid Intelligence by formulating four challenges: how do we develop AI systems that work in synergy with humans (Collaborative HI); how can these systems learn from and adapt to humans and their environment (Adaptive HI); how do we ensure that they behave ethically and responsibly (Responsible HI); and how can AI systems and humans share and explain their awareness, goals and strategies (Explainable HI). For each of these four challenges, we survey the state of the art and define a research agenda.
Published in IEEE Computer 53, pp. 18-28.
A research agenda for hybrid intelligence: Augmenting human intellect with collaborative, adaptive, responsible and explainable artificial intelligence
@ARTICLE{Linda2020b,
title = {A research agenda for hybrid intelligence: {A}ugmenting human intellect with collaborative, adaptive, responsible and explainable artificial intelligence},
journal = {{IEEE} Computer},
volume = {53},
author = {Akata, Z. and Ballier, D. and de Rijke, M. and Dignum, F. and Dignum, V. and Eiben, G. and Fokkens, A. and Grossi, D. and Hindriks, K. and Hoos, H. and Hung, H. and Jonker, C. and Monz, C. and Neerincx, M. and Oliehoek, F. and Prakken, H. and Schlobach, S. and van der Gaag, L.C. and van Harmelen, F. and van Hoof, H. and van Riemsdijk, B. and van Wynsberghe, A. and Verbrugge, R. and Berheij, B. and Vossen, P. and Welling, M.},
pages = {18-28},
year = {2020},
doi = {10.1109/MC.2020.2996587}
}
Download
Antonucci, A., Tiotto, T. (2020). Approximate MMAP by marginal search. In Brawner, K.W., Barták, R., Bell, E. (Eds), Proceedings of the Thirty-third International Florida Artificial Intelligence Research Society Conference (FLAIRS-33), AAAI Press, North Miami Beach, Florida, USA, pp. 181–184.
Approximate MMAP by marginal search
Authors: Antonucci, A. and Tiotto, T.
Year: 2020
Abstract: We present a heuristic strategy for marginal MAP (MMAP) queries in graphical models. The algorithm is based on a reduction of the task to a polynomial number of marginal inference computations. Given an input evidence, the marginals mass functions of the variables to be explained are computed. Marginal information gain is used to decide the variables to be explained first, and their most probable marginal states are consequently moved to the evidence. The sequential iteration of this procedure leads to a MMAP explanation and the minimum information gain obtained during the process can be regarded as a confidence measure for the explanation. Preliminary experiments show that the proposed confidence measure is properly detecting instances for which the algorithm is accurate and, for sufficiently high confidence levels, the algorithm gives the exact solution or an approximation whose Hamming distance from the exact one is small.
Published in Brawner, K.W., Barták, R., Bell, E. (Eds), Proceedings of the Thirty-third International Florida Artificial Intelligence Research Society Conference (FLAIRS-33), AAAI Press, North Miami Beach, Florida, USA, pp. 181–184.
Approximate MMAP by marginal search
@INPROCEEDINGS{antonucci2020a,
title = {Approximate {MMAP} by marginal search},
editor = {Brawner, K.W. and Bart\'ak, R. and Bell, E.},
publisher = {AAAI Press},
address = {North Miami Beach, Florida, USA},
booktitle = {Proceedings of the Thirty-{t}hird International Florida Artificial Intelligence Research Society Conference ({FLAIRS}-33)},
author = {Antonucci, A. and Tiotto, T.},
pages = {181--184},
year = {2020},
url = {https://www.aaai.org/Library/FLAIRS/flairs20contents.php}
}
Download
Azzimonti, L., Corani, G., Scutari, M. (2020). Structure learning from related data sets with a hierarchical Bayesian score. In Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Structure learning from related data sets with a hierarchical Bayesian score
Authors: Azzimonti, L. and Corani, G. and Scutari, M.
Year: 2020
Abstract: Score functions for learning the structure of Bayesian networks in the literature assume that data are a homogeneous set of observations; whereas it is often the case that they comprise different related, but not homogeneous, data sets collected in different ways. In this paper we propose a new Bayesian Dirichlet score, which we call Bayesian Hierarchical Dirichlet (BHD). The proposed score is based on a hierarchical model that pools information across data sets to learn a single encompassing network structure, while taking into account the differences in their probabilistic structures. We derive a closed-form expression for BHD using a variational approximation of the marginal likelihood and we study its performance using simulated data. We find that, when data comprise multiple related data sets, BHD outperforms the Bayesian Dirichlet equivalent uniform (BDeu) score in terms of reconstruction accuracy as measured by the Structural Hamming distance, and that it is as accurate as BDeu when data are homogeneous. Moreover, the estimated networks are sparser and therefore more interpretable than those obtained with BDeu, thanks to a lower number of false positive arcs.
Accepted in Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Structure learning from related data sets with a hierarchical Bayesian score
@INPROCEEDINGS{azzimonti2020a,
title = {Structure learning from related data sets with a hierarchical {B}ayesian score},
publisher = {JMLR.org},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models ({PGM} 2020)},
author = {Azzimonti, L. and Corani, G. and Scutari, M.},
year = {2020},
url = {https://pgm2020.cs.aau.dk/}
}
Download
Azzimonti, D., Rottondi, C., Giusti, A., Tornatore, M., Bianco, A. (2020). Active vs transfer learning approaches for qot estimation with small training datasets. In Optical Fiber Communication Conference (OFC 2020), Optical Society of America, M4E.1.
Active vs transfer learning approaches for qot estimation with small training datasets
Authors: Azzimonti, D. and Rottondi, C. and Giusti, A. and Tornatore, M. and Bianco, A.
Year: 2020
Abstract: We compare the level of accuracy achieved by active learning and domain adaptation approaches for quality of transmission estimation of an unestablished lightpath, in presence of small-sized training datasets.
Published in Optical Fiber Communication Conference (OFC 2020), Optical Society of America, M4E.1.
Active vs transfer learning approaches for qot estimation with small training datasets
@INPROCEEDINGS{azzimontid2020b,
title = {Active vs transfer learning approaches for qot estimation with small training datasets},
publisher = {Optical Society of America},
booktitle = {Optical Fiber Communication Conference ({OFC} 2020)},
author = {Azzimonti, D. and Rottondi, C. and Giusti, A. and Tornatore, M. and Bianco, A.},
pages = {M4E.1},
year = {2020},
doi = {10.1364/OFC.2020.M4E.1}
}
Download
Azzimonti, D., Rottondi, C., Tornatore, M. (2020). Reducing probes for quality of transmission estimation in optical networks with active learning. J. Opt. Commun. Netw. 12(1), pp. A38–A48.
Reducing probes for quality of transmission estimation in optical networks with active learning
Authors: Azzimonti, D. and Rottondi, C. and Tornatore, M.
Year: 2020
Abstract: Estimating the quality of transmission (QoT) of a lightpath before its establishment is a critical procedure for efficient design and management of optical networks. Recently, supervised machine learning (ML) techniques for QoT estimation have been proposed as an effective alternative to well-established, yet approximated, analytic models that often require the introduction of conservative margins to compensate for model inaccuracies and uncertainties. Unfortunately, to ensure high estimation accuracy, the training set (i.e., the set of historical field data, or “samples,” required to train these supervised ML algorithms) must be very large, while in real network deployments, the number of monitored/monitorable lightpaths is limited by several practical considerations. This is especially true for lightpaths with an above-threshold bit error rate (BER) (i.e., malfunctioning or wrongly dimensioned lightpaths), which are infrequently observed during network operation. Samples with above-threshold BERs can be acquired by deploying probe lightpaths, but at the cost of increased operational expenditures and wastage of spectral resources. In this paper, we propose to use active learning to reduce the number of probes needed for ML-based QoT estimation. We build an estimation model based on Gaussian processes, which allows iterative identification of those QoT instances that minimize estimation uncertainty. Numerical results using synthetically generated datasets show that, by using the proposed active learning approach, we can achieve the same performance of standard offline supervised ML methods, but with a remarkable reduction (at least 5% and up to 75%) in the number of training samples.
Published in J. Opt. Commun. Netw. 12(1), OSA, pp. A38–A48.
Reducing probes for quality of transmission estimation in optical networks with active learning
@ARTICLE{azzimontid2020a,
title = {Reducing probes for quality of transmission estimation in optical networks with active learning},
journal = {J. Opt. Commun. Netw.},
publisher = {OSA},
volume = {12},
author = {Azzimonti, D. and Rottondi, C. and Tornatore, M.},
number = {1},
pages = {A38--A48},
year = {2020},
doi = {10.1364/JOCN.12.000A38}
}
Download
Benavoli, A., Azzimonti, D., Piga, D. (2020). Skew gaussian processes for classification. Machine Learning 109.
Skew gaussian processes for classification
Authors: Benavoli, A. and Azzimonti, D. and Piga, D.
Year: 2020
Abstract: Gaussian processes (GPs) are distributions over functions, which provide a Bayesian nonparametric approach to regression and classification. In spite of their success, GPs have limited use in some applications, for example, in some cases a symmetric distribution with respect to its mean is an unreasonable model. This implies, for instance, that the mean and the median coincide, while the mean and median in an asymmetric (skewed) distribution can be different numbers. In this paper, we propose skew-Gaussian processes (SkewGPs) as a non-parametric prior over functions. A SkewGP extends the multivariate unified skew-normal distribution over finite dimensional vectors to a stochastic processes. The SkewGP class of distributions includes GPs and, therefore, SkewGPs inherit all good properties of GPs and increase their flexibility by allowing asymmetry in the probabilistic model. By exploiting the fact that SkewGP and probit likelihood are conjugate model, we derive closed form expressions for the marginal likelihood and predictive distribution of this new nonparametric classifier. We verify empirically that the proposed SkewGP classifier provides a better performance than a GP classifier based on either Laplace’s method or expectation propagation.
Published in Machine Learning 109.
Skew gaussian processes for classification
@ARTICLE{azzimontid2020c,
title = {Skew gaussian processes for classification},
journal = {Machine Learning},
volume = {109},
author = {Benavoli, A. and Azzimonti, D. and Piga, D.},
year = {2020},
doi = {10.1007/s10994-020-05906-3}
}
Download
Cabañas, R., Antonucci, A., Huber, D., Zaffalon, M. (2020). Credici: a java library for causal inference by credal networks. In Proceedings of the 10th International Conference on Probabilistic Graphical Models, Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
Credici: a java library for causal inference by credal networks
Authors: Cabañas, R. and Antonucci, A. and Huber, D. and Zaffalon, M.
Year: 2020
Abstract: We present CREDICI, a Java open-source tool for causal inference based on credal networks. Credal networks are an extension of Bayesian networks where local probability mass functions are only constrained to belong to given, so-called \emphcredal, sets. CREDICI is based on the recent work of Zaffalon et al. (2020), where an equivalence between Pearl's structural causal models and credal networks has been derived. This allows to reduce a counterfactual query in a causal model to a standard query in a credal network, even in the case of unidentifiable causal effects. The necessary transformations and data structures are implemented in CREDICI, while inferences are eventually computed by CREMA (Huber et al., 2020), a twin library for general credal network inference. Here we discuss the main implementation challenges and possible outlooks.
Published in Proceedings of the 10th International Conference on Probabilistic Graphical Models, Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
Credici: a java library for causal inference by credal networks
@INPROCEEDINGS{cabanas2020a,
title = {Credici: a java library for causal inference by credal networks},
publisher = {PMLR},
address = {Aalborg, Denmark},
series = {Proceedings of Machine Learning Research},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models},
author = {Caba\~nas, R. and Antonucci, A. and Huber, D. and Zaffalon, M.},
year = {2020}
}
Download
Cabañas, R., Cózar, J., Salmerón, A., Masegosa, A.R. (2020). Probabilistic graphical models with neural networks in inferpy. In Proceedings of the 10th International Conference on Probabilistic Graphical Models, Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
Probabilistic graphical models with neural networks in inferpy
Authors: Cabañas, R. and Cózar, J. and Salmerón, A. and Masegosa, A.R.
Year: 2020
Abstract: InferPy is an open-source Python package for variational inference in probabilistic models containing neural networks. Other similar libraries are often difficult for non-expert users. InferPy provides a much more compact and simple way to code such models, at the expense of slightly reducing expressibility and flexibility. The main objective of this package is to permit its use without having a strong theoretical background or thorough knowledge of the deep learning frameworks.
Published in Proceedings of the 10th International Conference on Probabilistic Graphical Models, Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
Probabilistic graphical models with neural networks in inferpy
@INPROCEEDINGS{cabanas2020b,
title = {Probabilistic graphical models with neural networks in inferpy},
publisher = {PMLR},
address = {Aalborg, Denmark},
series = {Proceedings of Machine Learning Research},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models},
author = {Caba\~nas, R. and C\'ozar, J. and Salmer\'on, A. and Masegosa, A.R.},
year = {2020}
}
Download
Cannelli, L., Facchinei, F., Scutari, G., Kungurtsev, V. (2020). Asynchronous optimization over graphs: linear convergence under error bound conditions. IEEE Transactions on Automatic Control.
Asynchronous optimization over graphs: linear convergence under error bound conditions
Authors: Cannelli, L. and Facchinei, F. and Scutari, G. and Kungurtsev, V.
Year: 2020
Abstract: We consider convex and nonconvex constrained optimization with a partially separable objective function: agents minimize the sum of local objective functions, each of which is known only by the associated agent and depends on the variables of that agent and those of a few others. This partitioned setting arises in several applications of practical interest. We propose what is, to the best of our knowledge, the first distributed, asynchronous algorithm with rate guarantees for this class of problems. When the objective function is nonconvex, the algorithm provably converges to a stationary solution at a sublinear rate whereas linear rate is achieved under the renowned Luo-Tseng error bound condition (which is less stringent than strong convexity). Numerical results on matrix completion and LASSO problems show the effectiveness of our method.
Accepted in IEEE Transactions on Automatic Control.
Asynchronous optimization over graphs: linear convergence under error bound conditions
@ARTICLE{cannelli2020a,
title = {Asynchronous optimization over graphs: linear convergence under error bound conditions},
journal = {{IEEE} Transactions on Automatic Control},
author = {Cannelli, L. and Facchinei, F. and Scutari, G. and Kungurtsev, V.},
year = {2020},
doi = {10.1109/TAC.2020.3033490}
}
Download
Casanova, A., Miranda, E., Zaffalon, M. (2020). Social pooling of beliefs and values with desirability. Proceedings of the 33rd International Flairs Conference (FLAIRS-33).
Social pooling of beliefs and values with desirability
Authors: Casanova, A. and Miranda, E. and Zaffalon, M.
Year: 2020
Abstract: The problem of aggregating beliefs and values of rational
subjects is treated with the formalism of sets of desirable
gambles. This leads on the one hand to a new perspective
of traditional results of social choice (in particular Arrow’s
theorem as well as sufficient conditions for the existence of
an oligarchy and democracy) and on the other hand to use the
same framework to create connections with opinion pooling.
In particular, we show that weak Pareto can be derived as a
coherence requirement and discuss the aggregation of state
independent beliefs.
Introduction
Accepted in Proceedings of the 33rd International Flairs Conference (FLAIRS-33).
Social pooling of beliefs and values with desirability
@ARTICLE{casanova2020a,
title = {Social pooling of beliefs and values with desirability},
journal = {Proceedings of the 33rd International Flairs Conference ({FLAIRS}-33)},
author = {Casanova, A. and Miranda, E. and Zaffalon, M.},
year = {2020},
url = {https://www.flairs-33.info/}
}
Download
Coletti, G., van der Gaag, L.C., Petturiti, D., Vantaggi, B. (2020). Detecting correlation between extreme probability events. International Journal of General Systems 49(1), pp. 64–87.
Detecting correlation between extreme probability events
Authors: Coletti, G. and van der Gaag, L.C. and Petturiti, D. and Vantaggi, B.
Year: 2020
Abstract: As classical definitions of correlation give rise to counterintuitive statements when extreme probability events are involved, we introduce enhanced notions of positive and negative correlation in the general framework of coherent conditional probability. These notions allow to handle extreme probability events in a principled way by accommodating the different levels of strength of the zero probabilities involved (namely, zero layers). Since the detection of correlations by means of zero layers is computationally challenging, we provide a full characterization relying on only conditional probability values.
Keywords: Conditional probability, Extreme probability event, Coherence, Correlation
Published in International Journal of General Systems 49(1), pp. 64–87.
Detecting correlation between extreme probability events
@ARTICLE{Linda2019e,
title = {Detecting correlation between extreme probability events},
journal = {International Journal of General Systems },
volume = {49},
author = {Coletti, G. and van der Gaag, L.C. and Petturiti, D. and Vantaggi, B.},
number = {1},
pages = {64--87},
year = {2020},
doi = {10.1080/03081079.2019.1692005}
}
Download
Corani, G., Azzimonti, D., Augusto, J.P.S.C., Zaffalon, M. (2020). Probabilistic reconciliation of hierarchical forecast via Bayes’ rule. In Proceedings ECML - PKDD 2020.
Probabilistic reconciliation of hierarchical forecast via Bayes’ rule
Authors: Corani, G. and Azzimonti, D. and Augusto, J.P.S.C. and Zaffalon, M.
Year: 2020
Abstract: We present a novel approach for reconciling hierarchical forecasts, based on Bayes’ rule. We define a prior distribution for the bottom time series of the hierarchy, based on the bottom base forecasts. Then we update their distribution via Bayes’ rule, based on the base forecasts
for the upper time series. Under the Gaussian assumption, we derive the updating in closed-form. We derive two algorithms, which differ as for the assumed independencies. We discuss their relation with the MinT reconciliation algorithm and with the Kalman filter, and we compare them experimentally.
Accepted in Proceedings ECML - PKDD 2020.
Probabilistic reconciliation of hierarchical forecast via Bayes’ rule
@INPROCEEDINGS{corani2020a,
title = {Probabilistic reconciliation of hierarchical forecast via {B}ayes’ rule},
booktitle = {Proceedings {ECML} - {PKDD} 2020},
author = {Corani, G. and Azzimonti, D. and Augusto, J.P.S.C. and Zaffalon, M.},
year = {2020}
}
Download
Cózar, J., Cabañas, R., Salmerón, A., Masegosa, A.R. (2020). Inferpy: probabilistic modeling with deep neural networks made easy. Neurocomputing 415, pp. 408–410.
Inferpy: probabilistic modeling with deep neural networks made easy
Authors: Cózar, J. and Cabañas, R. and Salmerón, A. and Masegosa, A.R.
Year: 2020
Abstract: InferPy is a Python package for probabilistic modeling with deep neural networks. It defines a user-friendly API that trades-off model complexity with ease of use, unlike other libraries whose focus is on dealing with very general probabilistic models at the cost of having a more complex API. In particular, this package allows to define, learn and evaluate general hierarchical probabilistic models containing deep neural networks in a compact and simple way. InferPy is built on top of Tensorflow Probability and Keras.
Published in Neurocomputing 415, pp. 408–410.
Inferpy: probabilistic modeling with deep neural networks made easy
@ARTICLE{cabanas2020c,
title = {Inferpy: probabilistic modeling with deep neural networks made easy},
journal = {Neurocomputing},
volume = {415},
author = {C\'ozar, J. and Caba\~nas, R. and Salmer\'on, A. and Masegosa, A.R.},
pages = {408--410},
year = {2020},
doi = {https://doi.org/10.1016/j.neucom.2020.07.117},
url = {http://www.sciencedirect.com/science/article/pii/S092523122031328X}
}
Download
Forgione, M., Piga, D. (2020). Model structures and fitting criteria for system identification with neural networks. In Proceedings of the 14th IEEE International Conference Application of Information and Communication Technologies (AICT 20).
Model structures and fitting criteria for system identification with neural networks
Authors: Forgione, M. and Piga, D.
Year: 2020
Abstract: This paper focuses on the identification of dynamical systems with tailor-made model structures, where neural networks are used to approximate uncertain components and domain knowledge is retained, if avail-able. These model structures are fitted to measured data using different criteria including a computationally efficient approach minimizing a regularized multi-step ahead simulation error. In this approach, the neural network parameters are estimated along with the initial conditions used to simulate the output signal in small-size subsequences. A regularization term is included in the fitting cost in order to enforce these initial conditions to be consistent with the estimated system dynamics. Pitfalls and limitations of naive one-step prediction and simulation error minimization are also discussed.
Published in Proceedings of the 14th IEEE International Conference Application of Information and Communication Technologies (AICT 20).
Model structures and fitting criteria for system identification with neural networks
@INPROCEEDINGS{forgione2020b,
title = {Model structures and fitting criteria for system identification with neural networks},
booktitle = {Proceedings of the 14th {IEEE} International Conference Application of Information and Communication Technologies ({AICT} 20)},
author = {Forgione, M. and Piga, D.},
year = {2020},
url = {https://arxiv.org/pdf/1911.13034.pdf}
}
Download
Forgione, M., Piga, D., Bemporad, A. (2020). Efficient Calibration of Embedded MPC. In Proceedings of the 21st IFAC World Congress (IFAC 20).
Efficient Calibration of Embedded MPC
Authors: Forgione, M. and Piga, D. and Bemporad, A.
Year: 2020
Abstract: Model Predictive Control (MPC) is a powerful and flexible design tool of high-performance controllers for physical systems in the presence of input and output constraints. A challenge for the practitioner applying MPC is the need of tuning a large number of parameters such as prediction and control horizons, weight matrices of the MPC cost function, and observer gains, according to different trade-offs. The MPC design task is even more involved when the control law has to be deployed to an embedded hardware unit endowed with limited computational resources. In this case, real-time system requirements limit the complexity of the applicable MPC configuration, engendering additional design tradeoffs and requiring to tune further parameters, such as the sampling time and the tolerances used in the on-line numerical solver. To take into account closed-loop performance and real-time requirements, in this paper we tackle the em-bedded MPC design problem using a global, data-driven, optimization approach We showcase the potential of this approach by tuning an MPC controller on two hardware platforms characterized by largely different computational capabilities
Published in Proceedings of the 21st IFAC World Congress (IFAC 20).
Efficient Calibration of Embedded MPC
@INPROCEEDINGS{forgione2020a,
title = {Efficient {C}alibration of {E}mbedded {MPC}},
booktitle = {Proceedings of the 21st {IFAC} World Congress ({IFAC} 20)},
author = {Forgione, M. and Piga, D. and Bemporad, A.},
year = {2020},
url = {https://arxiv.org/pdf/1911.13021.pdf}
}
Download
van der Gaag, L.C., Bolt, J.H. (2020). Poset representations for sets of elementary triplets. In Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Poset representations for sets of elementary triplets
Authors: van der Gaag, L.C. and Bolt, J.H.
Year: 2020
Abstract: Semi-graphoid independence relations, composed of independence triplets, are typically exponentially large in the number of variables involved. For compact representation of such a relation, just a subset of its triplets, called a basis, are listed explicitly, while its other triplets remain implicit through a set of derivation rules. Two types of basis were defined for this purpose, which are the dominant-triplet basis and the elementary-triplet basis, of which the latter is commonly assumed to be significantly larger in size in general. In this paper we introduce the elementary po-triplet as a compact representation of multiple elementary triplets, by using separating posets. By exploiting this new representation, the size of an elementary-triplet basis can be reduced considerably. For computing the elementary closure of a starting set of po-triplets, we present an elegant algorithm that operates on the least and largest elements of the separating posets involved.
Published in Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Poset representations for sets of elementary triplets
@INPROCEEDINGS{Linda2020c,
title = {Poset representations for sets of elementary triplets},
publisher = {JMLR.org},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models ({PGM} 2020)},
author = {van der Gaag, L.C. and Bolt, J.H.},
year = {2020},
url = {https://pgm2020.cs.aau.dk/index.php/accepted-papers/}
}
Download
van der Gaag, L.C., Renooij, S., Facchini, A. (2020). Building causal interaction models by recursive unfolding. In Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Building causal interaction models by recursive unfolding
Authors: van der Gaag, L.C. and Renooij, S. and Facchini, A.
Year: 2020
Abstract: Causal interaction models, such as the well-known noisy-or and leaky noisy-or models, have become quite popular as a means to parameterize conditional probability tables for Bayesian networks. In this paper we focus on the engineering of subnetworks to represent such models and present a novel technique called recursive unfolding for this purpose. This technique allows inserting, removing and merging cause variables in an interaction model at will, without affecting the underlying represented information. We detail the technique, with the recursion invariants involved, and illustrate its practical use for Bayesian-network engineering by means of a small example.
Accepted in Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Building causal interaction models by recursive unfolding
@INPROCEEDINGS{Linda2020a,
title = {Building causal interaction models by recursive unfolding},
publisher = {JMLR.org},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models ({PGM} 2020)},
author = {van der Gaag, L.C. and Renooij, S. and Facchini, A.},
year = {2020},
url = {https://pgm2020.cs.aau.dk/index.php/accepted-papers/}
}
Download
Geh, R., Mauá, D.D., Antonucci, A. (2020). Learning probabilistic sentential decision diagrams by sampling. In Proceedings of the Eight Symposium on Knowledge Discovery, Mining and Learning (KMILE 2020), SBC, Porto Alegre, RS, Brasil, pp. 129–136.
Learning probabilistic sentential decision diagrams by sampling
Authors: Geh, R. and Mauá, D.D. and Antonucci, A.
Year: 2020
Abstract: Probabilistic circuits are deep probabilistic models with neural-network-like semantics capable of accurately and efficiently answering probabilistic queries without sacrificing expressiveness. Probabilistic Sentential Decision Diagrams (PSDDs) are a subclass of probabilistic circuits able to embed logical constraints to the circuit’s structure. In doing so, they obtain extra expressiveness with empirical optimal performance. Despite achieving competitive performance compared to other state-of-the-art competitors, there have been very few attempts at learning PSDDs from a combination of both data and knowledge in the form of logical formulae. Our work investigates sampling random PSDDs consistent with domain knowledge and evaluating against state-of-the-art probabilistic models. We propose a method of sampling that retains important structural constraints on the circuit’s graph that guarantee query tractability. Finally, we show that these samples are able to achieve competitive performance even on larger domains.
Published in Proceedings of the Eight Symposium on Knowledge Discovery, Mining and Learning (KMILE 2020), SBC, Porto Alegre, RS, Brasil, pp. 129–136.
Learning probabilistic sentential decision diagrams by sampling
@INPROCEEDINGS{antonucci2020c,
title = {Learning probabilistic sentential decision diagrams by sampling},
publisher = {SBC},
address = {Porto Alegre, RS, Brasil},
booktitle = {Proceedings of the Eight Symposium on Knowledge Discovery, Mining and Learning ({KMILE} 2020)},
author = {Geh, R. and Mau\'a, D.D. and Antonucci, A.},
pages = {129--136},
year = {2020},
doi = {10.5753/kdmile.2020.11968}
}
Download
Huber, D., Cabañas, R., Antonucci, A., Zaffalon, M. (2020). CREMA: a Java library for credal network inference. In Jaeger, M., Nielsen, T.D. (Eds), Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
CREMA: a Java library for credal network inference
Authors: Huber, D. and Cabañas, R. and Antonucci, A. and Zaffalon, M.
Year: 2020
Abstract: We present CREMA (Credal Models Algorithms), a Java library for inference in credal networks. These models are analogous to Bayesian networks, but their local parameters are only constrained to vary in, so-called credal, sets. Inference in credal networks is intended as the computation of the bounds of a query with respect to those local variations. For credal networks the task is harder than in Bayesian networks, being NPPP-hard in general models. Yet, scalable approximate algorithms have been shown to provide good accuracies on large or dense models, while exact techniques can be designed to process small or sparse models. CREMA embeds these algorithms and also offers an API to build and query credal networks together with a specification format. This makes CREMA, whose features are discussed and described by a simple example, the most advanced tool for credal network modelling and inference developed so far.
Published in Jaeger, M., Nielsen, T.D. (Eds), Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
CREMA: a Java library for credal network inference
@INPROCEEDINGS{huber2020a,
title = {{CREMA}: a {J}ava library for credal network inference},
editor = {Jaeger, M. and Nielsen, T.D.},
publisher = {PMLR},
address = {Aalborg, Denmark},
series = {Proceedings of Machine Learning Research},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models ({PGM} 2020)},
author = {Huber, D. and Caba\~nas, R. and Antonucci, A. and Zaffalon, M.},
year = {2020},
url = {https://pgm2020.cs.aau.dk}
}
Download
Kanjirangat, V., Mellace, S., Antonucci, A. (2020). Temporal embeddings and transformer models for narrative text understanding. In Third International Workshop on Narrative Extraction from Texts (Text2Story 20), 42nd European Conference on Information Retrieval (ECIR 20), ceur.
Temporal embeddings and transformer models for narrative text understanding
Authors: Kanjirangat, V. and Mellace, S. and Antonucci, A.
Year: 2020
Abstract: We present two deep learning approaches to narrative text understanding for character relationship modelling. The temporal evolution of these relations is described by dynamic word embeddings, that are designed to learn semantic changes over time. An empirical analysis of the corresponding character trajectories shows that such approaches are effective in depicting dynamic evolution. A supervised learning approach based on the state-of-the-art transformer model BERT is used instead to detect static relations between characters. The empirical validation shows that such events (e.g., two characters belonging to the same family) might be spotted with good accuracy, even when using automatically annotated data. This provides a deeper understanding of narrative plots based on the identification of key facts. Standard clustering techniques are finally used for character de-aliasing, a necessary pre-processing step for both approaches. Overall, deep learning models appear to be suitable for narrative text understanding, while also providing a challenging and unexploited benchmark for general natural language understanding.
Published in Third International Workshop on Narrative Extraction from Texts (Text2Story 20), 42nd European Conference on Information Retrieval (ECIR 20), ceur.
Temporal embeddings and transformer models for narrative text understanding
@INPROCEEDINGS{vani2020b,
title = {Temporal embeddings and transformer models for narrative text understanding},
publisher = {ceur},
booktitle = {Third International Workshop on Narrative Extraction {f}rom Texts ({Text2Story} 20), 42nd European Conference on Information Retrieval ({ECIR} 20)},
author = {Kanjirangat, V. and Mellace, S. and Antonucci, A.},
year = {2020},
url = {http://ceur-ws.org/Vol-2593/paper9.pdf}
}
Download
Kanjirangat, V., Mitrovic, S., Antonucci, A., Rinaldi, F. (2020). SST-BERT at SemEval-2020 task 1: semantic shift tracing by clustering in BERT-based embedding spaces. In SemEval-2020 Task 1: Unsupervised Lexical Semantic Change Detection.To appear In Proceedings of the 14th International Workshop on Semantic Evaluation, Barcelona, Spain.
SST-BERT at SemEval-2020 task 1: semantic shift tracing by clustering in BERT-based embedding spaces
Authors: Kanjirangat, V. and Mitrovic, S. and Antonucci, A. and Rinaldi, F.
Year: 2020
Abstract: Lexical semantic change detection (also known as semantic shift tracing) is a task of identifying words that have changed their meaning over time. Unsupervised semantic shift tracing, focal point of SemEval2020, is particularly challenging. Given the unsupervised setup, in this work, we propose to identify clusters among different occurrences of each target word, considering these as representatives of different word meanings. As such, disagreements in obtained clusters naturally allow to quantify the level of semantic shift per each target word in four target languages. To leverage this idea, clustering is performed on contextualized (BERT-based) embeddings of word occurrences. The obtained results show that our approach performs well both measured separately (per language) and overall, where we surpass all provided SemEval baselines.
Published in SemEval-2020 Task 1: Unsupervised Lexical Semantic Change Detection.To appear In Proceedings of the 14th International Workshop on Semantic Evaluation, Barcelona, Spain.
SST-BERT at SemEval-2020 task 1: semantic shift tracing by clustering in BERT-based embedding spaces
@INPROCEEDINGS{vani2020semeval,
title = {{SST}-{BERT} at {SemEval}-2020 task 1: semantic shift tracing by clustering in {BERT}-based embedding spaces},
booktitle = {{SemEval}-2020 Task 1: Unsupervised Lexical Semantic Change Detection.To {a}ppear In Proceedings of the 14th International Workshop on Semantic Evaluation, Barcelona, Spain},
author = {Kanjirangat, V. and Mitrovic, S. and Antonucci, A. and Rinaldi, F.},
year = {2020},
url = {http://alt.qcri.org/semeval2020/}
}
Download
Kern, H., Corani, G., Huber, D., Vermes, N., Zaffalon, M., Varini, M., Wenzel, C., Fringer, A. (2020). Impact on place of death in cancer patients: a causal exploration in southern switzerland. BMC Palliative Care 19, 160.
Impact on place of death in cancer patients: a causal exploration in southern switzerland
Authors: Kern, H. and Corani, G. and Huber, D. and Vermes, N. and Zaffalon, M. and Varini, M. and Wenzel, C. and Fringer, A.
Year: 2020
Abstract:
Background
Most terminally ill cancer patients prefer to die at home, but a majority die in institutional settings. Research questions about this discrepancy have not been fully answered. This study applies artificial intelligence and machine learning techniques to explore the complex network of factors and the cause-effect relationships affecting the place of death, with the ultimate aim of developing policies favouring home-based end-of-life care.
Methods
A data mining algorithm and a causal probabilistic model for data analysis were developed with information derived from expert knowledge that was merged with data from 116 deceased cancer patients in southern Switzerland. This data set was obtained via a retrospective clinical chart review.
Results
Dependencies of disease and treatment-related decisions demonstrate an influence on the place of death of 13%. Anticancer treatment in advanced disease prevents or delays communication about the end of life between oncologists, patients and families. Unknown preferences for the place of death represent a great barrier to a home death. A further barrier is the limited availability of family caregivers for terminal home care. The family’s preference for the last place of care has a high impact on the place of death of 51%, while the influence of the patient’s preference is low, at 14%. Approximately one-third of family systems can be empowered by health care professionals to provide home care through open end-of-life communication and good symptom management. Such intervention has an influence on the place of death of 17%. If families express a convincing preference for home care, the involvement of a specialist palliative home care service can increase the probability of home deaths by 24%.
Conclusion
Concerning death at home, open communication about death and dying is essential. Furthermore, for the patient preference for home care to be respected, the family’s decision for the last place of care seems to be key. The early initiation of family-centred palliative care and the provision of specialist palliative home care for patients who wish to die at home are suggested.
Published in BMC Palliative Care 19, Research Square, 160.
Impact on place of death in cancer patients: a causal exploration in southern switzerland
@ARTICLE{Kern2020,
title = {Impact on place of death in cancer patients: a causal exploration in southern switzerland},
journal = {{BMC} Palliative Care},
publisher = {Research Square},
volume = {19},
author = {Kern, H. and Corani, G. and Huber, D. and Vermes, N. and Zaffalon, M. and Varini, M. and Wenzel, C. and Fringer, A.},
pages = {160},
year = {2020},
doi = {10.21203/rs.3.rs-29758/v3}
}
Download
Laurain, V., Tóth, R., Piga, D., Darwish, M.A.H. (2020). Sparse RKHS estimation via globally convex optimization and its application in LPV-IO identification. Automatica 115.
Sparse RKHS estimation via globally convex optimization and its application in LPV-IO identification
Authors: Laurain, V. and Tóth, R. and Piga, D. and Darwish, M.A.H.
Year: 2020
Abstract: Reproducing kernel Hilbert spaces, Elastic net, Support vector machines, Gaussian processes, Non-parametric estimation, Linear parameter-varying systems, Model order selection",
abstract = "Function estimation using the Reproducing Kernel Hilbert Space (RKHS) framework is a powerful tool for identification of a general class of nonlinear dynamical systems without requiring much a priori information on model orders and nonlinearities involved. However, the high degrees-of-freedom (DOFs) of RKHS estimators has its price, as in case of large scale function estimation problems, they often require a serious amount of data samples to explore the search space adequately for providing high-performance model estimates. In cases where nonlinear dynamic relations can be expressed as a sum of functions, the literature proposes solutions to this issue by enforcing sparsity for adequate restriction of the DOFs of the estimator, resulting in parsimonious model estimates. Unfortunately, all existing solutions are based on greedy approaches, leading to optimization schemes which cannot guarantee convergence to the global optimum. In this paper, we propose an L1-regularized non-parametric RKHS estimator which is the solution of a quadratic optimization problem. Effectiveness of the scheme is demonstrated on the non-parametric identification problem of LPV-IO models where the method solves simultaneously (i) the model order selection problem (in terms of number of input–output lags and input delay in the model structure) and (ii) determining the unknown functional dependency of the model coefficients on the scheduling variable directly from data. The paper also provides an extensive simulation study to illustrate effectiveness of the proposed scheme.
Published in Automatica 115.
Sparse RKHS estimation via globally convex optimization and its application in LPV-IO identification
@ARTICLE{piga2020b,
title = {Sparse {RKHS} estimation via globally convex optimization and its application in {LPV}-{IO} identification},
journal = {Automatica},
volume = {115},
author = {Laurain, V. and T\'oth, R. and Piga, D. and Darwish, M.A.H.},
year = {2020},
doi = {10.1016/j.automatica.2020.108914},
url = {http://www.sciencedirect.com/science/article/pii/S0005109820301126}
}
Download
Mangili, F., Broggini, D., Antonucci, A. (2020). Conversational recommender system by Bayesian methods. In Davis, Jesse, Tabia, Karim (Eds), Proceedings of the Fourteenth International Conference on Scalable Uncertainty Management (SUM 2020), Lecture Notes in Artificial Intelligence 12322, Springer, Cham, pp. 200–213.
Conversational recommender system by Bayesian methods
Authors: Mangili, F. and Broggini, D. and Antonucci, A.
Year: 2020
Abstract: We present a Bayesian approach to conversational recommender systems. After any interaction with the user, a probability mass function over the items is updated by the system. The conversational feature corresponds to a sequential discovery of the user preferences based on questions. Information-theoretic criteria are used to optimally shape the interactions and decide when the conversation ends. Most probable items are consequently recommended. Dedicated elicitation techniques for the prior probabilities of the parameters modelling the interactions are derived from basic structural judgements based on logical compatibility and symmetry assumptions. Such prior knowledge is combined with data for better item discrimination. Our Bayesian approach is validated against matrix factorization techniques for cold-start recommendations based on metadata using the popular benchmark data set MovieLens. Results show that the proposed approach allows to considerably reduce the number of interactions while maintaining good ranking performance.
Published in Davis, Jesse, Tabia, Karim (Eds), Proceedings of the Fourteenth International Conference on Scalable Uncertainty Management (SUM 2020), Lecture Notes in Artificial Intelligence 12322, Springer, Cham, pp. 200–213.
Conversational recommender system by Bayesian methods
@INPROCEEDINGS{mangili2020b,
title = {Conversational recommender system by {B}ayesian methods},
editor = {Davis, Jesse and Tabia, Karim},
publisher = {Springer, Cham},
series = {Lecture Notes in Artificial Intelligence},
volume = {12322},
booktitle = {Proceedings of the Fourteenth International Conference on Scalable Uncertainty Management ({SUM} 2020)},
author = {Mangili, F. and Broggini, D. and Antonucci, A.},
pages = {200--213},
year = {2020},
doi = {10.1007/978-3-030-58449-8_14}
}
Download
Mangili, F., Broggini, D., Antonucci, A., Alberti, M., Cimasoni, L. (2020). A Bayesian approach to conversational recommendation systems. AAAI 2020 Workshop on Interactive and Conversational Recommendation Systems (WICRS-20).
A Bayesian approach to conversational recommendation systems
Authors: Mangili, F. and Broggini, D. and Antonucci, A. and Alberti, M. and Cimasoni, L.
Year: 2020
Abstract: We present a conversational recommendation system based on a Bayesian approach. A probability mass function over the items is updated after any interaction with the user, with information-theoretic criteria optimally shaping the interaction and deciding when the conversation should be terminated and the most probable item consequently recommended. Dedicated elicitation techniques for the prior probabilities of the parameters modeling the interactions are derived from basic structural judgements. Such prior information can be combined with historical data to discriminate items with different recommendation histories. A case study based on the application of this approach to stagend.com, an online platform for booking entertainers, is finally discussed together with an empirical analysis showing the advantages in terms of recommendation quality and efficiency.
Accepted in AAAI 2020 Workshop on Interactive and Conversational Recommendation Systems (WICRS-20).
Note: Accepted for oral presentation
A Bayesian approach to conversational recommendation systems
@ARTICLE{mangili2020a,
title = {A {B}ayesian approach to conversational recommendation systems},
journal = {{AAAI} 2020 Workshop on Interactive and Conversational Recommendation Systems ({WICRS}-20)},
author = {Mangili, F. and Broggini, D. and Antonucci, A. and Alberti, M. and Cimasoni, L.},
year = {2020},
url = {https://sites.google.com/view/wicrs2020}
}
Download
Mattei, L., Antonucci, A., Mauá, D.D., Facchini, A., Llerena, J.V. (2020). Tractable inference in credal sentential decision diagrams. International Journal of Approximate Reasoning 125, pp. 26–48.
Tractable inference in credal sentential decision diagrams
Authors: Mattei, L. and Antonucci, A. and Mauá, D.D. and Facchini, A. and Llerena, J.V.
Year: 2020
Abstract: Probabilistic sentential decision diagrams are logic circuits where the inputs of disjunctive gates are annotated by probability values. They allow for a compact representation of joint probability mass functions defined over sets of Boolean variables, that are also consistent with the logical constraints defined by the circuit. The probabilities in such a model are usually “learned” from a set of observations. This leads to overconfident and prior-dependent inferences when data are scarce, unreliable or conflicting. In this work, we develop the credal sentential decision diagrams, a generalisation of their probabilistic counterpart that allows for replacing the local probabilities with (so-called credal) sets of mass functions. These models induce a joint credal set over the set of Boolean variables, that sharply assigns probability zero to states inconsistent with the logical constraints. Three inference algorithms are derived for these models. These allow to compute: (i) the lower and upper probabilities of an observation for an arbitrary number of variables; (ii) the lower and upper conditional probabilities for the state of a single variable given an observation; (iii) whether or not all the probabilistic sentential decision diagrams compatible with the credal specification have the same most probable explanation of a given set of variables given an observation of the other variables. These inferences are tractable, as all the three algorithms, based on bottom-up traversal with local linear programming tasks on the disjunctive gates, can be solved in polynomial time with respect to the circuit size. The first algorithm is always exact, while the remaining two might induce a conservative (outer) approximation in the case of multiply connected circuits. A semantics for this approximation together with an auxiliary algorithm able to decide whether or not the result is exact is also provided together with a brute-force characterization of the exact inference in these cases. For a first empirical validation, we consider a simple application based on noisy seven-segment display images. The credal models are observed to properly distinguish between easy and hard-to-detect instances and outperform other generative models not able to cope with logical constraints.
Published in International Journal of Approximate Reasoning 125, pp. 26–48.
Tractable inference in credal sentential decision diagrams
@ARTICLE{antonucci2020d,
title = {Tractable inference in credal sentential decision diagrams},
journal = {International Journal of Approximate Reasoning},
volume = {125},
author = {Mattei, L. and Antonucci, A. and Mau\'a, D.D. and Facchini, A. and Llerena, J.V.},
pages = {26--48},
year = {2020},
doi = {10.1016/j.ijar.2020.06.005}
}
Download
Mauà, D.D., Ribeiro, H., Katague, G., Antonucci, A. (2020). Two reformulation approaches to maximum-a-posteriori inference in sum-product networks. In Jaeger, M., Nielsen, T.D. (Eds), Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
Two reformulation approaches to maximum-a-posteriori inference in sum-product networks
Authors: Mauà, D.D. and Ribeiro, H. and Katague, G. and Antonucci, A.
Year: 2020
Abstract: There exists a dichotomy between classical probabilistic graphical models, such as Bayesian net- works (BNs), and modern tractable models, such as sum-product networks (SPNs). The former generally have intractable inference, but provide a high level of interpretability, while the latter admit a wide range of tractable inference routines, but are typically harder to interpret. Due to this dichotomy, tools to convert between BNs and SPNs are desirable. While one direction – compiling BNs into SPNs – is well discussed in Darwiche’s seminal work on arithmetic circuit compilation, the converse direction – decompiling SPNs into BNs – has received surprisingly little attention. In this paper, we fill this gap by proposing SPN2BN, an algorithm that decompiles an SPN into a BN. SPN2BN has several salient features when compared to the only other two works decompiling SPNs. Most significantly, the BNs returned by SPN2BN are minimal independence-maps that are more parsimonious with respect to the introduction of latent variables. Secondly, the output BN produced by SPN2BN can be precisely characterized with respect to a compiled BN. More specifically, a certain set of directed edges will be added to the input BN, giving what we will call the moral-closure. Lastly, it is established that our compilation-decompilation process is idempotent. This has practical significance as it limits the size of the decompiled SPN.
Published in Jaeger, M., Nielsen, T.D. (Eds), Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), Proceedings of Machine Learning Research, PMLR, Aalborg, Denmark.
Two reformulation approaches to maximum-a-posteriori inference in sum-product networks
@INPROCEEDINGS{antonucci2020b,
title = {Two reformulation approaches to maximum-a-posteriori inference in sum-product networks},
editor = {Jaeger, M. and Nielsen, T.D.},
publisher = {PMLR},
address = {Aalborg, Denmark},
series = {Proceedings of Machine Learning Research},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models ({PGM} 2020)},
author = {Mau\`a, D.D. and Ribeiro, H. and Katague, G. and Antonucci, A.},
year = {2020},
url = {https://pgm2020.cs.aau.dk}
}
Download
Mejari, M., Breschi, V., Naik, V.V., Piga, D. (2020). A bias-correction approach for the identification of piecewise affine output-error models. In 21st IFAC World Congress (IFAC 2020), Berlin, Germany.
A bias-correction approach for the identification of piecewise affine output-error models
Authors: Mejari, M. and Breschi, V. and Naik, V.V. and Piga, D.
Year: 2020
Abstract: The paper presents an algorithm for the identification of PieceWise Affine Output-Error (PWA-OE) models, which involves the estimation of the parameters defining affine submodels as well as a partition of the regressor space. For the estimation of affine submodel parameters, a bias-correction scheme is presented to correct the bias in the least-squares estimates which is caused by the output-error noise structure. The obtained bias-corrected estimates are proven to be consistent under suitable assumptions. The bias-correction method is then combined with a recursive estimation algorithm for clustering the regressors.
These clusters are used to compute a partition of the regressor space by employing linear multi-category discrimination. The effectiveness of the proposed methodology is demonstrated via a simulation case study.
Accepted in 21st IFAC World Congress (IFAC 2020), Berlin, Germany.
A bias-correction approach for the identification of piecewise affine output-error models
@INPROCEEDINGS{mejari2020a,
title = {A bias-correction approach for the identification of piecewise affine output-error models},
address = {Berlin, Germany},
booktitle = {21st {IFAC} World Congress ({IFAC} 2020)},
author = {Mejari, M. and Breschi, V. and Naik, V.V. and Piga, D.},
year = {2020},
url = {https://www.ifac2020.org}
}
Download
Mejari, M., Breschi, V., Piga, D. (2020). Recursive bias-correction method for identification of piecewise affine output-error models. IEEE Control Systems Letters 4, pp. 970–975.
Recursive bias-correction method for identification of piecewise affine output-error models
Authors: Mejari, M. and Breschi, V. and Piga, D.
Year: 2020
Abstract: Learning PieceWise Affine Output-Error (PWA-OE) models from data requires to estimate a finite set of affine output-error sub-models as well as a partition of the regressors space over which the sub-models are defined. For an output-error type noise structure, the algorithms based on ordinary least squares (LS) fail to compute a consistent estimate of the sub-model parameters. On the other hand, the prediction error methods (PEMs) provide a consistent parameter estimate, however, they require to solve a non-convex optimization problem for which the numerical algorithms may get trapped in a local minimum, leading to inaccurate estimates. In this paper, we propose a recursive bias-correction scheme for identifying PWA-OE models, retaining the computational efficiency of the standard LS algorithms while providing a consistent estimate of the sub-model parameters, under suitable assumptions. The proposed approach allows one to recursively update the estimates of the sub-models parameters and to cluster the regressors. Linear multi-category techniques are then employed to estimate a partition of the regressor space based on the estimated clusters. The performance of the proposed algorithm is demonstrated via an academic example.
Published in IEEE Control Systems Letters 4, pp. 970–975.
Recursive bias-correction method for identification of piecewise affine output-error models
@ARTICLE{mejari2020b,
title = {Recursive bias-correction method for identification of piecewise affine output-error models},
journal = {{IEEE} Control Systems Letters},
volume = {4},
author = {Mejari, M. and Breschi, V. and Piga, D.},
pages = {970--975},
year = {2020},
doi = {10.1109/LCSYS.2020.2998282}
}
Download
Miranda, E., Zaffalon, M. (2020). Compatibility, desirability, and the running intersection property. Artificial Intelligence 283, 103724.
Compatibility, desirability, and the running intersection property
Authors: Miranda, E., Zaffalon, M.
Year: 2020
Abstract: Compatibility is the problem of checking whether some given probabilistic assessments have a common joint probabilistic model. When the assessments are unconditional, the problem is well established in the literature and finds a solution through the running intersection property (RIP). This is not the case of conditional assessments. In this paper, we study the compatibility problem in a very general setting: any possibility space, unrestricted domains, imprecise (and possibly degenerate) probabilities. We extend the unconditional case to our setting, thus generalising most of previous results in the literature. The conditional case turns out to be fundamentally different from the unconditional one. For such a case, we prove that the problem can still be solved in general by RIP but in a more involved way: by constructing a junction tree and propagating information over it. Still, RIP does not allow us to optimally take advantage of sparsity: in fact, conditional compatibility can be simplified further by joining junction trees with coherence graphs.
Published in Artificial Intelligence 283, 103724.
Compatibility, desirability, and the running intersection property
@ARTICLE{zaffalon2020a,
title = {Compatibility, desirability, and the running intersection property},
journal = {Artificial Intelligence},
volume = {283},
author = {Miranda, E., Zaffalon, M.},
pages = {103724},
year = {2020},
doi = {10.1016/j.artint.2020.103274}
}
Download
Piga, D., Bemporad, A., Benavoli, A. (2020). Rao-Blackwellized sampling for batch and recursive Bayesian inference of piecewise affine models. Automatica 117, 109002.
Rao-Blackwellized sampling for batch and recursive Bayesian inference of piecewise affine models
Authors: Piga, D. and Bemporad, A. and Benavoli, A.
Year: 2020
Abstract: This paper addresses batch (offline) and recursive (online) Bayesian inference of Piecewise Affine (PWA) regression models. By exploiting the particular structure of PWA models, efficient Rao-Blackwellized Monte Carlo sampling algorithms are developed to approximate the joint posterior distribution of the model parameters. Only the marginal posterior of the parameters used to describe the regressor-space partition is approximated, either in a batch mode using a Metropolis–Hastings Markov-Chain Monte Carlo (MCMC) sampler, or sequentially using particle filters, while the conditional distribution of the other model parameters is computed analytically. Probability distributions for the predicted outputs given new test inputs are derived and modifications of the proposed approaches to address maximum-a-posteriori estimate are discussed. The performance of the proposed algorithms is shown via a numerical example and through a benchmark case study on data-driven modelling of the electronic component placement process in a pick-and-place machine.
Published in Automatica 117, 109002.
Rao-Blackwellized sampling for batch and recursive Bayesian inference of piecewise affine models
@ARTICLE{piga2020a,
title = {Rao-{B}lackwellized sampling for batch and recursive {B}ayesian inference of piecewise affine models},
journal = {Automatica},
volume = {117},
author = {Piga, D. and Bemporad, A. and Benavoli, A.},
pages = {109002},
year = {2020},
doi = {10.1016/j.automatica.2020.109002}
}
Download
Piga, D., Breschi, V., Bemporad, A. (2020). Estimation of jump box–jenkins models. Automatica 120, 109126.
Estimation of jump box–jenkins models
Authors: Piga, D. and Breschi, V. and Bemporad, A.
Year: 2020
Abstract: Jump Box–Jenkins (BJ) models are a collection of a finite set of linear dynamical submodels in BJ form that switch over time, according to a Markov chain. This paper addresses the problem of maximum-a-posteriori estimation of jump BJ models from a given training input/output dataset. The proposed solution method estimates the coefficients of the BJ submodels, the state transition probabilities of the Markov chain regulating the switching of operating modes, and the corresponding mode sequence hidden in the dataset. In particular, the posterior distribution of all the unknown variables characterizing the jump BJ model is derived and then maximized using a coordinate ascent algorithm. The resulting estimation algorithm alternates between Gauss–Newton optimization of the coefficients of the BJ submodels, a method derived based on an instance of prediction error methods tailored to BJ models with switching coefficients, and approximated dynamic programming for optimization of the sequence of active modes. The quality of the proposed estimation approach is evaluated on a numerical example based on synthetic data and in a case study related to segmentation of honeybee dances.
Published in Automatica 120, 109126.
Estimation of jump box–jenkins models
@ARTICLE{piga2020c,
title = {Estimation of jump box--jenkins models},
journal = {Automatica},
volume = {120},
author = {Piga, D. and Breschi, V. and Bemporad, A.},
pages = {109126},
year = {2020},
doi = {10.1016/j.automatica.2020.109126},
url = {http://www.sciencedirect.com/science/article/pii/S0005109820303241}
}
Download
Roveda, L., Bussolan, A., Braghin, F., Piga, D. (2020). 6D virtual sensor for wrench estimation in robotized interaction tasks exploiting extended Kalman filter. MDPI Machines.
6D virtual sensor for wrench estimation in robotized interaction tasks exploiting extended Kalman filter
Authors: Roveda, L. and Bussolan, A. and Braghin, F. and Piga, D.
Year: 2020
Abstract: Industrial robots are commonly used to perform interaction tasks (such as assemblies or
polishing), requiring the robot to be in contact with the surrounding environment. Such environments
are (partially) unknown to the robot controller. Therefore, there is the need to implement interaction
controllers capable of suitably reacting to the established contacts. Although standard force controllers
require force/torque measurements to close the loop, most of the industrial manipulators do not have
installed force/torque sensor(s). In addition, the integration of external sensors results in additional
costs and implementation effort, not affordable in many contexts/applications. To extend the use
of compliant controllers to sensorless interaction control, a model-based methodology is presented
in this paper for the online estimation of the interaction wrench, implementing a 6D virtual sensor.
Relying on sensorless Cartesian impedance control, an Extended Kalman Filter (EKF) is proposed
for the interaction wrench estimation. The described approach has been validated in simulations,
taking into account four different scenarios. In addition, experimental validation has been performed
employing a Franka EMIKA panda robot. A human–robot interaction scenario and an assembly task
have been considered to show the capabilities of the developed EKF, which is able to perform the
estimation with high bandwidth, achieving convergence with limited errors.
Published in MDPI Machines.
6D virtual sensor for wrench estimation in robotized interaction tasks exploiting extended Kalman filter
@ARTICLE{Roveda2020b,
title = {{6D} virtual sensor for wrench estimation in robotized interaction tasks exploiting extended {K}alman filter},
journal = {{MDPI} Machines},
author = {Roveda, L. and Bussolan, A. and Braghin, F. and Piga, D.},
year = {2020},
doi = {10.3390/machines8040067}
}
Download
Roveda, L., Castaman, N., Franceschi, P., Ghidoni, S., Pedrocchi, N. (2020). A control framework definition to overcome position/interaction dynamics uncertainties in force-controlled tasks. In IEEE International Conference on Robotics and Automation (ICRA) 2020.
A control framework definition to overcome position/interaction dynamics uncertainties in force-controlled tasks
Authors: Roveda, L. and Castaman, N. and Franceschi, P. and Ghidoni, S. and Pedrocchi, N.
Year: 2020
Abstract: Within the Industry 4.0 context, industrial robots need to show increasing autonomy. The manipulator has to be able to react to uncertainties/changes in the working environment, displaying a robust behavior. In this paper, a control framework is proposed to perform industrial interaction tasks in uncertain working scenes. The proposed methodology relies
on two components: i) a 6D pose estimation algorithm aiming
to recognize large and featureless parts; ii) a variable damping impedance controller (inner loop) enhanced by an adaptive saturation PI (outer loop) for high accuracy force control (i.e., zero steady-state force error and force overshoots avoidance). The proposed methodology allows to be robust w.r.t. task uncertainties (i.e., positioning errors and interaction dynamics). The proposed approach has been evaluated in an assembly task of a side-wall panel to be installed inside the aircraft cabin. As a test platform, the KUKA iiwa 14 R820 has been used together with the Microsoft Kinect 2.0 as RGB-D sensor. Experiments show the reliability in the 6D pose estimation and the high-performance in the force-tracking task, avoiding force overshoots while achieving the tracking of the reference force.
Published in IEEE International Conference on Robotics and Automation (ICRA) 2020.
A control framework definition to overcome position/interaction dynamics uncertainties in force-controlled tasks
@INPROCEEDINGS{Roveda2020f,
title = {A control framework definition to overcome position/interaction dynamics uncertainties in force-controlled tasks},
booktitle = {{IEEE} International Conference on Robotics and Automation ({ICRA}) 2020},
author = {Roveda, L. and Castaman, N. and Franceschi, P. and Ghidoni, S. and Pedrocchi, N.},
year = {2020},
doi = {10.1109/ICRA40945.2020.9197141}
}
Download
Roveda, L., Forgione, M., Piga, D. (2020). Robot control parameters auto-tuning in trajectory tracking applications. Control Engineering Practice 101, 104488.
Robot control parameters auto-tuning in trajectory tracking applications
Authors: Roveda, L. and Forgione, M. and Piga, D.
Year: 2020
Abstract: Autonomy is increasingly demanded to industrial manipulators. Robots have to be capable to regulate their behavior to different operational conditions, adapting to the specific task to be executed without requiring high time/resource-consuming human intervention. Achieving an automated tuning of the control parameters of a manipulator is still a challenging task, which involves modeling/identification of the robot dynamics. This usually results in an onerous procedure, both in terms of experimental and data-processing time. This paper addresses the problem of automated tuning of the manipulator controller for trajectory tracking, optimizing control parameters based on the specific trajectory to be executed. A Bayesian optimization algorithm is proposed to tune both the low-level controller parameters (i.e., the equivalent link-masses of the feedback linearizator and the feedforward controller) and the high-level controller parameters (i.e., the joint PID gains). The algorithm adapts the control parameters through a data-driven procedure, optimizing a user-defined trajectory-tracking cost. Safety constraints ensuring, e.g., closed-loop stability and bounds on the maximum joint position error are also included. The performance of proposed approach is demonstrated on a torque-controlled 7-degree-of-freedom FRANKA Emika robot manipulator. The 25 robot control parameters (i.e., 4 link-mass parameters and 21 PID gains) are tuned in less than 130 iterations, and comparable results with respect to the FRANKA Emika embedded position controller are achieved. In addition, the generalization capabilities of the proposed approach are shown exploiting the proper reference trajectory for the tuning of the control parameters.
Published in Control Engineering Practice 101, 104488.
Robot control parameters auto-tuning in trajectory tracking applications
@ARTICLE{roveda2020a,
title = {Robot control parameters auto-tuning in trajectory tracking applications},
journal = {Control Engineering Practice},
volume = {101},
author = {Roveda, L. and Forgione, M. and Piga, D.},
pages = {104488},
year = {2020},
doi = {10.1016/j.conengprac.2020.104488}
}
Download
Roveda, L., Forgione, M., Piga, D. (2020). One-stage auto-tuning procedure of robot dynamics and control parameters for trajectory tracking applications. In Ubiquitous Robots 2020.
One-stage auto-tuning procedure of robot dynamics and control parameters for trajectory tracking applications
Authors: Roveda, L. and Forgione, M. and Piga, D.
Year: 2020
Abstract: Autonomy is increasingly demanded by industrial
manipulators. Robots have to be capable to regulate their
behavior to different operational conditions, without requiring high time/resource-consuming human intervention. Achieving an automated tuning of the control parameters of a manipulator is still a challenging task. This paper addresses the problem of automated tuning of the manipulator controller for trajectory tracking. A Bayesian optimization algorithm is proposed to tune both the low-level controller parameters (i.e., robot dynamics compensation) and the high-level controller parameters (i.e., the joint PID gains). The algorithm adapts the control parameters through a data-driven procedure, optimizing a user-defined trajectory-tracking cost. Safety constraints ensuring, e.g., closed-loop stability and bounds on the maximum joint position errors, are also included. The performance of the proposed approach is demonstrated on a torque-controlled 7-degree-of-freedom FRANKA Emika robot manipulator. The 25 robot control parameters (i.e., 4 link-mass parameters and 21 PID gains) are tuned in 125 iterations, and comparable results with respect to the FRANKA Emika embedded position controller are achieved.
Published in Ubiquitous Robots 2020.
One-stage auto-tuning procedure of robot dynamics and control parameters for trajectory tracking applications
@INPROCEEDINGS{Roveda2020h,
title = {One-stage auto-tuning procedure of robot dynamics and control parameters for trajectory tracking applications},
booktitle = {Ubiquitous Robots 2020},
author = {Roveda, L. and Forgione, M. and Piga, D.},
year = {2020},
doi = {10.1109/UR49135.2020.9144761}
}
Download
Roveda, L., Magni, M., Cantoni, M., Piga, D., Bucca, G. (2020). Human–robot collaboration in sensorless assembly task learning enhanced by uncertainties adaptation via Bayesian optimization. Robotics and Autonomous Systems.
Human–robot collaboration in sensorless assembly task learning enhanced by uncertainties adaptation via Bayesian optimization
Authors: Roveda, L. and Magni, M. and Cantoni, M. and Piga, D. and Bucca, G.
Year: 2020
Abstract: Robots are increasingly exploited in production plants. Within the Industry 4.0 paradigm, the robot
complements the human’s capabilities, learning new tasks and adapting itself to compensate for
uncertainties. With this aim, the presented paper focuses on the investigation of machine learning
techniques to make a sensorless robot able to learn and optimize an industrial assembly task.
Relying on sensorless Cartesian impedance control, two main contributions are defined: (1) a task-
trajectory learning algorithm based on a few human’s demonstrations (exploiting Hidden Markov
Model approach), and (2) an autonomous optimization procedure of the task execution (exploiting
Bayesian Optimization). To validate the proposed methodology, an assembly task has been selected as
a reference application. The task consists of mounting a gear into its square-section shaft on a fixed
base to simulate the assembly of a gearbox. A Franka EMIKA Panda manipulator has been used as a
test platform, implementing the proposed methodology. The experiments, carried out on a population
of 15 subjects, show the effectiveness of the proposed strategy, making the robot able to learn and
optimize its behavior to accomplish the assembly task, even in the presence of task uncertainties.
Published in Robotics and Autonomous Systems.
Human–robot collaboration in sensorless assembly task learning enhanced by uncertainties adaptation via Bayesian optimization
@ARTICLE{Roveda2020j,
title = {Human--robot collaboration in sensorless assembly task learning enhanced by uncertainties adaptation via {B}ayesian optimization},
journal = {Robotics and Autonomous Systems},
author = {Roveda, L. and Magni, M. and Cantoni, M. and Piga, D. and Bucca, G.},
year = {2020},
doi = {10.1016/j.robot.2020.103711}
}
Download
Roveda, L., Magni, M., Cantoni, M., Piga, D., Bucca, G. (2020). Assembly task learning and optimization through Human’s demonstration and machine learning. In IEEE International Conference on Systems, Man, and Cybernetics.
Assembly task learning and optimization through Human’s demonstration and machine learning
Authors: Roveda, L. and Magni, M. and Cantoni, M. and Piga, D. and Bucca, G.
Year: 2020
Abstract: Robots are increasingly exploited in production plants, with the need to learn and to adapt themselves to new tasks. This paper focuses on the investigation of machine learning techniques to make a sensorless robot able to learn and optimize an industrial assembly task. Relying on sensorless Cartesian impedance control, a task-trajectory learning algorithm exploiting a limited number of human’s demonstrations (based on Hidden Markov Model), and an autonomous optimization procedure (based on Bayesian Optimization) are proposed to learn and optimize the assembly task. To validate the proposed methodology, an assembly task of a gear into its square-section shaft has been considered. A Franka EMIKA Panda manipulator has been used as a test platform. The experiments show the effectiveness of the proposed strategy, making the robot able to learn and optimize its behaviour to accomplish the assembly task, even in the presence of uncertainties.
Published in IEEE International Conference on Systems, Man, and Cybernetics.
Assembly task learning and optimization through Human’s demonstration and machine learning
@INPROCEEDINGS{Roveda2020k,
title = {Assembly task learning and optimization through {H}uman’s demonstration and machine learning},
booktitle = {{IEEE} International Conference on Systems, Man, and Cybernetics},
author = {Roveda, L. and Magni, M. and Cantoni, M. and Piga, D. and Bucca, G.},
year = {2020},
doi = {10.1109/SMC42975.2020.9282911}
}
Download
Roveda, L., Maskani, J., Franceschi, P., Arash, A., Braghin, F., Molinari Tosatti, L., Pedrocchi, N. (2020). Model-based reinforcement learning variable impedance control for human-robot collaboration. Journal of Intelligent & Robotic Systems.
Model-based reinforcement learning variable impedance control for human-robot collaboration
Authors: Roveda, L. and Maskani, J. and Franceschi, P. and Arash, A. and Braghin, F. and Molinari Tosatti, L. and Pedrocchi, N.
Year: 2020
Abstract: Industry 4.0 is taking human-robot collaboration at the center of the production environment. Collaborative robots
enhance productivity and flexibility while reducing human’s fatigue and the risk of injuries, exploiting advanced control
methodologies. However, there is a lack of real-time model-based controllers accounting for the complex human-robot
interaction dynamics. With this aim, this paper proposes a Model-Based Reinforcement Learning (MBRL) variable
impedance controller to assist human operators in collaborative tasks. More in details, an ensemble of Artificial Neural Networks (ANNs) is used to learn a human-robot interaction dynamic model, capturing uncertainties. Such a learned model is kept updated during collaborative tasks execution. In addition, the learned model is used by a Model Predictive Controller (MPC) with Cross-Entropy Method (CEM). The aim of the MPC+CEM is to online optimize the stiffness and damping impedance control parameters minimizing the human effort (i.e, minimizing the human-robot interaction forces). The proposed approach has been validated through an experimental procedure. A lifting task has been considered as the reference validation application (weight of the manipulated part: 10 kg unknown to the robot controller). A KUKA LBR iiwa 14 R820 has been used as a test platform. Qualitative performance (i.e, questionnaire on perceived collaboration) have been evaluated. Achieved results have been compared with previous developed offline model-free optimized controllers and with the robot manual guidance controller. The proposed MBRL variable impedance controller shows improved human-robot collaboration. The proposed controller is capable to actively assist the human in the target task, compensating for the unknown part weight. The human-robot interaction dynamic model has been trained with a few initial experiments (30 initial experiments). In addition, the possibility to keep the learning of the human-robot interaction dynamics active allows accounting for the adaptation of human motor system.
Published in Journal of Intelligent & Robotic Systems, Springer.
Model-based reinforcement learning variable impedance control for human-robot collaboration
@ARTICLE{Roveda2020c,
title = {Model-based reinforcement learning variable impedance control for human-robot collaboration},
journal = {Journal of Intelligent & Robotic Systems},
publisher = {Springer},
author = {Roveda, L. and Maskani, J. and Franceschi, P. and Arash, A. and Braghin, F. and Molinari Tosatti, L. and Pedrocchi, N.},
year = {2020},
doi = {10.1007/s10846- 020-01183-3}
}
Download
Roveda, L., Piga, D. (2020). Robust state dependent Riccati equation variable impedance control for robotic force-tracking tasks. International Journal of Intelligent Robotics and Applications.
Robust state dependent Riccati equation variable impedance control for robotic force-tracking tasks
Authors: Roveda, L. and Piga, D.
Year: 2020
Abstract: Industrial robots are increasingly used in highly flexible interaction tasks, where the intrinsic variability makes difficult to pre-program the manipulator for all the different scenarios. In such applications, interaction environments are commonly (partially) unknown to the robot, requiring the implemented controllers to take in charge for the stability of the interaction. While standard controllers are sensor-based, there is a growing need to make sensorless robots (i.e., most of the commercial robots are not equipped with force/torque sensors) able to sense the environment, properly reacting to the established interaction. This paper proposes a new methodology to sensorless force control manipulators. On the basis of sensorless Cartesian impedance control, an Extended Kalman Filter (EKF) is designed to estimate the interaction exchanged between the robot and the environment. Such an estimation is then used in order to close a robust high-performance force loop, designed exploiting a variable impedance control and a State Dependent Riccati Equation (SDRE) force controller. The described approach has been validated in simulations. A Franka EMIKA panda robot has been considered as a test platform. A probing task involving different materials (i.e., with different stiffness properties) has been considered to show the capabilities of the developed EKF (able to converge with limited errors) and controller (preserving stability and avoiding overshoots). The proposed controller has been compared with an LQR controller to show its improved performance.
Published in International Journal of Intelligent Robotics and Applications, Springer.
Robust state dependent Riccati equation variable impedance control for robotic force-tracking tasks
@ARTICLE{Roveda2020d,
title = {Robust state dependent {R}iccati equation variable impedance control for robotic force-tracking tasks},
journal = {International Journal of Intelligent Robotics and Applications},
publisher = {Springer},
author = {Roveda, L. and Piga, D.},
year = {2020},
doi = {10.1007/s41315-020-00153-0}
}
Download
Roveda, L., Piga, D. (2020). Interaction force computation exploiting environment stiffness estimation for sensorless robot applications. In IEEE Metrology for Industry 4.0 and IoT 2020.
Interaction force computation exploiting environment stiffness estimation for sensorless robot applications
Authors: Roveda, L. and Piga, D.
Year: 2020
Abstract: Industrial robots are increasingly used to perform tasks requiring an interaction with the surrounding environment.
However, standard controllers require force/torque measurements to close the loop. Most of the industrial manipulators do not have embedded force/torque sensor(s), requiring additional efforts (i.e., additional costs and implementation resources) for such integration in the robotic setup. To extend the use of compliant controllers to sensorless force control, a model-based methodology is presented in this paper. Relying on sensorless Cartesian impedance control, an Extended Kalman Filter (EKF) is proposed to estimate the interaction environment stiffness. Exploiting such estimation, the interaction force can be computed, e.g., to close the force loop, making the sensorless robot able to perform the target task (e.g., probing task, assembly task). The described approach has been validated with experiments. A Franka EMIKA panda robot has been used as a test platform. A probing task involving different materials (i.e., with different - unknown - stiffness properties) has been considered to show the capabilities of the developed EKF. The computed interaction force (on the basis of the estimated environment stiffness) has been compared with the Franka EMIKA panda robot force measurements to prove the effectiveness of the proposed approach.
Published in IEEE Metrology for Industry 4.0 and IoT 2020.
Interaction force computation exploiting environment stiffness estimation for sensorless robot applications
@INPROCEEDINGS{Roveda2020g,
title = {Interaction force computation exploiting environment stiffness estimation for sensorless robot applications},
booktitle = {{IEEE} Metrology for Industry 4.0 and {IoT} 2020},
author = {Roveda, L. and Piga, D.},
year = {2020},
doi = {10.1109/MetroInd4.0IoT48571.2020.9138189}
}
Download
Roveda, L., Savani, L., Arlati, S., Dinon, T., Legnani, G., Molinari Tosatti, L. (2020). Design methodology of an active back-support exoskeleton with adaptable backbone-based kinematics. International Journal of Industrial Ergonomics.
Design methodology of an active back-support exoskeleton with adaptable backbone-based kinematics
Authors: Roveda, L. and Savani, L. and Arlati, S. and Dinon, T. and Legnani, G. and Molinari Tosatti, L.
Year: 2020
Abstract: non-ergonomic conditions and to manipulate heavy parts. As a result, work-related musculoskeletal disorders are a major problem to tackle in workplace. In particular, back is one of the most affected regions. To solve such issue, many efforts have been made in the design and control of exoskeleton devices, relieving the human from the task load. Besides upper limbs and lower limbs exoskeletons, back-support exoskeletons have been also investigated, proposing both passive and active solutions. While passive solutions cannot empower the human’s capabilities, common active devices are rigid, without the possibility to track the human’s spine kinematics while executing the task. The here proposed paper describes a methodology to design an active back-support exoskeleton with backbone-based kinematics. On the basis of the (easily implementable) scissor hinge mechanism, a one-degree of freedom device has been designed. In particular, the resulting device allows tracking the motion of a reference vertebra, i.e., the vertebrae in the correspondence of the connection between the scissor hinge mechanism and the back of the operator. Therefore, the proposed device is capable to adapt to the human posture, guaranteeing the support while relieving the person from the task load. In addition, the proposed mechanism can be easily optimized and realized for different subjects, involving a subject-based design procedure, making possible to adapt its kinematics to track the spine motion of the specific user. A prototype of the proposed device has been 3D-printed to show the achieved kinematics. Preliminary tests for discomfort evaluation show the potential of the proposed methodology, foreseeing extensive subjects-based optimization, realization and testing of the device.
Published in International Journal of Industrial Ergonomics, Elsevier.
Design methodology of an active back-support exoskeleton with adaptable backbone-based kinematics
@ARTICLE{Roveda2020e,
title = {Design methodology of an active back-support exoskeleton with adaptable backbone-based kinematics},
journal = {International Journal of Industrial Ergonomics},
publisher = {Elsevier},
author = {Roveda, L. and Savani, L. and Arlati, S. and Dinon, T. and Legnani, G. and Molinari Tosatti, L.},
year = {2020},
doi = {10.1016/j.ergon.2020.102991}
}
Download
Schürch, M., Azzimonti, D., Benavoli, A., Zaffalon, M. (2020). Recursive estimation for sparse gaussian process regression. Automatica 120, 109127.
Recursive estimation for sparse gaussian process regression
Authors: Schürch, M. and Azzimonti, D. and Benavoli, A. and Zaffalon, M.
Year: 2020
Abstract: Gaussian Processes (GPs) are powerful kernelized methods for non-parameteric regression used in many applications. However, their use is limited to a few thousand of training samples due to their cubic time complexity. In order to scale GPs to larger datasets, several sparse approximations based on so-called inducing points have been proposed in the literature. In this work we investigate the connection between a general class of sparse inducing point GP regression methods and Bayesian recursive estimation which enables Kalman Filter like updating for online learning. The majority of previous work has focused on the batch setting, in particular for learning the model parameters and the position of the inducing points, here instead we focus on training with mini-batches. By exploiting the Kalman filter formulation, we propose a novel approach that estimates such parameters by recursively propagating the analytical gradients of the posterior over mini-batches of the data. Compared to state of the art methods, our method keeps analytic updates for the mean and covariance of the posterior, thus reducing drastically the size of the optimization problem. We show that our method achieves faster convergence and superior performance compared to state of the art sequential Gaussian Process regression on synthetic GP as well as real-world data with up to a million of data samples.
Published in Automatica 120, Elsevier, 109127.
Recursive estimation for sparse gaussian process regression
@ARTICLE{schurch2020a,
title = {Recursive estimation for sparse gaussian process regression},
journal = {Automatica},
publisher = {Elsevier},
volume = {120},
author = {Sch\"urch, M. and Azzimonti, D. and Benavoli, A. and Zaffalon, M.},
pages = {109127},
year = {2020}
}
Download
Shahid, A.A., Roveda, L., Piga, D., Braghin, F. (2020). Learning continuous control actions for robotic grasping with reinforcement learning. In IEEE International Conference on Systems, Man, and Cybernetics.
Learning continuous control actions for robotic grasping with reinforcement learning
Authors: Shahid, A.A. and Roveda, L. and Piga, D. and Braghin, F.
Year: 2020
Published in IEEE International Conference on Systems, Man, and Cybernetics.
Learning continuous control actions for robotic grasping with reinforcement learning
@INPROCEEDINGS{Roveda2020l,
title = {Learning continuous control actions for robotic grasping with reinforcement learning},
booktitle = {{IEEE} International Conference on Systems, Man, and Cybernetics},
author = {Shahid, A.A. and Roveda, L. and Piga, D. and Braghin, F.},
year = {2020},
doi = {10.1109/SMC42975.2020.9282951}
}
Download
Sorgini, F., Airò Farulla, G., Lukic, N., Danilov, I., Roveda, L., Milivojevic, M., Babu Pulikottil, T., Carrozza, M.C., Prinetto, P., Tolio, T., Oddo, C.M., Petrovic, P., Bojovic, B. (2020). Tactile sensing with gesture-controlled collaborative robot. In 2020 IEEE International Workshop on Metrology for Industry 4.0 & IoT (MetroInd4.0 & IoT).
Tactile sensing with gesture-controlled collaborative robot
Authors: Sorgini, F. and Airò Farulla, G. and Lukic, N. and Danilov, I. and Roveda, L. and Milivojevic, M. and Babu Pulikottil, T. and Carrozza, M.C. and Prinetto, P. and Tolio, T. and Oddo, C.M. and Petrovic, P. and Bojovic, B.
Year: 2020
Abstract: Sensors and human machine interfaces for
collaborative robotics will allow smooth interaction in contexts ranging from industry to tele-medicine and rescue. This paper introduces a bidirectional communication system to achieve multisensory telepresence during the gestural control of an industrial robotic arm. Force and motion from the robot are converted in neuromorphic haptic stimuli delivered on the user’s hand through a vibro-tactile glove. Untrained personnel participated in an experimental task benchmarking a pick-and-place operation. The robot end-effector was used to
sequentially press six buttons, illuminated according to a
random sequence, and comparing the tasks executed without
and with tactile feedback. The results demonstrated the
reliability of the hand tracking strategy developed for
controlling the robotic arm, and the effectiveness of a neuronal spiking model for encoding hand displacement and exerted forces in order to promote a fluid embodiment of the haptic interface and control strategy. The main contribution of this paper is in presenting a robotic arm under gesture-based remote control with multisensory telepresence, demonstrating for the first time that a spiking haptic interface can be used to effectively deliver on the skin surface a sequence of stimuli emulating the neural code of the mechanoreceptors beneath.
Published in 2020 IEEE International Workshop on Metrology for Industry 4.0 & IoT (MetroInd4.0 & IoT).
Tactile sensing with gesture-controlled collaborative robot
@INPROCEEDINGS{Roveda2020i,
title = {Tactile sensing with gesture-controlled collaborative robot},
booktitle = {2020 {IEEE} International Workshop on Metrology for Industry 4.0 & {IoT} ({MetroInd4}.0 & {IoT})},
author = {Sorgini, F. and Air\`o Farulla, G. and Lukic, N. and Danilov, I. and Roveda, L. and Milivojevic, M. and Babu Pulikottil, T. and Carrozza, M.C. and Prinetto, P. and Tolio, T. and Oddo, C.M. and Petrovic, P. and Bojovic, B.},
year = {2020},
doi = {10.1109/MetroInd4.0IoT48571.2020.9138183}
}
Download
Volpetti, C., Kanjirangat, V., Antonucci, A. (2020). Temporal word embeddings for narrative understanding. In 12th International Conference on Machine Learning and Computing (ICMLC 2020), ACM.
Temporal word embeddings for narrative understanding
Authors: Volpetti, C. and Kanjirangat, V. and Antonucci, A.
Year: 2020
Abstract: We propose temporal word embeddings as a suitable tool to study the evolution of characters and their sentiments across the plot of a narrative text. The dynamic evolution of instances within a narrative text is a challenging task, where complex behavioral evolutions and other characteristics specific to the narrative text
need to be inferred and interpreted. While starting from an existing approach to the learning of these models, we propose an alternative initialization procedure which seems to be especially suited for the case of narrative text. As a validation benchmark, we use the Harry Potter series of books as a challenging case study for such character
trait evolution. A benchmark data set based on temporal word
analogies related to the characters in the plot of the series is considered. The results are promising, and the empirical validation seems to support the working ideas behind this proposal.
Accepted in 12th International Conference on Machine Learning and Computing (ICMLC 2020), ACM.
Temporal word embeddings for narrative understanding
@INPROCEEDINGS{supsi2020a,
title = {Temporal word embeddings for narrative understanding},
publisher = {ACM},
booktitle = {12th International Conference on Machine Learning and Computing ({ICMLC} 2020)},
author = {Volpetti, C. and Kanjirangat, V. and Antonucci, A.},
year = {2020},
url = {http://www.icmlc.org/}
}
Download
Yoo, J., Kang, U., Scanagatta, M., Corani, G., Zaffalon, M. (2020). Sampling subgraphs with guaranteed treewidth for accurate and efficient graphical inference. In Proceedings of International Conference on Web Search and Data Mining (WSDM '20).
Sampling subgraphs with guaranteed treewidth for accurate and efficient graphical inference
Authors: Yoo, J. and Kang, U. and Scanagatta, M. and Corani, G. and Zaffalon, M.
Year: 2020
Abstract: How can we run graphical inference on large graphs efficiently and accurately? Many real-world networks are modeled as graphical models, and graphical inference is fundamental to understand the properties of those networks. In this work, we propose a novel approach for fast and accurate inference, which first samples a small subgraph and then runs inference over the subgraph instead of the given graph. This is done by the bounded treewidth (BTW) sampling, our novel algorithm that generates a subgraph with guar- anteed bounded treewidth while retaining as many edges as pos- sible. We first analyze the properties of BTW theoretically. Then, we evaluate our approach on node classification and compare it with the baseline which is to run loopy belief propagation (LBP) on the original graph. Our approach can be coupled with various inference algorithms: it shows higher accuracy up to 13.7% with the junction tree algorithm, and allows faster inference up to 23.8 times with LBP. We further compare BTW with previous graph sampling algorithms and show that it gives the best accuracy.
Accepted in Proceedings of International Conference on Web Search and Data Mining (WSDM '20).
Sampling subgraphs with guaranteed treewidth for accurate and efficient graphical inference
@INPROCEEDINGS{corani2019e,
title = {Sampling subgraphs with guaranteed treewidth for accurate and efficient graphical inference},
booktitle = {Proceedings of International Conference on Web Search and Data Mining ({WSDM} '20)},
author = {Yoo, J. and Kang, U. and Scanagatta, M. and Corani, G. and Zaffalon, M.},
year = {2020},
url = {https://dl.acm.org/doi/pdf/10.1145/3336191.3371815}
}
Download
Zaffalon, M., Antonucci, A., Cabañas, R. (2020). Structural causal models are (solvable by) credal networks. In Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Structural causal models are (solvable by) credal networks
Authors: Zaffalon, M. and Antonucci, A. and Cabañas, R.
Year: 2020
Abstract: A structural causal model is made of endogenous (manifest) and exogenous (latent) variables. We show that endogenous observations induce linear constraints on the probabilities of the exogenous variables. This allows to exactly map a causal model into a credal network. Causal inferences, such as interventions and counterfactuals, can consequently be obtained by standard algorithms for the updating of credal nets. These natively return sharp values in the identifiable case, while intervals corresponding to the exact bounds are produced for unidentifiable queries. A characterization of the causal models that allow the map above to be compactly derived is given, along with a discussion about the scalability for general models. This contribution should be regarded as a systematic approach to represent structural causal models by credal networks and hence to systematically compute causal inferences. A number of demonstrative examples is presented to clarify our methodology. Extensive experiments show that approximate algorithms for credal networks can immediately be used to do causal inference in real-size problems.
Accepted in Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020), JMLR.org.
Structural causal models are (solvable by) credal networks
@INPROCEEDINGS{zaffalon2020b,
title = {Structural causal models are (solvable by) credal networks},
publisher = {JMLR.org},
booktitle = {Proceedings of the 10th International Conference on Probabilistic Graphical Models ({PGM} 2020)},
author = {Zaffalon, M. and Antonucci, A. and Cabañas, R.},
year = {2020},
url = {https://pgm2020.cs.aau.dk/}
}
Download top2019
Antonucci, A. (2019). Reliable discretisation of deterministic equations in Bayesian networks. In Proceedings of the 32nd International Flairs Conference (FLAIRS-32), AAAI Press.
Reliable discretisation of deterministic equations in Bayesian networks
Authors: Antonucci, A.
Year: 2019
Abstract: We focus on the problem of modeling deterministic equations over continuous variables in discrete Bayesian networks. This is typically achieved by a discretisation of both input and output variables and a degenerate quantification of the corre- sponding conditional probability tables. This approach, based on classical probabilities, cannot properly model the information loss induced by the discretisation. We show that a reli- able modeling of such epistemic uncertainty can be instead achieved by credal sets, i.e., convex sets of probability mass functions. This transforms the original Bayesian network in a credal network, possibly returning interval-valued inferences, that are robust with respect to the information loss induced by the discretisation. Algorithmic strategies for an optimal choice of the discretisation bins are also discussed.
Published in Proceedings of the 32nd International Flairs Conference (FLAIRS-32), AAAI Press.
Note: Accepted for pubblication.
Reliable discretisation of deterministic equations in Bayesian networks
@INPROCEEDINGS{supsi2019c,
title = {Reliable discretisation of deterministic equations in {B}ayesian networks},
publisher = {AAAI Press},
booktitle = {Proceedings of the 32nd International Flairs Conference ({FLAIRS}-32)},
author = {Antonucci, A.},
year = {2019},
url = {https://aaai.org/ocs/index.php/FLAIRS/FLAIRS19/paper/view/18228/17346}
}
Download
Antonucci, A., Facchini, A., Mattei, L. (2019). Credal sentential decision diagrams. In Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications (ISIPTA '19).
Credal sentential decision diagrams
Authors: Antonucci, A. and Facchini, A. and Mattei, L.
Year: 2019
Abstract: Probabilistic sentential decision diagrams are logical circuits annotated by probability mass functions on the disjunctive gates. This allows for a compact representation of joint mass functions consistent with logical constraints. We propose a credal generalisation of the probabilistic quantification of these models, that allows to replace the local probabilities with (credal) sets of mass functions specified by linear constraints. This induces a joint credal set, that sharply assigns probability zero to states inconsistent with the constraints. These models can support cautious estimates of the local parameters when only small amounts of training data are available. Algorithmic strategies to compute lower and upper bounds of marginal and conditional queries are provided. The task can be achieved in lin- ear time with respect to the diagram size for marginal queries. The same can be done for conditional queries if the topology of the circuit is singly connected.
Published in Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications (ISIPTA '19).
Note: Accepted for pubblication.
Credal sentential decision diagrams
@INPROCEEDINGS{supsi2019b,
title = {Credal sentential decision diagrams},
booktitle = {Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications ({ISIPTA} '19)},
author = {Antonucci, A. and Facchini, A. and Mattei, L.},
year = {2019},
url = {http://www.isipta2019.ugent.be/contributions/antonucci19.pdf}
}
Download
Arnone, E., Azzimonti, L., Nobile, F., Sangalli, L.M. (2019). Modeling spatially dependent functional data via regression with differential regularization. Journal of Multivariate Analysis 170, pp. 275-295.
Modeling spatially dependent functional data via regression with differential regularization
Authors: Arnone, E. and Azzimonti, L. and Nobile, F. and Sangalli, L.M.
Year: 2019
Abstract: We propose a method for modeling spatially dependent functional data, based on regression with differential regularization. The regularizing term enables to include problem-specific information about the spatio-temporal variation of the phenomenon under study, formalized in terms of a time-dependent partial differential equation. The method is implemented using a discretization based on finite elements in space and finite differences in time. This non-tensor product basis allows to handle efficiently data distributed over complex domains and where the shape of the domain influences the phenomenon’s behavior. Moreover, the method can comply with specific conditions at the boundary of the domain of interest. Simulation studies compare the proposed model to available techniques for spatio-temporal data. The method is also illustrated via an application to the study of blood-flow velocity field in a carotid artery affected by atherosclerosis, starting from echo-color doppler and magnetic resonance imaging data.
Published in Journal of Multivariate Analysis 170, pp. 275-295.
Note: Special Issue on Functional Data Analysis and Related Topics
Modeling spatially dependent functional data via regression with differential regularization
@ARTICLE{azzimonti2018a,
title = {Modeling spatially dependent functional data via regression with differential regularization},
journal = {Journal of Multivariate Analysis},
volume = {170},
author = {Arnone, E. and Azzimonti, L. and Nobile, F. and Sangalli, L.M.},
pages = {275-295},
year = {2019},
doi = {10.1016/j.jmva.2018.09.006}
}
Download
Azzimonti, L., Corani, G., Zaffalon, M. (2019). Hierarchical estimation of parameters in Bayesian networks. Computational Statistics and Data Analysis 137, pp. 67–91.
Hierarchical estimation of parameters in Bayesian networks
Authors: Azzimonti, L. and Corani, G. and Zaffalon, M.
Year: 2019
Abstract: A novel approach for parameter estimation in Bayesian networks is presented. The main idea is to introduce a hyper-prior in the Multinomial-Dirichlet model, traditionally used for conditional distribution estimation in Bayesian networks. The resulting hierarchical model jointly estimates different conditional distributions belonging to the same conditional probability table, thus borrowing statistical strength from each other. An analytical study of the dependence structure a priori induced by the hierarchical model is performed and an ad hoc variational algorithm for fast and accurate inference is derived. The proposed hierarchical model yields a major performance improvement in classification with Bayesian networks compared to traditional models. The proposed variational algorithm reduces by two orders of magnitude the computational time, with the same accuracy in parameter estimation, compared to traditional MCMC methods.
Moreover, motivated by a real case study, the hierarchical model is applied to the estimation of Bayesian networks parameters by borrowing strength from related domains.
Published in Computational Statistics and Data Analysis 137, pp. 67–91.
Hierarchical estimation of parameters in Bayesian networks
@ARTICLE{azzimonti2019a,
title = {Hierarchical estimation of parameters in {B}ayesian networks},
journal = {Computational Statistics and Data Analysis},
volume = {137},
author = {Azzimonti, L. and Corani, G. and Zaffalon, M.},
pages = {67--91},
year = {2019},
doi = {10.1016/j.csda.2019.02.004}
}
Download
Azzimonti, D., Ginsbourger, D., Chevalier, C., Bect, J., Richet, Y. (2019). Adaptive design of experiments for conservative estimation of excursion sets. Technometrics, pp. 1–30.
Adaptive design of experiments for conservative estimation of excursion sets
Authors: Azzimonti, D. and Ginsbourger, D. and Chevalier, C. and Bect, J. and Richet, Y.
Year: 2019
Abstract: We consider the problem of estimating the set of all inputs that leads a system to some particular behavior. The system is modeled by an expensive-to-evaluate function, such as a computer experiment, and we are interested in its excursion set, i.e. the set of points where the function takes values above or below some prescribed threshold. The objective function is emulated with a Gaussian Process (GP) model based on an initial design of experiments enriched with evaluation results at (batch-) sequentially determined input points. The GP model provides conservative estimates for the excursion set, which control false positives while minimizing false negatives. We introduce adaptive strategies that sequentially select new evaluations of the function by reducing the uncertainty on conservative estimates. Following the Stepwise Uncertainty Reduction approach we obtain new evaluations by minimizing adapted criteria. Tractable formulae for the conservative criteria are derived, which allow more convenient optimization. The method is benchmarked on random functions generated under the model assumptions in different scenarios of noise and batch size. We then apply it to a reliability engineering test case. Overall, the proposed strategy of minimizing false negatives in conservative estimation achieves competitive performance both in terms of model-based and model-free indicators.
Published in Technometrics, Taylor & Francis, pp. 1–30.
Adaptive design of experiments for conservative estimation of excursion sets
@ARTICLE{azzimontid2019c,
title = {Adaptive design of experiments for conservative estimation of excursion sets},
journal = {Technometrics},
publisher = {Taylor & Francis},
author = {Azzimonti, D. and Ginsbourger, D. and Chevalier, C. and Bect, J. and Richet, Y.},
pages = {1--30},
year = {2019},
doi = {10.1080/00401706.2019.1693427}
}
Download
Azzimonti, D., Ginsbourger, D., Rohmer, J., Idier, D. (2019). Profile extrema for visualizing and quantifying uncertainties on excursion regions. Application to coastal flooding. Technometrics 61(4), pp. 474–493.
Profile extrema for visualizing and quantifying uncertainties on excursion regions. Application to coastal flooding
Authors: Azzimonti, D. and Ginsbourger, D. and Rohmer, J. and Idier, D.
Year: 2019
Abstract: We consider the problem of describing excursion sets of a real-valued function f, i.e. the set of inputs where f is above a fixed threshold. Such regions are hard to visualize if the input space dimension, d, is higher than 2. For a given projection matrix from the input space to a lower dimensional (usually 1, 2) subspace, we introduce profile sup (inf) functions that associate to each point in the projection’s image the sup (inf) of the function constrained over the pre-image of this point by the considered projection. Plots of profile extrema functions convey a simple, although intrinsically partial, visualization of the set. We consider expensive to evaluate functions where only a very limited number of evaluations, n, is available, e.g. n<100d, and we surrogate f with a posterior quantity of a Gaussian process (GP) model. We first compute profile extrema functions for the posterior mean given n evaluations of f. We quantify the uncertainty on such estimates by studying the distribution of GP profile extrema with posterior quasi-realizations obtained from an approximating process. We control such approximation with a bound inherited from the Borell-TIS inequality. The technique is applied to analytical functions (d = 2, 3) and to a 5-dimensional coastal flooding test case for a site located on the Atlantic French coast. Here f is a numerical model returning the area of flooded surface in the coastal region given some offshore conditions. Profile extrema functions allowed us to better understand which offshore conditions impact large flooding events.
Published in Technometrics 61(4), Taylor & Francis, pp. 474–493.
Profile extrema for visualizing and quantifying uncertainties on excursion regions. Application to coastal flooding
@ARTICLE{azzimontid2019a,
title = {Profile extrema for visualizing and quantifying uncertainties on excursion regions. Application to coastal flooding},
journal = {Technometrics},
publisher = {Taylor & Francis},
volume = {61},
author = {Azzimonti, D. and Ginsbourger, D. and Rohmer, J. and Idier, D.},
number = {4},
pages = {474--493},
year = {2019},
doi = {10.1080/00401706.2018.1562987}
}
Download
Azzimonti, D., Rottondi, C., Tornatore, M. (2019). Using active learning to decrease probes for QoT estimation in optical networks. In , Optical Society of America, Th1H.1.
Using active learning to decrease probes for QoT estimation in optical networks
Authors: Azzimonti, D. and Rottondi, C. and Tornatore, M.
Year: 2019
Abstract: We use active learning to reduce the number of probes needed for machine learning-based QoT estimation. When building an estimation model based on Gaussian processes, only QoT instances that minimize estimation uncertainty are iteratively requested.
Published in Optical Fiber Communication Conference (OFC) 2019, Optical Society of America, Th1H.1.
Using active learning to decrease probes for QoT estimation in optical networks
@INPROCEEDINGS{azzimontid2019b,
title = {Using active learning to decrease probes for {QoT} estimation in optical networks},
journal = {Optical Fiber Communication Conference ({OFC}) 2019},
publisher = {Optical Society of America},
author = {Azzimonti, D. and Rottondi, C. and Tornatore, M.},
pages = {Th1H.1},
year = {2019},
doi = {10.1364/OFC.2019.Th1H.1}
}
Download
Benavoli, A., Facchini, A., Zaffalon, M. (2019). Bernstein's socks, polynomial-time provable coherence and entanglement. In De Bock, J., de Campos, C., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), ISIPTA '19: Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications, PMLR 103, JMLR.org, pp. 23-31.
Bernstein's socks, polynomial-time provable coherence and entanglement
Authors: Benavoli, A. and Facchini, A. and Zaffalon, M.
Year: 2019
Abstract: We recently introduced a bounded rationality approach for the theory of desirable gambles. It is based on the unique requirement that being nonnegative for a gamble has to be defined so that it can be provable in polynomial time. In this paper we continue to investigate properties of this class of models. In particular we verify that the space of Bernstein polynomials in which nonnegativity is specified by the Krivine-Vasilescu certificate is yet another instance of this theory. As a consequence, we show how it is possible to construct in it a thought experiment uncovering entanglement with classical (hence non quantum) coins.
Accepted in De Bock, J., de Campos, C., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), ISIPTA '19: Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications, PMLR 103, JMLR.org, pp. 23-31.
Bernstein's socks, polynomial-time provable coherence and entanglement
@INPROCEEDINGS{zaffalon2019b,
title = {Bernstein's socks, polynomial-time provable coherence and entanglement},
editor = {De Bock, J. and de Campos, C. and de Cooman, G. and Quaeghebeur, E. and Wheeler, G.},
publisher = {JMLR.org},
series = {PMLR},
volume = {103},
booktitle = {{ISIPTA };'19: Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications},
author = {Benavoli, A. and Facchini, A. and Zaffalon, M.},
pages = {23-31},
year = {2019},
url = {http://proceedings.mlr.press/v103/benavoli19a/benavoli19a.pdf}
}
Download
Bolt, J.H., van der Gaag, L.C. (2019). On minimum elementary-triplet bases for independence relations. In De Bock, J., de Campos, C.P., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications (ISIPTA '19), PMLR 103, JMLR.org, pp. 32–37.
On minimum elementary-triplet bases for independence relations
Authors: Bolt, J.H. and van der Gaag, L.C.
Year: 2019
Abstract: A semi-graphoid independence relation is a set of independence statements, called triplets, and is typically exponentially large in the number of variables involved. For concise representation of such a relation, a subset of its triplets is listed in a so-called basis; its other triplets are defined implicitly through a set of axioms. An elementary-triplet basis for this purpose consists of all elementary triplets of a relation. Such a basis however, may include redundant information. In this paper we provide two lower bounds on the size of an elementary-triplet basis in general and an upper bound on the size of a minimum elementary-triplet basis. We further specify the construction of an elementary-triplet basis of minimum size for restricted relations.
Keywords: Independence relations, Axioms of independence, Elementary triplets, Basis representation.
Published in De Bock, J., de Campos, C.P., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications (ISIPTA '19), PMLR 103, JMLR.org, pp. 32–37.
On minimum elementary-triplet bases for independence relations
@INPROCEEDINGS{Linda2019c,
title = {On minimum elementary-triplet bases for independence relations},
editor = {De Bock, J. and de Campos, C.P. and de Cooman, G. and Quaeghebeur, E. and Wheeler, G.},
publisher = {JMLR.org},
series = {PMLR},
volume = {103},
booktitle = {Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications ({ISIPTA} '19)},
author = {Bolt, J.H. and van der Gaag, L.C.},
pages = {32--37},
year = {2019},
url = {http://www.sipta.org/isipta19/contributions/bolt19.pdf}
}
Download
Breschi, V., Piga, D., Bemporad, A. (2019). Online end-use energy disaggregation via jump linear models. Control Engineering Practice 89, pp. 30–42.
Online end-use energy disaggregation via jump linear models
Authors: Breschi, V. and Piga, D. and Bemporad, A.
Year: 2019
Abstract: This paper presents two iterative algorithms for non-intrusive appliance load monitoring, which aims to decompose the aggregate power consumption only measured at the household level into the contributions of the individual electric appliances. The approaches are based on modelling the total power consumption as a combination of jump linear sub-models, each of them describing the behaviour of the individual appliance. Dynamic-programming and multi-model Kalman filtering techniques are used to reconstruct the power consumptions at the single-appliance level from the aggregate power in an iterative way.
Published in Control Engineering Practice 89, pp. 30–42.
Online end-use energy disaggregation via jump linear models
@ARTICLE{piga2019b,
title = {Online end-use energy disaggregation via jump linear models},
journal = {Control Engineering Practice},
volume = {89},
author = {Breschi, V. and Piga, D. and Bemporad, A.},
pages = {30--42},
year = {2019},
doi = {10.1016/j.conengprac.2019.05.011}
}
Download
Bucher, D., Mangili, F., Cellina, F., Bonesana, C., Jonietz, D., Raubal, M. (2019). From location tracking to personalized eco-feedback: a framework for geographic information collection, processing and visualization to promote sustainable mobility behaviors. Travel Behaviour and Society 14, pp. 43-56.
From location tracking to personalized eco-feedback: a framework for geographic information collection, processing and visualization to promote sustainable mobility behaviors
Authors: Bucher, D. and Mangili, F. and Cellina, F. and Bonesana, C. and Jonietz, D. and Raubal, M.
Year: 2019
Abstract: Nowadays, most people carry around a powerful smartphone which is well suited to constantly monitor the location and sometimes even the activity of its user. This makes tracking prevalent and leads to a large number of projects concerned with trajectory data. One area of particular interest is transport and mobility, where data is important for urban planning and smart city-related activities, but can also be used to provide individual users with feedback and suggestions for personal behavior change. As part of a large-scale study based in Switzerland, we use activity tracking data to provide people with eco-feedback on their own mobility patterns and stimulate them to adopt more energy-efficient mobility choices. In this paper we explore the opportunities offered by smartphone based activity tracking, propose a general framework to exploit location data to foster more sustainable mobility behavior, describe the technical solutions chosen and discuss a range of outcomes in terms of user perception and sustainability potential. The presented approach extracts mobility patterns from users’ trajectories, computes credible alternative transport options, and presents the results in a concise and clear way. The resulting eco-feedback helps people to understand their mobility choices, discover the most non-ecological parts of their travel behavior, and explore feasible alternatives.
Published in Travel Behaviour and Society 14, pp. 43-56.
From location tracking to personalized eco-feedback: a framework for geographic information collection, processing and visualization to promote sustainable mobility behaviors
@ARTICLE{mangili2019a,
title = { From location tracking to personalized eco-feedback: a framework for geographic information collection, processing and visualization to promote sustainable mobility behaviors},
journal = {Travel Behaviour and Society},
volume = {14},
author = {Bucher, D. and Mangili, F. and Cellina, F. and Bonesana, C. and Jonietz, D. and Raubal, M.},
pages = {43-56},
year = {2019},
doi = {10.1016/j.tbs.2018.09.005}
}
Download
Carollo, V., Piga, D., Borri, C., Paggi, M. (2019). Identification of elasto-plastic and nonlinear fracture mechanics parameters of silver-plated copper busbars for photovoltaics. Engineering Fracture Mechanics 205, pp. 439–454.
Identification of elasto-plastic and nonlinear fracture mechanics parameters of silver-plated copper busbars for photovoltaics
Authors: Carollo, V. and Piga, D. and Borri, C. and Paggi, M.
Year: 2019
Abstract: Silver-plated copper busbars are screen printed onto silicon solar cells and have the key role to collect the electric current produced by the solar cell. Busbars of two adjacent solar cells are then connected by a soldered ribbon made of the same material. Due to mechanical and thermal loads, such a ribbon is subject to axial deformation that, often, causes plasticity and, in some cases, its breakage due to crack growth. A procedure based on the gradient-descent method and particle swarm optimization is herein proposed for the identification of elasto-plastic and nonlinear (cohesive zone model, CZM) fracture mechanics parameters of silver-plated copper busbars. The proposed method requires the experimental determination of the force-displacement curves from uniaxial tensile tests on busbar samples with and without initial notches. The inspection of in situ SEM images during the tests allows also the estimation of the crack opening, which is found to be an important local quantity to assess the reliability of different CZMs in simulating a crack growth process consistent with the real one.
Published in Engineering Fracture Mechanics 205, pp. 439–454.
Identification of elasto-plastic and nonlinear fracture mechanics parameters of silver-plated copper busbars for photovoltaics
@ARTICLE{piga2019d,
title = {Identification of elasto-plastic and nonlinear fracture mechanics parameters of silver-plated copper busbars for photovoltaics},
journal = {Engineering Fracture Mechanics},
volume = {205},
author = {Carollo, V. and Piga, D. and Borri, C. and Paggi, M.},
pages = {439--454},
year = {2019},
url = {https://www.sciencedirect.com/science/article/pii/S001379441830451X?via%3Dihub}
}
Download
Cellina, F., Bucher, D., Mangili, F., Simão, J.V., Rudel, R., Raubal, M. (2019). A large scale, app-based behaviour change experiment persuading sustainable mobility patterns: methods, results and lessons learnt. Sustainability 11(9), 2674.
A large scale, app-based behaviour change experiment persuading sustainable mobility patterns: methods, results and lessons learnt
Authors: Cellina, F. and Bucher, D. and Mangili, F. and Simão, J.V. and Rudel, R. and Raubal, M.
Year: 2019
Abstract: The present urban transportation system, mostly tailored for cars, has long shown its limitations. In many urban areas, public transportation and soft mobility would be able to effectively satisfy many travel needs. However, they tend to be neglected, due to a deep-rooted car dependency. How can we encourage people to make sustainable mobility choices, reducing car use and the related CO 2 emissions and energy consumption? Taking advantage of the wide availability of smartphone devices, we designed GoEco!, a smartphone application exploiting automatic mobility tracking, eco-feedback, social comparison and gamification elements to persuade individual modal change. We tested the effectiveness of GoEco! in two regions of Switzerland (Cantons Ticino and Zurich), in a large-scale, one year long randomized controlled trial. Notwithstanding a large drop-out rate experienced throughout the experiment, GoEco! was observed to produce a statistically significant impact (a decrease in CO 2 emissions and energy consumption per kilometer) for systematic routes in highly car-dependent urban areas, such as the Canton Ticino. In Zurich, instead, where high quality public transport is already available, no statistically significant effects were found. In this paper we present the GoEco! experiment and discuss its results and the lessons learnt, highlighting practical difficulties in performing randomized controlled trials in the field of mobility and providing recommendations for future research.
Published in mdpi (Ed), Sustainability 11(9), 2674.
Note: Special Issue Environmental and Behavioral Consequences of Interventions for Sustainable Travel
A large scale, app-based behaviour change experiment persuading sustainable mobility patterns: methods, results and lessons learnt
@ARTICLE{mangili2019b,
title = {A large scale, app-based behaviour change experiment persuading sustainable mobility patterns: methods, results and lessons learnt},
journal = {Sustainability},
editor = {mdpi},
volume = {11},
author = {Cellina, F. and Bucher, D. and Mangili, F. and Sim\~ao, J.V. and Rudel, R. and Raubal, M.},
number = {9},
pages = {2674},
year = {2019},
doi = {https://doi.org/10.3390/su11092674}
}
Download
Colic, N., Rinaldi, F. (2019). Improving spaCy dependency annotation and PoS tagging web service using independent NER services. Genomics Inform 17(2), e21.
Improving spaCy dependency annotation and PoS tagging web service using independent NER services
Authors: Colic, N. and Rinaldi, F.
Year: 2019
Abstract: Dependency parsing is often used as a component in many text analysis pipelines. However, performance, especially in specialized domains, suffers from the presence of complex terminology. Our hypothesis is that including named entity annotations can improve the speed and quality of dependency parses. As part of BLAH5, we built a web service delivering improved dependency parses by taking into account named entity annotations obtained by third party services. Our evaluation shows improved results and better speed.
Published in Genomics Inform 17(2), e21.
Improving spaCy dependency annotation and PoS tagging web service using independent NER services
@ARTICLE{rinaldi2019c,
title = {Improving {spaCy} dependency annotation and {PoS} tagging web service using independent {NER} services},
journal = {Genomics Inform},
volume = {17},
author = {Colic, N. and Rinaldi, F.},
number = {2},
pages = {e21},
year = {2019},
doi = {10.5808/GI.2019.17.2.e21}
}
Download
Correia, A.H.C., de Campos, C.P., van der Gaag, L.C. (2019). An experimental study of prior dependence in Bayesian network structure learning. In De Bock, J., de Campos, C.P., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications (ISIPTA '19), PMLR 103, JMLR.org, pp. 78–81.
An experimental study of prior dependence in Bayesian network structure learning
Authors: Correia, A.H.C. and de Campos, C.P. and van der Gaag, L.C.
Year: 2019
Abstract: The Bayesian Dirichlet equivalent uniform (BDeu) function is a popular score to evaluate the goodness of a Bayesian network structure given complete categorical data. Despite its interesting properties, such as likelihood equivalence, it does require a prior expressed via a user-defined parameter known as Equivalent Sample Size (ESS), which significantly affects the final structure. We study conditions to obtain prior independence in BDeu-based structure learning. We show in experiments that the amount of data needed to render the learning robust to different ESS values is prohibitively large, even in big data times.
Keywords: Robustness, Bayesian Networks, Structure Learning, BDeu
Published in De Bock, J., de Campos, C.P., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications (ISIPTA '19), PMLR 103, JMLR.org, pp. 78–81.
An experimental study of prior dependence in Bayesian network structure learning
@INPROCEEDINGS{Linda2910d,
title = {An experimental study of prior dependence in {B}ayesian network structure learning},
editor = {De Bock, J. and de Campos, C.P. and de Cooman, G. and Quaeghebeur, E. and Wheeler, G.},
publisher = {JMLR.org},
series = {PMLR},
volume = {103},
booktitle = {Proceedings of the Eleventh International Symposium on Imprecise Probability: Theories and Applications ({ISIPTA} '19)},
author = {Correia, A.H.C. and de Campos, C.P. and van der Gaag, L.C.},
pages = {78--81},
year = {2019},
url = {http://www.sipta.org/isipta19/contributions/correia19.pdf}
}
Download
Ellendorff, T., Furrer, L., Colic, N., Aepli, N., Rinaldi, F. (2019). Approaching SMM4H with merged models and multi-task learning. In Proceedings of the Fourth Social Media Mining for Health Applications (#smm4h) Workshop & Shared Task, Association for Computational Linguistics, pp. 58–61.
Approaching SMM4H with merged models and multi-task learning
Authors: Ellendorff, T. and Furrer, L. and Colic, N. and Aepli, N. and Rinaldi, F.
Year: 2019
Abstract: We describe our submissions to the 4th edi- tion of the Social Media Mining for Health Applications (SMM4H) shared task. Our team (UZH) participated in two sub-tasks: Automatic classifications of adverse effects mentions in tweets (Task 1) and Generaliz- able identification of personal health expe- rience mentions (Task 4). For our submis- sions, we exploited ensembles based on a pre- trained language representation with a neu- ral transformer architecture (BERT) (Tasks 1 and 4) and a CNN-BiLSTM(-CRF) network within a multi-task learning scenario (Task 1). These systems are placed on top of a carefully crafted pipeline of domain-specific pre- processing steps.
Published in Proceedings of the Fourth Social Media Mining for Health Applications (#smm4h) Workshop & Shared Task, Association for Computational Linguistics, pp. 58–61.
Approaching SMM4H with merged models and multi-task learning
@INPROCEEDINGS{rinaldi2019g,
title = {Approaching {SMM4H} with merged models and multi-task learning},
publisher = {Association for Computational Linguistics},
booktitle = {Proceedings of the Fourth Social Media Mining for Health Applications (\#smm4h) Workshop & Shared Task},
author = {Ellendorff, T. and Furrer, L. and Colic, N. and Aepli, N. and Rinaldi, F.},
pages = {58--61},
year = {2019},
doi = {10.18653/v1/W19-3208},
url = {https://www.aclweb.org/anthology/W19-3208}
}
Download
Furrer, L., Cornelius, J., Rinaldi, F. (2019). UZH@CRAFT-ST: a sequence-labeling approach to concept recognition. In Proceedings of the 5th Workshop on Bionlp Open Shared Tasks, Association for Computational Linguistics, pp. 185–195.
UZH@CRAFT-ST: a sequence-labeling approach to concept recognition
Authors: Furrer, L. and Cornelius, J. and Rinaldi, F.
Year: 2019
Abstract: As our submission to the CRAFT shared task 2019, we present two neural approaches to concept recognition. We propose two different systems for joint named entity recognition (NER) and normalization (NEN), both of which model the task as a sequence labeling problem. Our first system is a BiLSTM network with two separate outputs for NER and NEN trained from scratch, whereas the second system is an instance of BioBERT fine-tuned on the concept-recognition task. We exploit two strategies for extending concept coverage, ontology pretraining and backoff with a dictionary lookup. Our results show that the backoff strategy effectively tackles the problem of unseen concepts, addressing a major limitation of the chosen design. In the cross-system comparison, BioBERT proves to be a strong basis for creating a concept-recognition system, although some entity types are predicted more accurately by the BiLSTM-based system.
Published in Proceedings of the 5th Workshop on Bionlp Open Shared Tasks, Association for Computational Linguistics, pp. 185–195.
UZH@CRAFT-ST: a sequence-labeling approach to concept recognition
@INPROCEEDINGS{rinaldi2019h,
title = {{UZH@CRAFT}-{ST}: a sequence-labeling approach to concept recognition},
publisher = {Association for Computational Linguistics},
booktitle = {Proceedings of the 5th Workshop on Bionlp Open Shared Tasks},
author = {Furrer, L. and Cornelius, J. and Rinaldi, F.},
pages = {185--195},
year = {2019},
doi = {10.18653/v1/D19-5726}
}
Download
Furrer, L., Jancso, A., Colic, N., Rinaldi, F. (2019). Oger++: hybrid multi-type entity recognition. Journal of Cheminformatics 11(1), 7.
Oger++: hybrid multi-type entity recognition
Authors: Furrer, L. and Jancso, A. and Colic, N. and Rinaldi, F.
Year: 2019
Abstract: Background: We present a text-mining tool for recognizing biomedical entities in scientific literature. OGER++ is a hybrid system for named entity recognition and concept recognition (linking), which combines a dictionary-based annotator with a corpus-based disambiguation component. The annotator uses an efficient look-up strategy combined with a normalization method for matching spelling variants. The disambiguation classifier is implemented as a feed-forward neural network which acts as a postfilter to the previous step. Results: We evaluated the system in terms of processing speed and annotation quality. In the speed benchmarks, the OGER++ web service processes 9.7 abstracts or 0.9 full-text documents per second. On the CRAFT corpus, we achieved 71.4% and 56.7% F1 for named entity recognition and concept recognition, respectively. Conclusions: Combining knowledge-based and data-driven components allows creating a system with competitive performance in biomedical text mining.
Published in Journal of Cheminformatics 11(1), BioMed Central, 7.
Oger++: hybrid multi-type entity recognition
@ARTICLE{rinaldi2019d,
title = {Oger++: hybrid multi-type entity recognition},
journal = {Journal of Cheminformatics},
publisher = {BioMed Central},
volume = {11},
author = {Furrer, L. and Jancso, A. and Colic, N. and Rinaldi, F.},
number = {1},
pages = {7},
year = {2019},
doi = {10.1186/s13321-018-0326-3},
url = {https://doi.org/10.5167/uzh-162875}
}
Download
Kanjirangat, V., Antonucci, A. (2019). NOVEL2GRAPH: Visual summaries of narrative text enhanced by machine learning. Proceedings of the Text2StoryIR'19 Workshop, Cologne, Germany, 14-April-2019, pp. 29–37.
NOVEL2GRAPH: Visual summaries of narrative text enhanced by machine learning
Authors: Kanjirangat, V. and Antonucci, A.
Year: 2019
Abstract: A machine learning approach to the creation of visual summaries for
narrative text is presented. Standard natural language processing tools
for named entities recognition are used together with a clustering algorithm
to detect the characters of the novel and their aliases. The
most relevant ones and their relations are evaluated on the basis of
a simple statistical analysis. These characters are visually depicted
as nodes of an undirected graph whose edges describe relations with
other characters. Specialized sentiment analysis techniques based on
sentence embedding decide the colours of characters/nodes and their
relations/edges. Additional information about the characters (e.g., gender)
and their relations (e.g., siblings or partnerships) are returned by
binary classifiers and visually depicted in the graph. For those specialized
tasks, small amounts of manually annotated data are sucient
to achieve good accuracy. Compared to analogous tools, the machine
learning approach we present allows for a richer representation of texts
of this kind. A case study to demonstrate this approach for a series of
books is also reported.
Published in A. Jorge, R. Campos, A. Jatowt, S. Bhatia (Ed), Proceedings of the Text2StoryIR'19 Workshop, Cologne, Germany, 14-April-2019, ceur, pp. 29–37.
NOVEL2GRAPH: Visual summaries of narrative text enhanced by machine learning
@ARTICLE{vani2019a,
title = {{NOVEL2GRAPH}: {V}isual summaries of narrative text enhanced by machine learning},
journal = {Proceedings of the {Text2StoryIR'19} Workshop, Cologne, Germany, 14-April-2019},
editor = {A. Jorge, R. Campos, A. Jatowt, S. Bhatia},
publisher = {ceur},
author = {Kanjirangat, V. and Antonucci, A.},
pages = {29--37},
year = {2019},
url = {http://ceur-ws.org/Vol-2342/paper4.pdf}
}
Download
Kanjirangat, V., Oita, M., Oezdemir-Zaech, F. (2019). Semantically corroborating neural attention for biomedical question answering. In BioASQ: Large-Scale Biomedical Semantic Indexing and Question Answering: Workshop of ECML/PKDD 2019, Springer, Lecture Notes in Computer Science.
Semantically corroborating neural attention for biomedical question answering
Authors: Kanjirangat, V. and Oita, M. and Oezdemir-Zaech, F.
Year: 2019
Abstract: Biomedical question answering is a great challenge in NLP
due to complex scientific vocabulary and lack of massive annotated corpora, but, at the same time, is full of potential in optimizing in critical ways the biomedical practices. This paper describes the work carried out as a part of the BioASQ challenge (Task-7B Phase-B), and targets an integral step in the question answering process: extractive answer selection.This deals with the identification of the exact answer (words, phrases or sentences) from given article snippets that are related to the question at hand. We address this problem in the context of factoid and summarization question types, using a variety of deep learning and semantic methods, including various architectures (e.g, Dynamic Memory Networks and Bidirectional Attention Flow), transfer learning, biomedical
named entity recognition and corroboration of semantic evidence. On the top of candidate answer selection module, answer prediction to yes/no question types is also addressed by incorporating a sentiment analysis approach. The evaluation with respect to Rouge, MRR and F1 scores, in
relation to the type of question answering task being considered, exhibits the potential of this hybrid method in extracting the correct answer to a question. In addition, the proposed corroborating semantics module can be added on top of the typical QA pipeline to gain a measured 5% improvement in identifying the exact answer with respect to the gold standard.
Accepted in BioASQ: Large-Scale Biomedical Semantic Indexing and Question Answering: Workshop of ECML/PKDD 2019, Springer, Lecture Notes in Computer Science.
Semantically corroborating neural attention for biomedical question answering
@INPROCEEDINGS{supsi2019d,
title = {Semantically corroborating neural attention for biomedical question answering},
publisher = {Springer, Lecture Notes in Computer Science},
booktitle = {{BioASQ}: Large-Scale Biomedical Semantic Indexing and Question Answering: Workshop of {ECML/PKDD} 2019},
author = {Kanjirangat, V. and Oita, M. and Oezdemir-Zaech, F.},
year = {2019},
url = {http://bioasq.org/}
}
Download
Kim, J.D., Cohen, K.B., Collier, N., Lu, Z., Rinaldi, F. (2019). Introduction to BLAH5 special issue: recent progress on interoperability of biomedical text mining. Genomics Inform 17(2), e12.
Introduction to BLAH5 special issue: recent progress on interoperability of biomedical text mining
Authors: Kim, J.D. and Cohen, K.B. and Collier, N. and Lu, Z. and Rinaldi, F.
Year: 2019
Abstract: BLAH is organized annually by the Database Center for Life Science (DBCLS), Research Organization of Information and Systems (ROIS). The goal of the BLAH series is to enhance the interoperability of resources for biomedical text annotation and mining, which we believe is a key for the next breakthrough of biomedical text mining. This special issue delivers seven application notes and two mini reviews, under the theme, “biomedical text mining.” They are outcomes from the 5th Biomedical Linked Annotation Hackathon (BLAH5), which was held from 12th through 15th February 2019 in Kashiwa, Japan.
Published in Genomics Inform 17(2), e12.
Introduction to BLAH5 special issue: recent progress on interoperability of biomedical text mining
@ARTICLE{rinaldi2019b,
title = {Introduction to {BLAH5} special issue: recent progress on interoperability of biomedical text mining},
journal = {Genomics Inform},
volume = {17},
author = {Kim, J.D. and Cohen, K.B. and Collier, N. and Lu, Z. and Rinaldi, F.},
number = {2},
pages = {e12},
year = {2019},
doi = {10.5808/GI.2019.17.2.e12}
}
Download
Renooij, S., van der Gaag, L.C., Leray, Ph. (2019). On intercausal interactions in probabilistic relational models. In De Bock, J., de Campos, C.P., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), Proceedings of the Eleventh International Symposium on Imprecise Probabilities: Theories and Applications (ISIPTA '19), Proceedings of Machine Learning Research 103, pp. 327–329.
On intercausal interactions in probabilistic relational models
Authors: Renooij, S. and van der Gaag, L.C. and Leray, Ph.
Year: 2019
Abstract: Probabilistic relational models (PRMs) extend Bayesian networks beyond propositional expressiveness by allowing the representation of multiple interacting classes. For a specific instance of sets of concrete objects per class, a ground Bayesian network is composed by replicating parts of the PRM. The interactions between the objects that are thereby induced, are not always obvious from the PRM. We demonstrate in this paper that the replicative structure of the ground network in fact constrains the space of possible probability distributions and thereby the possibly patterns of intercausal interaction.
Keywords:
PRM instances, qualitative constraints on probability distributions, intercausal interaction.
Published in De Bock, J., de Campos, C.P., de Cooman, G., Quaeghebeur, E., Wheeler, G. (Eds), Proceedings of the Eleventh International Symposium on Imprecise Probabilities: Theories and Applications (ISIPTA '19), Proceedings of Machine Learning Research 103, pp. 327–329.
On intercausal interactions in probabilistic relational models
@INPROCEEDINGS{Linda2019b,
title = {On intercausal interactions in probabilistic relational models},
editor = {De Bock, J. and de Campos, C.P. and de Cooman, G. and Quaeghebeur, E. and Wheeler, G.},
series = {Proceedings of Machine Learning Research},
volume = {103},
booktitle = {Proceedings of the Eleventh International Symposium on Imprecise Probabilities: Theories and Applications ({ISIPTA} '19)},
author = {Renooij, S. and van der Gaag, L.C. and Leray, Ph.},
pages = {327--329},
year = {2019},
url = {http://www.sipta.org/isipta19/contributions/renooij19.pdf}
}
Download
Mattei, L., Soares, D.L., Antonucci, A., Mauà, D.D., Facchini, A. (2019). Exploring the space of probabilistic sentential decision diagrams. In Proceedings of the 3rd Tractable Probabilistic Modeling Workshop, 36th International Conference on Machine Learning.
Exploring the space of probabilistic sentential decision diagrams
Authors: Mattei, L. and Soares, D.L. and Antonucci, A. and Mauà, D.D. and Facchini, A.
Year: 2019
Abstract: Probabilistic sentential decision diagrams (PS- DDs) are annotated circuits providing a possibly compact specification of joint probability mass functions consistent with a formula over a set of propositional variables. PSDD inference is tractable in the sense that marginal queries can be achieved in linear time with respect to the circuit size by traversal algorithms. Unlike other probabilistic graphical models such as Bayesian networks, the problem of learning the structure for PSDDs received relatively little attention. We discuss some preliminary ideas related to the development of pure likelihood-score-based search methods for the learning of PSDD structures fit- ting a formula and a data set of consistent obser- vations. A sampling algorithm for these models is also provided.
Published in Proceedings of the 3rd Tractable Probabilistic Modeling Workshop, 36th International Conference on Machine Learning.
Note: Accepted for pubblication.
Exploring the space of probabilistic sentential decision diagrams
@INPROCEEDINGS{supsi2019a,
title = {Exploring the space of probabilistic sentential decision diagrams},
booktitle = {Proceedings of the 3rd Tractable Probabilistic Modeling Workshop, 36th International Conference on Machine Learning},
author = {Mattei, L. and Soares, D.L. and Antonucci, A. and Mau\`a, D.D. and Facchini, A.},
year = {2019},
url = {https://drive.google.com/file/d/1PkckPpeLUOP_Oeik_q4MprD6ZJeekqHZ/view}
}
Download
Mauri, A., Lettori, J., Fusi, G., Fausti, D., Mor, M., Braghin, F., Legnani, G., Roveda, L. (2019). Mechanical and control design of an industrial exoskeleton for advanced human empowering in heavy parts manipulation tasks. MDPI Robotics.
Mechanical and control design of an industrial exoskeleton for advanced human empowering in heavy parts manipulation tasks
Authors: Mauri, A. and Lettori, J. and Fusi, G. and Fausti, D. and Mor, M. and Braghin, F. and Legnani, G. and Roveda, L.
Year: 2019
Abstract: Exoskeleton robots are a rising technology in industrial contexts to assist humans in onerous applications. Mechanical and control design solutions are intensively investigated to achieve a high performance human-robot collaboration (e.g., transparency, ergonomics, safety, etc.). However, the most of the investigated solutions involve high-cost hardware, complex design solutions and standard actuation. Moreover, state-of-the-art empowering controllers do not allow for online assistance regulation and do not embed advanced safety rules. In the presented work, an industrial exoskeleton with high payload ratio for lifting and transportation of heavy parts is proposed. A low-cost mechanical design solution is described, exploiting compliant actuation at the shoulder joint to increase safety in human-robot cooperation. A hierarchic model-based controller with embedded safety rules is then proposed (including the modeling of the compliant actuator) to actively assist the human while executing the task. An inner optimal controller is proposed for trajectory tracking, while an outer safety-based fuzzy logic controller is proposed to online deform the task trajectory on the basis of the human’s intention of motion. A gain scheduler is also designed to calculate the inner optimal control gains on the basis of the performed trajectory. Simulations have been performed in order to validate the performance of the proposed device, showing promising results. The prototype is under realization.
Published in MDPI Robotics.
Mechanical and control design of an industrial exoskeleton for advanced human empowering in heavy parts manipulation tasks
@ARTICLE{Roveda2019b,
title = {Mechanical and control design of an industrial exoskeleton for advanced human empowering in heavy parts manipulation tasks},
journal = {{MDPI} Robotics},
author = {Mauri, A. and Lettori, J. and Fusi, G. and Fausti, D. and Mor, M. and Braghin, F. and Legnani, G. and Roveda, L.},
year = {2019},
doi = {10.3390/robotics8030065}
}
Download
Mejari, M., Piga, D., Toth, R., Bemporad, A. (2019). Kernelized identification of linear parameter-varying models with linear fractional representation. In 2019 European Control Conference (ecc), Naples, Italy.
Kernelized identification of linear parameter-varying models with linear fractional representation
Authors: Mejari, M. and Piga, D. and Toth, R. and Bemporad, A.
Year: 2019
Abstract: The article presents a method for the identification of Linear Parameter-Varying (LPV) models in a Linear Fractional Representation (LFR), which corresponds to a Linear Time-Invariant (LTI) model connected to a scheduling variable dependency via a feedback path. A two-stage identification approach is proposed. In the first stage, Kernelized Canonical Correlation Analysis (KCCA) is formulated to estimate the state sequence of the underlying LPV model. In the second stage, a non-linear least squares cost function is minimized by employing a coordinate descent algorithm to estimate latent variables characterizing the LFR and the unknown model
matrices of the LTI block by using the state estimates obtained at the first stage. Here, it is assumed that the structure of the scheduling variable dependent block in the feedback path is fixed. For a special case of affine dependence of the model on the feedback block, it is shown that the optimization problem in the second stage reduces to ordinary least-squares followed by a singular value decomposition.
Published in 2019 European Control Conference (ecc), Naples, Italy.
Kernelized identification of linear parameter-varying models with linear fractional representation
@INPROCEEDINGS{piga2019e,
title = {Kernelized identification of linear parameter-varying models with linear fractional representation},
address = {Naples, Italy},
booktitle = {2019 European Control Conference ({e}cc)},
author = {Mejari, M. and Piga, D. and Toth, R. and Bemporad, A.},
year = {2019},
doi = {10.23919/ECC.2019.8796150}
}
Download
Miranda, E., Zaffalon, M. (2019). Compatibility, coherence and the RIP. In Destercke, S.,Denoeux, T., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Ed), Uncertainty Modelling in Data Science, Advances in Intelligent Systems and Computing 832, Springer, pp. 166–174.
Compatibility, coherence and the RIP
Authors: Miranda, E. and Zaffalon, M.
Year: 2019
Abstract: We generalise the classical result on the compatibility of marginal, possible non-disjoint, assessments in terms of the running intersection property to the imprecise case, where our beliefs are modelled in terms of sets of desirable gambles. We consider the case where we have unconditional and conditional assessments, and show that the problem can be simplified via a tree decomposition.
Published in Destercke, S.,Denoeux, T., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Ed), Uncertainty Modelling in Data Science, Advances in Intelligent Systems and Computing 832, Springer, pp. 166–174.
Note: SMPS 2018: Proceedings of the 9th international conference on Soft Methods in Probability and Statistics
Compatibility, coherence and the RIP
@INCOLLECTION{zaffalon2018a,
title = {Compatibility, coherence and the {RIP}},
editor = {Destercke, S.,Denoeux, T., Gil, M. A., Grzegorzewski, P., Hryniewicz, O.},
publisher = {Springer},
series = {Advances in Intelligent Systems and Computing},
volume = {832},
booktitle = {Uncertainty Modelling in Data Science},
author = {Miranda, E. and Zaffalon, M.},
pages = {166--174},
year = {2019},
doi = {10.1007/978-3-319-97547-4_22}
}
Download
Musumeci, F., Rottondi, C.E.M., Corani, G., Shahkarami, S., Cugini, F., Tornatore, M. (2019). A tutorial on machine learning for failure management in optical networks. Journal of Lightwave Technology 37(16).
A tutorial on machine learning for failure management in optical networks
Authors: Musumeci, F. and Rottondi, C.E.M. and Corani, G. and Shahkarami, S. and Cugini, F. and Tornatore, M.
Year: 2019
Abstract: Failure management plays a role of capital importance in optical networks to avoid service disruptions and to satisfy customers' service level agreements. Machine Learning (ML) promises to revolutionize the (mostly manual and human-driven) approaches in which failure management in optical networks has been traditionally managed, by introducing automated methods for failure prediction, detection, localization and identification. This tutorial provides a gentle introduction to some ML techniques that have been recently applied in the field of optical-network failure management. It then introduces a taxonomy to classify failure-management tasks and discusses possible applications of ML for these failure management tasks. Finally, for a reader interested in more implementative details, we provide a step-by-step description of how to solve a representative example of a practical failure-management task.
Accepted in Journal of Lightwave Technology 37(16).
A tutorial on machine learning for failure management in optical networks
@ARTICLE{corani2019b,
title = {A tutorial on machine learning for failure management in optical networks},
journal = {Journal of Lightwave Technology},
volume = {37},
author = {Musumeci, F. and Rottondi, C.E.M. and Corani, G. and Shahkarami, S. and Cugini, F. and Tornatore, M.},
number = {16},
year = {2019},
doi = {10.1109/JLT.2019.2922586}
}
Download
Oita, M. (2019). Reverse engineering creativity into interpretable neural networks. In Future of Information and Communications, Lecture Notes in Networks and Systems 70, pp. 235-247.
Reverse engineering creativity into interpretable neural networks
Authors: Oita, M.
Year: 2019
Abstract: In the field of AI the ultimate goal is to achieve generic intelligence, also called True AI, but which depends on the successful enablement of imagination and creativity in artificial agents. To address this problem, this paper presents a novel deep learning framework for creativity, called INNGenuity. Pursuing an interdisciplinary implementation of creativity conditions, INNGenuity aims at the resolution of the various flaws of current AI learning architectures, which stem from the opacity of their models. Inspired by the neuroanatomy of the brain during creative cognition, the proposed framework's hybrid architecture blends both symbolic and connectionist AI, inline with Minsky's ”society of mind”. At its core, semantic gates are designed to facilitate an input/output flow of semantic structures and enable the usage of aligning mechanisms between neural activation clusters and semantic graphs. Having as goal alignment maximization, such a system would enable interpretability through the creation of labeled patterns of computation, and propose unaligned but relevant computation patterns as novel and useful, therefore creative.
Published in Future of Information and Communications, Lecture Notes in Networks and Systems 70, pp. 235-247.
Reverse engineering creativity into interpretable neural networks
@INPROCEEDINGS{oita2019innGenuity,
title = {Reverse engineering creativity into interpretable neural networks},
series = {Lecture Notes in Networks and Systems},
volume = {70},
booktitle = {Future of Information and Communications},
author = {Oita, M.},
pages = {235-247},
year = {2019},
doi = {10.1007/978-3-030-12385-7_19}
}
Download
Oita, M. (2019). Incremental alignment of metaphoric language model for poetry composition. In Computing Conference, Springer, "Advances in Intelligent Systems and Computing", pp. 834–845.
Incremental alignment of metaphoric language model for poetry composition
Authors: Oita, M.
Year: 2019
Abstract: The ability to automatically generate meaningful text with respect to a topic is an important AI mission. In particular, the automatic generation of content which is deemed creative is a great challenge. In this paper, poetry generation is approached through the lenses of a new architectural design for creativity that leverages semantics for the creation of variance and the preservation of the content coherency throughout the generation process. The fully implemented system is made available on github.
Accepted in Computing Conference, Springer, "Advances in Intelligent Systems and Computing", pp. 834–845.
Incremental alignment of metaphoric language model for poetry composition
@INPROCEEDINGS{oita2019poetryComposition,
title = {Incremental alignment of metaphoric language model for poetry composition},
publisher = {Springer, "Advances in Intelligent Systems and Computing"},
booktitle = {Computing Conference},
author = {Oita, M.},
pages = {834--845},
year = {2019},
doi = {10.1007/978-3-030-22871-2_59}
}
Download
Piga, D. (2019). Finite-horizon integration for continuous-time identification: bias analysis and application to variable stiffness actuators. International Journal of Control, pp. 1–14.
Finite-horizon integration for continuous-time identification: bias analysis and application to variable stiffness actuators
Authors: Piga, D.
Year: 2019
Abstract: Direct identification of continuous-time dynamical models from sampled data is now a mature discipline, which is known to have many advantages with respect to indirect approaches based on the identification of discretized models. This paper faces the problem of continuous-time identification of linear time-invariant systems through finite-horizon numerical integration and least-square estimation. The bias in the least-squares estimator due to the noise corrupting the signal observations is quantified, and the benefits of numerical integration in the attenuation of this bias are discussed. An extension of the approach which combines numerical integration, least-squares estimation and particle swarm optimization is proposed for the identification of nonlinear systems and nonlinear-in-the-parameter models, and then applied to the estimation of the torque-displacement characteristic of a commercial variable stiffness actuator driving a one-degree-of-freedom pendulum.
Accepted in International Journal of Control, Taylor & Francis, pp. 1–14.
Finite-horizon integration for continuous-time identification: bias analysis and application to variable stiffness actuators
@ARTICLE{piga2019c,
title = {Finite-horizon integration for continuous-time identification: bias analysis and application to variable stiffness actuators},
journal = {International Journal of Control},
publisher = {Taylor & Francis},
author = {Piga, D.},
pages = {1--14},
year = {2019},
doi = {10.1080/00207179.2018.1557348}
}
Download
Piga, D., Benavoli, A. (2019). Semialgebraic outer approximations for set-valued nonlinear filtering. In 2019 European Control Conference (ECC), Naples, Italy.
Semialgebraic outer approximations for set-valued nonlinear filtering
Authors: Piga, D. and Benavoli, A.
Year: 2019
Abstract: This paper addresses the set-valued filtering problem for discrete time-varying dynamical systems, whose process and measurement equations are polynomial functions of the system state. According to a set-membership framework, the process and measurement noises, as well as the initial state, are assumed to belong to bounded uncertainty regions, which are supposed to be generic semialgebraic sets described by polynomial inequalities. A sequential algorithm, based on sum-of-squares (SOS) representation of positive polynomials is proposed to compute a semialgebraic set described by an a-priori fixed number of polynomial constraints which is guaranteed to contain the true state of the system with certainty.
Published in 2019 European Control Conference (ECC), Naples, Italy.
Semialgebraic outer approximations for set-valued nonlinear filtering
@INPROCEEDINGS{piga2019f,
title = {Semialgebraic outer approximations for set-valued nonlinear filtering},
address = {Naples, Italy},
booktitle = {2019 European Control Conference ({ECC})},
author = {Piga, D. and Benavoli, A.},
year = {2019},
doi = {10.23919/ECC.2019.8795731}
}
Download
Piga, D., Forgione, M., Formentin, S., Bemporad, A. (2019). Performance-oriented model learning for data-driven MPC design. IEEE Control Systems Letters 3(3), pp. 577 - 582.
Performance-oriented model learning for data-driven MPC design
Authors: Piga, D. and Forgione, M. and Formentin, S. and Bemporad, A.
Year: 2019
Abstract: Model Predictive Control (MPC) is an enabling technology in applications requiring controlling physical processes in an optimized way under constraints on inputs and outputs. However, in MPC closed-loop performance is pushed to the limits only if the plant under control is accurately modeled; otherwise, robust architectures need to be employed, at the price of reduced performance due to worst-case conservative assumptions. In this paper, instead of adapting the controller to handle uncertainty, we adapt the learning procedure so that the prediction model is selected to provide the best closed-loop performance. More specifically, we apply for the first time the above “identification for control” rationale to hierarchical MPC using data-driven methods and Bayesian optimization.
Published in IEEE Control Systems Letters 3(3), pp. 577 - 582.
Performance-oriented model learning for data-driven MPC design
@ARTICLE{piga2019a,
title = {Performance-oriented model learning for data-driven {MPC} design},
journal = {{IEEE} Control Systems Letters},
volume = {3},
author = {Piga, D. and Forgione, M. and Formentin, S. and Bemporad, A.},
number = {3},
pages = {577 - 582},
year = {2019},
doi = {10.1109/LCSYS.2019.2913347}
}
Download
Renooij, S., van der Gaag, L.C. (2019). The hidden elegance of causal interaction models. In Ben Amor, N., Quost, B., Theobald, M (Eds), 13th International Conference on Scalable Uncertainty Management (SUM '19), Lecture Notes in Artificial Intelligence 11940, Springer, pp. 38–51.
The hidden elegance of causal interaction models
Authors: Renooij, S. and van der Gaag, L.C.
Year: 2019
Abstract: Causal interaction models such as the noisy-OR model, are used in Bayesian networks to simplify probability acquisition for variables with large numbers of modelled causes. These models essentially prescribe how to complete an exponentially large probability table from a linear number of parameters. Yet, typically the full probability tables are required for inference with Bayesian networks in which such interaction models are used, although inference algorithms tailored to specific types of network exist that can directly exploit the decomposition properties of the interaction models. In this paper we revisit these decomposition properties in view of general inference algorithms and demonstrate that they allow an alternative representation of causal interaction models that is quite concise, even with large numbers of causes involved. In addition to forestalling the need of tailored algorithms, our alternative representation brings engineering benefits beyond those widely recognised.
Keywords: Bayesian networks, causal interaction models, maintenance robustness
Published in Ben Amor, N., Quost, B., Theobald, M (Eds), 13th International Conference on Scalable Uncertainty Management (SUM '19), Lecture Notes in Artificial Intelligence 11940, Springer, pp. 38–51.
The hidden elegance of causal interaction models
@INPROCEEDINGS{linda2019a,
title = {The hidden elegance of causal interaction models},
editor = {Ben Amor, N. and Quost, B. and Theobald, M},
publisher = {Springer},
series = {Lecture Notes in Artificial Intelligence},
volume = {11940},
booktitle = {13th International Conference on Scalable Uncertainty Management ({SUM} '19)},
author = {Renooij, S. and van der Gaag, L.C.},
pages = {38--51},
year = {2019},
doi = {10.1007/978-3-030-35514-2_4}
}
Download
Rodriguez-Esteban, R., Vishnyakova, D., Rinaldi, F. (2019). Revisiting the decay of scientific email addresses. bioRxiv.
Revisiting the decay of scientific email addresses
Authors: Rodriguez-Esteban, R. and Vishnyakova, D. and Rinaldi, F.
Year: 2019
Abstract: Email is the primary means of communication for scientists. However, scientific authors change email address over time. Using a new method, we have calculated that approximately 18% of all authorstextquoteright contact email addresses in MEDLINE are invalid. While an unfortunate number, it is, however, lower than previously estimated. To mitigate this problem, institutions should provide email forwarding and scientific authors should use more stable email addresses. In fact, a steadily growing share already use free private email addresses: 32% of all new addresses in MEDLINE in 2018 were of this kind.
Published in bioRxiv, Cold Spring Harbor Laboratory.
Revisiting the decay of scientific email addresses
@ARTICLE{rinaldi2019j,
title = {Revisiting the decay of scientific email addresses},
journal = {{bioRxiv}},
publisher = {Cold Spring Harbor Laboratory},
author = {Rodriguez-Esteban, R. and Vishnyakova, D. and Rinaldi, F.},
year = {2019},
doi = {10.1101/633255},
url = {https://www.biorxiv.org/content/early/2019/05/12/633255}
}
Download
Roveda, L., Haghshenas, S., Caimmi, M., Pedrocchi, N., Molinari Tosatti, L. (2019). Assisting operators in heavy industrial tasks: on the design of an optimized cooperative impedance fuzzy-controller with embedded safety rules. Frontiers in Robotics and AI 6, 75.
Assisting operators in heavy industrial tasks: on the design of an optimized cooperative impedance fuzzy-controller with embedded safety rules
Authors: Roveda, L. and Haghshenas, S. and Caimmi, M. and Pedrocchi, N. and Molinari Tosatti, L.
Year: 2019
Abstract: Human-robot cooperation is increasingly demanded in industrial applications. Many tasks require the robot to enhance the capabilities of humans. In this scenario, safety also plays an important role in avoiding any accident involving humans, robots, and the environment. With this aim, the paper proposes a cooperative fuzzy-impedance control with embedded safety rules to assist human operators in heavy industrial applications while manipulating unknown weight parts. The proposed methodology is composed by four main components: (i) an inner Cartesian impedance controller (to achieve the compliant robot behavior), (ii) an outer fuzzy controller (to provide the assistance to the human operator), (iii) embedded safety rules (to limit force/velocity during the human-robot interaction enhancing safety), and (iv) a neural network approach (to optimize the control parameters for the human-robot collaboration on the basis of the target indexes of assistance performance defined for this purpose). The main achieved result refers to the capability of the controller to deal with uncertain payloads while assisting and empowering the human operator, both embedding in the controller safety features at force and velocity levels and minimizing the proposed performance indexes. The effectiveness of the proposed approach is verified with a KUKA iiwa 14 R820 manipulator in an experimental procedure where human subjects evaluate the robot performance in a collaborative lifting task of a 10 kg part.
Published in Frontiers in Robotics and AI 6, 75.
Assisting operators in heavy industrial tasks: on the design of an optimized cooperative impedance fuzzy-controller with embedded safety rules
@ARTICLE{Roveda2019a,
title = {Assisting operators in heavy industrial tasks: on the design of an optimized cooperative impedance fuzzy-controller with embedded safety rules},
journal = {Frontiers in Robotics and {AI}},
volume = {6},
author = {Roveda, L. and Haghshenas, S. and Caimmi, M. and Pedrocchi, N. and Molinari Tosatti, L.},
pages = {75},
year = {2019},
doi = {10.3389/frobt.2019.00075}
}
Download
Salani, M., Corbellini, G., Corani, G. (2019). Hybrid heuristic for the optimal design of photovoltaic installations considering mismatch loss effects. Computers & Operations Research 108, pp. 112–120.
Hybrid heuristic for the optimal design of photovoltaic installations considering mismatch loss effects
Authors: Salani, M. and Corbellini, G. and Corani, G.
Year: 2019
Abstract: We consider the Photovoltaic Installation Design Problem (PIDP). In this problem, photovoltaic modules must be connected in “strings” and wired to a set of electronic components. The aim is to minimize installation costs and maximize power production, which is affected by “mismatch losses” caused by non-uniform irradiation (shading) and is also directly related to design decisions. We relate the problem to the known class of location routing problems and we design a route-first/cluster-second heuristic. We propose an efficient machine learning approach to evaluate Photovoltaic (PV) string performances accounting for mismatch losses. We prove that our approach is effective in real-world instances provided by our industrial partner.
Published in Computers & Operations Research 108, pp. 112–120.
Hybrid heuristic for the optimal design of photovoltaic installations considering mismatch loss effects
@ARTICLE{corani2019a,
title = {Hybrid heuristic for the optimal design of photovoltaic installations considering mismatch loss effects},
journal = {Computers & Operations Research},
volume = {108},
author = {Salani, M. and Corbellini, G. and Corani, G.},
pages = {112--120},
year = {2019},
doi = {10.1016/j.cor.2019.04.009}
}
Download
Sechidis, K., Azzimonti, L., Pocock, A., Corani, G., Weatherall, J., Brown, G. (2019). Efficient feature selection using shrinkage estimators. Machine Learning 108(8), pp. 1261–1286.
Efficient feature selection using shrinkage estimators
Authors: Sechidis, K. and Azzimonti, L. and Pocock, A. and Corani, G. and Weatherall, J. and Brown, G.
Year: 2019
Abstract: Information theoretic feature selection methods quantify the importance of each feature by estimating mutual information terms to capture: the relevancy, the redundancy and the complementarity. These terms are commonly estimated by maximum likelihood, while an under-explored area of research is how to use shrinkage methods instead. Our work suggests a novel shrinkage method for data-efficient estimation of information theoretic terms. The small sample behaviour makes it particularly suitable for estimation of discrete distributions with large number of categories (bins). Using our novel estimators we derive a framework for generating feature selection criteria that capture any high-order feature interaction for redundancy and complementarity. We perform a thorough empirical study across datasets from diverse sources and using various evaluation measures. Our first finding is that our shrinkage based methods achieve better results, while they keep the same computational cost as the simple maximum likelihood based methods. Furthermore, under our framework we derive efficient novel high-order criteria that outperform state-of-the-art methods in various tasks.
Published in Machine Learning 108(8), pp. 1261–1286.
Efficient feature selection using shrinkage estimators
@ARTICLE{azzimonti2019b,
title = {Efficient feature selection using shrinkage estimators},
journal = {Machine Learning},
volume = {108},
author = {Sechidis, K. and Azzimonti, L. and Pocock, A. and Corani, G. and Weatherall, J. and Brown, G.},
number = {8},
pages = {1261--1286},
year = {2019},
doi = {10.1007/s10994-019-05795-1}
}
Download
Sheikhalishahi, S., Miotto, R., Dudley, J.T., Lavelli, A., Rinaldi, F., Osmani, V. (2019). Natural language processing of clinical notes on chronic diseases: systematic review. JMIR Med Inform 7(2), e12239.
Natural language processing of clinical notes on chronic diseases: systematic review
Authors: Sheikhalishahi, S. and Miotto, R. and Dudley, J.T. and Lavelli, A. and Rinaldi, F. and Osmani, V.
Year: 2019
Abstract: Background: Novel approaches that complement and go beyond evidence-based medicine are required in the domain of chronic diseases, given the growing incidence of such conditions on the worldwide population. A promising avenue is the secondary use of electronic health records (EHRs), where patient data are analyzed to conduct clinical and translational research. Methods based on machine learning to process EHRs are resulting in improved understanding of patient clinical trajectories and chronic disease risk prediction, creating a unique opportunity to derive previously unknown clinical insights. However, a wealth of clinical histories remains locked behind clinical narratives in free-form text. Consequently, unlocking the full potential of EHR data is contingent on the development of natural language processing (NLP) methods to automatically transform clinical text into structured clinical data that can guide clinical decisions and potentially delay or prevent disease onset. Objective: The goal of the research was to provide a comprehensive overview of the development and uptake of NLP methods applied to free-text clinical notes related to chronic diseases, including the investigation of challenges faced by NLP methodologies in understanding clinical narratives. Methods: Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines were followed and searches were conducted in 5 databases using “clinical notes,” “natural language processing,” and “chronic disease” and their variations as keywords to maximize coverage of the articles. Results: Of the 2652 articles considered, 106 met the inclusion criteria. Review of the included papers resulted in identification of 43 chronic diseases, which were then further classified into 10 disease categories using the International Classification of Diseases, 10th Revision. The majority of studies focused on diseases of the circulatory system (n=38) while endocrine and metabolic diseases were fewest (n=14). This was due to the structure of clinical records related to metabolic diseases, which typically contain much more structured data, compared with medical records for diseases of the circulatory system, which focus more on unstructured data and consequently have seen a stronger focus of NLP. The review has shown that there is a significant increase in the use of machine learning methods compared to rule-based approaches; however, deep learning methods remain emergent (n=3). Consequently, the majority of works focus on classification of disease phenotype with only a handful of papers addressing extraction of comorbidities from the free text or integration of clinical notes with structured data. There is a notable use of relatively simple methods, such as shallow classifiers (or combination with rule-based methods), due to the interpretability of predictions, which still represents a significant issue for more complex methods. Finally, scarcity of publicly available data may also have contributed to insufficient development of more advanced methods, such as extraction of word embeddings from clinical notes. Conclusions: Efforts are still required to improve (1) progression of clinical NLP methods from extraction toward understanding; (2) recognition of relations among entities rather than entities in isolation; (3) temporal extraction to understand past, current, and future clinical events; (4) exploitation of alternative sources of clinical knowledge; and (5) availability of large-scale, de-identified clinical corpora.
Published in JMIR Med Inform 7(2), e12239.
Natural language processing of clinical notes on chronic diseases: systematic review
@ARTICLE{rinaldi2019f,
title = {Natural language processing of clinical notes on chronic diseases: systematic review},
journal = {{JMIR} Med Inform},
volume = {7},
author = {Sheikhalishahi, S. and Miotto, R. and Dudley, J.T. and Lavelli, A. and Rinaldi, F. and Osmani, V.},
number = {2},
pages = {e12239},
year = {2019},
doi = {10.2196/12239},
url = {http://medinform.jmir.org/2019/2/e12239/}
}
Download
Vishnyakova, D., Rodriguez-Esteban, R., Rinaldi, F. (2019). A new approach and gold standard toward author disambiguation in MEDLINE. J Am Med Inform Assoc 26(10), pp. 1037–1045.
A new approach and gold standard toward author disambiguation in MEDLINE
Authors: Vishnyakova, D. and Rodriguez-Esteban, R. and Rinaldi, F.
Year: 2019
Abstract: Author-centric analyses of fast-growing biomedical reference databases are challenging due to author ambiguity. This problem has been mainly addressed through author disambiguation using supervised machine-learning algorithms. Such algorithms, however, require adequately designed gold standards that reflect the reference database properly. In this study we used MEDLINE to build the first unbiased gold standard in a reference database and improve over the existing state of the art in author disambiguation. Following a new corpus design method, publication pairs randomly picked from MEDLINE were evaluated by both crowdsourcing and expert curators. Because the latter showed higher accuracy than crowdsourcing, expert curators were tasked to create a full corpus. The corpus was then used to explore new features that could improve state-of-the-art author disambiguation algorithms that would not have been discoverable with previously existing gold standards. We created a gold standard based on 1900 publication pairs that shows close similarity to MEDLINE in terms of chronological distribution and information completeness. A machine-learning algorithm that includes new features related to the ethnic origin of authors showed significant improvements over the current state of the art and demonstrates the necessity of realistic gold standards to further develop effective author disambiguation algorithms. An unbiased gold standard can give a more accurate picture of the status of author disambiguation research and help in the discovery of new features for machine learning. The principles and methods shown here can be applied to other reference databases beyond MEDLINE.
Published in J Am Med Inform Assoc 26(10), pp. 1037–1045.
A new approach and gold standard toward author disambiguation in MEDLINE
@ARTICLE{rinaldi2019a,
title = {A new approach and gold standard toward author disambiguation in {MEDLINE}},
journal = {J Am Med Inform Assoc},
volume = {26},
author = {Vishnyakova, D. and Rodriguez-Esteban, R. and Rinaldi, F.},
number = {10},
pages = {1037--1045},
year = {2019},
doi = {10.1093/jamia/ocz028}
}
Download
Zaffalon, M., Miranda, E. (2019). Desirability foundations of robust rational decision making. Synthese.
Desirability foundations of robust rational decision making
Authors: Zaffalon, M. and Miranda, E.
Year: 2019
Abstract: Recent work has formally linked the traditional axiomatisation of incomplete preferences à la Anscombe-Aumann with the theory of desirability developed in the context of imprecise probability, by showing in particular that they are the very same theory. The equivalence has been established under the constraint that the set of possible prizes is finite. In this paper, we relax such a constraint, thus de facto creating one of the most general theories of rationality and decision making available today. We provide the theory with a sound interpretation and with basic notions, and results, for the separation of beliefs and values, and for the case of complete preferences. Moreover, we discuss the role of conglomerability for the presented theory, arguing that it should be a rationality requirement under very broad conditions.
Accepted in Synthese, Springer.
Desirability foundations of robust rational decision making
@ARTICLE{zaffalon2019a,
title = {Desirability foundations of robust rational decision making},
journal = {Synthese},
publisher = {Springer},
author = {Zaffalon, M. and Miranda, E.},
year = {2019},
doi = {10.1007/s11229-018-02010-x}
}
Download top2018
Antonucci, A., Facchini, A. (2018). A credal extension of independent choice logic. In Proceedings of the 12th International Conference on Scalable Uncertainty Management (SUM 2018), pp. 35-49.
A credal extension of independent choice logic
Authors: Antonucci, A. and Facchini, A.
Year: 2018
Abstract: We propose an extension of Poole’s independent choice logic based on a relaxation of the underlying independence assumptions. A credal semantics involving multiple joint probability mass functions over the possible worlds is adopted. This represents a conservative approach to probabilistic logic program- ming achieved by considering all the mass functions consistent with the prob- abilistic facts. This allows to model tasks for which independence among some probabilistic choices cannot be assumed, and a specific dependence model cannot be assessed. Preliminary tests on an object ranking application show that, despite the loose underlying assumptions, informative inferences can be extracted.
Published in Proceedings of the 12th International Conference on Scalable Uncertainty Management (SUM 2018), pp. 35-49.
A credal extension of independent choice logic
@INPROCEEDINGS{antonucci2018c,
title = {A credal extension of independent choice logic},
booktitle = {Proceedings of the 12th International Conference on Scalable Uncertainty Management ({SUM} 2018)},
author = {Antonucci, A. and Facchini, A.},
pages = {35-49},
year = {2018},
doi = {10.1007/978-3-030-00461-3_3},
url = {https://arxiv.org/abs/1806.08298}
}
Download
Antonucci, A., Facchini, A. (2018). Set-valued probabilistic sentential decision diagrams. In Proceedings of the 5th Workshop on Probabilistic Logic Programming.
Set-valued probabilistic sentential decision diagrams
Authors: Antonucci, A. and Facchini, A.
Year: 2018
Abstract: Probabilistic sentential decision diagrams are a class of arithmetic circuits locally annotated by probability mass functions. This al- lows for a compact representation of joint mass functions in the presence of logical constraints. Here we propose a set-valued generalisation of the probabilistic quantification in these models, that allows to replace the sharp specification of the local probabilities with linear constraints over them. This induces a (convex) set of joint probability mass functions, all consistent with the assigned logical constraints. These models should be adopted for a cautious learning of the local parameters if only small amount of data are available. Algorithmic strategies to compute the lower and upper bounds of marginal and conditional queries with respect to these sets of joint mass functions are sketched. The task can be achieved in linear time with respect to the diagram size for marginal queries and, if the diagram is singly connected, the same can be done for conditional queries.
Published in Proceedings of the 5th Workshop on Probabilistic Logic Programming.
Set-valued probabilistic sentential decision diagrams
@INPROCEEDINGS{antonucci2018d,
title = {Set-valued probabilistic sentential decision diagrams},
booktitle = {Proceedings of the 5th Workshop on Probabilistic Logic Programming},
author = {Antonucci, A. and Facchini, A.},
year = {2018},
url = {http://stoics.org.uk/plp/plp2018/}
}
Download
Bemporad, A., Breschi, V., Piga, D., Boyd, S. (2018). Fitting jump models. Automatica 96, pp. 11–21.
Fitting jump models
Authors: Bemporad, A. and Breschi, V. and Piga, D. and Boyd, S.
Year: 2018
Abstract: We describe a new framework for fitting jump models to a sequence of data. The key idea is to alternate between minimizing a loss function to fit multiple model parameters, and minimizing a discrete loss function to determine which set of model parameters is active at each data point. The framework is quite general and encompasses popular classes of models, such as hidden Markov models and piecewise affine models. The shape of the chosen loss functions to minimize determines the shape of the resulting jump model.
Published in Automatica 96, pp. 11–21.
Fitting jump models
@ARTICLE{piga2018e,
title = {Fitting jump models},
journal = {Automatica},
volume = {96},
author = {Bemporad, A. and Breschi, V. and Piga, D. and Boyd, S.},
pages = {11--21},
year = {2018},
doi = {10.1016/j.automatica.2018.06.022}
}
Download
Breschi, V., Bemporad, A., Piga, D., Boyd, S. (2018). Prediction error methods in learning jump ARMAX models. In 2018 IEEE Conference on Decision and Control (cdc), pp. 2247–2252.
Prediction error methods in learning jump ARMAX models
Authors: Breschi, V. and Bemporad, A. and Piga, D. and Boyd, S.
Year: 2018
Published in 2018 IEEE Conference on Decision and Control (cdc), pp. 2247–2252.
Prediction error methods in learning jump ARMAX models
@INPROCEEDINGS{piga2018i,
title = {Prediction error methods in learning jump {ARMAX} models},
booktitle = {2018 {IEEE} Conference on Decision and Control ({c}dc)},
author = {Breschi, V. and Bemporad, A. and Piga, D. and Boyd, S.},
pages = {2247--2252},
year = {2018},
doi = {10.1109/CDC.2018.8619819}
}
Download
Breschi, V., Piga, D., Bemporad, A. (2018). Kalman filtering for energy disaggregation. In Proc. of the 1st IFAC Workshop on Integrated Assessment Modelling for Environmental Systems 51(5), pp. 108–113.
Kalman filtering for energy disaggregation
Authors: Breschi, V. and Piga, D. and Bemporad, A.
Year: 2018
Abstract: Providing the users information on the energy consumed in the household at the appliance level is of major importance for increasing their awareness of their consumption behavior. In this paper, we propose a technique based on Kalman filters to estimate the devices' consumption patterns from aggregate readings, i.e., to solve the so called disaggregation problem. The method is suited for on-line disaggregation and the proposed results show that it is robust against modelling errors and unmodelled appliances.
Published in Proc. of the 1st IFAC Workshop on Integrated Assessment Modelling for Environmental Systems IFAC-PapersOnLine 51(5), pp. 108–113.
Kalman filtering for energy disaggregation
@INPROCEEDINGS{piga2018b,
title = {Kalman filtering for energy disaggregation},
journal = {{IFAC}-{PapersOnLine}},
volume = {51},
booktitle = {Proc. {o}f the 1st {IFAC} Workshop on Integrated Assessment Modelling for Environmental Systems},
author = {Breschi, V. and Piga, D. and Bemporad, A.},
number = {5},
pages = {108--113},
year = {2018},
doi = {10.1016/j.ifacol.2018.06.219}
}
Download
Breschi, V., Piga, D., Bemporad, A. (2018). Jump model learning and filtering for energy end-use disaggregation. In Proc. of the 18th IFAC Symposium on System Identification 51(15), pp. 275–280.
Jump model learning and filtering for energy end-use disaggregation
Authors: Breschi, V. and Piga, D. and Bemporad, A.
Year: 2018
Abstract: Energy disaggregation aims at reconstructing the power consumed by each electric appliance available in a household from the aggregate power readings collected by a single-point smart meter. With the ultimate goal of fully automatizing this procedure, we first estimate a set of jump models, each of them describing the consumption behaviour of each electric appliance. By representing the total power consumed at the household level as the sum of the outputs of the estimated jump models, a filtering algorithm, based on dynamic programming, is then employed to reconstruct, in an iterative way, the power consumption at an individual appliance level.
Published in Proc. of the 18th IFAC Symposium on System Identification 51(15), pp. 275–280.
Jump model learning and filtering for energy end-use disaggregation
@INPROCEEDINGS{piga2018f,
title = {Jump model learning and filtering for energy end-use disaggregation},
volume = {51},
booktitle = {Proc. {o}f the 18th {IFAC} Symposium on System Identification},
author = {Breschi, V. and Piga, D. and Bemporad, A.},
number = {15},
pages = {275--280},
year = {2018},
doi = {10.1016/j.ifacol.2018.09.147}
}
Download
de Campos, C.P., Scanagatta, M., Corani, G., Zaffalon, M. (2018). Entropy-based pruning for learning Bayesian networks using BIC. Artificial Intelligence 260, pp. 42-50.
Entropy-based pruning for learning Bayesian networks using BIC
Authors: de Campos, C.P. and Scanagatta, M. and Corani, G. and Zaffalon, M.
Year: 2018
Abstract: For decomposable score-based structure learning of Bayesian networks, existing approaches first compute a collection of candidate parent sets for each variable and then optimize over this collection by choosing one parent set for each variable without creating directed cycles while maximizing the total score. We target the task of constructing the collection of candidate parent sets when the score of choice is the Bayesian Information Criterion (BIC). We provide new non-trivial results that can be used to prune the search space of candidate parent sets of each node. We analyze how these new results relate to previous ideas in the literature both theoretically and empirically. We show in experiments with UCI data sets that gains can be significant. Since the new pruning rules are easy to implement and have low computational costs, they can be promptly integrated into all state-of-the-art methods for structure learning of Bayesian networks.
Published in Artificial Intelligence 260, pp. 42-50.
Entropy-based pruning for learning Bayesian networks using BIC
@ARTICLE{deCampos2018a,
title = {Entropy-based pruning for learning {B}ayesian networks using {BIC}},
journal = {Artificial Intelligence},
volume = {260},
author = {de Campos, C.P. and Scanagatta, M. and Corani, G. and Zaffalon, M.},
pages = {42-50},
year = {2018},
doi = {10.1016/j.artint.2018.04.002}
}
Download
Kern, H., Corani, G., Huber, D., Vermes, N., Zaffalon, M. (2018). What interplay of factors influences the place of death in cancer patients? an innovative probabilistic approach sheds light on a well-known question. Journal of Pain and Symptom Management 56(6), e25.
What interplay of factors influences the place of death in cancer patients? an innovative probabilistic approach sheds light on a well-known question
Authors: Kern, H. and Corani, G. and Huber, D. and Vermes, N. and Zaffalon, M.
Year: 2018
Abstract: Most terminally ill cancer patients would prefer to die at home although a majority die in institutional settings. Questions about this discrepancy are not yet fully answered. This study applies machine learning techniques to explore the complicated network of factors and the cause-effect relationships affecting the place of death with the ultimate aim of developing policies.
Published in Journal of Pain and Symptom Management Journal of Pain and Symptom Management 56(6), Elsevier, e25.
What interplay of factors influences the place of death in cancer patients? an innovative probabilistic approach sheds light on a well-known question
@ARTICLE{kern2018a,
title = {What interplay of factors influences the place of death in cancer patients? {a}n innovative probabilistic approach sheds light on a well-known question},
journal = {Journal of Pain and Symptom Management},
publisher = {Elsevier},
volume = {56},
booktitle = {Journal of Pain and Symptom Management},
author = {Kern, H. and Corani, G. and Huber, D. and Vermes, N. and Zaffalon, M.},
number = {6},
pages = {e25},
year = {2018},
doi = {10.1016/j.jpainsymman.2018.10.016}
}
Download
Marchetti, S., Antonucci, A. (2018). Reliable uncertain evidence modeling in Bayesian networks by credal networks. In Proceedings of the 31st International Flairs Conference (FLAIRS-31), AAAI Press, pp. 513–518.
Reliable uncertain evidence modeling in Bayesian networks by credal networks
Authors: Marchetti, S. and Antonucci, A.
Year: 2018
Abstract: A reliable modeling of uncertain evidence in Bayesian networks based on a set-valued quantification is proposed. Both soft and virtual evidences are considered. We show that evidence propagation in this setup can be reduced to standard updating in an augmented credal network, equivalent to a set of consistent Bayesian networks. A characterization of the computational complexity for this task is derived together with an efficient exact procedure for a subclass of instances. In the case of multiple uncertain evidences over the same variable, the proposed procedure can provide a set-valued version of the geometric approach to opinion pooling.
Published in Proceedings of the 31st International Flairs Conference (FLAIRS-31), AAAI Press, pp. 513–518.
Reliable uncertain evidence modeling in Bayesian networks by credal networks
@INPROCEEDINGS{antonucci2018a,
title = {Reliable uncertain evidence modeling in {B}ayesian networks by credal networks},
publisher = {AAAI Press},
booktitle = {Proceedings of the 31st International Flairs Conference ({FLAIRS}-31)},
author = {Marchetti, S. and Antonucci, A.},
pages = {513--518},
year = {2018},
url = { https://aaai.org/ocs/index.php/FLAIRS/FLAIRS18/paper/download/17696/16792}
}
Download
Marchetti, S., Antonucci, A. (2018). Imaginary kinematics. In Proceedings of the 34th Conference on Uncertainty in Artificial Intelligence, AUAI Press, pp. 104-113.
Imaginary kinematics
Authors: Marchetti, S. and Antonucci, A.
Year: 2018
Abstract: We introduce a novel class of adjustment rules for a collection of beliefs. This is an extension of Lewis’ imaging to absorb probabilistic evidence in generalized settings. Unlike standard tools for belief revision, our proposal may be used when information is inconsistent with an agent’s belief base. We show that the functionals we introduce are based on the imaginary counterpart of probability kinematics for standard belief revision, and prove that, under certain conditions, all standard postulates for belief revision are satisfied.
Published in Proceedings of the 34th Conference on Uncertainty in Artificial Intelligence, AUAI Press, pp. 104-113.
Imaginary kinematics
@INPROCEEDINGS{antonucci2018b,
title = {Imaginary kinematics},
publisher = {AUAI Press},
booktitle = {Proceedings of the 34th Conference on Uncertainty in Artificial Intelligence},
author = {Marchetti, S. and Antonucci, A.},
pages = {104-113},
year = {2018},
url = {http://auai.org/uai2018/proceedings/papers/42.pdf}
}
Download
Mejari, M., Naik, V.V., Piga, D., Bemporad, A. (2018). Regularized moving-horizon PWA regression for LPV system identification. In Proc. of the 18th IFAC Symposium on System Identification 51(15), pp. 1092–1097.
Regularized moving-horizon PWA regression for LPV system identification
Authors: Mejari, M. and Naik, V.V. and Piga, D. and Bemporad, A.
Year: 2018
Abstract: This paper addresses the identification of Linear Parameter-Varying (LPV) models through regularized moving-horizon PieceWise Affine (PWA) regression. Specifically, the scheduling-variable space is partitioned into polyhedral regions, where each region is assigned to a PWA function describing the local affine dependence of the LPV model coefficients on the scheduling variable. The regression approach consists of two stages. In the first stage, the data samples are processed iteratively, and a Mixed-Integer Quadratic Programming (MIQP) problem is solved to cluster the scheduling variable observations and simultaneously fit the model parameters to the training data, within a relatively short moving-horizon window of the past. At the second stage, the polyhedral partition of the scheduling-variable space is computed by separating the estimated clusters through linear multi-category discrimination.
Published in Proc. of the 18th IFAC Symposium on System Identification 51(15), pp. 1092–1097.
Note: 18th IFAC Symposium on System Identification SYSID 2018
Regularized moving-horizon PWA regression for LPV system identification
@INPROCEEDINGS{piga2018d,
title = {Regularized moving-horizon {PWA} regression for {LPV} system identification},
volume = {51},
booktitle = {Proc. {o}f the 18th {IFAC} Symposium on System Identification},
author = {Mejari, M. and Naik, V.V. and Piga, D. and Bemporad, A.},
number = {15},
pages = {1092--1097},
year = {2018},
doi = {10.1016/j.ifacol.2018.09.048}
}
Download
Mejari, M., Naik, V.V., Piga, D., Bemporad, A. (2018). Energy disaggregation using piecewise affine regression and binary quadratic programming. In 2018 IEEE Conference on Decision and Control (cdc), pp. 3116–3121.
Energy disaggregation using piecewise affine regression and binary quadratic programming
Authors: Mejari, M. and Naik, V.V. and Piga, D. and Bemporad, A.
Year: 2018
Published in 2018 IEEE Conference on Decision and Control (cdc), pp. 3116–3121.
Energy disaggregation using piecewise affine regression and binary quadratic programming
@INPROCEEDINGS{piga2018h,
title = {Energy disaggregation using piecewise affine regression and binary quadratic programming},
booktitle = {2018 {IEEE} Conference on Decision and Control ({c}dc)},
author = {Mejari, M. and Naik, V.V. and Piga, D. and Bemporad, A.},
pages = {3116--3121},
year = {2018},
doi = {10.1109/CDC.2018.8619175}
}
Download
Mejari, M., Piga, D., Bemporad, A. (2018). A bias-correction method for closed-loop identification of linear parameter-varying systems. Automatica 87, pp. 128–141.
A bias-correction method for closed-loop identification of linear parameter-varying systems
Authors: Mejari, M. and Piga, D. and Bemporad, A.
Year: 2018
Abstract: Due to safety constraints and unstable open-loop dynamics, system identification of many real-world processes often requiresgathering data from closed-loop experiments. In this paper, we present a bias-correction scheme for closed-loop identification of Linear Parameter-Varying Input–Output (LPV-IO) models, which aims at correcting the bias caused by the correlation between the input signal exciting the process and output noise. The proposed identification algorithm provides a consistent estimate of the open-loop model parameters when both the output signal and the scheduling variable are corrupted by measurement noise. The effectiveness of the proposed methodology is tested in two simulation case studies.
Published in Automatica 87, pp. 128–141.
A bias-correction method for closed-loop identification of linear parameter-varying systems
@ARTICLE{piga2018c,
title = {A bias-correction method for closed-loop identification of linear parameter-varying systems},
journal = {Automatica},
volume = {87},
author = {Mejari, M. and Piga, D. and Bemporad, A.},
pages = {128--141},
year = {2018},
doi = {10.1016/j.automatica.2017.09.014}
}
Download
Piga, D., Formentin, S., Bemporad, A. (2018). Direct data-driven control of constrained systems. IEEE Transactions on Control Systems Technology 26(4), pp. 1422–1429.
Direct data-driven control of constrained systems
Authors: Piga, D. and Formentin, S. and Bemporad, A.
Year: 2018
Abstract: In model-based control design, one often has to describe the plant by a linear model. Deriving such a model poses issues of parameterization, estimation, and validation of the model before designing the controller. In this paper, a direct data-driven control method is proposed for designing controllers that can handle constraints without deriving a model of the plant and directly from data. A hierarchical control architecture is used, in which an inner linear time-invariant for linear parameter-varying controller is first designed to match a simple and a priori specified closed-loop model. Then, an outer model predictive controller is synthesized to handle input/output constraints and to enhance the performance of the inner loop. The effectiveness of the approach is illustrated by means of a simulation and an experimental example. Practical implementation issues are also discussed.
Published in IEEE Transactions on Control Systems Technology 26(4), pp. 1422–1429.
Direct data-driven control of constrained systems
@ARTICLE{piga2018a,
title = {Direct data-driven control of constrained systems},
journal = {{IEEE} Transactions on Control Systems Technology},
volume = {26},
author = {Piga, D. and Formentin, S. and Bemporad, A.},
number = {4},
pages = {1422--1429},
year = {2018},
doi = {10.1109/TCST.2017.2702118}
}
Download
Scanagatta, M., Corani, G., de Campos, C.P., Zaffalon, M. (2018). Approximate structure learning for large Bayesian networks. Machine Learning 107(8-10), pp. 1209--1227.
Approximate structure learning for large Bayesian networks
Authors: Scanagatta, M. and Corani, G. and de Campos, C.P. and Zaffalon, M.
Year: 2018
Abstract: We present approximate structure learning algorithms for Bayesian networks. We discuss on the two main phases of the task: the preparation of the cache of the scores and structure optimization, both with bounded and unbounded treewidth. We improve on state-of-the-art methods that rely on an ordering-based search by sampling more effectively the space of the orders. This allows for a remarkable improvement in learning Bayesian networks from thousands of variables.
We also present a thorough study of the accuracy and the running time of inference, comparing bounded-treewidth and unbounded-treewidth models.
Published in Machine Learning 107(8-10), Springer, pp. 1209--1227.
Approximate structure learning for large Bayesian networks
@ARTICLE{scanagatta2018b,
title = {Approximate structure learning for large {B}ayesian networks},
journal = {Machine Learning},
publisher = {Springer},
volume = {107},
author = {Scanagatta, M. and Corani, G. and de Campos, C.P. and Zaffalon, M.},
number = {8-10},
pages = {1209--1227},
year = {2018},
doi = {10.1007/s10994-018-5701-9}
}
Download
Scanagatta, M., Corani, G., Zaffalon, M., Yoo, J., Kang, U. (2018). Efficient learning of bounded-treewidth Bayesian networks from complete and incomplete data sets. International Journal of Approximate Reasoning 95, pp. 152–166.
Efficient learning of bounded-treewidth Bayesian networks from complete and incomplete data sets
Authors: Scanagatta, M. and Corani, G. and Zaffalon, M. and Yoo, J. and Kang, U.
Year: 2018
Abstract: Learning a Bayesian networks with bounded treewidth is important for reducing the complexity of the inferences. We present a novel anytime algorithm (k-MAX) method for this task, which scales up to thousands of variables. Through
extensive experiments we show that it consistently yields higher-scoring structures than its competitors on complete data sets. We then consider the problem
of structure learning from incomplete data sets. This can be addressed by structural EM, which however is computationally very demanding. We thus adopt
the novel k-MAX algorithm in the maximization step of structural EM, obtaining an efficient computation of the expected sufficient statistics. We test the
resulting structural EM method on the task of imputing missing data, comparing it against the state-of-the-art approach based on random forests. Our approach achieves the same imputation accuracy of the competitors, but in about
one tenth of the time. Furthermore we show that it has worst-case complexity linear in the input size, and that it is easily parallelizable.
Published in International Journal of Approximate Reasoning 95, pp. 152–166.
Efficient learning of bounded-treewidth Bayesian networks from complete and incomplete data sets
@ARTICLE{scanagatta2018a,
title = {Efficient learning of bounded-treewidth {B}ayesian networks from complete and incomplete data sets},
journal = {International Journal of Approximate Reasoning},
volume = {95},
author = {Scanagatta, M. and Corani, G. and Zaffalon, M. and Yoo, J. and Kang, U.},
pages = {152--166},
year = {2018},
doi = {10.1016/j.ijar.2018.02.004}
}
Download
Selvi, D., Piga, D., Bemporad, A. (2018). Towards direct data-driven model-free design of optimal controllers. In 2018 European Control Conference (ecc), pp. 2836–2841.
Towards direct data-driven model-free design of optimal controllers
Authors: Selvi, D. and Piga, D. and Bemporad, A.
Year: 2018
Abstract: The most critical step in modern direct data-driven control design approaches, such as virtual reference feedback tuning and non-iterative correlation-based tuning, is the choice of an adequate closed-loop reference model. Indeed, the chosen reference model should reflect the desired closed-loop performance but also be reproducible by the underlying unknown process when in closed loop with the synthesized controller. In this paper, we propose a novel approach to compute, directly from data, an “optimal” reference model along with the corresponding controller. The performance index used to define the optimality of the reference model measures the tracking error and the actuator efforts (as it is typical in performance-driven controllers such as linear-quadratic Gaussian control and model predictive control), along with a term penalizing the expected mismatch between the reference model and the actual closed-loop system. The performance index depends on the variables used to parametrize the reference model and the controller, which are optimized through a suitable combination of particle swarm optimization and virtual reference feedback tuning.
Published in 2018 European Control Conference (ecc), pp. 2836–2841.
Towards direct data-driven model-free design of optimal controllers
@INPROCEEDINGS{piga2018g,
title = {Towards direct data-driven model-free design of optimal controllers},
booktitle = {2018 European Control Conference ({e}cc)},
author = {Selvi, D. and Piga, D. and Bemporad, A.},
pages = {2836--2841},
year = {2018},
doi = {10.23919/ECC.2018.8550184}
}
Download top2017
Arnone, E., Azzimonti, L., Nobile, F., Sangalli, L.M. (2017). A time-dependent PDE regularization to model functional data defined over spatio-temporal domains. In Aneiros G., Bongiorno E.G., Cao R., Vieu P. (Ed), Functional Statistics and Related Fields, Springer International Publishing, pp. 41–44.
A time-dependent PDE regularization to model functional data defined over spatio-temporal domains
Authors: Arnone, E. and Azzimonti, L. and Nobile, F. and Sangalli, L.M.
Year: 2017
Abstract: We propose a method for the analysis of functional data defined over spatio-temporal domains when prior knowledge on the phenomenon under study is available. The model is based on regression with Partial Differential Equations (PDE) penalization. The PDE formalizes the information on the phenomenon and models the regularity of the field in space and time.
Published in Aneiros G., Bongiorno E.G., Cao R., Vieu P. (Ed), Functional Statistics and Related Fields, Springer International Publishing, pp. 41–44.
A time-dependent PDE regularization to model functional data defined over spatio-temporal domains
@INBOOK{azzimonti2017b,
title = {A time-dependent {PDE} regularization to model functional data defined over spatio-temporal domains},
editor = {Aneiros G., Bongiorno E.G., Cao R., Vieu P. },
publisher = {Springer International Publishing},
booktitle = {Functional Statistics and Related Fields},
author = {Arnone, E. and Azzimonti, L. and Nobile, F. and Sangalli, L.M.},
pages = {41--44},
year = {2017},
doi = {10.1007/978-3-319-55846-2_6}
}
Download
Azzimonti, L., Corani, G., Zaffalon, M. (2017). Hierarchical Multinomial-Dirichlet model for the estimation of conditional probability tables. In Raghavan, V., Aluru, S., Karypis, G., Miele, L., Wu, X. (Ed), 2017 IEEE 17th International Conference on Data Mining (ICDM), pp. 739--744.
Hierarchical Multinomial-Dirichlet model for the estimation of conditional probability tables
Authors: Azzimonti, L. and Corani, G. and Zaffalon, M.
Year: 2017
Abstract: We present a novel approach for estimating conditional probability tables, based on a joint, rather than independent, estimate of the conditional distributions belonging to the same table. We derive exact analytical expressions for the estimators and we analyse their properties both analytically and via simulation. We then apply this method to the estimation of parameters in a Bayesian network. Given the structure of the network, the proposed approach better estimates the joint distribution and significantly improves the classification performance with respect to traditional approaches.
Published in Raghavan, V., Aluru, S., Karypis, G., Miele, L., Wu, X. (Ed), 2017 IEEE 17th International Conference on Data Mining (ICDM), pp. 739--744.
Hierarchical Multinomial-Dirichlet model for the estimation of conditional probability tables
@INPROCEEDINGS{azzimonti2017c,
title = {Hierarchical {M}ultinomial-{D}irichlet model for the estimation of conditional probability tables},
editor = {Raghavan, V., Aluru, S., Karypis, G., Miele, L., Wu, X.},
booktitle = {2017 {IEEE} 17th International Conference on Data Mining ({ICDM})},
author = {Azzimonti, L. and Corani, G. and Zaffalon, M.},
pages = {739--744},
year = {2017},
doi = {10.1109/ICDM.2017.85}
}
Download
Balleri, A., Farina, A., Benavoli, A. (2017). Coordination of optimal guidance law and adaptive radiated waveform for interception and rendezvous problems. IET Radar, Sonar & Navigation 11(7), pp. 1132 - 139.
Coordination of optimal guidance law and adaptive radiated waveform for interception and rendezvous problems
Authors: Balleri, A. and Farina, A. and Benavoli, A.
Year: 2017
Abstract: We present an algorithm that allows an interceptor aircraft equipped with an airborne radar to meet another air target (the intercepted) by developing a guidance law and automatically adapting and optimising the transmitted waveform on a pulse to pulse basis. The algorithm uses a Kalman filter to predict the relative position and speed of the interceptor with respect to the target. The transmitted waveform is automatically selected based on its ambiguity function and accuracy properties along the approaching path. For each pulse, the interceptor predicts its position and velocity with respect to the target, takes a measurement of range and radial velocity and, with the Kalman filter, refines the relative range and range rate estimates. These are fed into a Linear Quadratic Gaussian (LQG) controller that ensures the interceptor reaches the target automatically and successfully with minimum error and with the minimum guidance energy consumption.
Published in IET Radar, Sonar & Navigation 11(7), Institution of Engineering and Technology, pp. 1132 - 139.
Coordination of optimal guidance law and adaptive radiated waveform for interception and rendezvous problems
@ARTICLE{benavoli2017a,
title = {Coordination of optimal guidance law and adaptive radiated waveform for interception and rendezvous problems},
journal = {{IET} Radar, Sonar & Navigation},
publisher = {Institution of Engineering and Technology},
volume = {11},
author = {Balleri, A. and Farina, A. and Benavoli, A.},
number = {7},
pages = {1132 - 139},
year = {2017},
doi = {10.1049/iet-rsn.2016.0547}
}
Download
Benavoli, A., Corani, G., Demsar, J., Zaffalon, M. (2017). Time for a change: a tutorial for comparing multiple classifiers through Bayesian analysis. Journal of Machine Learning Research 18(77), pp. 1–36.
Time for a change: a tutorial for comparing multiple classifiers through Bayesian analysis
Authors: Benavoli, A. and Corani, G. and Demsar, J. and Zaffalon, M.
Year: 2017
Abstract: The machine learning community adopted the use of null hypothesis significance testing (NHST) in order to ensure the statistical validity of results. Many scientific fields however realized the shortcomings of frequentist reasoning and in the most radical cases even banned its use in publications. We should do the same: just as we have embraced the Bayesian paradigm in the development of new machine learning methods, so we should also use it in the analysis of our own results. We argue for abandonment of NHST by exposing its fallacies and, more importantly, offer better - more sound and useful - alternatives for it.
Published in Journal of Machine Learning Research 18(77), pp. 1–36.
Time for a change: a tutorial for comparing multiple classifiers through Bayesian analysis
@ARTICLE{benavoli2016e,
title = {Time for a change: a tutorial for comparing multiple classifiers through {B}ayesian analysis},
journal = {Journal of Machine Learning Research},
volume = {18},
author = {Benavoli, A. and Corani, G. and Demsar, J. and Zaffalon, M.},
number = {77},
pages = {1--36},
year = {2017},
url = {http://jmlr.org/papers/v18/16-305.html}
}
Download
Benavoli, A., Facchini, A., Vicente-Perez, J., Zaffalon, M. (2017). A polarity theory for sets of desirable gambles. In Proc. ISIPTA '17 Int. Symposium on Imprecise Probability: Theories and Applications, PMLR, pp. 1-12.
A polarity theory for sets of desirable gambles
Authors: Benavoli, A. and Facchini, A. and Vicente-Perez, J. and Zaffalon, M.
Year: 2017
Abstract: Coherent sets of almost desirable gambles and credal sets are known to be equivalent models. That is, there exists a bijection between the two collections of sets preserving the usual operations, e.g. conditioning. Such a correspondence is based on the polarity theory for closed convex cones. Learning from this simple observation, in this paper we introduce a new (lexicographic) polarity theory for general convex cones and then we apply it in order to establish an analogous correspondence between coherent sets of desirable gambles and convex sets of lexicographic probabilities.
Published in Proc. ISIPTA '17 Int. Symposium on Imprecise Probability: Theories and Applications, PMLR, pp. 1-12.
A polarity theory for sets of desirable gambles
@INPROCEEDINGS{Benavoli2017c,
title = {A polarity theory for sets of desirable gambles},
publisher = {PMLR},
booktitle = {Proc. {ISIPTA} '17 Int. Symposium on Imprecise Probability: Theories and Applications},
author = {Benavoli, A. and Facchini, A. and Vicente-Perez, J. and Zaffalon, M.},
pages = {1-12},
year = {2017},
url = {http://proceedings.mlr.press/v62/benavoli17b/benavoli17b.pdf}
}
Download
Benavoli, A., Facchini, A., Piga, D., Zaffalon, M. (2017). SOS for bounded rationality. In Proceedings of Machine Learning Research 62, PMLR, pp. 25–36.
SOS for bounded rationality
Authors: Benavoli, A. and Facchini, A. and Piga, D. and Zaffalon, M.
Year: 2017
Abstract: In the gambling foundation of probability theory, rationality requires that a subject should always (never) find desirable all nonnegative (negative) gambles, because no matter the result of the exper- iment the subject never (always) decreases her money. Evaluating the nonnegativity of a gamble in infinite spaces is a difficult task. In fact, even if we restrict the gambles to be polynomials in R n , the problem of determining nonnegativity is NP-hard. The aim of this paper is to develop a computable theory of desirable gambles. Instead of requiring the subject to accept all nonnegative gambles, we only require her to accept gambles for which she can efficiently determine the nonnegativity (in particular SOS polynomials). We call this new criterion bounded rationality.
Published in Proceedings of Machine Learning Research 62, PMLR, pp. 25–36.
Note: ISIPTA '17 Int. Symposium on Imprecise Probability: Theories and Applications
SOS for bounded rationality
@INPROCEEDINGS{Benavoli2017b,
title = {{SOS} for bounded rationality},
publisher = {PMLR},
volume = {62},
booktitle = {Proceedings of Machine Learning Research},
author = {Benavoli, A. and Facchini, A. and Piga, D. and Zaffalon, M.},
pages = {25--36},
year = {2017},
url = {http://proceedings.mlr.press/v62/benavoli17a/benavoli17a.pdf}
}
Download
Benavoli, A., Facchini, A., Zaffalon, M. (2017). Bayes + Hilbert = Quantum Mechanics. In Proceedings of the 14th Interational Conference on Quantum Physics and Logic (qpl 2017), Nijmegen, the Netherlands, 3-7 July.
Bayes + Hilbert = Quantum Mechanics
Authors: Benavoli, A. and Facchini, A. and Zaffalon, M.
Year: 2017
Published in Proceedings of the 14th Interational Conference on Quantum Physics and Logic (qpl 2017), Nijmegen, the Netherlands, 3-7 July.
Bayes + Hilbert = Quantum Mechanics
@INPROCEEDINGS{Benavoli2017m,
title = {Bayes + {H}ilbert = {Q}uantum {M}echanics},
booktitle = {Proceedings of the 14th Interational Conference on Quantum Physics and Logic ({q}pl 2017), Nijmegen, the Netherlands, 3-7 July},
author = {Benavoli, A. and Facchini, A. and Zaffalon, M.},
year = {2017},
url = {http://qpl.science.ru.nl/papers/QPL_2017_paper_4.pdf}
}
Download
Bucher, D., Mangili, F., Bonesana, C., Jonietz, D., Cellina, F., Raubal, M. (2017). Demo abstract: extracting eco-feedback information from automatic activity tracking to promote energy-efficient individual mobility behavior. In 33(1), pp. 1-2.
Demo abstract: extracting eco-feedback information from automatic activity tracking to promote energy-efficient individual mobility behavior
Authors: Bucher, D. and Mangili, F. and Bonesana, C. and Jonietz, D. and Cellina, F. and Raubal, M.
Year: 2017
Abstract: Nowadays, most people own a smartphone which
is well suited to constantly record the movement of its user.
One use of the gathered mobility data is to provide users
with feedback and suggestions for personal behavior change.
Such eco-feedback on mobility patterns may stimulate users
to adopt more energy-efficient mobility choices. In this paper,
we present a methodology to extract mobility patterns from
users’ trajectories, compute alternative transport options, and
aggregate and present them in an intuitive way. The resulting
eco-feedback helps people understand their mobility choices
and explore sustainable alternatives.
Published in Computer Science - Research and Development 33(1), pp. 1-2.
Demo abstract: extracting eco-feedback information from automatic activity tracking to promote energy-efficient individual mobility behavior
@INPROCEEDINGS{mangili2017c,
title = {Demo abstract: extracting eco-feedback information from automatic activity tracking to promote energy-efficient individual mobility behavior},
journal = {Computer Science - Research and Development},
volume = {33},
author = {Bucher, D. and Mangili, F. and Bonesana, C. and Jonietz, D. and Cellina, F. and Raubal, M.},
number = {1},
pages = {1-2},
year = {2017},
doi = {10.1007/s00450-017-0375-2}
}
Download
Corani, G., Benavoli, A., Demšar, J., Mangili, F., Zaffalon, M. (2017). Statistical comparison of classifiers through Bayesian hierarchical modelling. Machine Learning 106(11), pp. 1817–1837.
Statistical comparison of classifiers through Bayesian hierarchical modelling
Authors: Corani, G. and Benavoli, A. and Demšar, J. and Mangili, F. and Zaffalon, M.
Year: 2017
Abstract: Usually one compares the accuracy of two competing classifiers using null hypothesis significance tests (nhst).
Yet the nhst tests suffer from important shortcomings, which
can be overcome by switching to Bayesian hypothesis testing.
We propose a Bayesian hierarchical model that jointly analyzes the cross-validation results obtained by two classifiers on multiple data sets. The model estimates more accurately the difference between classifiers on the individual data sets than the traditional approach of averaging, independently on each data set, the cross-validation results.
It does so by jointly analyzing the results obtained on all data sets, and applying shrinkage to the estimates.
The model eventually returns the posterior probability of the accuracies of the two classifiers being practically equivalent or significantly different.
Published in Machine Learning 106(11), pp. 1817–1837.
Statistical comparison of classifiers through Bayesian hierarchical modelling
@ARTICLE{corani2017a,
title = {Statistical comparison of classifiers through {B}ayesian hierarchical modelling},
journal = {Machine Learning},
volume = {106},
author = {Corani, G. and Benavoli, A. and Demšar, J. and Mangili, F. and Zaffalon, M.},
number = {11},
pages = {1817--1837},
year = {2017},
doi = {10.1007/s10994-017-5641-9}
}
Download
Cruder, C., Falla, D., Mangili, F., Azzimonti, L., Araújo, L., Williamon, A., Barbero, M. (2017). Profiling the location and extent of musicians' pain using digital pain drawings. PAIN Practice 18(1), 53–66.
Profiling the location and extent of musicians' pain using digital pain drawings
Authors: Cruder, C. and Falla, D. and Mangili, F. and Azzimonti, L. and Araújo, L. and Williamon, A. and Barbero, M.
Year: 2017
Abstract: BACKGROUND AND AIMS:
According to the existing literature, musicians are at risk to experience a range of musculoskeletal painful conditions. Recently, digital technology has been developed to investigate pain location and pain extent. The aim of this study was to describe pain location and pain extent in musicians using a digital method for pain drawing analysis. Additionally, the association between pain drawing (PD) variables and clinical features in musicians with pain were explored.
MATERIALS AND METHODS:
One hundred fifty-eight musicians (90 women and 68 men; age 22.4±3.6 years) were recruited from Swiss and UK conservatoires. Participants were asked to complete a survey including both background musical information and clinical features, the Quick Dash (QD) questionnaire and the digital PDs.
RESULTS:
Of the 158 participants, 126 musicians (79.7%) reported having pain, with higher prevalence in the areas of the neck and shoulders, the lower back and the right arm. The mean of pain extent was 3.1% ±6.5. The mean of QD was larger for musicians showing the presence of pain than for those without pain. Additionally, the results indicated a positive correlation between QD score and pain extent, and there were significant correlations between age and pain intensity, as well as between pain extent and pain intensity.
CONCLUSIONS:
The high prevalence of pain among musicians has been confirmed using a digital PD. In addition, positive correlations between pain extent and upper limb disability has been demonstrated. Our findings highlight the need for effective prevention and treatment strategies for musicians. This article is protected by copyright. All rights reserved.
Published in PAIN Practice 18(1), Wiley, 53–66.
Profiling the location and extent of musicians' pain using digital pain drawings
@ARTICLE{mangili2017a,
title = {Profiling the location and extent of musicians' pain using digital pain drawings},
journal = {{PAIN} Practice},
publisher = {Wiley},
volume = {18},
author = {Cruder, C. and Falla, D. and Mangili, F. and Azzimonti, L. and Ara\'ujo, L. and Williamon, A. and Barbero, M.},
number = {1},
pages = {53--66},
year = {2017},
doi = {10.1111/papr.12581}
}
Download
Gorini, F., Azzimonti, L., Delfanti, G., Scarfò, L., Scielzo, C., Bertilaccio, M.T., Ranghetti, P., Gulino, A., Doglioni, C., Napoli, A.D., Capri, M., Franceschi, C., Calligaris-Cappio, F., Ghia, P., Bellone, M., Dellabona, P., Casorati, G., de Lalla, C. (2017). Invariant NKT cells contribute to Chronic Lymphocytic Leukemia surveillance and prognosis. Blood 129(26), pp. 3440-3451.
Invariant NKT cells contribute to Chronic Lymphocytic Leukemia surveillance and prognosis
Authors: Gorini, F. and Azzimonti, L. and Delfanti, G. and Scarfò, L. and Scielzo, C. and Bertilaccio, M.T. and Ranghetti, P. and Gulino, A. and Doglioni, C. and Napoli, A.D. and Capri, M. and Franceschi, C. and Calligaris-Cappio, F. and Ghia, P. and Bellone, M. and Dellabona, P. and Casorati, G. and de Lalla, C.
Year: 2017
Abstract: Chronic lymphocytic leukemia (CLL) is characterized by the expansion of malignant CD5+ B lymphocytes in blood, bone marrow, and lymphoid organs. CD1d-restricted invariant natural killer T (iNKT) cells are innate-like T lymphocytes strongly implicated in tumor surveillance. We investigated the impact of iNKT cells in the natural history of the disease in the Eμ-Tcl1 (Tcl1) CLL mouse model and 68 CLL patients. We found that Tcl1-CLL cells express CD1d and that iNKT cells critically delay disease onset but become functionally impaired upon disease progression. In patients, disease progression correlates with high CD1d expression on CLL cells and impaired iNKT cells. Conversely, disease stability correlates with negative or low CD1d expression on CLL cells and normal iNKT cells, suggesting indirect leukemia control. iNKT cells indeed hinder CLL survival in vitro by restraining CD1d-expressing nurse-like cells, a relevant proleukemia macrophage population. Multivariable analysis identified iNKT cell frequency as an independent predictor of disease progression. Together, these results support the contribution of iNKT cells to CLL immune surveillance and highlight iNKT cell frequency as a prognostic marker for disease progression.
Published in Blood 129(26), pp. 3440-3451.
Invariant NKT cells contribute to Chronic Lymphocytic Leukemia surveillance and prognosis
@ARTICLE{azzimonti2017a,
title = {Invariant {NKT} cells contribute to {C}hronic {L}ymphocytic {L}eukemia surveillance and prognosis},
journal = {Blood},
volume = {129},
author = {Gorini, F. and Azzimonti, L. and Delfanti, G. and Scarf\`o, L. and Scielzo, C. and Bertilaccio, M.T. and Ranghetti, P. and Gulino, A. and Doglioni, C. and Napoli, A.D. and Capri, M. and Franceschi, C. and Calligaris-Cappio, F. and Ghia, P. and Bellone, M. and Dellabona, P. and Casorati, G. and de Lalla, C.},
number = {26},
pages = {3440-3451},
year = {2017},
doi = {10.1182/blood-2016-11-751065}
}
Download
Mangili, F., Bonesana, C., Antonucci, A. (2017). Reliable knowledge-based adaptive tests by credal networks. In Antonucci, A., Cholvy, L., Papini, O. (Eds), Symbolic and Quantitative Approaches to Reasoning with Uncertainty. ECSQARU 2017, Lecture Notes in Computer Science 10369, Springer, Cham, pp. 282–291.
Reliable knowledge-based adaptive tests by credal networks
Authors: Mangili, F. and Bonesana, C. and Antonucci, A.
Year: 2017
Abstract: An adaptive test is a computer-based testing technique which adjusts the sequence of questions on the basis of the estimated ability level of the test taker. We suggest the use of credal networks, a generalization of Bayesian networks based on sets of probability mass functions, to implement adaptive tests exploiting the knowledge of the test developer instead of training on databases of answers. Compared to Bayesian networks, these models might offer higher expressiveness and hence a more reliable modeling of the qualitative expert knowledge. The counterpart is a less straightforward identification of the information-theoretic measure controlling the question-selection and the test-stopping criteria. We elaborate on these issues and propose a sound and computationally feasible procedure. Validation against a Bayesian-network approach on a benchmark about German language proficiency assessments suggests that credal networks can be reliable in assessing the student level and effective in reducing the number of questions required to do it.
Published in Antonucci, A., Cholvy, L., Papini, O. (Eds), Symbolic and Quantitative Approaches to Reasoning with Uncertainty. ECSQARU 2017, Lecture Notes in Computer Science 10369, Springer, Cham, pp. 282–291.
Reliable knowledge-based adaptive tests by credal networks
@INPROCEEDINGS{mangili2017b,
title = {Reliable knowledge-based adaptive tests by credal networks},
editor = {Antonucci, A. and Cholvy, L. and Papini, O.},
publisher = {Springer, Cham},
series = {Lecture Notes in Computer Science},
volume = {10369},
booktitle = {Symbolic and Quantitative Approaches to Reasoning {w}ith Uncertainty. {ECSQARU} 2017 },
author = {Mangili, F. and Bonesana, C. and Antonucci, A.},
pages = {282--291},
year = {2017},
doi = {10.1007/978-3-319-61581-3_26}
}
Download
Piga, D., Benavoli, A. (2017). A unified framework for deterministic and probabilistic d-stability analysis of uncertain polynomial matrices. IEEE Transactions on Automatic Control PP(99).
A unified framework for deterministic and probabilistic d-stability analysis of uncertain polynomial matrices
Authors: Piga, D. and Benavoli, A.
Year: 2017
Abstract: In control theory, we are often interested in robust D-stability analysis, which aims at verifying if all the eigenvalues of an uncertain matrix lie in a given region D. Although many algorithms have been developed to provide conditions for an uncertain matrix to be robustly D-stable, the problem of computing the probability of an uncertain matrix to be D-stable is still unexplored. The goal of this paper is to fill this gap in two directions. First, the only constraint on the stability region D that we impose is that its complement is a semialgebraic set. This comprises many important cases in robust control theory. Second, the D-stability analysis problem is formulated in a probabilistic framework, by assuming that only few probabilistic information is available on the uncertain parameters, such as support and some moments. We will show how to compute the minimum probability that the matrix is D-stable by using convex relaxations based on the theory of moments.
Published in IEEE Transactions on Automatic Control PP(99).
A unified framework for deterministic and probabilistic d-stability analysis of uncertain polynomial matrices
@ARTICLE{piga2017a,
title = {A unified framework for deterministic and probabilistic d-stability analysis of uncertain polynomial matrices},
journal = {{IEEE} Transactions on Automatic Control},
volume = {PP},
author = {Piga, D. and Benavoli, A.},
number = {99},
year = {2017},
doi = {10.1109/TAC.2017.2699281}
}
Download
Scanagatta, M., Corani, G., Zaffalon, M. (2017). Improved local search in Bayesian networks structure learning. In Antti Hyttinen, Joe Suzuki, Brandon Malone (Eds), Proceedings of The 3rd International Workshop on Advanced Methodologies for Bayesian Networks (AMBN), Proceedings of Machine Learning Research 73, PMLR, pp. 45-56.
Improved local search in Bayesian networks structure learning
Authors: Scanagatta, M. and Corani, G. and Zaffalon, M.
Year: 2017
Abstract: We present a novel approach for score-based structure learning of Bayesian network, which couples an existing ordering-based algorithm for structure optimization with a novel operator for exploring the neighborhood of a given order in the space of the orderings. Our approach achieves state-of-the-art performances in data sets containing thousands of variables.
Published in Antti Hyttinen, Joe Suzuki, Brandon Malone (Eds), Proceedings of The 3rd International Workshop on Advanced Methodologies for Bayesian Networks (AMBN), Proceedings of Machine Learning Research 73, PMLR, pp. 45-56.
Improved local search in Bayesian networks structure learning
@INPROCEEDINGS{scanagatta2017,
title = {Improved local search in {B}ayesian networks structure learning},
editor = {Antti Hyttinen and Joe Suzuki and Brandon Malone},
publisher = {PMLR},
series = {Proceedings of Machine Learning Research},
volume = {73},
booktitle = {Proceedings of The 3rd International Workshop on Advanced Methodologies for Bayesian Networks ({AMBN})},
author = {Scanagatta, M. and Corani, G. and Zaffalon, M.},
pages = {45-56},
year = {2017},
url = {http://proceedings.mlr.press/v73/scanagatta17a.html}
}
Download
Soullard, Y., Antonucci, A., Destercke, S. (2017). Technical gestures recognition by set-valued hidden Markov models with prior knowledge. In Ferraro, M. B., Giordani, P., Vantaggi, B., Gagolewski, M., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Eds), Soft Methods for Data Science, Advances in Intelligent Systems and Computing 456, Springer, pp. 455–462.
Technical gestures recognition by set-valued hidden Markov models with prior knowledge
Authors: Soullard, Y. and Antonucci, A. and Destercke, S.
Year: 2017
Abstract: Hidden Markov models are popular tools for gesture recognition. Once the generative processes of gestures have been identified, an observation sequence is usually classified as the gesture having the highest likelihood, thus ignoring possible prior information. In this paper, we consider two potential improvements of such methods: the inclusion of prior information, and the possibility of considering convex sets of probabilities (in the likelihoods and the prior) to infer imprecise, but more reliable, predictions when information is insufficient. We apply the proposed approach to technical gestures, typically characterized by severe class imbalance. By modelling such imbalances as a prior information, we achieve more accurate results, while the imprecise quantification is shown to produce more reliable estimates.
Published in Ferraro, M. B., Giordani, P., Vantaggi, B., Gagolewski, M., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Eds), Soft Methods for Data Science, Advances in Intelligent Systems and Computing 456, Springer, pp. 455–462.
Note: SMPS 2016: Proceedings of the 8th international conference on Soft Methods in Probability and Statistics
Technical gestures recognition by set-valued hidden Markov models with prior knowledge
@INCOLLECTION{antonucci2016a,
title = {Technical gestures recognition by set-valued hidden {M}arkov models with prior knowledge},
editor = {Ferraro, M. B. and Giordani, P. and Vantaggi, B. and Gagolewski, M. and Gil, M. A. and Grzegorzewski, P. and Hryniewicz, O.},
publisher = {Springer},
series = {Advances in Intelligent Systems and Computing},
volume = {456},
booktitle = {Soft Methods for Data Science},
author = {Soullard, Y. and Antonucci, A. and Destercke, S.},
pages = {455--462},
year = {2017},
doi = {10.1007/978-3-319-42972-4_56}
}
Download
Miranda, E., Zaffalon, M. (2017). Full conglomerability, continuity and marginal extension. In Ferraro, M. B., Giordani, P., Vantaggi, B., Gagolewski, M., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Eds), Soft Methods for Data Science, Advances in Intelligent Systems and Computing 456, Springer, pp. 355-362.
Full conglomerability, continuity and marginal extension
Authors: Miranda, E., Zaffalon, M.
Year: 2017
Abstract: We investigate fully conglomerable coherent lower previsions in the sense of Walley, and some particular cases of interest: envelopes of fully conglomerable linear previsions, envelopes of countably additive linear previsions and fully disintegrable linear previsions. We study the connections with continuity and countable super-additivity, and show that full conglomerability can be characterised in terms of a supremum of marginal extension models.
Accepted in Ferraro, M. B., Giordani, P., Vantaggi, B., Gagolewski, M., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Eds), Soft Methods for Data Science, Advances in Intelligent Systems and Computing 456, Springer, pp. 355-362.
Note: SMPS 2016: Proceedings of the 8th international conference on Soft Methods in Probability and Statistics
Full conglomerability, continuity and marginal extension
@INCOLLECTION{zaffalon2017a,
title = {Full conglomerability, continuity and marginal extension},
editor = {Ferraro, M. B. and Giordani, P. and Vantaggi, B. and Gagolewski, M. and Gil, M. A. and Grzegorzewski, P. and Hryniewicz, O.},
publisher = {Springer},
series = {Advances in Intelligent Systems and Computing},
volume = {456},
booktitle = {Soft Methods for Data Science},
author = {Miranda, E., Zaffalon, M.},
pages = {355-362},
year = {2017},
doi = {10.1007/978-3-319-42972-4}
}
Download
Miranda, E., Zaffalon, M. (2017). Full conglomerability. Journal of Statistical Theory and Practice 11(4), pp. 634--669.
Full conglomerability
Authors: Miranda, E., Zaffalon, M.
Year: 2017
Abstract: We do a thorough mathematical study of the notion of full conglomerability, that is, conglomerability with respect to all the partitions of an infinite possibility space, in the sense considered by Peter Walley (1991). We consider both the cases of precise and imprecise probability (sets of probabilities). We establish relations between conglomerability and countable additivity, continuity, super-additivity and marginal extension. Moreover, we discuss the special case where a model is conglomerable with respect to a subset of all the partitions, and try to sort out the different notions of conglomerability present in the literature. We conclude that countable additivity, which is routinely used to impose full conglomerability in the precise case, appears to be the most well-behaved way to do so in the imprecise case as well by taking envelopes of countably additive probabilities. Moreover, we characterise these envelopes by means of a number of necessary and sufficient conditions.
Published in Journal of Statistical Theory and Practice 11(4), pp. 634--669.
Full conglomerability
@ARTICLE{zaffalon2017b,
title = {Full conglomerability},
journal = {Journal of Statistical Theory and Practice},
volume = {11},
author = {Miranda, E., Zaffalon, M.},
number = {4},
pages = {634--669},
year = {2017},
doi = {10.1080/15598608.2017.1295890}
}
Download
Zaffalon, M., Miranda, E. (2017). Axiomatising incomplete preferences through sets of desirable gambles. Journal of Artificial Intelligence Research 60, pp. 1057–1126.
Axiomatising incomplete preferences through sets of desirable gambles
Authors: Zaffalon, M. and Miranda, E.
Year: 2017
Abstract: We establish the equivalence of two very general theories: the first is the decision-theoretic formalisation of incomplete preferences based on the mixture independence axiom; the second is the theory of coherent sets of desirable gambles (bounded variables) developed in the context of imprecise probability and extended here to vector-valued gambles. Such an equivalence allows us to analyse the theory of incomplete preferences from the point of view of desirability. Among other things, this leads us to uncover an unexpected and clarifying relation: that the notion of `state independence'---the traditional assumption that we can have separate models for beliefs (probabilities) and values (utilities)---coincides with that of `strong independence' in imprecise probability; this connection leads us also to propose much weaker, and arguably more realistic, notions of state independence. Then we simplify the treatment of complete beliefs and values by putting them on a more equal footing. We study the role of the Archimedean condition---which allows us to actually talk of expected utility---, identify some weaknesses and propose alternatives that solve these. More generally speaking, we show that desirability is a valuable alternative foundation to preferences for decision theory that streamlines and unifies a number of concepts while preserving great generality. In addition, the mentioned equivalence shows for the first time how to extend the theory of desirability to imprecise non-linear utility, thus enabling us to formulate one of the most powerful self-consistent theories of reasoning and decision-making available today.
Published in Journal of Artificial Intelligence Research 60, pp. 1057–1126.
Axiomatising incomplete preferences through sets of desirable gambles
@ARTICLE{zaffalon2017c,
title = {Axiomatising incomplete preferences through sets of desirable gambles},
journal = {Journal of Artificial Intelligence Research},
volume = {60},
author = {Zaffalon, M. and Miranda, E.},
pages = {1057--1126},
year = {2017},
doi = {10.1613/jair.5230}
}
Download top2016
Antonucci, A., Corani, G. (2016). The multilabel naive credal classifier. International Journal of Approximate Reasoning 83, pp. 320-336.
The multilabel naive credal classifier
Authors: Antonucci, A. and Corani, G.
Year: 2016
Abstract: A credal classifier for multilabel data is presented. This is obtained as an extension of Zaffalon's naive credal classifier to the case of non-exclusive class labels. The dependence relations among the labels are shaped with a tree topology. The classifier, based on a polynomial-time algorithm to compute whether or not a class label is optimal, returns a compact description of the set of optimal sequences of labels. Extensive experiments on real multilabel data show that the classifier gives more robust predictions than its Bayesian counterpart. In practice, when multiple sequences are returned in output, the Bayesian model is more likely to be inaccurate, while the sequences returned by the credal classifier are more likely to include the correct one.
Published in International Journal of Approximate Reasoning 83, pp. 320-336.
The multilabel naive credal classifier
@ARTICLE{antonucci2016c,
title = {The multilabel naive credal classifier},
journal = {International Journal of Approximate Reasoning},
volume = {83},
author = {Antonucci, A. and Corani, G.},
pages = {320-336},
year = {2016},
doi = {10.1016/j.ijar.2016.10.006}
}
Download
Benavoli, A., de Campos, C.P. (2016). Bayesian dependence tests for continuous, binary and mixed continuous-binary variables. Entropy 18(9), pp. 1-24.
Bayesian dependence tests for continuous, binary and mixed continuous-binary variables
Authors: Benavoli, A. and de Campos, C.P.
Year: 2016
Abstract: Tests for dependence of continuous, discrete and mixed continuous-discrete variables are ubiquitous in science. The goal of this paper is to derive Bayesian alternatives to frequentist null hypothesis significance tests for dependence. In particular, we will present three Bayesian tests for dependence of binary, continuous and mixed variables. These tests are nonparametric and based on the Dirichlet Process, which allows us to use the same prior model for all of them. Therefore, the tests are “consistent” among each other, in the sense that the probabilities that variables are dependent computed with these tests are commensurable across the different types of variables being tested. By means of simulations with artificial data, we show the effectiveness of the new tests.
Published in Entropy 18(9), Multidisciplinary Digital Publishing Institute, pp. 1-24.
Bayesian dependence tests for continuous, binary and mixed continuous-binary variables
@ARTICLE{benavoli2016g,
title = {Bayesian dependence tests for continuous, binary and mixed continuous-binary variables},
journal = {Entropy},
publisher = {Multidisciplinary Digital Publishing Institute},
volume = {18},
author = {Benavoli, A. and de Campos, C.P.},
number = {9},
pages = {1-24},
year = {2016},
doi = {10.3390/e18090326},
url = {http://www.mdpi.com/1099-4300/18/9/326}
}
Download
Benavoli, A., Corani, G., Mangili, F. (2016). Should we really use post-hoc tests based on mean-ranks?. Journal of Machine Learning Research 17(5), pp. 1–10.
Should we really use post-hoc tests based on mean-ranks?
Authors: Benavoli, A. and Corani, G. and Mangili, F.
Year: 2016
Abstract: The statistical comparison of multiple algorithms over multiple data sets is fundamental in machine learning. This is typically carried out by the Friedman test. When the Friedman test rejects the null hypothesis, multiple comparisons are carried out to establish which are the significant differences among algorithms. The multiple comparisons are usually performed using the mean-ranks test. The aim of this technical note is to discuss the inconsistencies of the mean-ranks post-hoc test with the goal of discouraging its use in machine learning as well as in medicine, psychology, etc.. We show that the outcome of the mean-ranks test depends on the pool of algorithms originally included in the experiment. In other words, the outcome of the comparison between algorithms A and B depends also on the performance of the other algorithms included in the original experiment. This can lead to paradoxical situations. For instance the difference between A and B could be declared significant if the pool comprises algorithms C, D, E and not significant if the pool comprises algorithms F, G, H. To overcome these issues, we suggest instead to perform the multiple comparison using a test whose outcome only depends on the two algorithms being compared, such as the sign-test or the Wilcoxon signed-rank test.
Published in Journal of Machine Learning Research 17(5), pp. 1–10.
Should we really use post-hoc tests based on mean-ranks?
@ARTICLE{benavoli2015c,
title = {Should we really use post-hoc tests based on mean-ranks?},
journal = {Journal of Machine Learning Research},
volume = {17},
author = {Benavoli, A. and Corani, G. and Mangili, F.},
number = {5},
pages = {1--10},
year = {2016},
url = {http://jmlr.org/papers/volume17/benavoli16a/benavoli16a.pdf}
}
Download
Benavoli, A., Facchini, A., Zaffalon, M. (2016). Quantum mechanics: the Bayesian theory generalized to the space of hermitian matrices. Phys. Rev. A 94, 042106.
Quantum mechanics: the Bayesian theory generalized to the space of hermitian matrices
Authors: Benavoli, A. and Facchini, A. and Zaffalon, M.
Year: 2016
Abstract: We consider the problem of gambling on a quantum experiment and enforce rational behavior by a few rules. These rules yield, in the classical case, the Bayesian theory of probability via duality theorems. In our quantum setting, they yield the Bayesian theory generalized to the space of Hermitian matrices. This very theory is quantum mechanics: in fact, we derive all its four postulates from the generalized Bayesian theory. This implies that quantum mechanics is self-consistent. It also leads us to reinterpret the main operations in quantum mechanics as probability rules: Bayes' rule (measurement), marginalization (partial tracing), independence (tensor product). To say it with a slogan, we obtain that quantum mechanics is the Bayesian theory in the complex numbers.
Published in Phys. Rev. A 94, American Physical Society, 042106.
Quantum mechanics: the Bayesian theory generalized to the space of hermitian matrices
@ARTICLE{benavoli2016d,
title = {Quantum mechanics: the {B}ayesian theory generalized to the space of hermitian matrices},
journal = {Phys. Rev. A},
publisher = {American Physical Society},
volume = {94},
author = {Benavoli, A. and Facchini, A. and Zaffalon, M.},
pages = {042106},
year = {2016},
doi = {10.1103/PhysRevA.94.042106},
url = {http://arxiv.org/abs/1605.08177}
}
Download
Benavoli, A., Facchini, A., Zaffalon, M. (2016). Quantum rational preferences and desirability. In Proceedings of the 1st International Workshop on “imperfect Decision Makers: Admitting Real-world Rationality”, Nips 2016.
Quantum rational preferences and desirability
Authors: Benavoli, A. and Facchini, A. and Zaffalon, M.
Year: 2016
Abstract: We develop a theory of quantum rational decision making in the tradition of Anscombe and Aumann's axiomatisation of preferences on horse lotteries. It is essentially the Bayesian decision theory generalised to the space of Hermitian matrices. Among other things, this leads us to give a representation theorem showing that quantum complete rational preferences are obtained by means of expected utility considerations.
Published in Proceedings of the 1st International Workshop on “imperfect Decision Makers: Admitting Real-world Rationality”, Nips 2016 ArXiv e-prints 1610.06764.
Quantum rational preferences and desirability
@INPROCEEDINGS{benavoli2016h,
title = {Quantum rational preferences and desirability},
journal = {{ArXiv} {e}-{p}rints 1610.06764},
booktitle = {Proceedings of the 1st International Workshop on ``imperfect Decision Makers: Admitting Real-{w}orld Rationality'', Nips 2016},
author = {Benavoli, A. and Facchini, A. and Zaffalon, M.},
year = {2016},
url = {http://arxiv.org/abs/1610.06764}
}
Download
Benavoli, A., Piga, D. (2016). A probabilistic interpretation of set-membership filtering: application to polynomial systems through polytopic bounding. Automatica 70, pp. 158–172.
A probabilistic interpretation of set-membership filtering: application to polynomial systems through polytopic bounding
Authors: Benavoli, A. and Piga, D.
Year: 2016
Abstract: Set-membership estimation is usually formulated in the context of set-valued calculus and no probabilistic calculations are necessary. In this paper, we show that set-membership estimation can be equivalently formulated in the probabilistic setting by employing sets of probability measures. Inference in set-membership estimation is thus carried out by computing expectations with respect to the updated set of probability measures P as in the probabilistic case. In particular, it is shown that inference can be performed by solving a particular semi-infinite linear programming problem, which is a special case of the truncated moment problem in which only the zero-th order moment is known (i.e., the support). By writing the dual of the above semi- infinite linear programming problem, it is shown that, if the nonlinearities in the measurement and process equations are polynomial and if the bounding sets for initial state, process and measurement noises are described by polynomial inequalities, then an approximation of this semi-infinite linear programming problem can efficiently be obtained by using the theory of sum- of-squares polynomial optimization. We then derive a smart greedy procedure to compute a polytopic outer-approximation of the true membership-set, by computing the minimum-volume polytope that outer-bounds the set that includes all the means computed with respect to P.
Published in Automatica 70, pp. 158–172.
A probabilistic interpretation of set-membership filtering: application to polynomial systems through polytopic bounding
@ARTICLE{benavoli2016a,
title = {A probabilistic interpretation of set-membership filtering: application to polynomial systems through polytopic bounding},
journal = {Automatica},
volume = {70},
author = {Benavoli, A. and Piga, D.},
pages = {158--172},
year = {2016},
doi = {10.1016/j.automatica.2016.03.021}
}
Download
Bucher, D., Cellina, F., Mangili, F., Raubal, M., Rudel, R., Rizzoli, A.E., Elabed, O. (2016). Exploiting fitness apps for sustainable mobility - challenges deploying the GoEco! App. In Proceedings of the 2016 conference ICT for Sustainability, Advances in Computer Science Research, pp. 89 - 98.
Exploiting fitness apps for sustainable mobility - challenges deploying the GoEco! App
Authors: Bucher, D. and Cellina, F. and Mangili, F. and Raubal, M. and Rudel, R. and Rizzoli, A.E. and Elabed, O.
Year: 2016
Abstract: The large interest in analyzing one’s own fitness led
to the development of more and more powerful smartphone applications. Most are capable of tracking a user’s position and mode of locomotion, data that do not only reflect personal health, but also mobility choices. A large field of research is concerned with mobility analysis and planning for a variety of reasons, including sustainable transport. Collecting data on mobility behavior using fitness tracker apps is a tempting choice, because they include many of the desired functions, most people own a smartphone and installing a fitness tracker is quick and convenient. However, as their original focus is on measuring fitness behavior, there are a number of difficulties in their usage for mobility tracking. In this paper we denote the various challenges we faced when deploying GoEco! Tracker (an app using the Moves R fitness tracker to collect mobility measurements), and provide an analysis on how to best overcome them. Finally, we summarize findings after one
month of large scale testing with a few hundred users within the GoEco! living lab performed in Switzerland.
Published in Proceedings of the 2016 conference ICT for Sustainability, Advances in Computer Science Research, pp. 89 - 98.
Exploiting fitness apps for sustainable mobility - challenges deploying the GoEco! App
@INPROCEEDINGS{mangili2016c,
title = {Exploiting fitness apps for sustainable mobility - challenges deploying the {GoEco}! App},
series = {Advances in Computer Science Research},
booktitle = {Proceedings of the 2016 {c}onference {ICT} for Sustainability},
author = {Bucher, D. and Cellina, F. and Mangili, F. and Raubal, M. and Rudel, R. and Rizzoli, A.E. and Elabed, O.},
pages = {89 - 98},
year = {2016},
doi = {doi:10.2991/ict4s-16.2016.11}
}
Download
Cabañas, R., Antonucci, A., Cano, A., Gómez-Olmedo, M. (2016). Evaluating interval-valued influence diagrams. International Journal of Approximate Reasoning 80, pp. 393-411.
Evaluating interval-valued influence diagrams
Authors: Cabañas, R. and Antonucci, A. and Cano, A. and Gómez-Olmedo, M.
Year: 2016
Abstract: Influence diagrams are probabilistic graphical models used to represent and solve sequential decision problems under uncertainty. Sharp numerical values are required to quantify probabilities and utilities. This might be an issue with real models, whose parameters are typically obtained from expert judgements or partially reliable data. We consider an interval-valued quantification of the parameters to gain realism in the modeling and evaluate the sensitivity of the inferences with respect to perturbations in the sharp values of the parameters. An extension of the classical influence diagrams formalism to support such interval-valued potentials is presented. The variable elimination and arc reversal inference algorithms are generalized to cope with these models. At the price of an outer approximation, the extension keeps the same complexity as with sharp values. Numerical experiments show improved performances with respect to previous methods. As a natural application, we propose these models for practical sensitivity analysis in traditional influence diagrams. The maximum perturbation level on single or multiple parameters preserving the optimal strategy can be computed. This allows the identification of the parameters deserving a more careful elicitation.
Published in International Journal of Approximate Reasoning 80, pp. 393-411.
Evaluating interval-valued influence diagrams
@ARTICLE{antonucci2016b,
title = {Evaluating interval-valued influence diagrams},
journal = {International Journal of Approximate Reasoning},
volume = {80},
author = {Caba\~nas, R. and Antonucci, A. and Cano, A. and G\'omez-Olmedo, M.},
pages = {393-411},
year = {2016},
doi = {10.1016/j.ijar.2016.05.004}
}
Download
de Campos, C.P., Benavoli, A. (2016). Joint analysis of multiple algorithms and performance measures. New Generation Computing, pp. 1–18.
Joint analysis of multiple algorithms and performance measures
Authors: de Campos, C.P. and Benavoli, A.
Year: 2016
Abstract: There has been an increasing interest in the development of new methods using Pareto optimality to deal with multi-objective criteria (for example, accuracy and time complexity). Once one has developed an approach to a problem of interest, the problem is then how to compare it with the state of art. In machine learning, algorithms are typically evaluated by comparing their performance on different data sets by means of statistical tests. Standard tests used for this purpose are able to consider jointly neither performance measures nor multiple competitors at once. The aim of this paper is to resolve these issues by developing statistical procedures that are able to account for multiple competing measures at the same time and to compare multiple algorithms altogether. In particular, we develop two tests: a frequentist procedure based on the generalized likelihood ratio test and a Bayesian procedure based on a multinomial-Dirichlet conjugate model. We further extend them by discovering conditional independences among measures to reduce the number of parameters of such models, as usually the number of studied cases is very reduced in such comparisons. Data from a comparison among general purpose classifiers are used to show a practical application of our tests.
Published in New Generation Computing, pp. 1–18.
Joint analysis of multiple algorithms and performance measures
@ARTICLE{deCampos2016,
title = {Joint analysis of multiple algorithms and performance measures},
journal = {New Generation Computing},
author = {de Campos, C.P. and Benavoli, A.},
pages = {1--18},
year = {2016},
doi = {10.1007/s00354-016-0005-8},
url = {http://people.idsia.ch/~alessio/decampos-benavoli-ngc2016.pdf}
}
Download
de Campos, C.P., Corani, G., Scanagatta, M., Cuccu, M., Zaffalon, M. (2016). Learning extended tree augmented naive structures. International Journal of Approximate Reasoning. 68, pp. 153–163.
Learning extended tree augmented naive structures
Authors: de Campos, C.P. and Corani, G. and Scanagatta, M. and Cuccu, M. and Zaffalon, M.
Year: 2016
Abstract: This work proposes an extended version of the well-known tree-augmented naive Bayes (TAN) classifier where the structure learning step is performed without requiring features to be connected to the class. Based on a modification of Edmonds' algorithm, our structure learning procedure explores a superset of the structures that are considered by TAN, yet achieves global optimality of the learning score function in a very efficient way (quadratic in the number of features, the same complexity as learning TANs). We enhance our procedure with a new score function that only takes into account arcs that are relevant to predict the class, as well as an optimization over the equivalent sample size during learning. These ideas may be useful for structure learning of Bayesian networks in general. A range of experiments show that we obtain models with better prediction accuracy than Naive Bayes and TAN, and comparable to the accuracy of the state-of-the-art classifier averaged one-dependence estimator (AODE). We release our implementation of ETAN so that it can be easily installed and run within Weka.
Accepted in International Journal of Approximate Reasoning. 68, pp. 153–163.
Learning extended tree augmented naive structures
@ARTICLE{decampos2015a,
title = {Learning extended tree augmented naive structures},
journal = {International Journal of Approximate Reasoning.},
volume = {68},
author = {de Campos, C.P. and Corani, G. and Scanagatta, M. and Cuccu, M. and Zaffalon, M.},
pages = {153--163},
year = {2016},
doi = {10.1016/j.ijar.2015.04.006}
}
Download
Corani, G., Scanagatta, M. (2016). Air pollution prediction via multi-label classification. Environmental Modelling & Software 80, pp. 259–264.
Air pollution prediction via multi-label classification
Authors: Corani, G. and Scanagatta, M.
Year: 2016
Abstract: A Bayesian network classifier can be used to estimate the probability of an air pollutant overcoming a certain threshold. Yet multiple predictions are typically required regarding variables which are stochastically dependent, such as ozone measured in multiple stations or assessed according to by different indicators. The common practice (independent approach) is to devise an independent classifier for each class variable being predicted; yet this approach overlooks the dependencies among the class variables. By appropriately modeling such dependencies one can improve the accuracy of the forecasts. We address this problem by designing a multi-label classifier, which simultaneously predict multiple air pollution variables. To this end we design a multi-label classifier based on Bayesian networks and learn its structure through structural learning. We present experiments in three different case studies regarding the prediction of PM2.5 and ozone. The multi-label classifier outperforms the independent approach, allowing to take better decisions.
Published in Environmental Modelling & Software 80, pp. 259–264.
Air pollution prediction via multi-label classification
@ARTICLE{corani2016a,
title = {Air pollution prediction via multi-label classification},
journal = {Environmental Modelling & Software},
volume = {80},
author = {Corani, G. and Scanagatta, M.},
pages = {259--264},
year = {2016},
doi = {10.1016/j.envsoft.2016.02.030}
}
Download
Fu, S. (2016). Hierarchical Bayesian LASSO for a negative binomial regression. Journal of Statistical Computation and Simulation.
Hierarchical Bayesian LASSO for a negative binomial regression
Authors: Fu, S.
Year: 2016
Abstract: Numerous researches have been carried out to explain the relationship between the count data y and numbers of covariates x through a generalized linear model (GLM). This paper proposes a hierarchical Bayesian least absolute shrinkage and selection operator (LASSO) solution using six different prior models to the negative binomial regression. Latent variables Z have been introduced to simplify the GLM to a standard linear regression model. The proposed models regard two conjugate zero-mean Normal priors for the regression parameters and three independent priors for the variance: the Exponential, Inverse-Gamma and Scaled Inverse- distributions. Different types of priors result in different amounts of shrinkage. A Metropolis–Hastings-within-Gibbs algorithm is used to compute the posterior distribution of the parameters of interest through a data augmentation process. Based on the posterior samples, an original double likelihood ratio test statistic have been proposed to choose the most relevant covariates and shrink the insignificant coefficients to zero. Numerical experiments on a real-life data set prove that Bayesian LASSO methods achieved significantly better predictive accuracy and robustness than the classical maximum likelihood estimation and the standard Bayesian inference.
Published in Journal of Statistical Computation and Simulation.
Hierarchical Bayesian LASSO for a negative binomial regression
@ARTICLE{shuaiFu2015a,
title = {Hierarchical {B}ayesian {LASSO} for a negative binomial regression},
journal = {Journal of Statistical Computation and Simulation},
author = {Fu, S.},
year = {2016},
doi = {10.1080/00949655.2015.1106541}
}
Download
Guerciotti, B., Vergara, C., Azzimonti, L., Forzenigo, L., Buora, A., Biondetti, P., Domanin, M. (2016). Computational study of the fluid-dynamics in carotids before and after endarterectomy. Journal of Biomechanics 49(1), pp. 26–38.
Computational study of the fluid-dynamics in carotids before and after endarterectomy
Authors: Guerciotti, B. and Vergara, C. and Azzimonti, L. and Forzenigo, L. and Buora, A. and Biondetti, P. and Domanin, M.
Year: 2016
Abstract: In this work, we provide a computational study of the effects of carotid endarterectomy (CEA) on the fluid-dynamics at internal carotid bifurcations. We perform numerical simulations in real geometries of the same patients before and after CEA, using patient-specific boundary data obtained by Echo-Color Doppler measurements. We analyze four patients with a primary closure and other four where a patch was used to close arteriotomies. The results show that (i) CEA is able to restore physiological fluid-dynamic conditions; (ii) among the post-operative cases, the presence of patch leads to local hemodynamic conditions which might imply a higher risk of restenosis in comparison with the cases without patch.
Published in Elsevier (Ed), Journal of Biomechanics 49(1), pp. 26–38.
Computational study of the fluid-dynamics in carotids before and after endarterectomy
@ARTICLE{azzimonti2016a,
title = {Computational study of the fluid-dynamics in carotids before and after endarterectomy},
journal = {Journal of Biomechanics},
editor = {Elsevier},
volume = {49},
author = {Guerciotti, B. and Vergara, C. and Azzimonti, L. and Forzenigo, L. and Buora, A. and Biondetti, P. and Domanin, M.},
number = {1},
pages = {26--38},
year = {2016},
doi = {10.1016/j.jbiomech.2015.11.009}
}
Download
Mangili, F. (2016). A prior near-ignorance Gaussian process model for nonparametric regression. International Journal of Approximate Reasoning.
A prior near-ignorance Gaussian process model for nonparametric regression
Authors: Mangili, F.
Year: 2016
Abstract: Abstract This paper proposes a prior near-ignorance model for regression based on a set of Gaussian Processes (GP). GPs are natural prior distributions for Bayesian regression. They offer a great modeling flexibility and have found widespread application in many regression problems. However, a GP requires the prior elicitation of its mean function, which represents our prior belief about the shape of the regression function, and of the covariance between any two function values. In the absence of prior information, it may be difficult to fully specify these infinite dimensional parameters. In this work, by modeling the prior mean of the GP as a linear combination of a set of basis functions and assuming as prior for the combination coefficients a set of conjugate distributions obtained as limits of truncate exponential priors, we have been able to model prior ignorance about the mean of the GP. The resulting model satisfies translation invariance, learning and, under some constraints, convergence, which are desirable properties for a prior near-ignorance model. Moreover, it is shown in this paper how this model can be extended to allow for a weaker specification of the GP covariance between function values, by letting each basis function to vary in a set of functions. Application to hypothesis testing has shown how the use of this model induces the capability of automatically detecting when a reliable decision cannot be made based on the available data.
Published in International Journal of Approximate Reasoning.
A prior near-ignorance Gaussian process model for nonparametric regression
@ARTICLE{mangili2016b,
title = {A prior near-ignorance {G}aussian process model for nonparametric regression},
journal = {International Journal of Approximate Reasoning },
author = {Mangili, F.},
year = {2016},
doi = {http://dx.doi.org/10.1016/j.ijar.2016.07.005}
}
Download
Mangili, F., Bonesana, C., Antonucci, A., Zaffalon, M., Rubegni, E., Addimando, L. (2016). Adaptive testing by Bayesian networks with application to language assessment. In Micarelli, Alessandro, Stamper, John, Panourgia, Kitty (Eds), Intelligent Tutoring Systems: 13th International Conference, ITS 2016, Zagreb, Croatia, June 7-10, 2016. Proceedings, Lecture Notes in Computer Science, pp. 471–472.
Adaptive testing by Bayesian networks with application to language assessment
Authors: Mangili, F. and Bonesana, C. and Antonucci, A. and Zaffalon, M. and Rubegni, E. and Addimando, L.
Year: 2016
Abstract: We present a general procedure for computerized adaptive testing based on probabilistic graphical models, and show on a real-world benchmark how this procedure can increase the internal consistency of the test and reduce the number of questions without affecting accuracy.
Published in Micarelli, Alessandro, Stamper, John, Panourgia, Kitty (Eds), Intelligent Tutoring Systems: 13th International Conference, ITS 2016, Zagreb, Croatia, June 7-10, 2016. Proceedings, Lecture Notes in Computer Science, pp. 471–472.
Adaptive testing by Bayesian networks with application to language assessment
@INPROCEEDINGS{mangili2016a,
title = {Adaptive testing by {B}ayesian networks with application to language assessment},
editor = {Micarelli, Alessandro and Stamper, John and Panourgia, Kitty},
series = {Lecture Notes in Computer Science},
booktitle = {Intelligent Tutoring Systems: 13th International Conference, {ITS} 2016, Zagreb, Croatia, June 7-10, 2016. Proceedings},
author = {Mangili, F. and Bonesana, C. and Antonucci, A. and Zaffalon, M. and Rubegni, E. and Addimando, L.},
pages = {471--472},
year = {2016},
url = {http://link.springer.com/content/pdf/bbm%3A978-3-319-39583-8%2F1.pdf}
}
Download
Mauá, D.D., Antonucci, A., de Campos, C.P. (2016). Hidden Markov models with set-valued parameters. Neurocomputing 180, pp. 94–107.
Hidden Markov models with set-valued parameters
Authors: Mauá, D.D. and Antonucci, A. and de Campos, C.P.
Year: 2016
Abstract: Hidden Markov models (HMMs) are widely used probabilistic models of sequential data. As with other probabilistic models, they require the speci- fication of local conditional probability distributions, whose assessment can be too difficult and error-prone, especially when data are scarce or costly to acquire. The imprecise HMM (iHMM) generalizes HMMs by allowing the quantification to be done by sets of, instead of single, probability dis- tributions. iHMMs have the ability to suspend judgment when there is not enough statistical evidence, and can serve as a sensitivity analysis tool for standard non-stationary HMMs. In this paper, we consider iHMMs under the strong independence interpretation, for which we develop efficient infer- ence algorithms to address standard HMM usage such as the computation of likelihoods and most probable explanations, as well as performing filter- ing and predictive inference. Experiments with real data show that iHMMs produce more reliable inferences without compromising the computational efficiency.
Published in Neurocomputing 180, pp. 94–107.
Hidden Markov models with set-valued parameters
@ARTICLE{antonucci2015c,
title = {Hidden {M}arkov models with set-valued parameters},
journal = {Neurocomputing},
volume = {180},
author = {Mau\'a, D.D. and Antonucci, A. and de Campos, C.P.},
pages = {94--107},
year = {2016},
doi = {doi:10.1016/j.neucom.2015.08.095}
}
Download
Miranda, E., Zaffalon, M. (2016). Conformity and independence with coherent lower previsions. International Journal of Approximate Reasoning 78, pp. 125--137.
Conformity and independence with coherent lower previsions
Authors: Miranda, E. and Zaffalon, M.
Year: 2016
Abstract: We define the conformity of marginal and conditional models with a joint model within Walley's theory of coherent lower previsions. Loosely speaking, conformity means that the joint can reproduce the marginal and conditional models we started from. By studying conformity with and without additional assumptions of epistemic irrelevance and independence, we establish connections with a number of prominent models in Walley's theory: the marginal extension, the irrelevant natural extension, the independent natural extension and the strong product.
Published in International Journal of Approximate Reasoning 78, pp. 125--137.
Conformity and independence with coherent lower previsions
@ARTICLE{zaffalon2016c,
title = {Conformity and independence with coherent lower previsions},
journal = {International Journal of Approximate Reasoning},
volume = {78},
author = {Miranda, E. and Zaffalon, M.},
pages = {125--137},
year = {2016},
doi = {10.1016/j.ijar.2016.07.004}
}
Download
Rancoita, P.M.V., Zaffalon, M., Zucca, E., Bertoni, F., de Campos, C.P. (2016). Bayesian network data imputation with application to survival tree analysis. Computational Statistics and Data Analysis 93, pp. 373–387.
Bayesian network data imputation with application to survival tree analysis
Authors: Rancoita, P.M.V. and Zaffalon, M. and Zucca, E. and Bertoni, F. and de Campos, C.P.
Year: 2016
Abstract: Retrospective clinical datasets are often characterized by a relatively small sample size and many missing data. In this case, a common way for handling the missingness consists in discarding from the analysis patients with missing covariates, further reducing the sample size. Alternatively, if the mechanism that generated the missing allows, incomplete data can be imputed on the basis of the observed data, avoiding the reduction of the sample size and allowing methods to deal with complete data later on. Moreover, methodologies for data imputation might depend on the particular purpose and might achieve better results by considering specific characteristics of the domain. The problem of missing data treatment is studied in the context of survival tree analysis for the estimation of a prognostic patient stratification. Survival tree methods usually address this problem by using surrogate splits, that is, splitting rules that use other variables yielding similar results to the original ones. Instead, our methodology consists in modeling the dependencies among the clinical variables with a Bayesian network, which is then used to perform data imputation, thus allowing the survival tree to be applied on the completed dataset. The Bayesian network is directly learned from the incomplete data using a structural expectation-maximization (EM) procedure in which the maximization step is performed with an exact anytime method, so that the only source of approximation is due to the EM formulation itself. On both simulated and real data, our proposed methodology usually outperformed several existing methods for data imputation and the imputation so obtained improved the stratification estimated by the survival tree (especially with respect to using surrogate splits).
Published in Computational Statistics and Data Analysis 93, pp. 373–387.
Bayesian network data imputation with application to survival tree analysis
@ARTICLE{zaffalon2015b,
title = {Bayesian network data imputation with application to survival tree analysis},
journal = {Computational Statistics and Data Analysis},
volume = {93},
author = {Rancoita, P.M.V. and Zaffalon, M. and Zucca, E. and Bertoni, F. and de Campos, C.P.},
pages = {373--387},
year = {2016},
doi = {10.1016/j.csda.2014.12.008}
}
Download
Scanagatta, M., Corani, G., de Campos, C.P., Zaffalon, M. (2016). Learning treewidth-bounded Bayesian networks with thousands of variables. In Daniel D. Lee, Masashi Sugiyama, Ulrike V. Luxburg, Isabelle Guyon, Roman Garnett (Eds), NIPS 2016: Advances in Neural Information Processing Systems 29.
Learning treewidth-bounded Bayesian networks with thousands of variables
Authors: Scanagatta, M. and Corani, G. and de Campos, C.P. and Zaffalon, M.
Year: 2016
Abstract: We present a method for learning treewidth-bounded Bayesian networks from data sets containing thousands of variables. Bounding the treewidth of a Bayesian network greatly reduces the complexity of inferences. Yet, being a global property of the graph, it considerably increases the difficulty of the learning process. Our novel algorithm accomplishes this task, scaling both to large domains and to large treewidths. Our novel approach consistently outperforms the state of the art on experiments with up to thousands of variables.
Published in Daniel D. Lee, Masashi Sugiyama, Ulrike V. Luxburg, Isabelle Guyon, Roman Garnett (Eds), NIPS 2016: Advances in Neural Information Processing Systems 29 NIPS.
Learning treewidth-bounded Bayesian networks with thousands of variables
@INPROCEEDINGS{scanagatta2016a,
title = {Learning treewidth-bounded {B}ayesian networks with thousands of variables},
journal = {{NIPS}},
editor = {Daniel D. Lee and Masashi Sugiyama and Ulrike V. Luxburg and Isabelle Guyon and Roman Garnett},
booktitle = {{NIPS} 2016: Advances in Neural Information Processing Systems 29},
author = {Scanagatta, M. and Corani, G. and de Campos, C.P. and Zaffalon, M.},
year = {2016},
url = {http://papers.nips.cc/paper/6232-learning-treewidth-bounded-bayesian-networks-with-thousands-of-variables}
}
Download top2015
Antonucci, A., Corani, G. (2015). The multilabel naive credal classifier. In Augustin, T., Doria, S., Miranda, E., Quaeghebeur, E. (Eds), ISIPTA '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 27–36.
The multilabel naive credal classifier
Authors: Antonucci, A. and Corani, G.
Year: 2015
Abstract: We present a credal classifier for multilabel data. The model generalizes the naive credal classifier to the multilabel case. An imprecise-probabilistic quantifi- cation is achieved by means of the imprecise Dirichlet model in its global formulation. A polynomial-time algorithm to compute whether or not a label is opti- mal according to the maximality criterion is derived. Experimental results show the importance of robust predictions in multilabel problems.
Published in Augustin, T., Doria, S., Miranda, E., Quaeghebeur, E. (Eds), ISIPTA '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 27–36.
The multilabel naive credal classifier
@INPROCEEDINGS{antonucci2015a,
title = {The multilabel naive credal classifier},
editor = {Augustin, T. and Doria, S. and Miranda, E. and Quaeghebeur, E.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications},
author = {Antonucci, A. and Corani, G.},
pages = {27--36},
year = {2015},
url = {http://www.sipta.org/isipta15/data/paper/32.pdf}
}
Download
Antonucci, A., de Rosa, R., Giusti, A., Cuzzolin, F. (2015). Robust classification of multivariate time series by imprecise hidden Markov models. International Journal of Approximate Reasoning 56(B), pp. 249–263.
Robust classification of multivariate time series by imprecise hidden Markov models
Authors: Antonucci, A. and de Rosa, R. and Giusti, A. and Cuzzolin, F.
Year: 2015
Abstract: A novel technique to classify time series with imprecise hidden Markov models is presented. The learning of these models is achieved by coupling the EM algorithm with the imprecise Dirichlet model. In the stationarity limit, each model corresponds to an imprecise mixture of Gaussian densities, this reducing the problem to the classification of static, imprecise-probabilistic, information. Two classifiers, one based on the expected value of the mixture, the other on the Bhattacharyya distance between pairs of mixtures, are developed. The computation of the bounds of these descriptors with respect to the imprecise quantification of the parameters is reduced to, respectively, linear and quadratic optimization tasks, and hence efficiently solved. Classification is performed by extending the k-nearest neighbors approach to interval-valued data. The classifiers are credal, this means that multiple class labels can be returned in the output. Experiments on benchmark datasets for computer vision show that these methods achieve the required robustness whilst outperforming other precise and imprecise methods.
Published in International Journal of Approximate Reasoning 56(B), pp. 249–263.
Robust classification of multivariate time series by imprecise hidden Markov models
@ARTICLE{antonucci2014c,
title = {Robust classification of multivariate time series by imprecise hidden {M}arkov models},
journal = {International Journal of Approximate Reasoning},
volume = {56},
author = {Antonucci, A. and de Rosa, R. and Giusti, A. and Cuzzolin, F.},
number = {B},
pages = {249--263},
year = {2015},
doi = {10.1016/j.ijar.2014.07.005}
}
Download
Antonucci, A., Scanagatta, M., Mauà, D.D., de Campos, C.P. (2015). Early classification of time series by hidden Markov models with set-valued parameters . In Proceedings of the NIPS Time Series Workshop 2015.
Early classification of time series by hidden Markov models with set-valued parameters
Authors: Antonucci, A. and Scanagatta, M. and Mauà, D.D. and de Campos, C.P.
Year: 2015
Abstract: Hidden Markov models are popular tools for the modeling of multivariate time series. A set-valued quantification of the parameters might increase realism in the description of non-stationarity. A recent work shows that the computation of the bounds of the likelihood of a sequence with respect to such imprecise quantifica- tion can be performed in the same polynomial time of a precise computation with sharp values. This is the basis for a credal classifier of time series. For each training sequence we learn a set-valued model and compute the interval likelihood of the test sequence. The returned classes are those of the models associated to the undominated intervals. The approach is particularly accurate on the instances for which single classes are returned. We therefore apply this method to early classi- fication of streaming data. As soon as the credal classifier returns a single output we assign a class label even if the stream is not terminated. Tests on a speech recognition benchmark suggest that the proposed approach might outperform a thresholding of the precise likelihoods with the classical models.
Published in Proceedings of the NIPS Time Series Workshop 2015.
Early classification of time series by hidden Markov models with set-valued parameters
@INPROCEEDINGS{antonucci2015d,
title = {Early classification of time series by hidden {M}arkov models with set-valued parameters },
booktitle = {Proceedings of the {NIPS} Time Series Workshop 2015},
author = {Antonucci, A. and Scanagatta, M. and Mau\`a, D.D. and de Campos, C.P.},
year = {2015},
url = {https://sites.google.com/site/nipsts2015/home}
}
Download
Azzimonti, L., Sangalli, L.M., Secchi, P., Domanin, M., Nobile, F. (2015). Blood flow velocity field estimation via spatial regression with PDE penalization. Journal of the American Statistical Association, Theory and Methods Section 110(511), pp. 1057–1071.
Blood flow velocity field estimation via spatial regression with PDE penalization
Authors: Azzimonti, L. and Sangalli, L.M. and Secchi, P. and Domanin, M. and Nobile, F.
Year: 2015
Abstract: We propose an innovative method for the accurate estimation of surfaces and spatial fields when prior knowledge of the phenomenon under study is available. The prior knowledge included in the model derives from physics, physiology, or mechanics of the problem at hand, and is formalized in terms of a partial differential equation governing the phenomenon behavior, as well as conditions that the phenomenon has to satisfy at the boundary of the problem domain. The proposed models exploit advanced scientific computing techniques and specifically make use of the finite element method. The estimators have a penalized regression form and the usual inferential tools are derived. Both the pointwise and the areal data frameworks are considered. The driving application concerns the estimation of the blood flow velocity field in a section of a carotid artery, using data provided by echo-color Doppler. This applied problem arises within a research project that aims at studying atherosclerosis pathogenesis. Supplementary materials for this article are available online.
Published in Journal of the American Statistical Association, Theory and Methods Section 110(511), pp. 1057–1071.
Blood flow velocity field estimation via spatial regression with PDE penalization
@ARTICLE{azzimonti2015a,
title = {Blood flow velocity field estimation via spatial regression with {PDE} penalization},
journal = {Journal of the American Statistical Association, Theory and Methods Section},
volume = {110},
author = {Azzimonti, L. and Sangalli, L.M. and Secchi, P. and Domanin, M. and Nobile, F.},
number = {511},
pages = {1057--1071},
year = {2015},
doi = {10.1080/01621459.2014.946036}
}
Download
Benavoli, A., de Campos, C.P. (2015). Statistical tests for joint analysis of performance measures. In Suzuki, J., Ueno, M. (Eds), Advanced Methodologies for Bayesian Networks: Second International Workshop, Ambn 2015, Yokohama, Japan, November 16-18, 2015. Proceedings 9505, Springer International Publishing, Cham, pp. 76–92.
Statistical tests for joint analysis of performance measures
Authors: Benavoli, A. and de Campos, C.P.
Year: 2015
Abstract: Recently there has been an increasing interest in the development of new methods using Pareto optimality to deal with multi-objective criteria (for example, accuracy and architectural complexity). Once one has learned a model based on their devised method, the problem is then how to compare it with the state of art. In machine learning, algorithms are typically evaluated by comparing their performance on different data sets by means of statistical tests. Unfortunately, the standard tests used for this purpose are not able to jointly consider performance measures. The aim of this paper is to resolve this issue by developing statistical procedures that are able to account for multiple competing measures at the same time. In particular, we develop two tests: a frequentist procedure based on the generalized likelihood-ratio test and a Bayesian procedure based on a multinomial-Dirichlet conjugate model. We further extend them by discovering conditional independences among measures to reduce the number of parameter of such models, as usually the number of studied cases is very reduced in such comparisons. Real data from a comparison among general purpose classifiers is used to show a practical application of our tests.
Published in Suzuki, J., Ueno, M. (Eds), Advanced Methodologies for Bayesian Networks: Second International Workshop, Ambn 2015, Yokohama, Japan, November 16-18, 2015. Proceedings 9505, Springer International Publishing, Cham, pp. 76–92.
Statistical tests for joint analysis of performance measures
@INBOOK{Benavoli2015e,
title = {Statistical tests for joint analysis of performance measures},
editor = {Suzuki, J. and Ueno, M.},
publisher = {Springer International Publishing},
address = {Cham},
volume = {9505},
booktitle = {Advanced Methodologies for Bayesian Networks: Second International Workshop, Ambn 2015, Yokohama, Japan, November 16-18, 2015. Proceedings},
author = {Benavoli, A. and de Campos, C.P.},
pages = {76--92},
year = {2015},
doi = {10.1007/978-3-319-28379-1_6}
}
Download
Benavoli, A., Corani, G., Mangili, F., Zaffalon, M. (2015). A Bayesian nonparametric procedure for comparing algorithms. In Francis Bach, David Blei (Eds), Proceedings of the 32th International Conference on Machine Learning (ICML 2015), pp. 1264–1272.
A Bayesian nonparametric procedure for comparing algorithms
Authors: Benavoli, A. and Corani, G. and Mangili, F. and Zaffalon, M.
Year: 2015
Abstract: A fundamental task in machine learning is to compare the performance of multiple algorithms. This is usually performed by the frequentist Friedman test followed by multiple comparisons. This implies dealing with the well-known short- comings of null hypothesis significance tests. We propose a Bayesian approach to overcome these problems. We provide three main contributions. First, we propose a nonparametric Bayesian ver- sion of the Friedman test using a Dirichlet pro- cess (DP) based prior. We show that, from a Bayesian perspective, the Friedman test is an in- ference for a multivariate mean based on an ellip- soid inclusion test. Second, we derive a joint pro- cedure for the multiple comparisons which ac- counts for their dependencies and which is based on the posterior probability computed through the DP. The proposed approach allows verifying the null hypothesis, not only rejecting it. Third, as a practical application we show the results in our algorithm for racing, i.e. identifying the best algorithm among a large set of candidates se- quentially assessed. Our approach consistently outperforms its frequentist counterpart.
Published in Francis Bach, David Blei (Eds), Proceedings of the 32th International Conference on Machine Learning (ICML 2015), pp. 1264–1272.
A Bayesian nonparametric procedure for comparing algorithms
@INPROCEEDINGS{benavoli2015d,
title = {A {B}ayesian nonparametric procedure for comparing algorithms},
editor = { Francis Bach and David Blei},
booktitle = {Proceedings of the 32th International Conference on Machine Learning ({ICML} 2015)},
author = {Benavoli, A. and Corani, G. and Mangili, F. and Zaffalon, M.},
pages = {1264--1272},
year = {2015}
}
Download
Benavoli, A., Mangili, F. (2015). Gaussian processes for Bayesian hypothesis tests on regression functions. In Proceedings of the 18th International Conference on Artificial Intelligence (AISTAT 2015), pp. 74–82.
Gaussian processes for Bayesian hypothesis tests on regression functions
Authors: Benavoli, A. and Mangili, F.
Year: 2015
Abstract: Gaussian processes have been used in different application domains such as classification, regression etc. In this paper we show that they can also be employed as a universal tool for developing a large variety of Bayesian statistical hypothesis tests for regression functions. In particular, we will use GPs for testing whether (i) two functions are equal; (ii) a function is monotone (even accounting for seasonality effects); (iii) a function is periodic; (iv) two functions are proportional. By simulation studies, we will show that, beside being more flexible, GP tests are also competitive in terms of performance with state-of-art algorithms.
Published in Proceedings of the 18th International Conference on Artificial Intelligence (AISTAT 2015), pp. 74–82.
Gaussian processes for Bayesian hypothesis tests on regression functions
@INPROCEEDINGS{benavoli2015a,
title = {Gaussian processes for {B}ayesian hypothesis tests on regression functions},
booktitle = {Proceedings of the 18th International Conference on Artificial Intelligence ({AISTAT} 2015)},
author = {Benavoli, A. and Mangili, F.},
pages = {74--82},
year = {2015},
url = {http://proceedings.mlr.press/v38/benavoli15.html}
}
Download
Benavoli, A., Mangili, F., Ruggeri, F., Zaffalon, M. (2015). Imprecise Dirichlet process with application to the hypothesis test on the probability that X≤Y. Journal of Statistical Theory and Practice 9, pp. 658–684.
Imprecise Dirichlet process with application to the hypothesis test on the probability that X≤Y
Authors: Benavoli, A. and Mangili, F. and Ruggeri, F. and Zaffalon, M.
Year: 2015
Abstract: The Dirichlet process (DP) is one of the most popular Bayesian nonparametric models. An open problem with the DP is how to choose its infinite-dimensional parameter (base measure) in case of lack of prior information. In this work we present the Imprecise DP (IDP)—a prior near-ignorance DP-based model that does not require any choice of this probability measure. It consists of a class of DPs obtained by letting the normalized base measure of the DP vary in the set of all probability measures. We discuss the tight connections of this approach with Bayesian robustness and in particular prior near-ignorance modeling via sets of probabilities. We use this model to perform a Bayesian hypothesis test on the probability P(X≤Y). We study the theoretical properties of the IDP test (e.g., asymptotic consistency), and compare it with the frequentist Mann-Whitney-Wilcoxon rank test that is commonly employed as a test on P(X≤Y). In particular we will show that our method is more robust, in the sense that it is able to isolate instances in which the aforementioned test is virtually guessing at random.
Published in Journal of Statistical Theory and Practice 9, pp. 658–684.
Imprecise Dirichlet process with application to the hypothesis test on the probability that X≤Y
@ARTICLE{benavoli2015b,
title = {Imprecise {D}irichlet process with application to the hypothesis test on the probability that {X&le};{Y}},
journal = {Journal of Statistical Theory and Practice},
volume = {9},
author = {Benavoli, A. and Mangili, F. and Ruggeri, F. and Zaffalon, M.},
pages = {658--684},
year = {2015},
doi = {10.1080/15598608.2014.985997}
}
Download
Cabañas, R., Antonucci, A., Cano, A., Gómez-Olmedo, M. (2015). Variable elimination for interval-valued influence diagrams. In Destercke, S., Denoeux, T. (Eds), Proceedings of the 13th European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty (ECSQARU 2015), Lecture Notes in Computer Science 9161, pp. 541–551.
Variable elimination for interval-valued influence diagrams
Authors: Cabañas, R. and Antonucci, A. and Cano, A. and Gómez-Olmedo, M.
Year: 2015
Abstract: Influence diagrams are probabilistic graphical models used to represent and solve decision problems under uncertainty. Sharp nu- merical values are required to quantify probabilities and utilities. Yet, real models are based on data streams provided by partially reliable sen- sors or experts. We propose an interval-valued quantification of these parameters to gain realism in the modelling and to analyse the sensitiv- ity of the inferences with respect to perturbations of the sharp values. An extension of the classical influence diagrams formalism to support interval-valued potentials is provided. Moreover, a variable elimination algorithm especially designed for these models is developed and evalu- ated in terms of complexity and empirical performances.
Published in Destercke, S., Denoeux, T. (Eds), Proceedings of the 13th European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty (ECSQARU 2015), Lecture Notes in Computer Science 9161, pp. 541–551.
Variable elimination for interval-valued influence diagrams
@INCOLLECTION{antonucci2015b,
title = {Variable elimination for interval-valued influence diagrams},
editor = {Destercke, S. and Denoeux, T.},
series = {Lecture Notes in Computer Science},
volume = {9161},
booktitle = {Proceedings of the 13th European Conference on Symbolic and Quantitative Approaches to Reasoning {w}ith Uncertainty ({ECSQARU} 2015)},
author = {Caba\~nas, R. and Antonucci, A. and Cano, A. and G\'omez-Olmedo, M.},
pages = {541--551},
year = {2015},
chapter = {Symbolic and Quantitative Approaches to Reasoning with Uncertainty},
doi = {10.1007/978-3-319-20807-7_49}
}
Download
de Campos, C.P., Antonucci, A. (2015). Imprecision in machine learning and AI. In The IEEE Intelligent Informatics Bulletin 16(1), IEEE Computer Society, pp. 20–23.
Imprecision in machine learning and AI
Authors: de Campos, C.P. and Antonucci, A.
Year: 2015
Published in The IEEE Intelligent Informatics Bulletin 16(1), IEEE Computer Society, pp. 20–23.
Imprecision in machine learning and AI
@INCOLLECTION{antonucci2015e,
title = {Imprecision in machine learning and {AI}},
publisher = {IEEE Computer Society},
volume = {16},
booktitle = {The {IEEE} Intelligent Informatics Bulletin},
author = {de Campos, C.P. and Antonucci, A.},
number = {1},
pages = {20--23},
year = {2015},
url = {http://www.comp.hkbu.edu.hk/~cib/2015/Dec/iib_vol16no1.pdf}
}
Download
Corani, G., Benavoli, A. (2015). A Bayesian approach for comparing cross-validated algorithms on multiple data sets. Machine Learning 100(2), pp. 285–304.
A Bayesian approach for comparing cross-validated algorithms on multiple data sets
Authors: Corani, G. and Benavoli, A.
Year: 2015
Abstract: We present a Bayesian approach for making statistical inference about the accuracy (or any other score) of two competing algorithms which have been assessed via cross-validation on multiple data sets. The approach is constituted by two pieces. The first is a novel correlated Bayesian t-test for the analysis of the cross-validation results on a single data set which accounts for the correlation due to the overlapping training sets. The second piece merges the posterior probabilities computed by the Bayesian correlated t test on the different data sets to make inference on multiple data sets. It does so by adopting a Poisson-binomial model. The inferences on multiple data sets account for the different uncertainty of the cross-validation results on the different data sets. It is the first test able to achieve this goal. It is generally more powerful than the signed-rank test if ten runs of cross-validation are performed, as it is anyway generally recommended.
Accepted in Machine Learning 100(2), pp. 285–304.
A Bayesian approach for comparing cross-validated algorithms on multiple data sets
@ARTICLE{corani2015b,
title = {A {B}ayesian approach for comparing cross-validated algorithms on multiple data sets},
journal = {Machine Learning},
volume = {100},
author = {Corani, G. and Benavoli, A.},
number = {2},
pages = {285--304},
year = {2015},
doi = {10.1007/s10994-015-5486-z}
}
Download
Corani, G., Benavoli, A., Mangili, F., Zaffalon, M. (2015). Bayesian hypothesis testing in machine learning. In Proc. ECML PKDD 2015 (European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases), pp. 199–202.
Bayesian hypothesis testing in machine learning
Authors: Corani, G. and Benavoli, A. and Mangili, F. and Zaffalon, M.
Year: 2015
Abstract: Most hypothesis testing in machine learning is done using the frequentist null-hypothesis significance test, which has severe drawbacks. We review recent Bayesian tests which overcome the drawbacks of the frequentist ones.
Published in Proc. ECML PKDD 2015 (European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases), pp. 199–202.
Bayesian hypothesis testing in machine learning
@INPROCEEDINGS{corani2015c,
title = {Bayesian hypothesis testing in machine learning},
booktitle = {Proc. {ECML} {PKDD} 2015 (European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases)},
author = {Corani, G. and Benavoli, A. and Mangili, F. and Zaffalon, M.},
pages = {199--202},
year = {2015},
doi = {10.1007/978-3-319-23461-8_13}
}
Download
Corani, G., Mignatti, A. (2015). Credal model averaging for classification: representing prior ignorance and expert opinions.. International Journal of Approximate Reasoning 56(B), pp. 264–277.
Credal model averaging for classification: representing prior ignorance and expert opinions.
Authors: Corani, G. and Mignatti, A.
Year: 2015
Abstract: Bayesian model averaging (BMA) is the state of the art approach for overcoming model uncertainty. Yet, especially on small data sets, the results yielded by BMA might be sensitive to the prior over the models. Credal model averaging (CMA) addresses this problem by substituting the single prior over the models by a set of priors (credal set). Such approach solves the problem of how to choose the prior over the models and automates sensitivity analysis. We discuss various CMA algorithms for building an ensemble of logistic regressors characterized by different sets of covariates. We show how CMA can be appropriately tuned to the case in which one is prior-ignorant and to the case in which instead domain knowledge is available. CMA detects prior-dependent instances, namely instances in which a different class is more probable depending on the prior over the models. On such instances CMA suspends the judgment, returning multiple classes. We thoroughly compare different BMA and CMA variants on a real case study, predicting presence of Alpine marmot burrows in an Alpine valley. We find that BMA is almost a random guesser on the instances recognized as prior-dependent by CMA.
Published in International Journal of Approximate Reasoning 56(B), pp. 264–277.
Credal model averaging for classification: representing prior ignorance and expert opinions.
@ARTICLE{corani2014a,
title = {Credal model averaging for classification: representing prior ignorance and expert opinions.},
journal = {International Journal of Approximate Reasoning},
volume = {56},
author = {Corani, G. and Mignatti, A.},
number = {B},
pages = {264--277},
year = {2015},
doi = {10.1016/j.ijar.2014.07.001}
}
Download
Corani, G., Mignatti, A. (2015). Robust Bayesian model averaging for the analysis of presence–absence data. Environmental and Ecological Statistics 22(3), pp. 513–534.
Robust Bayesian model averaging for the analysis of presence–absence data
Authors: Corani, G. and Mignatti, A.
Year: 2015
Abstract: When developing a species distribution model, usually one tests several competing models such as logistic regressions characterized by different sets of covariates. Yet, there is an exponential number of subsets of covariates to choose from. This generates the problem of model uncertainty. Bayesian model averaging (BMA) is a state-of-the-art approach to deal with model uncertainty. BMA weights the inferences of multiple models. However, the results yielded by BMA depend on the prior probability assigned to the models. Credal model averaging (CMA) extends BMA towards robustness. It substitutes the single prior over the models by a set of priors. The CMA inferences (e.g., posterior probability of inclusion of a covariate, coefficient of a covariate, probability of presence) are intervals. The interval shows the sensitivity of the BMA estimate on the prior over the models. CMA detects the prior-dependent instances, namely cases in which the most probable outcome becomes presence or absence depending on the adopted prior over the models. On such prior-dependent instances, BMA behaves almost as a random guesser. The weakness of BMA on the prior-dependent instances is to our knowledge pointed out for the first time in the ecological literature. On the prior-dependent instances CMA avoids random guessing acknowledging undecidability. In this way it stimulates the decision maker to convey further information before taking the decision. We provide thorough experiments on different data sets.
Accepted in Environmental and Ecological Statistics 22(3), pp. 513–534.
Robust Bayesian model averaging for the analysis of presence–absence data
@ARTICLE{corani2015a,
title = {Robust {B}ayesian model averaging for the analysis of presence--absence data},
journal = {Environmental and Ecological Statistics},
volume = {22},
author = {Corani, G. and Mignatti, A.},
number = {3},
pages = {513--534},
year = {2015},
doi = {10.1007/s10651-014-0308-1}
}
Download
Fu, S. (2015). A hierarchical Bayesian approach to negative binomial regression. Methods and Applications of Analysis 22(4), pp. 409–428.
A hierarchical Bayesian approach to negative binomial regression
Authors: Fu, S.
Year: 2015
Abstract: There is a growing interest in establishing the relationship between the count data y and numerous covariates x through a generalized linear model (GLM), such as explaining the road crash counts from the geometry and environmental factors. This paper proposes a hierarchical Bayesian method to deal with the negative binomial GLM. The Negative Binomial distribution is preferred for modeling nonnegative overdispersed data. The Bayesian inference is chosen to account for prior expert knowledge on regression coefficients in a small sample size setting and the hierarchical structure allows to consider the dependence among the subsets. A Metropolis-Hastings-within-Gibbs algorithm is used to compute the posterior distribution of the parameters of interest through a data augmentation process. The Bayesian approach highly over-performs the classical maximum likelihood estimation in terms of goodness of fit, especially when the sample size decreases and the model complexity increases. Their respective performances have been examined in both the simulated and real-life case studies.
Published in Methods and Applications of Analysis 22(4), pp. 409–428.
A hierarchical Bayesian approach to negative binomial regression
@ARTICLE{shuaiFu2015b,
title = {A hierarchical {B}ayesian approach to negative binomial regression},
journal = {Methods and Applications of Analysis},
volume = {22},
author = {Fu, S.},
number = {4},
pages = {409--428},
year = {2015},
doi = {10.4310/MAA.2015.v22.n4.a4}
}
Download
Mangili, F. (2015). A prior near-ignorance Gaussian Process model for nonparametric regression. In Augustin,T., Doria, S., Miranda, E., Quaeghebeur, E. (Eds), ISIPTA '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 187–196.
A prior near-ignorance Gaussian Process model for nonparametric regression
Authors: Mangili, F.
Year: 2015
Abstract: A Gaussian Process (GP) defines a distribution over functions and thus it is a natural prior distribution for learning real-valued functions from a set of noisy data.
GPs offer a great modeling flexibility and have found widespread application in many regression problems.
A GP is fully defined by a mean function that represents our prior belief about the shape of the regression function and a covariance function, relating the function values at different covariates.
In the absence of prior information, one typically assumes a GP with zero mean function.
Therefore, a priori, it is assumed that the regression function is constantly equal to zero.
The aim of this paper is to model a situation of prior near-ignorance about the GP mean function. For this we consider the set of all GPs with fixed covariance function and constrant mean function free to vary from $-\infty$ to $+\infty$.
We apply the model with constant mean function to hypothesis testing; in particular we test the equality of two regression functions and show that the use of a prior near-ignorance model allows the test to automatically detect when a reliable decision cannot be made based on the available data.
Finally, we propose a generalization of this model that allows considering other sets of prior mean functions.
Published in Augustin,T., Doria, S., Miranda, E., Quaeghebeur, E. (Eds), ISIPTA '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 187–196.
Note: 20-24 July, Pescara, Italy
A prior near-ignorance Gaussian Process model for nonparametric regression
@INPROCEEDINGS{mangili2015b,
title = {A prior near-ignorance {G}aussian {P}rocess model for nonparametric regression},
editor = {Augustin,T. and Doria, S. and Miranda, E. and Quaeghebeur, E.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications},
author = {Mangili, F.},
pages = {187--196},
year = {2015},
url = {http://www.sipta.org/isipta15/data/paper/15.pdf}
}
Download
Mangili, F., Benavoli, A. (2015). New prior near-ignorance models on the simplex. International Journal of Approximate Reasoning 56(Part B), pp. 278–306.
New prior near-ignorance models on the simplex
Authors: Mangili, F. and Benavoli, A.
Year: 2015
Abstract: The aim of this paper is to derive new near-ignorance models on the probabil-
ity simplex, which do not directly involve the Dirichlet distribution and, thus,
are alternative to the Imprecise Dirichlet Model (IDM). We focus our investi-
gation on a particular class of distributions on the simplex which is known as
the class of Normalized Infinitely Divisible (NID) distributions; it includes the
Dirichlet distribution as a particular case. For this class it is possible to derive
general formulae for prior and posterior predictive inferences, by exploiting
the Lévy-Khintchine representation theorem. This allows us to generally char-
acterize the near-ignorance properties of the NID class. After deriving these
general properties, we focus our attention on three members of this class. We
will show that one of these near-ignorance models satisfies the representation
invariance principle and, for a given value of the prior strength, always pro-
vides inferences that encompass those of the IDM. The other two models do
not satisfy this principle, but their imprecision depends linearly or almost lin-
early on the number of observed categories; we argue that this is sometimes
a desirable property for a predictive model.
Published in International Journal of Approximate Reasoning 56(Part B), pp. 278–306.
New prior near-ignorance models on the simplex
@ARTICLE{Mangili2014a,
title = {New prior near-ignorance models on the simplex},
journal = {International Journal of Approximate Reasoning},
volume = {56},
author = {Mangili, F. and Benavoli, A.},
number = {Part B},
pages = {278--306},
year = {2015},
doi = {10.1016/j.ijar.2014.08.005}
}
Download
Mangili, F., Benavoli, A., de Campos, C.P., Zaffalon, M. (2015). Reliable survival analysis based on the Dirichlet Process. Biometrical Journal 57(6), pp. 1002–1019.
Reliable survival analysis based on the Dirichlet Process
Authors: Mangili, F. and Benavoli, A. and de Campos, C.P. and Zaffalon, M.
Year: 2015
Abstract: We present a robust Dirichlet process for estimating survival functions from samples with right-censored
data. It adopts a prior near-ignorance approach to avoid almost any assumption about the distribution of the
population lifetimes, as well as the need of eliciting an infinite dimensional parameter (in case of lack of
prior information), as it happens with the usual Dirichlet process prior. We show how such model can be
used to derive robust inferences from right-censored lifetime data. Robustness is due to the identification of
the decisions that are prior-dependent, and can be interpreted as an analysis of sensitivity with respect to the
hypothetical inclusion of fictitious new samples in the data. In particular, we derive a nonparametric estimator
of the survival probability and a hypothesis test about the probability that the lifetime of an individual
from one population is shorter than the lifetime of an individual from another. We evaluate these ideas on
simulated data and on the Australian AIDS survival dataset. The methods are publicly available through an
easy-to-use R package.
Published in Biometrical Journal 57(6), pp. 1002–1019.
Reliable survival analysis based on the Dirichlet Process
@ARTICLE{mangili2015a,
title = {Reliable survival analysis based on the {D}irichlet {P}rocess},
journal = {Biometrical Journal},
volume = {57},
author = {Mangili, F. and Benavoli, A. and de Campos, C.P. and Zaffalon, M.},
number = {6},
pages = {1002--1019},
year = {2015},
doi = {10.1002/bimj.201500062}
}
Download
Miranda, E., Zaffalon, M. (2015). On the problem of computing the conglomerable natural extension. International Journal of Approximate Reasoning 56(A), pp. 1–27.
On the problem of computing the conglomerable natural extension
Authors: Miranda, E. and Zaffalon, M.
Year: 2015
Abstract: Embedding conglomerability as a rationality requirement in
probability was among the aims of Walley's behavioural theory of
coherent lower previsions. However, recent work has shown that this
attempt has only been partly successful. If we focus in particular
on the extension of given assessments to a rational and
conglomerable model (in the least-committal way), we have that the
procedure used in Walley's theory, the natural extension,
provides only an approximation to the model that is actually sought
for: the so-called conglomerable natural extension. In this
paper we consider probabilistic assessments in the form of a
coherent lower prevision P, which is another name for a lower
expectation functional, and make an in-depth mathematical study of
the problem of computing the conglomerable natural extension for
this case: that is, where it is defined as the smallest coherent
lower prevision F ≥ P that is conglomerable, in case it exists. Past work has shown that F can be approximated by an increasing sequence (En)n∈ℕ of coherent lower previsions. We solve an open problem by showing that this sequence can consist of infinitely many distinct elements. Moreover, we give sufficient conditions, of quite broad applicability, to make sure that the point-wise limit of the sequence is F in case P is the lower envelope of finitely many linear previsions. In addition, we study the question of the existence of F and its relationship with the notion of marginal extension.
Published in International Journal of Approximate Reasoning 56(A), pp. 1–27.
On the problem of computing the conglomerable natural extension
@ARTICLE{zaffalon2014b,
title = {On the problem of computing the conglomerable natural extension},
journal = {International Journal of Approximate Reasoning},
volume = {56},
author = {Miranda, E. and Zaffalon, M.},
number = {A},
pages = {1--27},
year = {2015},
doi = {10.1016/j.ijar.2014.09.003}
}
Download
Miranda, E., Zaffalon, M. (2015). Independent products in infinite spaces. Journal of Mathematical Analysis and Applications 425(1), pp. 460–488.
Independent products in infinite spaces
Authors: Miranda, E. and Zaffalon, M.
Year: 2015
Abstract: Probabilistic independence, intended as the mutual irrelevance of given variables, can be solidly founded on a notion of self-consistency of an uncertainty model, in particular when probabilities go imprecise. There is nothing in this approach that prevents it from being adopted in very general setups, and yet it has mostly been detailed for variables taking finitely many values. In this mathematical study, we complement previous research by exploring the extent to which such an approach can be generalised. We focus in particular on the independent products of two variables. We characterise the main notions, including some of factorisation and productivity, in the general case where both spaces can be infinite and show that, however, there are situations---even in the case of precise probability---where no independent product exists. This is not the case as soon as at least one space is finite. We study in depth this case at the frontiers of good-behaviour detailing the relations among the most important notions; we show for instance that being an independent product is equivalent to a
certain productivity condition. Then we step back to the general case: we give conditions for the existence of independent products and study ways to get around its inherent limitations.
Published in Journal of Mathematical Analysis and Applications 425(1), pp. 460–488.
Independent products in infinite spaces
@ARTICLE{zaffalon2015a,
title = {Independent products in infinite spaces},
journal = {Journal of Mathematical Analysis and Applications},
volume = {425},
author = {Miranda, E. and Zaffalon, M.},
number = {1},
pages = {460--488},
year = {2015},
doi = {10.1016/j.jmaa.2014.12.049}
}
Download
Miranda, E., Zaffalon, M. (2015). Conformity and independence with coherent lower previsions. In Augustin,T., Doria, S., Miranda, E., Quaeghebeur, E. (Eds), ISIPTA '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 197–206.
Conformity and independence with coherent lower previsions
Authors: Miranda, E. and Zaffalon, M.
Year: 2015
Abstract: We study the conformity of marginal unconditional and conditional models with a joint model under assumptions of epistemic irrelevance and independence, within Walley's theory of coherent lower previsions. By doing so, we make a link with a number of prominent models within this theory: the marginal extension, the irrelevant natural extension, the independent natural extension and the strong product.
Published in Augustin,T., Doria, S., Miranda, E., Quaeghebeur, E. (Eds), ISIPTA '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 197–206.
Conformity and independence with coherent lower previsions
@INPROCEEDINGS{zaffalon2015c,
title = {Conformity and independence with coherent lower previsions},
editor = {Augustin,T. and Doria, S. and Miranda, E. and Quaeghebeur, E.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '15: Proceedings of the Ninth International Symposium on Imprecise Probability: Theories and Applications},
author = {Miranda, E. and Zaffalon, M.},
pages = {197--206},
year = {2015},
url = {http://www.sipta.org/isipta15/data/paper/16.pdf}
}
Download
Scanagatta, M., de Campos, C.P., Corani, G., Zaffalon, M. (2015). Learning Bayesian networks with thousands of variables. In Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, Roman Garnett (Eds), NIPS 2015: Advances in Neural Information Processing Systems 28.
Learning Bayesian networks with thousands of variables
Authors: Scanagatta, M. and de Campos, C.P. and Corani, G. and Zaffalon, M.
Year: 2015
Abstract: We present a method for learning Bayesian networks from data sets containing
thousands of variables without the need for structure constraints. Our approach
is made of two parts. The first is a novel algorithm that effectively explores the
space of possible parent sets of a node. It guides the exploration towards the
most promising parent sets on the basis of an approximated score function that
is computed in constant time. The second part is an improvement of an existing
ordering-based algorithm for structure optimization. The new algorithm provably
achieves a higher score compared to its original formulation. On very large data
sets containing up to ten thousand nodes, our novel approach consistently outper-
forms the state of the art.
Published in Corinna Cortes, Neil D. Lawrence, Daniel D. Lee, Masashi Sugiyama, Roman Garnett (Eds), NIPS 2015: Advances in Neural Information Processing Systems 28 NIPS.
Learning Bayesian networks with thousands of variables
@INPROCEEDINGS{scanagatta2015a,
title = {Learning {B}ayesian networks with thousands of variables},
journal = {{NIPS}},
editor = {Corinna Cortes and Neil D. Lawrence and Daniel D. Lee and Masashi Sugiyama and Roman Garnett},
booktitle = {{NIPS} 2015: Advances in Neural Information Processing Systems 28},
author = {Scanagatta, M. and de Campos, C.P. and Corani, G. and Zaffalon, M.},
year = {2015},
url = {http://papers.nips.cc/paper/5803-learning-bayesian-networks-with-thousands-of-variables}
}
Download top2014
Antonucci, A., de Campos, C.P., Huber, D., Zaffalon, M. (2014). Approximate credal network updating by linear programming with applications to decision making. International Journal of Approximate Reasoning 58, pp. 25–38.
Approximate credal network updating by linear programming with applications to decision making
Authors: Antonucci, A. and de Campos, C.P. and Huber, D. and Zaffalon, M.
Year: 2014
Abstract: Credal nets are probabilistic graphical models which extend Bayesian nets to cope with sets of distributions. An algorithm for approximate credal network updating is presented. The problem in its general formulation is a multilinear optimization task, which can be linearized by an appropriate rule for fixing all the local models apart from those of a single variable. This simple idea can be iterated and quickly leads to accurate inferences. A transformation is also derived to reduce decision making in credal networks based on the maximality criterion to updating. The decision task is proved to have the same complexity of standard inference, being NPPP-complete for general credal nets and NP-complete for polytrees. Similar results are derived for the E-admissibility criterion. Numerical experiments confirm a good performance of the method.
Published in International Journal of Approximate Reasoning 58, pp. 25–38.
Approximate credal network updating by linear programming with applications to decision making
@ARTICLE{antonucci2014e,
title = {Approximate credal network updating by linear programming with applications to decision making},
journal = {International Journal of Approximate Reasoning},
volume = {58},
author = {Antonucci, A. and de Campos, C.P. and Huber, D. and Zaffalon, M.},
pages = {25--38},
year = {2014},
doi = {10.1016/j.ijar.2014.10.003}
}
Download
Antonucci, A., de Campos, C.P., Zaffalon, M. (2014). Probabilistic graphical models. In Augustin, T., Coolen, F., de Cooman, G., Troffaes, M. (Eds), Introduction to Imprecise Probabilities, Wiley, pp. 207–229.
Probabilistic graphical models
Authors: Antonucci, A. and de Campos, C.P. and Zaffalon, M.
Year: 2014
Abstract: This report presents probabilistic graphical models that are based on imprecise probabilities using a simplified language. In particular, the discussion is focused on credal networks and discrete domains. It describes the building blocks of credal networks, algorithms to perform inference, and discusses on complexity results and related work. The goal is to have an easy-to-follow introduction to the topic.
Published in Augustin, T., Coolen, F., de Cooman, G., Troffaes, M. (Eds), Introduction to Imprecise Probabilities, Wiley, pp. 207–229.
Probabilistic graphical models
@INBOOK{antonucci2014a,
title = {Probabilistic graphical models},
editor = {Augustin, T. and Coolen, F. and de Cooman, G. and Troffaes, M.},
publisher = {Wiley},
booktitle = {Introduction to Imprecise Probabilities},
author = {Antonucci, A. and de Campos, C.P. and Zaffalon, M.},
pages = {207--229},
year = {2014},
chapter = {9},
doi = {10.1002/9781118763117.ch9}
}
Download
Antonucci, A., Karlsson, A., Sundgren, D. (2014). Decision making with hierarchical credal sets. In Laurent, A., Strauss, O., Bouchon-Meunier, B., Yager, R.R. (Eds), Information Processing and Management of Uncertainty in Knowledge-based Systems, Communications in Computer and Information Science 444, Springer, pp. 456–465.
Decision making with hierarchical credal sets
Authors: Antonucci, A. and Karlsson, A. and Sundgren, D.
Year: 2014
Abstract: We elaborate on hierarchical credal sets, which are sets of probability mass functions paired with second-order distributions. A new criterion to make decisions based on these models is proposed. This is achieved by sampling from the set of mass functions and considering the Kullback-Leibler divergence from the weighted center of mass of the set. We evaluate this criterion in a simple classification scenario: the results show performance improvements when compared to a credal classifier where the second-order distribution is not taken into account.
Published in Laurent, A., Strauss, O., Bouchon-Meunier, B., Yager, R.R. (Eds), Information Processing and Management of Uncertainty in Knowledge-based Systems, Communications in Computer and Information Science 444, Springer, pp. 456–465.
Decision making with hierarchical credal sets
@INPROCEEDINGS{antonucci2014b,
title = {Decision making with hierarchical credal sets},
editor = {Laurent, A. and Strauss, O. and Bouchon-Meunier, B. and Yager, R.R.},
publisher = {Springer},
series = {Communications in Computer and Information Science},
volume = {444},
booktitle = {Information Processing and Management of Uncertainty in Knowledge-{b}ased Systems},
author = {Antonucci, A. and Karlsson, A. and Sundgren, D.},
pages = {456--465},
year = {2014},
doi = {10.1007/978-3-319-08852-5_47}
}
Download
Azzimonti, L., Nobile, F., Sangalli, L.M., Secchi, P. (2014). Mixed finite elements for spatial regression with PDE penalization. SIAM/ASA Journal on Uncertainty Quantification 2(1), pp. 305-335.
Mixed finite elements for spatial regression with PDE penalization
Authors: Azzimonti, L. and Nobile, F. and Sangalli, L.M. and Secchi, P.
Year: 2014
Abstract: We study a class of models at the interface between statistics and numerical analysis. Specifically, we consider nonparametric regression models for the estimation of spatial fields from pointwise and noisy observations, which account for problem-specific prior information, described in terms of a partial differential equation governing the phenomenon under study. The prior information is incorporated in the model via a roughness term using a penalized regression framework. We prove the well-posedness of the estimation problem, and we resort to a mixed equal order finite element method for its discretization. Moreover, we prove the well-posedness and the optimal convergence rate of the proposed discretization method. Finally the smoothing technique is extended to the case of areal data, particularly interesting in many applications.
Published in SIAM/ASA Journal on Uncertainty Quantification 2(1), pp. 305-335.
Mixed finite elements for spatial regression with PDE penalization
@ARTICLE{azzimonti2014a,
title = {Mixed finite elements for spatial regression with {PDE} penalization},
journal = {{SIAM/ASA} Journal on Uncertainty Quantification},
volume = {2},
author = {Azzimonti, L. and Nobile, F. and Sangalli, L.M. and Secchi, P.},
number = {1},
pages = {305-335},
year = {2014},
doi = {10.1137/130925426}
}
Download
Benavoli, A. (2014). Belief function and multivalued mapping robustness in statistical estimation. International Journal of Approximate Reasoning 55(1, Part 3), pp. 311–329.
Belief function and multivalued mapping robustness in statistical estimation
Authors: Benavoli, A.
Year: 2014
Abstract: We consider the case in which the available knowledge does not allow to specify a precise probabilistic model for the prior and/or likelihood in statistical estimation. We assume that this imprecision can be represented by belief functions models. Thus, we exploit the mathematical structure of belief functions and their equivalent representation in terms of closed convex sets of probabilities to derive robust posterior inferences using Walley theory of imprecise probabilities. Then, we apply these robust models to practical inference problems and we show the connections of the proposed inference method with interval estimation and statistical inference with missing data.
Published in International Journal of Approximate Reasoning 55(1, Part 3), pp. 311–329.
Belief function and multivalued mapping robustness in statistical estimation
@ARTICLE{benavoli2013a,
title = {Belief function and multivalued mapping robustness in statistical estimation},
journal = {International Journal of Approximate Reasoning},
volume = {55},
author = {Benavoli, A.},
number = {1, Part 3},
pages = {311--329},
year = {2014},
doi = {10.1016/j.ijar.2013.04.014}
}
Download
Benavoli, A., Mangili, F., Corani, G., Zaffalon, M., Ruggeri, F. (2014). A Bayesian Wilcoxon signed-rank test based on the Dirichlet process. In Eric P. Xing, Tony Jebara (Eds), Proceedings of the 31st International Conference on Machine Learning (ICML 2014), pp. 1026–1034.
A Bayesian Wilcoxon signed-rank test based on the Dirichlet process
Authors: Benavoli, A. and Mangili, F. and Corani, G. and Zaffalon, M. and Ruggeri, F.
Year: 2014
Abstract: Bayesian methods are ubiquitous in machine learning.
Nevertheless, the analysis of empirical results is
typically performed by frequentist tests. This implies dealing with
null hypothesis significance tests and p-values, even though the shortcomings of such methods
are well known.
We propose a nonparametric Bayesian version of the Wilcoxon signed-rank test using a Dirichlet process (DP) based prior.
We address in two different ways the problem of how to choose the infinite dimensional parameter that characterizes the DP.
The proposed test has all the traditional strengths of the Bayesian approach; for instance, unlike the frequentist tests, it allows verifying the null hypothesis, not only rejecting it, and taking decisions which minimize the expected loss.
Moreover, one of the solutions proposed to model the infinite-dimensional parameter of the DP
allows isolating instances in which the traditional frequentist test is guessing at random.
We show results dealing with the comparison of two classifiers using real and simulated data.
Published in Eric P. Xing, Tony Jebara (Eds), Proceedings of the 31st International Conference on Machine Learning (ICML 2014), pp. 1026–1034.
A Bayesian Wilcoxon signed-rank test based on the Dirichlet process
@INPROCEEDINGS{benavoli2014a,
title = {A {B}ayesian {W}ilcoxon signed-rank test based on the {D}irichlet process},
editor = { Eric P. Xing and Tony Jebara},
booktitle = {Proceedings of the 31st International Conference on Machine Learning ({ICML} 2014)},
author = {Benavoli, A. and Mangili, F. and Corani, G. and Zaffalon, M. and Ruggeri, F.},
pages = {1026--1034},
year = {2014},
url = {http://proceedings.mlr.press/v32/benavoli14.html}
}
Download
Benavoli, A., Zaffalon, M. (2014). Prior near ignorance for inferences in the k-parameter exponential family. Statistics 49, pp. 1104–1140.
Prior near ignorance for inferences in the k-parameter exponential family
Authors: Benavoli, A. and Zaffalon, M.
Year: 2014
Abstract: This paper proposes a model of prior ignorance about a multivariate variable based on a set of distributions . In particular, we discuss four minimal properties that a model of prior ignorance should satisfy: invariance, near ignorance, learning and convergence. Near ignorance and invariance ensure that our prior model behaves as a vacuous model with respect to some statistical inferences (e.g. mean, credible intervals, etc.) and some transformation of the parameter space. Learning and convergence ensure that our prior model can learn from data and, in particular, that the influence of on the posterior inferences vanishes with increasing numbers of observations. We show that these four properties can all be satisfied by a set of conjugate priors in the multivariate exponential families if the set includes finitely additive probabilities obtained as limits of truncated exponential functions. The obtained set is a model of prior ignorance with respect to the functions (queries) that are commonly used for statistical inferences and, because of conjugacy, it is tractable and easy to elicit. Applications of the model to some practical statistical problems show the effectiveness of the approach.
Published in Statistics 49, pp. 1104–1140.
Prior near ignorance for inferences in the k-parameter exponential family
@ARTICLE{benavoli2014b,
title = {Prior near ignorance for inferences in the k-parameter exponential family},
journal = {Statistics},
volume = {49},
author = {Benavoli, A. and Zaffalon, M.},
pages = {1104--1140},
year = {2014},
doi = {10.1080/02331888.2014.960869}
}
Download
De Bock, J., de Campos, C.P., Antonucci, A. (2014). Global sensitivity analysis for MAP inference in graphical models. In Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D. , Weinberger, K.Q. (Eds), Advances in Neural Information Processing Systems 27 (NIPS 2014), Curran Associates, Inc., pp. 2690–2698.
Global sensitivity analysis for MAP inference in graphical models
Authors: De Bock, J. and de Campos, C.P. and Antonucci, A.
Year: 2014
Abstract: We study the sensitivity of a MAP configuration of a discrete probabilistic graph- ical model with respect to perturbations of its parameters. These perturbations are global, in the sense that simultaneous perturbations of all the parameters (or any chosen subset of them) are allowed. Our main contribution is an exact algorithm that can check whether the MAP configuration is robust with respect to given per- turbations. Its complexity is essentially the same as that of obtaining the MAP configuration itself, so it can be promptly used with minimal effort. We use our algorithm to identify the largest global perturbation that does not induce a change in the MAP configuration, and we successfully apply this robustness measure in two practical scenarios: the prediction of facial action units with posed images and the classification of multiple real public data sets. A strong correlation between the proposed robustness measure and accuracy is verified in both scenarios.
Published in Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D. , Weinberger, K.Q. (Eds), Advances in Neural Information Processing Systems 27 (NIPS 2014), Curran Associates, Inc., pp. 2690–2698.
Global sensitivity analysis for MAP inference in graphical models
@INCOLLECTION{antonucci2014f,
title = {Global sensitivity analysis for {MAP} inference in graphical models},
editor = {Ghahramani, Z. and Welling, M. and Cortes, C. and Lawrence, N.D. and Weinberger, K.Q. },
publisher = {Curran Associates, Inc.},
booktitle = {Advances in Neural Information Processing Systems 27 ({NIPS} 2014)},
author = {De Bock, J. and de Campos, C.P. and Antonucci, A.},
pages = {2690--2698},
year = {2014},
url = {http://papers.nips.cc/paper/5472-global-sensitivity-analysis-for-map-inference-in-graphical-models.pdf}
}
Download
de Campos, C.P., Cuccu, M., Corani, G., Zaffalon, M. (2014). Extended tree augmented naive classifier. In van der Gaag, L., Feelders, A. (Ed), PGM'14: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models, Lecture Notes in Artificial Intelligence 8754, Springer, pp. 176–189.
Extended tree augmented naive classifier
Authors: de Campos, C.P. and Cuccu, M. and Corani, G. and Zaffalon, M.
Year: 2014
Abstract: This work proposes an extended version of the well-known tree-augmented naive Bayes (TAN) classifier where the structure learning step is performed without requiring features to be connected to the class. Based on a modification of Edmonds' algorithm, our structure learning procedure explores a superset of the structures that are considered by TAN, yet achieves global optimality of the learning score function in a very efficient way (quadratic in the number of features, the same complexity as learning TANs). A range of experiments show that we obtain models with better accuracy than TAN and comparable to the accuracy of the state-of-the-art classifier averaged one-dependence estimator.
Published in van der Gaag, L., Feelders, A. (Ed), PGM'14: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models, Lecture Notes in Artificial Intelligence 8754, Springer, pp. 176–189.
Extended tree augmented naive classifier
@INPROCEEDINGS{decampos2014a,
title = {Extended tree augmented naive classifier},
editor = {van der Gaag, L., Feelders, A.},
publisher = {Springer},
series = {Lecture Notes in Artificial Intelligence},
volume = {8754},
booktitle = {{PGM'14}: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models},
author = {de Campos, C.P. and Cuccu, M. and Corani, G. and Zaffalon, M.},
pages = {176--189},
year = {2014},
doi = {10.1007/978-3-319-11433-0_12}
}
Download
Corani, G., Abellán, J., Masegosa, A., Moral, S., Zaffalon, M. (2014). Classification. In Augustin,T., Coolen,F., de Cooman,G., Troffaes,M. (Eds), Introduction to Imprecise Probabilities, Wiley, pp. 261–285.
Classification
Authors: Corani, G. and Abellán, J. and Masegosa, A. and Moral, S. and Zaffalon, M.
Year: 2014
Abstract: This report presents an introduction to credal classification. The discussion focuses on the naive credal classifier and its extensions as well as on credal classifiers based on classification trees. In addition we discuss the metrics suitable for scoring credal classifiers and present some experiments. The goal is to have an easy-to-follow introduction to the topic.
Published in Augustin,T., Coolen,F., de Cooman,G., Troffaes,M. (Eds), Introduction to Imprecise Probabilities, Wiley, pp. 261–285.
Classification
@INCOLLECTION{corani2013b,
title = {Classification},
editor = {Augustin,T. and Coolen,F. and de Cooman,G. and Troffaes,M.},
publisher = {Wiley},
booktitle = {Introduction to Imprecise Probabilities},
author = {Corani, G. and Abell\'an, J. and Masegosa, A. and Moral, S. and Zaffalon, M.},
pages = {261--285},
year = {2014},
chapter = {10},
url = {http://eu.wiley.com/WileyCDA/WileyTitle/productCd-0470973811.html}
}
Download
Corani, G., Antonucci, A. (2014). Credal Ensembles of Classifiers. Computational Statistics & Data Analysis 71, pp. 818–831.
Credal Ensembles of Classifiers
Authors: Corani, G. and Antonucci, A.
Year: 2014
Abstract: It is studied how to aggregate the probabilistic predictions generated by different SPODE (Super-Parent-One-Dependence Estimators) classifiers. It is shown that aggregating such predictions via compression-based weights achieves a slight but consistent improvement of performance over previously existing aggregation methods, including Bayesian Model Averaging and simple average (the approach adopted by the AODE algorithm). Then, attention is given to the problem of choosing the prior probability distribution over the models; this is an important issue in any Bayesian ensemble of models. To robustly deal with the choice of the prior, the single prior over the models is substituted by a set of priors over the models (credal set), thus obtaining a credal
ensemble of Bayesian classifiers. The credal ensemble recognizes the prior-dependent instances, namely the instances whose most probable class varies when different prior over the models are considered. When faced with prior-dependent instances, the credal ensemble remains reliable by returning a set of classes rather than a single class. Two
credal ensembles of SPODEs are developed; the first generalizes the Bayesian Model Averaging and the second the compression-based aggregation. Extensive experiments show that the novel ensembles compare favorably to traditional methods for aggregating SPODEs and also to previous credal classifiers.
Published in Computational Statistics & Data Analysis 71, pp. 818–831.
Credal Ensembles of Classifiers
@ARTICLE{corani2012f,
title = {Credal {E}nsembles of {C}lassifiers},
journal = {Computational Statistics & Data Analysis},
volume = {71},
author = {Corani, G. and Antonucci, A.},
pages = {818--831},
year = {2014},
doi = {10.1016/j.csda.2012.11.010}
}
Download
Corani, G., Antonucci, A., Mauá, D., Gabaglio, S. (2014). Trading off Speed and Accuracy in Multilabel Classification. In van der Gaag, L., Feelders, A. (Eds), PGM'14: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models, Lecture Notes in Artificial Intelligence 8754, Springer, pp. 145–159.
Trading off Speed and Accuracy in Multilabel Classification
Authors: Corani, G. and Antonucci, A. and Mauá, D. and Gabaglio, S.
Year: 2014
Abstract: In previous work, we devised an approach for multilabel clas-
sification based on an ensemble of Bayesian networks. It was characterized by an efficient structural learning and by high accuracy. Its shortcoming was the high computational complexity of the MAP inference, necessary to identify the most probable joint configuration of all classes.
In this work, we switch from the ensemble approach to the single model approach. This allows important computational savings. The reduction of inference times is exponential in the difference between the treewidth of the single model and the number of classes. We adopt moreover a more sophisticated approach for the structural learning of the class subgraph.
The proposed single models outperforms alternative approaches for multilabel classification such as binary relevance and ensemble of classifier chains.
Published in van der Gaag, L., Feelders, A. (Eds), PGM'14: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models Proc. 7th European Workshop on Probabilistic Graphical Models (PGM '14), Lecture Notes in Artificial Intelligence 8754, Springer, pp. 145–159.
Trading off Speed and Accuracy in Multilabel Classification
@INPROCEEDINGS{corani2014b,
title = {Trading off {S}peed and {A}ccuracy in {M}ultilabel {C}lassification},
journal = {Proc. 7th European Workshop on Probabilistic Graphical Models ({PGM} '14)},
editor = {van der Gaag, L. and Feelders, A.},
publisher = {Springer},
series = {Lecture Notes in Artificial Intelligence},
volume = {8754},
booktitle = {{PGM'14}: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models},
author = {Corani, G. and Antonucci, A. and Mau\'a, D. and Gabaglio, S.},
pages = {145--159},
year = {2014},
doi = {10.1007/978-3-319-11433-0_10}
}
Download
Cozman, F.G., de Campos, C.P. (2014). Kuznetsov independence for interval-valued expectations and sets of probability distributions: Properties and algorithms. International Journal of Approximate Reasoning 55(2), pp. 666–682.
Kuznetsov independence for interval-valued expectations and sets of probability distributions: Properties and algorithms
Authors: Cozman, F.G. and de Campos, C.P.
Year: 2014
Abstract: Kuznetsov independence of variables X and Y means that, for any pair of bounded functions f(X) and g(Y), E[f(X)g(Y)]=E[f(X)] *times* E[g(Y)], where E[.] denotes interval-valued expectation and *times* denotes interval multiplication. We present properties of Kuznetsov independence for several variables, and connect it with other concepts of independence in the literature; in particular we show that strong extensions are always included in sets of probability distributions whose lower and upper expectations satisfy Kuznetsov independence. We introduce an algorithm that computes lower expectations subject to judgments of Kuznetsov independence by mixing column generation techniques with nonlinear programming. Finally, we define a concept of conditional Kuznetsov independence, and study its graphoid properties.
Published in International Journal of Approximate Reasoning 55(2), Elsevier, pp. 666–682.
Note: Appeared online in Sept/2013
Kuznetsov independence for interval-valued expectations and sets of probability distributions: Properties and algorithms
@ARTICLE{cozman2013ijar,
title = {Kuznetsov independence for interval-valued expectations and sets of probability distributions: {P}roperties and algorithms},
journal = {International Journal of Approximate Reasoning},
publisher = {Elsevier},
volume = {55},
author = {Cozman, F.G. and de Campos, C.P.},
number = {2},
pages = {666--682},
year = {2014},
doi = {10.1016/j.ijar.2013.09.013}
}
Download
Mauá, D.D., de Campos, C.P., Antonucci, A. (2014). Hidden Markov models with imprecisely specified parameters. In Proceedings of the Brazilian Conference on Intelligent Systems.
Hidden Markov models with imprecisely specified parameters
Authors: Mauá, D.D. and de Campos, C.P. and Antonucci, A.
Year: 2014
Abstract: Hidden Markov models (HMMs) are widely used models for sequential data. As with other probabilistic graphical models, they require the specification of precise probability values, which can be too restrictive for some domains, espe- cially when data are scarce or costly to acquire. We present a generalized version of HMMs, whose quantification can be done by sets of, instead of single, probability distributions. Our models have the ability to suspend judgment when there is not enough statistical evidence, and can serve as a sensitivity analysis tool for standard non-stationary HMMs. Efficient inference algorithms are developed to address standard HMM usage such as the computation of likelihoods and most probable explanations. Experiments with real data show that the use of imprecise probabilities leads to more reliable inferences without compromising efficiency.
Published in Proceedings of the Brazilian Conference on Intelligent Systems.
Hidden Markov models with imprecisely specified parameters
@INPROCEEDINGS{antonucci2014d,
title = {Hidden {M}arkov models with imprecisely specified parameters},
booktitle = {Proceedings of the Brazilian Conference on Intelligent Systems},
author = {Mau\'a, D.D. and de Campos, C.P. and Antonucci, A.},
year = {2014}
}
Download
Mauá, D.D., de Campos, C.P., Benavoli, A., Antonucci, A. (2014). Probabilistic inference in credal networks: new complexity results. Journal of Artifical Intelligence Research 50, pp. 603–637.
Probabilistic inference in credal networks: new complexity results
Authors: Mauá, D.D. and de Campos, C.P. and Benavoli, A. and Antonucci, A.
Year: 2014
Abstract: Credal networks are graph-based statistical models whose parameters take values in a
set, instead of being sharply specified as in traditional statistical models (e.g., Bayesian
networks). The computational complexity of inferences on such models depends on the
irrelevance/independence concept adopted. In this paper, we study inferential complexity
under the concepts of epistemic irrelevance and strong independence. We show that infer-
ences under strong independence are NP-hard even in trees with binary variables except
for a single ternary one. We prove that under epistemic irrelevance the polynomial-time
complexity of inferences in credal trees is not likely to extend to more general models
(e.g., singly connected topologies). These results clearly distinguish networks that admit
efficient inferences and those where inferences are most likely hard, and settle several open
questions regarding their computational complexity. We show that these results remain
valid even if we disallow the use of zero probabilities. We also show that the computation
of bounds on the probability of the future state in a hidden Markov model is the same
whether we assume epistemic irrelevance or strong independence, and we prove a similar
result for inference in naive Bayes structures. These inferential equivalences are important
for practitioners, as hidden Markov models and naive Bayes structures are used in real
applications of imprecise probability
Published in Journal of Artifical Intelligence Research 50, pp. 603–637.
Probabilistic inference in credal networks: new complexity results
@ARTICLE{maua14jair,
title = {Probabilistic inference in credal networks: new complexity results},
journal = {Journal of Artifical Intelligence Research},
volume = {50},
author = {Mau\'a, D.D. and de Campos, C.P. and Benavoli, A. and Antonucci, A.},
pages = {603--637},
year = {2014},
doi = {10.1613/jair.4355}
}
Download
Polpo, A., de Campos, C.P., Sinha, D., Lipsitz, S., Lin, J. (2014). Transform both sides model: a parametric approach. Computational Statistics and Data Analysis 71, pp. 903–913.
Transform both sides model: a parametric approach
Authors: Polpo, A. and de Campos, C.P. and Sinha, D. and Lipsitz, S. and Lin, J.
Year: 2014
Abstract: A parametric regression model for right-censored data with a log-linear median regression function and a transformation in both response and regression parts, named parametric Transform-Both-Sides (TBS) model, is presented. The TBS model has a parameter that handles data asymmetry while allowing various different distributions for the error, as long as they are unimodal symmetric distributions centered at zero. The discussion is focused on the estimation procedure with five important error distributions (normal, double-exponential, Student's t, Cauchy and logistic) and presents properties, associated functions (that is, survival and hazard functions) and estimation methods based on maximum likelihood and on the Bayesian paradigm. These procedures are implemented in TBSSurvival, an open-source fully documented R package. The use of the package is illustrated and the performance of the model is analyzed using both simulated and real data sets.
Published in Computational Statistics and Data Analysis 71, Elsevier, pp. 903–913.
Note: Appeared online in Jul/2013
Transform both sides model: a parametric approach
@ARTICLE{decampos2013b,
title = {Transform both sides model: a parametric approach},
journal = {Computational Statistics and Data Analysis},
publisher = {Elsevier},
volume = {71},
author = {Polpo, A. and de Campos, C.P. and Sinha, D. and Lipsitz, S. and Lin, J.},
pages = {903--913},
year = {2014},
doi = {10.1016/j.csda.2013.07.023}
}
Download
Scanagatta, M., de Campos, C.P., Zaffalon, M. (2014). Min-BDeu and max-BDeu scores for learning Bayesian networks. In van der Gaag, L., Feelders, A. (Eds), PGM'14: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models, Lecture Notes in Artificial Intelligence 8754, Springer, pp. 426–441.
Min-BDeu and max-BDeu scores for learning Bayesian networks
Authors: Scanagatta, M. and de Campos, C.P. and Zaffalon, M.
Year: 2014
Abstract: This work presents two new score functions based on
the Bayesian Dirichlet equivalent uniform (BDeu) score for learning
Bayesian network structures. They consider the sensitivity of BDeu
to varying parameters of the Dirichlet prior. The scores take on the
most adversary and the most beneficial priors among those within a
contamination set around the symmetric one. We build these scores in
such way that they are decomposable and can be computed
efficiently. Because of that, they can be integrated into any
state-of-the-art structure learning method that explores the space
of directed acyclic graphs and allows decomposable scores.
Empirical results suggest that our scores outperform the standard
BDeu score in terms of the likelihood of unseen data and in terms of
edge discovery with respect to the true network, at least when
the training sample size is small. We discuss the relation between
these new scores and the accuracy of inferred models. Moreover, our
new criteria can be used to identify the amount of data after which
learning is saturated, that is, additional data are of little help
to improve the resulting model.
Published in van der Gaag, L., Feelders, A. (Eds), PGM'14: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models, Lecture Notes in Artificial Intelligence 8754, Springer, pp. 426–441.
Min-BDeu and max-BDeu scores for learning Bayesian networks
@INPROCEEDINGS{scanagatta2014a,
title = {Min-{BDeu} and max-{BDeu} scores for learning {B}ayesian networks},
editor = {van der Gaag, L. and Feelders, A.},
publisher = {Springer},
series = {Lecture Notes in Artificial Intelligence},
volume = {8754},
booktitle = {{PGM'14}: Proceedings of the Seventh European Workshop on Probabilistic Graphical Models},
author = {Scanagatta, M. and de Campos, C.P. and Zaffalon, M.},
pages = {426--441},
year = {2014},
doi = {10.1007/978-3-319-11433-0_28}
}
Download
Zaffalon, M., Corani, G. (2014). Comments on "Imprecise probability models for learning multinomial distributions from data. Applications to learning credal networks" by Andrés R. Masegosa and Serafín Moral. International Journal of Approximate Reasoning 55(7), pp. 1597–1600.
Comments on "Imprecise probability models for learning multinomial distributions from data. Applications to learning credal networks" by Andrés R. Masegosa and Serafín Moral
Authors: Zaffalon, M. and Corani, G.
Year: 2014
Abstract: We briefly overview the problem of learning probabilities from data using imprecise probability models that express very weak prior beliefs. Then we comment on the new contributions to this question given in the paper by Masegosa and Moral and provide some insights about the performance of their models in data mining experiments of classification.
Published in International Journal of Approximate Reasoning 55(7), pp. 1597–1600.
Comments on "Imprecise probability models for learning multinomial distributions from data. Applications to learning credal networks" by Andrés R. Masegosa and Serafín Moral
@ARTICLE{zaffalon2014a,
title = {Comments on {"Imprecise} probability models for learning multinomial distributions from data. Applications to learning credal networks" by {A}ndr\'es {R}. Masegosa and {S}eraf\'in {M}oral},
journal = {International Journal of Approximate Reasoning},
volume = {55},
author = {Zaffalon, M. and Corani, G.},
number = {7},
pages = {1597--1600},
year = {2014},
doi = {10.1016/j.ijar.2014.05.001}
}
Download top2013
Antonucci, A., de Campos, C.P., Huber, D., Zaffalon, M. (2013). Approximating credal network inferences by linear programming. In van der Gaag, L. C. (Ed), Proceedings of the 12th European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty, Lecture Notes in Artificial Intelligence 7958, Springer, Berlin Heidelberg, pp. 13–25.
Approximating credal network inferences by linear programming
Authors: Antonucci, A. and de Campos, C.P. and Huber, D. and Zaffalon, M.
Year: 2013
Abstract: An algorithm for approximate credal network updating is presented. The problem in its general formulation is a multilinear optimization task, which can be linearized by an appropriate rule for fixing all the local models apart from those of a single variable. This simple idea can be iterated and quickly leads to very accurate inferences. The approach can also be specialized to classification with credal networks based on the maximality criterion. A complexity analysis for both the problem and the algorithm is reported together with numerical experiments, which confirm the good performance of the method. While the inner approximation produced by the algorithm gives rise to a classifier which might return a subset of the optimal class set, preliminary empirical results suggest that the accuracy of the optimal class set is seldom affected by the approximate probabilities.
Published in van der Gaag, L. C. (Ed), Proceedings of the 12th European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty, Lecture Notes in Artificial Intelligence 7958, Springer, Berlin Heidelberg, pp. 13–25.
Approximating credal network inferences by linear programming
@INPROCEEDINGS{antonucci2013a,
title = {Approximating credal network inferences by linear programming},
editor = {van der Gaag, L. C.},
publisher = {Springer},
address = {Berlin Heidelberg},
series = {Lecture Notes in Artificial Intelligence},
volume = {7958},
booktitle = {Proceedings of the 12th European Conference on Symbolic and Quantitative Approaches to Reasoning {w}ith Uncertainty},
author = {Antonucci, A. and de Campos, C.P. and Huber, D. and Zaffalon, M.},
pages = {13--25},
year = {2013},
doi = {10.1007/978-3-642-39091-3_2}
}
Download
Antonucci, A., Corani, G., Mauá, D.D., Gabaglio, S. (2013). An ensemble of Bayesian networks for multilabel classification. In Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI-13), pp. 1220–1225.
An ensemble of Bayesian networks for multilabel classification
Authors: Antonucci, A. and Corani, G. and Mauá, D.D. and Gabaglio, S.
Year: 2013
Abstract: We present a novel approach for multilabel classification based on an ensemble of Bayesian networks. The class variables are connected by a tree; each model of the ensemble uses a different class as root of the tree. We assume the features to be conditionally independent given the classes, thus generalizing the naive Bayes assumption to the multiclass case. This assumption allows us to optimally
identify the correlations between classes and features; such correlations are moreover shared across all models of the ensemble. Inferences are drawn from the ensemble via logarithmic opinion pooling. To minimize Hamming loss, we compute the marginal probability of the classes by running standard inference on each Bayesian network in the ensemble, and then pooling the inferences. To instead minimize the subset 0/1 loss, we pool the joint distributions of each model and cast the problem as a MAP inference in the corresponding graphical
model. Experiments show that the approach is competitive with state-of-the-art methods for multilabel classification.
Published in Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI-13), pp. 1220–1225.
An ensemble of Bayesian networks for multilabel classification
@INPROCEEDINGS{antonucci2013d,
title = {An ensemble of {B}ayesian networks for multilabel classification},
booktitle = {Proceedings of the 23rd International Joint Conference on Artificial Intelligence ({IJCAI}-13)},
author = {Antonucci, A. and Corani, G. and Mau\'a, D.D. and Gabaglio, S.},
pages = {1220--1225},
year = {2013}
}
Download
Antonucci, A., Huber, D., Zaffalon, M., Luginbuehl, P., Chapman, I., Ladouceur, R. (2013). CREDO: a military decision-support system based on credal networks. In Proceedings of the 16th Conference on Information Fusion (FUSION 2013), pp. 1–8.
CREDO: a military decision-support system based on credal networks
Authors: Antonucci, A. and Huber, D. and Zaffalon, M. and Luginbuehl, P. and Chapman, I. and Ladouceur, R.
Year: 2013
Abstract: A software tool especially designed for military domains to create and query decision-support systems is presented. Credal networks, which are Bayesian networks whose parameters have the freedom to vary in convex sets, are used to model the relations among the system variables. A novel elicitation procedure of these sets, which allows the military experts to report their knowledge by purely qualitative judgements, is proposed. Two high-level fusion procedures to cope with multiple experts in this framework are also derived. All these features are supported by the software and demonstrated in an application to space security tested during the last NATO multinational experiment.
Published in Proceedings of the 16th Conference on Information Fusion (FUSION 2013), pp. 1–8.
CREDO: a military decision-support system based on credal networks
@INPROCEEDINGS{antonucci2013c,
title = {{CREDO}: a military decision-support system based on credal networks},
booktitle = {Proceedings of the 16th Conference on Information Fusion ({FUSION} 2013)},
author = {Antonucci, A. and Huber, D. and Zaffalon, M. and Luginbuehl, P. and Chapman, I. and Ladouceur, R.},
pages = {1--8},
year = {2013}
}
Download
Antonucci, A., de Rosa, R., Giusti, A., Cuzzolin, F. (2013). Temporal data classification by imprecise dynamical models. In Cozman, F.G., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 13–22.
Temporal data classification by imprecise dynamical models
Authors: Antonucci, A. and de Rosa, R. and Giusti, A. and Cuzzolin, F.
Year: 2013
Abstract: We propose a new methodology to classify temporal data with imprecise hidden Markov models. For each sequence we learn a different model by coupling the EM algorithm with the imprecise Dirichlet model. As a model descriptor, we consider the expected value of the observable variable in the limit of stationar- ity of the Markov chain. In the imprecise case, only the bounds of this descriptor can be evaluated. In practice the sequence, which can be regarded as a trajectory in the feature space, is summarized by a hyperbox in the same space. We classify these static but interval-valued data by a credal generalization of the k-nearest neighbors algorithm. Experiments on benchmark datasets for computer vision show that the method achieves the required robustness whilst outperforming other precise and imprecise methods.
Published in Cozman, F.G., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 13–22.
Temporal data classification by imprecise dynamical models
@INPROCEEDINGS{antonucci2013b,
title = {Temporal data classification by imprecise dynamical models},
editor = {Cozman, F.G. and Denoeux, T. and Destercke, S. and Seidenfeld, T.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications},
author = {Antonucci, A. and de Rosa, R. and Giusti, A. and Cuzzolin, F.},
pages = {13--22},
year = {2013},
url = {http://www.sipta.org/isipta13/proceedings/papers/s002.pdf}
}
Download
Azzimonti, L., Ieva, F., Paganoni, A.M. (2013). Nonlinear nonparametric mixed-effects models for unsupervised classification. Computational Statistics 28(4), pp. 1549–1570.
Nonlinear nonparametric mixed-effects models for unsupervised classification
Authors: Azzimonti, L. and Ieva, F. and Paganoni, A.M.
Year: 2013
Abstract: In this work we propose a novel EM method for the estimation of nonlinear nonparametric mixed-effects models, aimed at unsupervised classification. We perform simulation studies in order to evaluate the algorithm performance and we apply this new procedure to a real dataset.
Published in Computational Statistics 28(4), pp. 1549–1570.
Nonlinear nonparametric mixed-effects models for unsupervised classification
@ARTICLE{azzimonti2013a,
title = {Nonlinear nonparametric mixed-effects models for unsupervised classification},
journal = {Computational Statistics},
volume = {28},
author = {Azzimonti, L. and Ieva, F. and Paganoni, A.M.},
number = {4},
pages = {1549--1570},
year = {2013},
doi = {10.1007/s00180-012-0366-5}
}
Download
Benavoli, A. (2013). The generalised moment-based filter. Automatic Control, IEEE Transactions on 58(10), pp. 2642–2647.
The generalised moment-based filter
Authors: Benavoli, A.
Year: 2013
Abstract: Can we solve the filtering problem from the only knowledge of few moments of the noise terms? In this paper, by exploiting set of distributions based filtering, we solve this problem without introducing additional assumptions on the distributions of the noises (e.g., Gaussianity) or on the final form of the estimator (e.g., linear estimator). Given the moments (e.g., mean and variance) of random variable X, it is possible to define the set of all distributions that are compatible with the moments information. This set can be equivalently characterized by its extreme distributions: a family of mixtures of Dirac deltas. The lower and upper expectation of any function g of X are obtained in correspondence of these extremes and can be computed by solving a linear programming problem. The filtering problem can then be solved by running iteratively this linear programming problem. In this paper, we discuss theoretical properties of this filter, we show the connection with set-membership estimation and its practical applications.
Accepted in Automatic Control, IEEE Transactions on 58(10), pp. 2642–2647.
The generalised moment-based filter
@ARTICLE{benavoli2013b,
title = {The generalised moment-based filter},
journal = {Automatic Control, {IEEE} Transactions on},
volume = {58},
author = {Benavoli, A.},
number = {10},
pages = {2642--2647},
year = {2013},
doi = {10.1109/TAC.2013.2255971}
}
Download
Benavoli, A. (2013). Imprecise hierarchical Dirichlet model with applications. In Information Fusion (fusion), 2013 Proc. Of the 16th International Conference on, pp. 1918–1925.
Imprecise hierarchical Dirichlet model with applications
Authors: Benavoli, A.
Year: 2013
Abstract: Many estimation problems in data fusion involve multiple parameters that can be related in some way by the structure of the problem. This implies that a joint probabilistic model for these parameters should reflect this dependence. In parametric estimation, a Bayesian way to account for this possible dependence is to use hierarchical models, in which data depends on hidden parameters that in turn depend on hyperprior parameters. An issue in this analysis is how to choose the hyperprior in case of lack of prior information. This paper focuses on parametric estimation problems involving multinomial-Dirichlet models and presents a model of prior ignorance for the hyperparameters. This model consists to a set of Dirichlet distributions that expresses a condition of prior ignorance. We analyse the theoretical properties of this model and we apply it to practical fusion problems: (i) the estimate of the packet drop rate in a centralized sensor network; (ii) the estimate of the transition probabilities for a multiple-model algorithm.
Published in Information Fusion (fusion), 2013 Proc. Of the 16th International Conference on, pp. 1918–1925.
Imprecise hierarchical Dirichlet model with applications
@INPROCEEDINGS{benavoli2013d,
title = {Imprecise hierarchical {D}irichlet model with applications},
booktitle = {Information Fusion ({f}usion), 2013 Proc. Of the 16th International Conference on},
author = {Benavoli, A.},
pages = {1918--1925},
year = {2013},
url = {http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=6641239&isnumber=6641065}
}
Download
Benavoli, A., Papi, F. (2013). Set-membership PHD filter. In Information Fusion (fusion), 2013 Proc. Of the 16th International Conference on, pp. 1722–1729.
Set-membership PHD filter
Authors: Benavoli, A. and Papi, F.
Year: 2013
Abstract: The paper proposes a novel Probability Hypothesis Density (PHD) filter for linear system in which initial state, process and measurement noises are only known to be bounded (they can vary on compact sets, e.g., polytopes). This means that no probabilistic assumption is imposed on the distributions of initial state and noises besides the knowledge of their supports. These are the same assumptions that are used in set-membership estimation. By exploiting a formulation of set-membership estimation in terms of set of probability measures, we derive the equations of the set-membership PHD filter, which consist in propagating in time compact sets that include with guarantee the targets' states. Numerical simulations show the effectiveness of the proposed approach and the comparison with a sequential Monte Carlo PHD filter which instead assumes that initial state and noises have uniform distributions.
Published in Information Fusion (fusion), 2013 Proc. Of the 16th International Conference on, pp. 1722–1729.
Set-membership PHD filter
@INPROCEEDINGS{benavoli2013c,
title = {Set-membership {PHD} filter},
booktitle = {Information Fusion ({f}usion), 2013 Proc. Of the 16th International Conference on},
author = {Benavoli, A. and Papi, F.},
pages = {1722--1729},
year = {2013},
url = {http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=6641211&isnumber=6641065}
}
Download
Benavoli, A., Zaffalon, M. (2013). Density-ratio robustness in dynamic state estimation. Mechanical Systems and Signal Processing 37(1–2), pp. 54–75.
Density-ratio robustness in dynamic state estimation
Authors: Benavoli, A. and Zaffalon, M.
Year: 2013
Abstract: The filtering problem is addressed by taking into account imprecision in the knowledge about the probabilistic relationships involved. Imprecision is modelled in this paper by a particular closed convex set of probabilities that is known with the name of density ratio class or constant odds-ratio (COR) model. The contributions of this paper are the following. First, we shall define an optimality criterion based on the squared-loss function for the estimates derived from a general closed convex set of distributions. Second, after revising the properties of the density ratio class in the context of parametric estimation, we shall extend these properties to state estimation accounting for system dynamics. Furthermore, for the case in which the nominal density of the COR model is a multivariate Gaussian, we shall derive closed-form solutions for the set of optimal estimates and for the credible region. Third, we discuss how to perform Monte Carlo integrations to compute lower and upper expectations from a COR set of densities. Then we shall derive a procedure that, employing Monte Carlo sampling techniques, allows us to propagate in time both the lower and upper state expectation functionals and, thus, to derive an efficient solution of the filtering problem. Finally, we empirically compare the proposed estimator with the Kalman filter. This shows that our solution is more robust to the presence of modelling errors in the system and that, hence, appears to be a more realistic approach than the Kalman filter in such a case.
Published in Mechanical Systems and Signal Processing 37(1–2), pp. 54–75.
Density-ratio robustness in dynamic state estimation
@ARTICLE{benavoli2012g,
title = {Density-ratio robustness in dynamic state estimation},
journal = {Mechanical Systems and Signal Processing},
volume = {37},
author = {Benavoli, A. and Zaffalon, M.},
number = {1--2},
pages = {54--75},
year = {2013},
doi = {10.1016/j.ymssp.2012.09.004},
url = {http://www.idsia.ch/~alessio/benavoli2012g.pdf}
}
Download
de Campos, C.P., Cozman, F.G. (2013). Complexity of inferences in polytree-shaped semi-qualitative probabilistic networks. In Proceedings of the 27th AAAI Conference on Advances in Artificial Intelligence (AAAI), pp. 217–223.
Complexity of inferences in polytree-shaped semi-qualitative probabilistic networks
Authors: de Campos, C.P. and Cozman, F.G.
Year: 2013
Abstract: Semi-qualitative probabilistic networks (SQPNs) merge two important graphical model formalisms: Bayesian networks and qualitative probabilistic networks. They provide a very general modeling framework by allowing the combination of numeric and qualitative assessments over a discrete domain, and can be compactly encoded by exploiting the same factorization of joint probability distributions that are behind the Bayesian networks. This paper explores the computational complexity of semi-qualitative probabilistic networks, and takes the polytree-shaped networks as its main target. We show that the inference problem is coNP-Complete for binary polytrees with multiple observed nodes. We also show that inferences can be performed in linear time if there is a single observed node, which is a relevant practical case. Because our proof is constructive, we obtain an efficient linear time algorithm for SQPNs under such assumptions. To the best of our knowledge, this is the first exact polynomial-time algorithm for SQPNs. Together these results provide a clear picture of the inferential complexity in polytree-shaped SQPNs.
Published in Proceedings of the 27th AAAI Conference on Advances in Artificial Intelligence (AAAI), pp. 217–223.
Complexity of inferences in polytree-shaped semi-qualitative probabilistic networks
@INPROCEEDINGS{decampos2013a,
title = {Complexity of inferences in polytree-shaped semi-qualitative probabilistic networks},
booktitle = {Proceedings of the 27th {AAAI} Conference on Advances in Artificial Intelligence ({AAAI})},
author = {de Campos, C.P. and Cozman, F.G.},
pages = {217--223},
year = {2013}
}
Download
de Campos, C.P., Rancoita, P.M.V., Kwee, I., Zucca, E., Zaffalon, M., Bertoni, F. (2013). Discovering subgroups of patients from DNA copy number data using NMF on compacted matrices. PLoS ONE 8(11), e79720.
Discovering subgroups of patients from DNA copy number data using NMF on compacted matrices
Authors: de Campos, C.P. and Rancoita, P.M.V. and Kwee, I. and Zucca, E. and Zaffalon, M. and Bertoni, F.
Year: 2013
Abstract: In the study of complex genetic diseases, the identification of subgroups of patients sharing similar genetic characteristics represents a challenging task, for example, to improve treatment decision. One type of genetic lesion, frequently investigated in such disorders, is the change of the DNA copy number (CN) at specific genomic traits. Non-negative Matrix Factorization (NMF) is a standard technique to reduce the dimensionality of a data set and to cluster data samples, while keeping its most relevant information in meaningful components. Thus, it can be used to discover subgroups of patients from CN profiles. It is however computationally impractical for very high dimensional data, such as CN microarray data. Deciding the most suitable number of subgroups is also a challenging problem. The aim of this work is to derive a procedure to compact high dimensional data, in order to improve NMF applicability without compromising the quality of the clustering. This is particularly important for analyzing high-resolution microarray data. Many commonly used quality measures, as well as our own measures, are employed to decide the number of subgroups and to assess the quality of the results. Our measures are based on the idea of identifying robust subgroups, inspired by biologically/clinically relevance instead of simply aiming at well-separated clusters. We evaluate our procedure using four real independent data sets. In these data sets, our method was able to find accurate subgroups with individual molecular and clinical features and outperformed the standard NMF in terms of accuracy in the factorization fitness function. Hence, it can be useful for the discovery of subgroups of patients with similar CN profiles in the study of heterogeneous diseases.
Published in PLoS ONE 8(11), e79720.
Discovering subgroups of patients from DNA copy number data using NMF on compacted matrices
@ARTICLE{decampos2013d,
title = {Discovering subgroups of patients from {DNA} copy number data using {NMF} on compacted matrices},
journal = {{PLoS} {ONE}},
volume = {8},
author = {de Campos, C.P. and Rancoita, P.M.V. and Kwee, I. and Zucca, E. and Zaffalon, M. and Bertoni, F.},
number = {11},
pages = {e79720},
year = {2013},
doi = {10.1371/journal.pone.0079720}
}
Download
Corani, G., Magli, M., Giusti, A., Gianaroli, L., Gambardella, L. (2013). A Bayesian network model for predicting pregnancy after in vitro fertilization. Computers in Biology and Medicine 43(11), pp. 1783–1792.
A Bayesian network model for predicting pregnancy after in vitro fertilization
Authors: Corani, G. and Magli, M. and Giusti, A. and Gianaroli, L. and Gambardella, L.
Year: 2013
Abstract: We present a Bayesian network model for predicting the outcome of in vitro
fertilization (IVF). The problem is characterized by a particular missingness process; we propose a simple but effective averaging approach which improves parameter estimates compared to the traditional MAP estimation. We present results with generated data and the analysis of a real data set. Moreover, we assess by means of a simulation study the effectiveness of the model in supporting the selection of the embryos to be transferred.
Published in Computers in Biology and Medicine 43(11), pp. 1783–1792.
A Bayesian network model for predicting pregnancy after in vitro fertilization
@ARTICLE{corani2013c,
title = {A {B}ayesian network model for predicting pregnancy after in vitro fertilization},
journal = {Computers in Biology and Medicine},
volume = {43},
author = {Corani, G. and Magli, M. and Giusti, A. and Gianaroli, L. and Gambardella, L.},
number = {11},
pages = {1783--1792},
year = {2013},
doi = {10.1016/j.compbiomed.2013.07.035}
}
Download
Corani, G., Mignatti, A. (2013). Credal model averaging of logistic regression for modeling the distribution of marmot burrows. In Cozman, F.G., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds),, pp. 233–243.
Credal model averaging of logistic regression for modeling the distribution of marmot burrows
Authors: Corani, G. and Mignatti, A.
Year: 2013
Abstract: Bayesian model averaging (BMA) weights the inferences produced by a set of competing models, using as weights the models posterior probabilities. An open problem of BMA is how to set the prior probability of the models. Credal model averaging (CMA) is a credal ensemble of Bayesian models, which generalizes BMA by substituting the single prior over the models by a set of priors. The base models of the ensemble are learned in a Bayesian fashion. We use CMA to ensemble base classiers which are Bayesian logistic regressors, characterized by dierent sets of covariates. CMA returns indeterminate classications when the classication is prior-dependent, namely when the most probable class depends on the prior probability assigned to the different models. We apply CMA for modelling the presence and absence of marmot burrows in an Alpine valley in Italy and show that it compares favorably to BMA.
Published in Cozman, F.G., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), Proceedings of ISIPTA '13 (the Eighth International Symposium on Imprecise Probability: Theories and Applications), pp. 233–243.
Credal model averaging of logistic regression for modeling the distribution of marmot burrows
@INPROCEEDINGS{corani2013a,
title = {Credal model averaging of logistic regression for modeling the distribution of marmot burrows},
journal = {Proceedings of {ISIPTA} '13 (the Eighth International Symposium on Imprecise Probability: Theories and Applications)},
editor = {Cozman, F.G. and Denoeux, T. and Destercke, S. and Seidenfeld, T. },
author = {Corani, G. and Mignatti, A.},
pages = {233--243},
year = {2013}
}
Download
Gianaroli, L., Magli, M.C., Gambardella, L., Giusti, A., Grugnetti, C., Corani, G. (2013). Objective way to support embryo transfer: a probabilistic decision. Human Reproduction 28(5), pp. 1210–1220.
Objective way to support embryo transfer: a probabilistic decision
Authors: Gianaroli, L. and Magli, M.C. and Gambardella, L. and Giusti, A. and Grugnetti, C. and Corani, G.
Year: 2013
Abstract: study question: Is it feasible to identify factors that significantly affect the clinical outcome of IVF-ICSI cycles and use them to reliably
design a predictor of implantation?
summary answer: The Bayesian network (BN) identified top-history embryos, female age and the insemination technique as the
most relevant factors for predicting the occurrence of pregnancy (AUC, area under curve, of 0.72). In addition, it could discriminate
between no implantation and single or twin implantations in a prognostic model that can be used prospectively.
what is known already: The key requirement for achieving a single live birth in an IVF-ICSI cycle is the capacity to estimate
embryo viability in relation to maternal receptivity. Nevertheless, the lack of a strong predictor imposes several restrictions on this strategy.
study design, size, duration: Medical histories, laboratory data and clinical outcomes of all fresh transfer cycles performed at the International Institute for Reproductive Medicine of Lugano, Switzerland, in the period 2006–2008 (n 1⁄4 388 cycles), were retrospectively evaluated and analyzed.
participants/materials, setting, methods: Patients were unselected for age, sperm parameters or other infertility cri-
teria. Before being admitted to treatment, uterine anomalies were excluded by diagnostic hysteroscopy.
To evaluate the factors possibly related to embryo viability and maternal receptivity, the class variable was categorized as pregnancy versus
no pregnancy and the features included: female age, number of previous cycles, insemination technique, sperm of proven fertility, the number
of transferred top-history embryos, the number of transferred top-quality embryos, the number of follicles .14 mm and the level of estradiol
on the day of HCG administration. To assess the classifier, the indicators of performance were computed by cross-validation. Two statistical
models were used: the decision tree and the BN.
main results and the role of choice: The decision tree identified the number of transferred top-history embryos, female
age and the insemination technique as the features discriminating between pregnancy and no pregnancy. The model achieved an accuracy of
81.5% that was significantly higher in comparison with the trivial classifier, but the increase was so modest that the model was clinically
useless for predictions of pregnancy. The BN could more reliably predict the occurrence of pregnancy with an AUC of 0.72, and confirmed
the importance of top-history embryos, female age and insemination technique in determining implantation. In addition, it could discriminate
between no implantation, single implantation and twin implantation with the AUC of 0.72, 0.64 and 0.83, respectively.
limitations, reasons for caution: The relatively small sample of the study did not permit the inclusion of more features that
could also have a role in determining the clinical outcome. The design of this study was retrospective to identify the relevant features; a
prospective study is now needed to verify the validity of the model.
wider implications of the findings: The resulting predictive model can discriminate with reasonable reliability between
pregnancy and no pregnancy, and can also predict the occurrence of a single pregnancy or multiple pregnancy. This could represent an ef-
fective support for deciding how many embryos and which embryos to transfer for each couple. Due to its flexibility, the number of variables
in the predictor can easily be increased to include other features that may affect implantation.
study funding/competing interests: This study was supported by a grant, CTI Medtech Project Number: 9707.1 PFLS-L,
Swiss Confederation. No competing interests are declared.
Published in Human Reproduction 28(5), pp. 1210–1220.
Objective way to support embryo transfer: a probabilistic decision
@ARTICLE{corani2013d,
title = {Objective way to support embryo transfer: a probabilistic decision},
journal = {Human Reproduction},
volume = {28},
author = {Gianaroli, L. and Magli, M.C. and Gambardella, L. and Giusti, A. and Grugnetti, C. and Corani, G.},
number = {5},
pages = {1210--1220},
year = {2013},
doi = {10.1093/humrep/det030}
}
Download
von Hohenstaufen, K.A., Conconi, A., de Campos, C.P., Franceschetti, S., Bertoni, F., Margiotta Casaluci, G., Stathis, A., Ghielmini, M., Stussi, G., Cavalli, F., Gaidano, G., Zucca, E. (2013). Prognostic impact of monocyte count at presentation in mantle cell lymphoma. British Journal of Haematology 162(4), pp. 465–473.
Prognostic impact of monocyte count at presentation in mantle cell lymphoma
Authors: von Hohenstaufen, K.A. and Conconi, A. and de Campos, C.P. and Franceschetti, S. and Bertoni, F. and Margiotta Casaluci, G. and Stathis, A. and Ghielmini, M. and Stussi, G. and Cavalli, F. and Gaidano, G. and Zucca, E.
Year: 2013
Abstract: Increased number of circulating monocytes at presentation has been recently associated with shorter survival in Hodgkin lymphoma, follicular lymphoma and diffuse large B cell lymphoma. This study aimed to assess the prognostic impact of the absolute monocyte count (AMC) at diagnosis in mantle cell lymphoma (MCL). From the series of MCL cases recorded on the databases of the Oncology Institute of Southern Switzerland in Bellinzona (Switzerland) and the Division of Haematology of the Amedeo Avogadro University of Eastern Piedmont in Novara (Italy), the AMC at diagnosis was available in 97 cases. Cox regression was used for both univariate and multivariate analysis. With a median follow up of 7 years, the 5-year overall survival (OS) was 29% for patients with AMC >500/ul and 62% for patients with AMC <= 500/ul (p=0.006). Elevated AMC and beta-2 microglobulin at diagnosis remained independent outcome predictors at multivariate analysis and might be used to build a simple prognostic scoring system. Survival was significantly shorter in patients with both AMC and beta-2 microglobulin above the upper limit of normal but the MCL international prognostic index (MIPI) remained the strongest survival predictor in this series. In this relatively small and heterogeneous series an increased AMC identified poor-risk patients. Our results suggest that AMC in conjunction with the beta-2 microglobulin level might provide an inexpensive way to stratify the MCL patient risk as a complement to the MIPI, which was confirmed to be a very powerful prognostic tool.
Published in British Journal of Haematology 162(4), Blackwell Publishing Ltd, pp. 465–473.
Prognostic impact of monocyte count at presentation in mantle cell lymphoma
@ARTICLE{decampos2013c,
title = {Prognostic impact of monocyte count at presentation in mantle cell lymphoma},
journal = {British Journal of Haematology},
publisher = {Blackwell Publishing Ltd},
volume = {162},
author = {von Hohenstaufen, K.A. and Conconi, A. and de Campos, C.P. and Franceschetti, S. and Bertoni, F. and Margiotta Casaluci, G. and Stathis, A. and Ghielmini, M. and Stussi, G. and Cavalli, F. and Gaidano, G. and Zucca, E.},
number = {4},
pages = {465--473},
year = {2013},
doi = {10.1111/bjh.12409},
url = {http://onlinelibrary.wiley.com/doi/10.1111/bjh.12409/pdf}
}
Download
Mangili, F., Benavoli, A. (2013). New prior near-ignorance models on the simplex. In Cozman, F.G., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, Compiegne (FR), pp. 1–9.
New prior near-ignorance models on the simplex
Authors: Mangili, F. and Benavoli, A.
Year: 2013
Abstract: The aim of this paper is to derive new near-ignorance models on the probability simplex, which do not directly involve the Dirichlet distribution and, thus, that are alternative to the Imprecise Dirichlet Model. We focus our investigation to a particular class of distributions on the simplex which is known as the class of Normalized Infinitely Divisible distributions; it includes the Dirichlet distribution as a particular case. Starting from three members of this class, which admit a closed-form expression for the probability density function, we derive three new near-ignorance prior models on the simplex, we analyse their properties and compare them with the Imprecise Dirichlet Model.
Published in Cozman, F.G., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, Compiegne (FR), pp. 1–9.
New prior near-ignorance models on the simplex
@INPROCEEDINGS{mangili2013a,
title = {New prior near-ignorance models on the simplex},
editor = {Cozman, F.G. and Denoeux, T. and Destercke, S. and Seidenfeld, T.},
publisher = {SIPTA},
address = {Compiegne (FR)},
booktitle = {{ISIPTA };'13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications},
author = {Mangili, F. and Benavoli, A.},
pages = {1--9},
year = {2013}
}
Download
Mauá, D.D., de Campos, C.P., Benavoli, A., Antonucci, A. (2013). On the complexity of strong and epistemic credal networks. In Proceedings of the 29th Conference on Uncertainty in Artificial Intelligence, AUAI Press, pp. 391–400.
On the complexity of strong and epistemic credal networks
Authors: Mauá, D.D. and de Campos, C.P. and Benavoli, A. and Antonucci, A.
Year: 2013
Abstract: Credal networks are graph-based statistical models whose parameters take values in a set, instead of being sharply specified as in traditional statistical models (e.g., Bayesian networks). The computational complexity of inferences on such models depends on the irrelevance/independence concept adopted. In this paper, we study inferential complexity under the concepts of epistemic irrelevance and strong independence. We show that inferences under strong independence are NP-hard even in trees with ternary variables. We prove that under epistemic irrelevance the polynomial time complexity of inferences in credal trees is not likely to extend to more general models (e.g. singly connected networks). These results clearly distinguish networks that admit efficient inferences and those where inferences are most likely hard, and settle several open questions regarding computational complexity.
Published in Proceedings of the 29th Conference on Uncertainty in Artificial Intelligence, AUAI Press, pp. 391–400.
On the complexity of strong and epistemic credal networks
@INPROCEEDINGS{maua2013a,
title = {On the complexity of strong and epistemic credal networks},
publisher = {AUAI Press},
booktitle = {Proceedings of the 29th Conference on Uncertainty in Artificial Intelligence},
author = {Mau\'a, D.D. and de Campos, C.P. and Benavoli, A. and Antonucci, A.},
pages = {391--400},
year = {2013}
}
Download
Mauá, D.D., de Campos, C.P., Zaffalon, M. (2013). On the complexity of solving polytree-shaped limited memory influence diagrams with binary variables. Artificial Intelligence 205, pp. 30–38.
On the complexity of solving polytree-shaped limited memory influence diagrams with binary variables
Authors: Mauá, D.D. and de Campos, C.P. and Zaffalon, M.
Year: 2013
Abstract: Influence diagrams are intuitive and concise representations of structured decision problems. When the problem is non-Markovian, an optimal strategy can be exponentially large in the size of the diagram. We can avoid the inherent intractability by constraining the size of admissible strategies, giving rise to limited memory influence diagrams. A valuable question is then how small do strategies need to be to enable efficient optimal planning. Arguably, the smallest strategies one can conceive simply prescribe an action for each time step, without considering past decisions or observations. Previous work has shown that finding such optimal strategies even for polytree-shaped diagrams with ternary variables and a single value node is NP-hard, but the case of binary variables was left open. In this paper we address such a case, by first noting that optimal strategies can be obtained in polynomial time for polytree-shaped diagrams with binary variables and a single value node. We then show that the same problem is NP-hard if the diagram has multiple value nodes. These two results close the fixed-parameter complexity analysis of optimal strategy selection in influence diagrams parametrized by the shape of the diagram, the number of value nodes and the maximum variable cardinality.
Published in Artificial Intelligence 205, pp. 30–38.
On the complexity of solving polytree-shaped limited memory influence diagrams with binary variables
@ARTICLE{maua2013b,
title = {On the complexity of solving polytree-shaped limited memory influence diagrams with binary variables},
journal = {Artificial Intelligence},
volume = {205},
author = {Mau\'a, D.D. and de Campos, C.P. and Zaffalon, M.},
pages = {30--38},
year = {2013},
doi = {10.1016/j.artint.2013.10.002}
}
Download
Miranda, E., Zaffalon, M. (2013). Conglomerable coherent lower previsions. In Kruse, R., Berthold, M. R., Moewes, C., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Eds), Synergies of Soft Computing and Statistics for Intelligent Data Analysis, Advances in Intelligent and Soft Computing 190, Springer Berlin Heidelberg, pp. 419–427.
Conglomerable coherent lower previsions
Authors: Miranda, E. and Zaffalon, M.
Year: 2013
Abstract: Walley's theory of coherent lower previsions builds upon the former theory by Williams with the explicit aim to make it deal with conglomerability. We show that such a construction has been only partly successful because Walley's founding axiom of joint coherence does not entirely capture the implications of conglomerability. As a way to fully achieve Walley's original aim, we propose then the new theory of conglomerable coherent lower previsions. We show that Walley's theory coincides with ours when all conditioning events have positive lower probability, or when conditioning partitions are nested.
Published in Kruse, R., Berthold, M. R., Moewes, C., Gil, M. A., Grzegorzewski, P., Hryniewicz, O. (Eds), Synergies of Soft Computing and Statistics for Intelligent Data Analysis, Advances in Intelligent and Soft Computing 190, Springer Berlin Heidelberg, pp. 419–427.
Conglomerable coherent lower previsions
@INCOLLECTION{zaffalon2012a,
title = {Conglomerable coherent lower previsions},
editor = {Kruse, R. and Berthold, M. R. and Moewes, C. and Gil, M. A. and Grzegorzewski, P. and Hryniewicz, O.},
publisher = {Springer Berlin Heidelberg},
series = {Advances in Intelligent and Soft Computing},
volume = {190},
booktitle = {Synergies of Soft Computing and Statistics for Intelligent Data Analysis},
author = {Miranda, E. and Zaffalon, M.},
pages = {419--427},
year = {2013},
doi = {10.1007/978-3-642-33042-1_45}
}
Download
Miranda, E., Zaffalon, M. (2013). Conglomerable coherence. International Journal of Approximate Reasoning 54(9), pp. 1322–1350.
Conglomerable coherence
Authors: Miranda, E. and Zaffalon, M.
Year: 2013
Abstract: We contrast Williams' and Walley's theories of coherent lower previsions in the light of conglomerability. These are two of the most credited approaches to a behavioural theory of imprecise probability. Conglomerability is the notion that distinguishes them the most: Williams' theory does not consider it, while Walley aims at embedding it in his theory. This question is important, as conglomerability is a major point of disagreement at the foundations of probability, since it was first defined by de Finetti in 1930. We show that Walley's notion of joint coherence (which is the single axiom of his theory) for conditional lower previsions does not take all the implications of conglomerability into account. Considered also some previous results in the literature, we deduce that Williams' theory should be the one to use when conglomerability is not required; for the opposite case, we define the new theory of conglomerably coherent lower previsions, which is arguably the one to use, and of which Walley's theory can be understood as an approximation. We show that this approximation is exact in two important cases: when all conditioning events have positive lower probability, and when conditioning partitions are nested.
Published in International Journal of Approximate Reasoning 54(9), pp. 1322–1350.
Conglomerable coherence
@ARTICLE{zaffalon2013b,
title = {Conglomerable coherence},
journal = {International Journal of Approximate Reasoning},
volume = {54},
author = {Miranda, E. and Zaffalon, M.},
number = {9},
pages = {1322--1350},
year = {2013},
doi = {10.1016/j.ijar.2013.04.016}
}
Download
Miranda, E., Zaffalon, M. (2013). Computing the conglomerable natural extension. In Cozman, F., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 255–264.
Computing the conglomerable natural extension
Authors: Miranda, E. and Zaffalon, M.
Year: 2013
Abstract: Given a coherent lower prevision P, we consider the problem of computing the smallest coherent lower prevision F ≥ P that is conglomerable, in case it exists. F is called the conglomerable natural extension. Past work has showed that F can be approximated by an increasing sequence (En)n∈ℕ of coherent lower previsions. We close an open problem by showing that this sequence can be made of infinitely many distinct elements. Moreover, we give sufficient conditions, of quite broad applicability, to make sure that the point-wise limit of the sequence is F in case P is the lower envelope of finitely many linear previsions. In addition, we study the question of the existence of F and its relationship with the notion of marginal extension.
Published in Cozman, F., Denoeux, T., Destercke, S., Seidenfeld, T. (Eds), ISIPTA '13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 255–264.
Computing the conglomerable natural extension
@INPROCEEDINGS{zaffalon2013c,
title = {Computing the conglomerable natural extension},
editor = {Cozman, F. and Denoeux, T. and Destercke, S. and Seidenfeld, T.},
publisher = {SIPTA},
booktitle = {{ISIPTA };'13: Proceedings of the Eighth International Symposium on Imprecise Probability: Theories and Applications},
author = {Miranda, E. and Zaffalon, M.},
pages = {255--264},
year = {2013},
url = {http://www.sipta.org/isipta13/proceedings/papers/s025.pdf}
}
Download
Zaffalon, M., Miranda, E. (2013). Probability and time. Artificial Intelligence 198, pp. 1–51.
Probability and time
Authors: Zaffalon, M. and Miranda, E.
Year: 2013
Abstract: Probabilistic reasoning is often attributed a temporal meaning, in which conditioning is regarded as a normative rule to compute future beliefs out of current beliefs and observations. However, the well-established ‘updating interpretation’ of conditioning is not concerned with beliefs that evolve in time, and in particular with future beliefs. On the other hand, a temporal justification of conditioning was proposed already by De Moivre and Bayes, by requiring that current and future beliefs be consistent. We reconsider the latter approach while dealing with a generalised version of the problem, using a behavioural theory of imprecise probability in the form of coherent lower previsions as well as of coherent sets of desirable gambles, and letting the possibility space be finite or infinite. We obtain that using conditioning is normative, in the imprecise case, only if one establishes future behavioural commitments at the same time of current beliefs. In this case it is also normative that present beliefs be conglomerable, which is a result that touches on a long-term controversy at the foundations of probability. In the remaining case, where one commits to some future behaviour after establishing present beliefs, we characterise the several possibilities to define consistent future assessments; this shows in particular that temporal consistency does not preclude changes of mind. And yet, our analysis does not support that rationality requires consistency in general, even though pursuing consistency makes sense and is useful, at least as a way to guide and evaluate the assessment process. These considerations narrow down in the special case of precise probability, because this formalism cannot distinguish the two different situations illustrated above: it turns out that the only consistent rule is conditioning and moreover that it is not rational to be willing to stick to precise probability while using a rule different from conditioning to compute future beliefs; rationality requires in addition the disintegrability of the present-time probability.
Published in Artificial Intelligence 198, pp. 1–51.
Probability and time
@ARTICLE{zaffalon2013a,
title = {Probability and time},
journal = {Artificial Intelligence},
volume = {198},
author = {Zaffalon, M. and Miranda, E.},
pages = {1--51},
year = {2013},
doi = {10.1016/j.artint.2013.02.005}
}
Download top2012
Antonucci, A. (2012). An interval-valued dissimilarity measure for belief functions based on credal semantics. In Denoeux, T., Masson, M.H. (Eds), Belief Functions: Theory and Applications, Advances in Intelligent and Soft Computing 164, Springer Berlin / Heidelberg, pp. 37–44.
An interval-valued dissimilarity measure for belief functions based on credal semantics
Authors: Antonucci, A.
Year: 2012
Abstract: Evidence theory extends Bayesian probability theory by allowing for a
more expressive model of subjective uncertainty. Besides standard interpretation of belief functions, where uncertainty corresponds to probability masses which might refer to whole subsets of the possibility space, credal semantics can be also considered. Accordingly, a belief function can be identified with the whole set of probability mass functions consistent with the beliefs induced by the masses. Following
this interpretation, a novel, set-valued, dissimilarity measure with a clear behavioral interpretation can be defined. We describe the main features of this new measure and comment the relation with other measures proposed in the literature.
Published in Denoeux, T., Masson, M.H. (Eds), Belief Functions: Theory and Applications, Advances in Intelligent and Soft Computing 164, Springer Berlin / Heidelberg, pp. 37–44.
An interval-valued dissimilarity measure for belief functions based on credal semantics
@INPROCEEDINGS{antonucci2012a,
title = {An interval-valued dissimilarity measure for belief functions based on credal semantics},
editor = {Denoeux, T. and Masson, M.H.},
publisher = {Springer Berlin / Heidelberg},
series = {Advances in Intelligent and Soft Computing},
volume = {164},
booktitle = {Belief Functions: Theory and Applications},
author = {Antonucci, A.},
pages = {37--44},
year = {2012},
doi = {10.1007/978-3-642-29461-7_4}
}
Download
Antonucci, A., Cattaneo, M.E.V.G., Corani, G. (2012). Likelihood-based robust classification with Bayesian networks. In Communications in Computer and Information Science, Advances in Computational Intelligence 299(5), Springer Berlin / Heidelberg, pp. 491–500.
Likelihood-based robust classification with Bayesian networks
Authors: Antonucci, A. and Cattaneo, M.E.V.G. and Corani, G.
Year: 2012
Abstract: Bayesian networks are commonly used for classification: a structural learning algorithm determines the network graph, while standard approaches estimate the model parameters from data. Yet, with few data the corresponding assessments can be unreliable. To gain robustness in this phase, we consider a likelihood-based learning approach, which takes all the model quantifications whose likelihood exceeds a given threshold. A new classification algorithm based on this approach is presented. Notably, this is a credal classifier, i.e., more than a single class can be returned in output. This is the case when the Bayesian networks consistent with the threshold constraint assign different class labels to a test instance. This is the first classifier of this kind for general topologies. Experiments show how this approach provide the desired robustness.
Published in Communications in Computer and Information Science, Advances in Computational Intelligence 299(5), Springer Berlin / Heidelberg, pp. 491–500.
Likelihood-based robust classification with Bayesian networks
@INPROCEEDINGS{antonucci2012b,
title = {Likelihood-based robust classification with {B}ayesian networks},
publisher = {Springer Berlin / Heidelberg},
series = {Advances in Computational Intelligence},
volume = {299},
booktitle = {Communications in Computer and Information Science},
author = {Antonucci, A. and Cattaneo, M.E.V.G. and Corani, G.},
number = {5},
pages = {491--500},
year = {2012},
doi = {10.1007/978-3-642-31718-7_51}
}
Download
Antonucci, A., Corani, G., Gabaglio, S. (2012). Active learning by the naive credal classifier. In Cano, A., Gomez-Olmedo, M., Nielsen, T. (Eds), Proc. of the 6th European Workshop on Probabilistic Graphical Models (PGM 2012), pp. 3–10.
Active learning by the naive credal classifier
Authors: Antonucci, A. and Corani, G. and Gabaglio, S.
Year: 2012
Abstract: In standard classication a training set of supervised instances is given. In a more general setup, some supervised instances are available, while further ones should be chosen from an unsupervised set and then annotated. As the annotation step is costly, active learning algorithms are used to select which instances to annotate to maximally increase the classication performance while annotating only a limited number of them. Several active learning algorithms are based on the naive Bayes classier. We work instead with the naive credal classier, namely an extension of naive Bayes to imprecise probability. We propose two novel methods for active learning based on the naive credal classier. Empirical
comparisons show performance comparable or slightly superior to that of approaches solely based on the naive Bayes.
Published in Cano, A., Gomez-Olmedo, M., Nielsen, T. (Eds), Proc. of the 6th European Workshop on Probabilistic Graphical Models (PGM 2012), pp. 3–10.
Active learning by the naive credal classifier
@INPROCEEDINGS{antonucci2012c,
title = {Active learning by the naive credal classifier},
editor = {Cano, A. and Gomez-Olmedo, M. and Nielsen, T.},
booktitle = {Proc. {o}f the 6th European Workshop on Probabilistic Graphical Models ({PGM} 2012)},
author = {Antonucci, A. and Corani, G. and Gabaglio, S.},
pages = {3--10},
year = {2012}
}
Download
Battistelli, G., Benavoli, A., Chisci, L. (2012). State estimation with remote sensors and intermittent transmissions. Systems & Control Letters 61(1), pp. 155–164.
State estimation with remote sensors and intermittent transmissions
Authors: Battistelli, G. and Benavoli, A. and Chisci, L.
Year: 2012
Abstract: This paper deals with the problem of estimating the state of a discrete-time linear stochastic dynamical system on the basis of data collected from multiple sensors subject to a limitation on the communication rate from the remote sensor units. The optimal probabilistic measurement-independent strategy for deciding when to transmit estimates from each sensor is derived. Simulation results show that the derived strategy yields certain advantages in terms of worst-case time-averaged performance with respect to periodic ones when coordination among sensors is not possible.
Published in Systems & Control Letters 61(1), pp. 155–164.
State estimation with remote sensors and intermittent transmissions
@ARTICLE{benavoli2012d,
title = {State estimation with remote sensors and intermittent transmissions},
journal = {Systems & Control Letters},
volume = {61},
author = {Battistelli, G. and Benavoli, A. and Chisci, L.},
number = {1},
pages = {155--164},
year = {2012},
doi = {10.1016/j.sysconle.2011.10.005}
}
Download
Battistelli, G., Benavoli, A., Chisci, L. (2012). Data-driven communication for state estimation with sensor networks. Automatica 48(5), pp. 926–935.
Data-driven communication for state estimation with sensor networks
Authors: Battistelli, G. and Benavoli, A. and Chisci, L.
Year: 2012
Abstract: This paper deals with the problem of estimating the state of a discrete-time linear stochastic dynamical system on the basis of data collected from multiple sensors subject to a limitation on the communication rate from the sensors. More specifically, the attention is devoted to a centralized sensor network consisting of: (1) multiple remote nodes which collect measurements of the given system, compute state estimates at the full measurement rate and transmit data (either raw measurements or estimates) at a reduced communication rate; (2) a fusion node that, based on received data, provides an estimate of the system state at the full rate. Local data-driven transmission strategies are considered and issues related to the stability and performance of such strategies are investigated. Simulation results confirm the effectiveness of the proposed strategies.
Published in Automatica 48(5), pp. 926–935.
Data-driven communication for state estimation with sensor networks
@ARTICLE{benavoli2012c,
title = {Data-driven communication for state estimation with sensor networks},
journal = {Automatica},
volume = {48},
author = {Battistelli, G. and Benavoli, A. and Chisci, L.},
number = {5},
pages = {926--935},
year = {2012},
doi = {10.1016/j.automatica.2012.02.028}
}
Download
Battistelli, G., Benavoli, A., Chisci, L. (2012). Data-driven strategies for selective data transmission in sensor networks. In CDC 2012, Proc. of the 51st Ieee Conference on Decision and Control, Maui, Usa, pp. 1–6.
Data-driven strategies for selective data transmission in sensor networks
Authors: Battistelli, G. and Benavoli, A. and Chisci, L.
Year: 2012
Abstract: Energy efficiency is a crucial issue for any task involving wireless sensor networks. The present paper addresses nonlinear state estimation over a centralized sensor network, i.e. a set of sensor nodes communicating with a central information fusion unit, and proposes smart data-driven strategies by which sensors decide which data transmit to the central unit so as to reduce data communication, and thus avoid congestion problems as well as prolong the network lifetime, while providing enhanced performance with respect to periodic transmission. Both measurement and estimate transmission strategies are developed. To cope with nonlinear sensors that cannot fully observe the state, suitable nonlinear observability decompositions are employed. A bearing-only tracking simulation case-study is presented in order to demonstrate the effectiveness of the proposed approach.
Published in CDC 2012, Proc. of the 51st Ieee Conference on Decision and Control, Maui, Usa, pp. 1–6.
Data-driven strategies for selective data transmission in sensor networks
@INPROCEEDINGS{benavoli2012f,
title = {Data-driven strategies for selective data transmission in sensor networks},
booktitle = {{CDC} 2012, Proc. {o}f the 51st Ieee Conference on Decision and Control, Maui, Usa},
author = {Battistelli, G. and Benavoli, A. and Chisci, L.},
pages = {1--6},
year = {2012}
}
Download
Benavoli, A. (2012). Belief function robustness in estimation. In Denoeux, T., Masson, M.-H. (Eds), Belief Functions: Theory and Applications, Advances in Intelligent and Soft Computing 164, Springer Berlin / Heidelberg, pp. 375–383.
Belief function robustness in estimation
Authors: Benavoli, A.
Year: 2012
Abstract: We consider the case in which the available knowledge does not allow to specify a precise probabilistic model for the prior and/or likelihood in statistical estimation. We assume that this imprecision can be represented by belief functions. Thus, we exploit the mathematical structure of belief functions and their equivalent representation in terms of closed convex sets of probability measures to derive robust posterior inferences.
Published in Denoeux, T., Masson, M.-H. (Eds), Belief Functions: Theory and Applications, Advances in Intelligent and Soft Computing 164, Springer Berlin / Heidelberg, pp. 375–383.
Belief function robustness in estimation
@INCOLLECTION{benavoli2012a,
title = {Belief function robustness in estimation},
editor = {Denoeux, T. and Masson, M.-H.},
publisher = {Springer Berlin / Heidelberg},
series = {Advances in Intelligent and Soft Computing},
volume = {164},
booktitle = {Belief Functions: Theory and Applications},
author = {Benavoli, A.},
pages = {375--383},
year = {2012},
doi = {10.1007/978-3-642-29461-7_44}
}
Download
Benavoli, A., Noack, B. (2012). Pushing Kalman's idea to the extremes. In In Information Fusion (FUSION), 2012 Proc. of the 15th International Conference on, pp. 1–8.
Pushing Kalman's idea to the extremes
Authors: Benavoli, A. and Noack, B.
Year: 2012
Abstract: The paper focuses on the fundamental idea of Kalman's seminal paper: how to solve the filtering problem from the only knowledge of the first two moments of the noise terms. In this paper, by exploiting set of distributions based filtering, we solve this problem without introducing additional assumptions on the distributions of the noise terms (e.g., Gaussianity) or on the final form of the estimator (e.g., linear estimator). Given the moments (e.g., mean and variance) of random variable X, it is possible to define the set of all distributions that are compatible with the moments information. This set of distributions can be equivalently characterized by its extreme distributions which is a family of mixtures of Dirac's deltas. The lower and upper expectation of any function g of X are obtained in correspondence of these extremes and can be computed by solving a linear programming problem. The filtering problem can then be solved by running iteratively this linear programming problem.
Accepted in In Information Fusion (FUSION), 2012 Proc. of the 15th International Conference on, pp. 1–8.
Pushing Kalman's idea to the extremes
@INPROCEEDINGS{benavoli2012e,
title = {Pushing {K}alman's idea to the extremes},
booktitle = {In Information Fusion ({FUSION}), 2012 Proc. {o}f the 15th International Conference on},
author = {Benavoli, A. and Noack, B.},
pages = {1--8},
year = {2012}
}
Download
Benavoli, A., Zaffalon, M. (2012). A model of prior ignorance for inferences in the one-parameter exponential family. Journal of Statistical Planning and Inference 142(7), pp. 1960–1979.
A model of prior ignorance for inferences in the one-parameter exponential family
Authors: Benavoli, A. and Zaffalon, M.
Year: 2012
Abstract: This paper proposes a model of prior ignorance about a scalar variable based on a set of distributions M. In particular, a set of minimal properties that a set M of distributions should satisfy to be a model of prior ignorance without producing vacuous inferences is defined. In the case the likelihood model corresponds to a one-parameter exponential family of distributions, it is shown that the above minimal properties are equivalent to a special choice of the domains for the parameters of the conjugate exponential prior. This makes it possible to define the largest (that is, the least-committal) set of conjugate priors M that satisfies the above properties. The obtained set M is a model of prior ignorance with respect to the functions (queries) that are commonly used for statistical inferences; it is easy to elicit and, because of conjugacy, tractable; it encompasses frequentist and the so-called objective Bayesian inferences with improper priors. An application of the model to a problem of inference with count data is presented.
Published in Journal of Statistical Planning and Inference 142(7), pp. 1960–1979.
A model of prior ignorance for inferences in the one-parameter exponential family
@ARTICLE{benavoli2012b,
title = {A model of prior ignorance for inferences in the one-parameter exponential family},
journal = {Journal of Statistical Planning and Inference},
volume = {142},
author = {Benavoli, A. and Zaffalon, M.},
number = {7},
pages = {1960--1979},
year = {2012},
doi = {10.1016/j.jspi.2012.01.023}
}
Download
Corani, G., Antonucci, A., De Rosa, R. (2012). Compression-based AODE classifiers. In De Raedt, L. et al. (Ed), Proc. 20th European Conference on Artificial Intelligence (ECAI 2012), pp. 264–269.
Compression-based AODE classifiers
Authors: Corani, G. and Antonucci, A. and De Rosa, R.
Year: 2012
Abstract: We propose the COMP-AODE classifier, which adopts the compression-based approach to average the posterior probabilities computed by different non-naive classifiers (SPODEs). COMP-AODE improves classification performance over the well-known AODE model. COMP-AODE assumes a uniform prior over the SPODEs; we then develop the credal classifier COMP-AODE*, substituting the uniform prior by a set of priors. COMP-AODE* returns more classes when the classification is prior-dependent, namely if the most probable class varies with the prior adopted over the SPODEs. COMP-AODE* achieves higher classification utility than both COMP-AODE and AODE.
Published in De Raedt, L. et al. (Ed), Proc. 20th European Conference on Artificial Intelligence (ECAI 2012), pp. 264–269.
Compression-based AODE classifiers
@INPROCEEDINGS{corani2012d,
title = {Compression-based {AODE} classifiers},
editor = {De Raedt, L. et al. },
booktitle = {Proc. 20th European Conference on Artificial Intelligence ({ECAI} 2012)},
author = {Corani, G. and Antonucci, A. and De Rosa, R.},
pages = {264--269},
year = {2012}
}
Download
Corani, G., Antonucci, A., Zaffalon, M. (2012). Bayesian networks with imprecise probabilities: theory and application to classification. In Holmes, D.E., Jain, L.C. (Eds), Data Mining: Foundations and Intelligent Paradigms, Intelligent Systems Reference Library 23, Springer, Berlin / Heidelberg, pp. 49–93.
Bayesian networks with imprecise probabilities: theory and application to classification
Authors: Corani, G. and Antonucci, A. and Zaffalon, M.
Year: 2012
Abstract: Bayesian networks are powerful probabilistic graphical models for modelling uncertainty. Among others, classification represents an important application: some of the most used classifiers are based on Bayesian networks. Bayesian networks are precise models: exact numeric values should be provided for quantification. This requirement is sometimes too narrow. Sets instead of single distributions can provide a more realistic description in these cases. Bayesian networks can be generalized to cope with sets of distributions. This leads to a novel class of imprecise probabilistic graphical models, called credal networks . In particular, classifiers based on Bayesian networks are generalized to so-called credal classifiers . Unlike Bayesian classifiers, which always detect a single class as the one maximizing the posterior class probability, a credal classifier may eventually be unable to discriminate a single class. In other words, if the available information is not sufficient, credal classifiers allow for indecision between two or more classes, this providing a less informative but more robust conclusion than Bayesian classifiers.
Published in Holmes, D.E., Jain, L.C. (Eds), Data Mining: Foundations and Intelligent Paradigms, Intelligent Systems Reference Library 23, Springer, Berlin / Heidelberg, pp. 49–93.
Bayesian networks with imprecise probabilities: theory and application to classification
@INCOLLECTION{corani2012c,
title = {Bayesian networks with imprecise probabilities: theory and application to classification},
editor = {Holmes, D.E. and Jain, L.C.},
publisher = {Springer, Berlin / Heidelberg},
series = {Intelligent Systems Reference Library},
volume = {23},
booktitle = {Data Mining: Foundations and Intelligent Paradigms},
author = {Corani, G. and Antonucci, A. and Zaffalon, M.},
pages = {49--93},
year = {2012},
doi = {10.1007/978-3-642-23166-7_4}
}
Download
Corani, G., Magli, C., Giusti, A., Gianaroli, L., Gambardella, L. (2012). A Bayesian network model for predicting the outcome of in vitro fertilization. In Cano, A., Gomez-Olmedo, M., Nielsen, T. (Eds), Proc. of the 6th European Workshop on Probabilistic Graphical Models (PGM 2012), pp. 75–82.
A Bayesian network model for predicting the outcome of in vitro fertilization
Authors: Corani, G. and Magli, C. and Giusti, A. and Gianaroli, L. and Gambardella, L.
Year: 2012
Abstract: We present a Bayesian network model for predicting the outcome of in-vitro fertilization (IVF). The problem is characterized by a peculiar missingness process, and we propose a simple but effective averaging approach which improves parameter estimates compared to the traditional MAP estimation. The model can provide relevant insights to IVF experts.
Published in Cano, A., Gomez-Olmedo, M., Nielsen, T. (Eds), Proc. of the 6th European Workshop on Probabilistic Graphical Models (PGM 2012), pp. 75–82.
A Bayesian network model for predicting the outcome of in vitro fertilization
@INPROCEEDINGS{corani2012e,
title = {A {B}ayesian network model for predicting the outcome of in vitro fertilization},
editor = {Cano, A. and Gomez-Olmedo, M. and Nielsen, T.},
booktitle = {Proc. {o}f the 6th European Workshop on Probabilistic Graphical Models ({PGM} 2012)},
author = {Corani, G. and Magli, C. and Giusti, A. and Gianaroli, L. and Gambardella, L.},
pages = {75--82},
year = {2012}
}
Download
Magli, C., Corani, G., Giusti, A., Castelletti, E., Gambardella, L., Gianaroli, L. (2012). A prognostic model for multiple-embryo transfers. Human Reproduction (Supplement: Abstract book, Proc. Annual Meeting ESHRE 2012) 27(2), pp. ii162–ii205.
A prognostic model for multiple-embryo transfers
Authors: Magli, C. and Corani, G. and Giusti, A. and Castelletti, E. and Gambardella, L. and Gianaroli, L.
Year: 2012
Abstract: The EU prognostic model is composed by two sub-models: the E sub-model estimates the probability of an embryo being viable; the U sub-model estimates the probability of the woman to sustain a viable embryo (maternal receptivity). The establishment of pregnancy requires one or more viable embryos and a receptive maternal environment. There is however no general consensus on which variable to included in the E and U sub-models. On a related topic, defining a strong predictor of embryo viability is still an open problem. Morphological criteria are used to identify top-quality embryos; some studies suggest that better predictivity can be achieved combining into a single score the grades obtained by the embryo in different stages. We thus scored the embryos as non-top, top and top-history, the latter score being assigned to embryos judged of top quality in several observations. In this study we analyzed 352 IVF cycles (average patients age: 36 years; average number of transferred embryos: 2; clinical pregnancy rate: 21.5%) through the EU model, adopting a statistical criterion to choose the variables in the E and U sub-model, assessed the viabilities of non-top, top and top-history embryos and measured the predictive ability of the model.
Published in Human Reproduction (Supplement: Abstract book, Proc. Annual Meeting ESHRE 2012) 27(2), pp. ii162–ii205.
A prognostic model for multiple-embryo transfers
@ARTICLE{corani2012b,
title = {A prognostic model for multiple-embryo transfers},
journal = {Human Reproduction (Supplement: Abstract {b}ook, Proc. Annual Meeting {ESHRE} 2012)},
volume = {27},
author = {Magli, C. and Corani, G. and Giusti, A. and Castelletti, E. and Gambardella, L. and Gianaroli, L.},
number = {2},
pages = {ii162--ii205},
year = {2012},
doi = {10.1093/humrep/27.s2.77}
}
Download
Mauá, D.D., de Campos, C.P., Zaffalon, M. (2012). Solving limited memory influence diagrams. Journal of Artificial Intelligence Research 44, pp. 97–140.
Solving limited memory influence diagrams
Authors: Mauá, D.D. and de Campos, C.P. and Zaffalon, M.
Year: 2012
Abstract: We present a new algorithm for exactly solving decision making problems represented as influence diagrams. We do not require the usual assumptions of no forgetting and regularity; this allows us to solve problems with simultaneous decisions and limited information. The algorithm is empirically shown to outperform a state-of-the-art algorithm on randomly generated problems of up to 150 variables and 1064 solutions. We show that these problems are NP-hard even if the underlying graph structure of the problem has low treewidth and the variables take on a bounded number of states, and that they admit no provably good approximation if variables can take on an arbitrary number of states.
Published in Journal of Artificial Intelligence Research 44, pp. 97–140.
Solving limited memory influence diagrams
@ARTICLE{maua2012a,
title = {Solving limited memory influence diagrams},
journal = {Journal of Artificial Intelligence Research},
volume = {44},
author = {Mau\'a, D.D. and de Campos, C.P. and Zaffalon, M.},
pages = {97--140},
year = {2012},
url = {http://www.jair.org/media/3625/live-3625-6282-jair.pdf}
}
Download
Mauá, D.D., de Campos, C.P., Zaffalon, M. (2012). Updating credal networks is approximable in polynomial time. International Journal of Approximate Reasoning 53(8), pp. 1183–1199.
Updating credal networks is approximable in polynomial time
Authors: Mauá, D.D. and de Campos, C.P. and Zaffalon, M.
Year: 2012
Abstract: Credal networks relax the precise probability requirement of Bayesian networks, enabling a richer representation of uncertainty in the form of closed convex sets of probability measures. The increase in expressiveness comes at the expense of higher computational costs. In this paper, we present a new variable elimination algorithm for exactly computing posterior inferences in extensively specified credal networks, which is empirically shown to outperform a state-of-the-art algorithm. The algorithm is then turned into a provably good approximation scheme, that is, a procedure that for any input is guaranteed to return a solution not worse than the optimum by a given factor. Remarkably, we show that when the networks have bounded treewidth and bounded number of states per variable the approximation algorithm runs in time polynomial in the input size and in the inverse of the error factor, thus being the first known fully polynomial-time approximation scheme for inference in credal networks.
Published in International Journal of Approximate Reasoning 53(8), pp. 1183–1199.
Updating credal networks is approximable in polynomial time
@ARTICLE{maua2012d,
title = {Updating credal networks is approximable in polynomial time},
journal = {International Journal of Approximate Reasoning},
volume = {53},
author = {Mau\'a, D.D. and de Campos, C.P. and Zaffalon, M.},
number = {8},
pages = {1183--1199},
year = {2012},
doi = {10.1016/j.ijar.2012.06.014},
url = {http://www.sciencedirect.com/science/article/pii/S0888613X12000904?v=s5}
}
Download
Mauá, D.D., de Campos, C.P. (2012). Anytime marginal map inference. In Proceedings of the 28th International Conference on Machine Learning (ICML 2012), pp. 1471–1478.
Anytime marginal map inference
Authors: Mauá, D.D. and de Campos, C.P.
Year: 2012
Abstract: This paper presents a new anytime algorithm for the marginal MAP problem in graphical models. The algorithm is described in detail, its complexity and convergence rate are studied, and relations to previous theoretical results for the problem are discussed. It is shown that the algorithm runs in polynomial-time if the underlying graph of the model has bounded tree-width, and that it provides guarantees to the lower and upper bounds obtained within a fixed amount of computational resources. Experiments with both real and synthetic generated models highlight its main characteristics and show that it compares favorably against Park and Darwiche's systematic search, particularly in the case of problems with many MAP variables and moderate tree-width.
Published in Proceedings of the 28th International Conference on Machine Learning (ICML 2012), pp. 1471–1478.
Anytime marginal map inference
@INPROCEEDINGS{maua2012b,
title = {Anytime marginal map inference},
booktitle = {Proceedings of the 28th International Conference on Machine Learning ({ICML} 2012)},
author = {Mau\'a, D.D. and de Campos, C.P.},
pages = {1471--1478 },
year = {2012},
url = {http://icml.cc/2012/papers/728.pdf}
}
Download
Mauá, D.D., de Campos, C.P., Zaffalon, M. (2012). The complexity of approximately solving influence diagrams. In Proceedings of the 28th Conference on Uncertainty in Artificial Intelligence (UAI 2012), pp. 604–613.
The complexity of approximately solving influence diagrams
Authors: Mauá, D.D. and de Campos, C.P. and Zaffalon, M.
Year: 2012
Abstract: Influence diagrams allow for intuitive and yet precise description of complex situations involving decision making under uncertainty. Unfortunately, most of the problems described by influence diagrams are hard to solve. In this paper we discuss the complexity of approximately solving influence diagrams. We do not assume no-forgetting or regularity, which makes the class of problems we address very broad. Remarkably, we show that when both the treewidth and the cardinality of the variables are bounded the problem admits a fully polynomial-time approximation scheme.
Published in Proceedings of the 28th Conference on Uncertainty in Artificial Intelligence (UAI 2012), pp. 604–613.
The complexity of approximately solving influence diagrams
@INPROCEEDINGS{maua2012c,
title = {The complexity of approximately solving influence diagrams},
booktitle = {Proceedings of the 28th Conference on Uncertainty in Artificial Intelligence ({UAI} 2012)},
author = {Mau\'a, D.D. and de Campos, C.P. and Zaffalon, M.},
pages = {604--613},
year = {2012},
url = {http://www.auai.org/uai2012/papers/166.pdf}
}
Download
Mignatti, A., Corani, G., Rizzoli, A.E. (2012). Credal model averaging: dealing robustly with model uncertainty on small data sets. In Proc. 6th International Congress on Environmental Modelling and Software (iEMSs 2012), pp. 163-170.
Credal model averaging: dealing robustly with model uncertainty on small data sets
Authors: Mignatti, A. and Corani, G. and Rizzoli, A.E.
Year: 2012
Abstract: Datasets of population dynamics are typically characterized by a short temporal extension. In this condition, several alternative models typically achieve close accuracy, though returning quite different predictions (model uncertainty ). Bayesian model averaging (BMA) addresses this issue by averaging the prediction of the different models, using as weights the posterior probability of the models. However, an open problem of BMA is the choice of the prior probability of the models, which can largely impact on the inferences, especially when data are scarce. We present Credal Model Averaging (CMA), which addresses this problem by simultaneously considering a set of prior probability distributions over the models. This allows to represent very weak prior knowledge about the appropriateness of the different models and also to easily accommodate expert judgments, considering that in many cases the expert is not willing to commit himself to a single prior probability distribution. The predictions generated by CMA are intervals whose lengths shows the sensitivity of the predictions on the choice of the prior over the models.
Published in Proc. 6th International Congress on Environmental Modelling and Software (iEMSs 2012), pp. 163-170.
Credal model averaging: dealing robustly with model uncertainty on small data sets
@INCOLLECTION{corani2012a,
title = {Credal model averaging: dealing robustly with model uncertainty on small data sets},
booktitle = {Proc. 6th International Congress on Environmental Modelling and Software ({iEMSs} 2012)},
author = {Mignatti, A. and Corani, G. and Rizzoli, A.E.},
pages = {163-170},
year = {2012},
url = {http://www.iemss.org/iemss2012/proceedings/A3_0707_Mignatti_et_al.pdf}
}
Download
Miranda, E., Zaffalon, M., de Cooman, G. (2012). Conglomerable natural extension. International Journal of Approximate Reasoning 53(8), pp. 1200–1227.
Conglomerable natural extension
Authors: Miranda, E. and Zaffalon, M. and de Cooman, G.
Year: 2012
Abstract: At the foundations of probability theory lies a question that has been open since de Finetti framed it in 1930: whether or not an uncertainty model should be required to be conglomerable. Conglomerability is related to accepting infinitely many conditional bets. Walley is one of the authors who have argued in favor of conglomerability, while de Finetti rejected the idea. In this paper we study the extension of the conglomerability condition to two types of uncertainty models that are more general than the ones envisaged by de Finetti: sets of desirable gambles and coherent lower previsions. We focus in particular on the weakest (i.e., the least-committal) of those extensions, which we call the conglomerable natural extension. The weakest extension that does not take conglomerability into account is simply called the natural extension. We show that taking the natural extension of assessments after imposing conglomerability—the procedure adopted in Walley's theory—does not yield, in general, the conglomerable natural extension (but it does so in the case of the marginal extension). Iterating this process of imposing conglomerability and taking the natural extension produces a sequence of models that approach the conglomerable natural extension, although it is not known, at this point, whether this sequence converges to it. We give sufficient conditions for this to happen in some special cases, and study the differences between working with coherent sets of desirable gambles and coherent lower previsions. Our results indicate that it is necessary to rethink the foundations of Walley's theory of coherent lower previsions for infinite partitions of conditioning events.
Published in International Journal of Approximate Reasoning 53(8), pp. 1200–1227.
Conglomerable natural extension
@ARTICLE{zaffalon2012b,
title = {Conglomerable natural extension},
journal = {International Journal of Approximate Reasoning},
volume = {53},
author = {Miranda, E. and Zaffalon, M. and de Cooman, G.},
number = {8},
pages = {1200--1227},
year = {2012},
doi = {10.1016/j.ijar.2012.06.015}
}
Download
Zaffalon, M., Corani, G., Mauá, D.D. (2012). Evaluating credal classifiers by utility-discounted predictive accuracy. International Journal of Approximate Reasoning 53(8), pp. 1282–1301.
Evaluating credal classifiers by utility-discounted predictive accuracy
Authors: Zaffalon, M. and Corani, G. and Mauá, D.D.
Year: 2012
Abstract: Predictions made by imprecise-probability models are often indeterminate (that is, set-valued). Measuring the quality of an indeterminate prediction by a single number is important to fairly compare different models, but a principled approach to this problem is currently missing. In this paper we derive, from a set of assumptions, a metric to evaluate the predictions of credal classifiers. These are supervised learning models that issue set-valued predictions. The metric turns out to be made of an objective component, and another that is related to the decision-maker's degree of risk aversion to the variability of predictions. We discuss when the measure can be rendered independent of such a degree, and provide insights as to how the comparison of classifiers based on the new measure changes with the number of predictions to be made. Finally, we make extensive empirical tests of credal, as well as precise, classifiers by using the new metric. This shows the practical usefulness of the metric, while yielding a first insightful and extensive comparison of credal classifiers.
Published in International Journal of Approximate Reasoning 53(8), pp. 1282–1301.
Evaluating credal classifiers by utility-discounted predictive accuracy
@ARTICLE{zaffalon2012c,
title = {Evaluating credal classifiers by utility-discounted predictive accuracy},
journal = {International Journal of Approximate Reasoning},
volume = {53},
author = {Zaffalon, M. and Corani, G. and Mau\'a, D.D.},
number = {8},
pages = {1282--1301},
year = {2012},
doi = {10.1016/j.ijar.2012.06.022}
}
Download top2011
Antonucci, A. (2011). The imprecise noisy-or gate. In FUSION 2011: Proceedings of the 14th International Conference on Information Fusion, IEEE, pp. 709–715.
The imprecise noisy-or gate
Authors: Antonucci, A.
Year: 2011
Abstract: The noisy-OR gate is an important tool for a compact elicitation of the conditional probabilities of a Bayesian network. An imprecise-probabilistic version of this model, where sets instead of single distributions are used to model uncertainty about the inhibition of the causal factors, is proposed. This transforms the original Bayesian network into a so-called credal network. Despite the higher computational complexity generally characterizing inference on credal networks, it is possible to prove that, exactly as for Bayesian networks, the local complexity to update probabilities on an imprecise noisy-OR gate takes only linear, instead of exponential, time in the number of causes. This result is also extended to fault tree analysis and allows for a fast fusion of the causal effects on models with an imprecise-probabilistic quantification of the initiating events.
Published in FUSION 2011: Proceedings of the 14th International Conference on Information Fusion, IEEE, pp. 709–715.
The imprecise noisy-or gate
@INPROCEEDINGS{antonucci2011c,
title = {The imprecise noisy-or gate},
publisher = {IEEE},
booktitle = {{FUSION} 2011: Proceedings of the 14th International Conference on Information Fusion},
author = {Antonucci, A.},
pages = {709--715},
year = {2011}
}
Download
Antonucci, A., de Campos, C.P. (2011). Decision making by credal nets. In Proceedings of the International Conference on Intelligent Human-machine Systems and Cybernetics (IHMSC 2011) 1, IEEE, Hangzhou (China), pp. 201–204.
Decision making by credal nets
Authors: Antonucci, A. and de Campos, C.P.
Year: 2011
Abstract: Credal nets are probabilistic graphical models which extend Bayesian nets to cope with sets of distributions. This feature makes the model particularly suited for the implementation of classifiers and knowledge-based systems. When working with sets of (instead of single) probability distributions, the identification of the optimal option can be based on different criteria, some of them eventually leading to multiple choices. Yet, most of the inference algorithms for credal nets are designed to compute only the bounds of the posterior probabilities. This prevents some of the existing criteria from being used. To overcome this limitation, we present two simple transformations for credal nets which make it possible to compute decisions based on the maximality and E-admissibility criteria without any modification in the inference algorithms. We also prove that these decision problems have the same complexity of standard inference, being NP^PP-hard for general credal nets and NP-hard for polytrees.
Published in Proceedings of the International Conference on Intelligent Human-machine Systems and Cybernetics (IHMSC 2011) 1, IEEE, Hangzhou (China), pp. 201–204.
Decision making by credal nets
@INPROCEEDINGS{antonucci2011d,
title = {Decision making by credal nets},
publisher = {IEEE},
address = {Hangzhou (China)},
volume = {1},
booktitle = {Proceedings of the International Conference on Intelligent Human-{m}achine Systems and Cybernetics ({IHMSC} 2011)},
author = {Antonucci, A. and de Campos, C.P.},
pages = {201--204},
year = {2011}
}
Download
Antonucci, A., Cattaneo, M., Corani, G. (2011). Likelihood-based naive credal classifier. In ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 21–30.
Likelihood-based naive credal classifier
Authors: Antonucci, A. and Cattaneo, M. and Corani, G.
Year: 2011
Abstract: The naive credal classifier extends the classical naive Bayes classifier to imprecise probabilities, substituting the uniform prior by the imprecise Dirichlet model. As an alternative to the naive credal classifier, we present a hierarchical likelihood-based approach, which extends in a novel way the naive Bayes towards imprecise probabilities; in particular, it considers any possible quantification (each one defining a naive Bayes classifier) apart from those assigning to the available data a probability below a given threshold level. Besides the available supervised data, in the likelihood evaluation we also consider the instance to be classified, for which the value of the class variable is assumed missing-at-random. We obtain a closed formula to compute the dominance according to the maximality criterion for any threshold level. As there are currently no well-established metrics for comparing credal classifiers which have considerably different determinacy, we compare the two classifiers when they have comparable determinacy, finding that in those cases they generate almost equivalent classifications.
Published in ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 21–30.
Likelihood-based naive credal classifier
@INPROCEEDINGS{antonucci2011a,
title = {Likelihood-based naive credal classifier},
publisher = {SIPTA},
booktitle = {{ISIPTA} '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications},
author = {Antonucci, A. and Cattaneo, M. and Corani, G.},
pages = {21--30},
year = {2011},
url = {http://www.sipta.org/isipta11/proceedings/papers/s032.pdf}
}
Download
Antonucci, A., de Rosa, R., Giusti, A. (2011). Action recognition by imprecise hidden Markov models. In Proceedings of the 2011 International Conference on Image Processing, Computer Vision and Pattern Recognition, IPCV 2011, CSREA Press, pp. 474–478.
Action recognition by imprecise hidden Markov models
Authors: Antonucci, A. and de Rosa, R. and Giusti, A.
Year: 2011
Abstract: Hidden Markov models (HMMs) are powerful tools to capture the dynamics of a human action by providing a sufficient level of abstraction to recognise what two video sequences, depicting the same kind of action, have in common. If the sequence is short and hence only few data are available, the EM algorithm, which is generally employed to learn HMMs, might return unreliable estimates. As a possible solution to this problem, a robust version of the EM algorithm, which provides an interval-valued quantification of the HMM probabilities is provided. This takes place in an imprecise-probabilistic framework, where action recognition can be based on the (bounds of the) likelihood assigned by an imprecise HMM to the considered video sequence. Experiments show that this approach is quite effective in discriminating the hard-to-recognise sequences from the easy ones. In practice, either the recognition algorithm returns a set of action labels, which typically includes the right one, either a single answer, which is very likely to be correct, is provided.
Published in Proceedings of the 2011 International Conference on Image Processing, Computer Vision and Pattern Recognition, IPCV 2011, CSREA Press, pp. 474–478.
Action recognition by imprecise hidden Markov models
@INPROCEEDINGS{antonucci2011b,
title = {Action recognition by imprecise hidden {M}arkov models},
publisher = {CSREA Press},
booktitle = {Proceedings of the 2011 International Conference on Image Processing, Computer Vision and Pattern Recognition, {IPCV} 2011},
author = {Antonucci, A. and de Rosa, R. and Giusti, A.},
pages = {474--478},
year = {2011},
url = {http://www.lidi.info.unlp.edu.ar/WorldComp2011-Mirror/IPC5150.pdf}
}
Download
Benavoli, A., Chisci, L. (2011). Robust stochastic control based on imprecise probabilities. In Proc. of the 18th IFAC World Congress, pp. 4606–4613.
Robust stochastic control based on imprecise probabilities
Authors: Benavoli, A. and Chisci, L.
Year: 2011
Abstract: This paper deals with the optimal quadratic control problem for non Gaussian discrete-time linear stochastic systems from the perspective of imprecise probabilities. The adopted philosophy is to use a convex set of probability distributions to characterize the imprecision in the knowledge about the probabilistic relationships present in the system to be controlled. In particular, an uncertain system model, named Linear Gaussian Vacuous Mixture (LGVM), in which disturbances and initial state uncertainty are described as convex combinations (mixtures) of nominal Gaussian distributions and unknown vacuous distributions, is adopted. A novel control approach is then derived, according to a worst-case paradigm, by minimizing the upper expectation of a finite-horizon quadratic cost functional with respect to all admissible probability distributions and exploiting a receding horizon strategy. Simulation experiments demonstrate its robustness in presence of large unexpected impulsive disturbances.
Published in Proc. of the 18th IFAC World Congress, pp. 4606–4613.
Robust stochastic control based on imprecise probabilities
@INPROCEEDINGS{benavoli2011a,
title = {Robust stochastic control based on imprecise probabilities},
booktitle = {Proc. {o}f the 18th {IFAC} World Congress},
author = {Benavoli, A. and Chisci, L.},
pages = {4606--4613},
year = {2011}
}
Download
Benavoli, A., Ristic, B. (2011). Classification with imprecise likelihoods: a comparison of TBM, random set and imprecise probability approach. In Information Fusion (FUSION), 2011 Proc. of the 14th International Conference on, pp. 1–8.
Classification with imprecise likelihoods: a comparison of TBM, random set and imprecise probability approach
Authors: Benavoli, A. and Ristic, B.
Year: 2011
Abstract: The problem is target classification in the circumstances where the likelihood models are imprecise. The paper highlights the differences between three suitable solutions: the Transferrable Belief model (TBM), the random set approach and the imprecise probability approach. The random set approach produces identical results to those obtained using the TBM classifier, provided that equivalent measurement models are employed. Similar classification results are also obtained using the imprecise probability theory, although the latter is more general and provides more robust framework for reasoning under uncertainty.
Published in Information Fusion (FUSION), 2011 Proc. of the 14th International Conference on, pp. 1–8.
Classification with imprecise likelihoods: a comparison of TBM, random set and imprecise probability approach
@INPROCEEDINGS{benavoli2011b,
title = {Classification with imprecise likelihoods: a comparison of {TBM}, random set and imprecise probability approach},
booktitle = {Information Fusion ({FUSION}), 2011 Proc. {o}f the 14th International Conference on},
author = {Benavoli, A. and Ristic, B.},
pages = {1--8},
year = {2011}
}
Download
Benavoli, A., Zaffalon, M. (2011). A discussion on learning and prior ignorance for sets of priors in the one-parameter exponential family. In ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, Innsbruck (AU), pp. 1–10.
A discussion on learning and prior ignorance for sets of priors in the one-parameter exponential family
Authors: Benavoli, A. and Zaffalon, M.
Year: 2011
Abstract: For a conjugate likelihood-prior model in the one parameter exponential family of distributions, we show that, by letting the parameters of the conjugate exponential prior vary in suitable sets, it is possible to define a set of conjugate priors M that guarantees prior near-ignorance without producing vacuous inferences. This result is obtained following both a behavioural and a sensitivity analysis interpretation of prior near-ignorance. We also discuss the problem of the incompatibility of learning and prior near-ignorance for sets of priors in the one-parameter exponential family of distributions in the case of imperfect observations. In particular, we prove that learning and prior near-ignorance are compatible under an imperfect observation mechanismif and only if the support of the priors inM is the whole real axis.
Published in ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, Innsbruck (AU), pp. 1–10.
A discussion on learning and prior ignorance for sets of priors in the one-parameter exponential family
@INPROCEEDINGS{benavoli2011c,
title = {A discussion on learning and prior ignorance for sets of priors in the one-parameter exponential family},
address = {Innsbruck (AU)},
booktitle = {{ISIPTA} '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications},
author = {Benavoli, A. and Zaffalon, M.},
pages = {1--10},
year = {2011},
url = {http://www.sipta.org/isipta11/proceedings/papers/s027.pdf}
}
Download
Benavoli, A., Zaffalon, M., Miranda, E. (2011). Robust filtering through coherent lower previsions. Automatic Control, IEEE Transactions on 56(7), pp. 1567–1581.
Robust filtering through coherent lower previsions
Authors: Benavoli, A. and Zaffalon, M. and Miranda, E.
Year: 2011
Abstract: The classical filtering problem is re-examined to take into account imprecision in the knowledge about the probabilistic relationships involved. Imprecision is modeled in this paper by closed convex sets of probabilities. We derive a solution of the state estimation problem under such a framework that is very general: it can deal with any closed convex set of probability distributions used to characterize uncertainty in the prior, likelihood, and state transition models. This is made possible by formulating the theory directly in terms of coherent lower previsions, that is, of the lower envelopes of the expectations obtained from the set of distributions. The general solution is specialized to two particular classes of coherent lower previsions. The first consists of a family of Gaussian distributions whose means are only known to belong to an interval. The second is the so-called linear-vacuous mixture model, which is a family made of convex combinations of a known nominal distribution (e.g., a Gaussian) with arbitrary distributions. For the latter case, we empirically compare the proposed estimator with the Kalman filter. This shows that our solution is more robust to the presence of modelling errors in the system and that, hence, appears to be a more realistic approach than the Kalman filter in such a case.
Published in Automatic Control, IEEE Transactions on 56(7), pp. 1567–1581.
Robust filtering through coherent lower previsions
@ARTICLE{benavoli2011d,
title = {Robust filtering through coherent lower previsions},
journal = {Automatic Control, {IEEE} Transactions on},
volume = {56},
author = {Benavoli, A. and Zaffalon, M. and Miranda, E.},
number = {7},
pages = {1567--1581},
year = {2011},
doi = {10.1109/TAC.2010.2090707}
}
Download
de Campos, C.P. (2011). New complexity results for MAP in Bayesian networks. In International Joint Conference on Artificial Intelligence (IJCAI), AAAI Press, pp. 2100–2106.
New complexity results for MAP in Bayesian networks
Authors: de Campos, C.P.
Year: 2011
Abstract: This paper presents new results for the (partial) maximum a posteriori (MAP) problem in Bayesian networks, which is the problem of querying the most probable state configuration of some of the network variables given evidence. It is demonstrated that the problem remains hard even in networks with very simple topology, such as binary polytrees and simple trees (including the Naive Bayes structure), which extends previous complexity results. Furthermore, a Fully Polynomial Time Approximation Scheme for MAP in networks with bounded treewidth and bounded number of states per variable is developed. Approximation schemes were thought to be impossible, but here it is shown otherwise under the assumptions just mentioned, which are adopted in most applications.
Published in International Joint Conference on Artificial Intelligence (IJCAI), AAAI Press, pp. 2100–2106.
New complexity results for MAP in Bayesian networks
@INPROCEEDINGS{decampos2011c,
title = {New complexity results for {MAP} in {B}ayesian networks},
publisher = {AAAI Press},
booktitle = {International Joint Conference on Artificial Intelligence ({IJCAI})},
author = {de Campos, C.P.},
pages = {2100--2106},
year = {2011},
url = {http://ijcai.org/papers11/Papers/IJCAI11-351.pdf}
}
Download
de Campos, C.P., Benavoli, A. (2011). Inference with multinomial data: why to weaken the prior strength. In International Joint Conference on Artificial Intelligence (IJCAI), AAAI Press, pp. 2107–2112.
Inference with multinomial data: why to weaken the prior strength
Authors: de Campos, C.P. and Benavoli, A.
Year: 2011
Abstract: This paper considers inference from multinomial data and addresses the problem of choosing the strength of the Dirichlet prior under a mean-squared error criterion. We compare the Maxi-mum Likelihood Estimator (MLE) and the most commonly used Bayesian estimators obtained by assuming a prior Dirichlet distribution with non-informative prior parameters, that is, the parameters of the Dirichlet are equal and altogether sum up to the so called strength of the prior. Under this criterion, MLE becomes more preferable than the Bayesian estimators at the increase of the number of categories k of the multinomial, because non-informative Bayesian estimators induce a region where they are dominant that quickly shrinks with the increase of k. This can be avoided if the strength of the prior is not kept constant but decreased with the number of categories. We argue that the strength should decrease at least k times faster than usual estimators do.
Published in International Joint Conference on Artificial Intelligence (IJCAI), AAAI Press, pp. 2107–2112.
Inference with multinomial data: why to weaken the prior strength
@INPROCEEDINGS{decampos2011e,
title = {Inference with multinomial data: why to weaken the prior strength},
publisher = {AAAI Press},
booktitle = {International Joint Conference on Artificial Intelligence ({IJCAI})},
author = {de Campos, C.P. and Benavoli, A.},
pages = {2107--2112},
year = {2011},
url = {http://ijcai.org/papers11/Papers/IJCAI11-352.pdf}
}
Download
de Campos, C.P., Ji, Q. (2011). Efficient structure learning of Bayesian networks using constraints. Journal of Machine Learning Research 12, pp. 663–689.
Efficient structure learning of Bayesian networks using constraints
Authors: de Campos, C.P. and Ji, Q.
Year: 2011
Abstract: This paper addresses the problem of learning Bayesian network structures from data based on score functions that are decomposable. It describes properties that strongly reduce the time and memory costs of many known methods without losing global optimality guarantees. These properties are derived for different score criteria such as Minimum Description Length (or Bayesian Information Criterion), Akaike Information Criterion and Bayesian Dirichlet Criterion. Then a branch-and-bound algorithm is presented that integrates structural constraints with data in a way to guarantee global optimality. As an example, structural constraints are used to map the problem of structure learning in Dynamic Bayesian networks into a corresponding augmented Bayesian network. Finally, we show empirically the benefits of using the properties with state-of-the-art methods and with the new algorithm, which is able to handle larger data sets than before.
Published in Journal of Machine Learning Research 12, pp. 663–689.
Efficient structure learning of Bayesian networks using constraints
@ARTICLE{decampos2011a,
title = {Efficient structure learning of {B}ayesian networks using constraints},
journal = {Journal of Machine Learning Research},
volume = {12},
author = {de Campos, C.P. and Ji, Q.},
pages = {663--689},
year = {2011},
url = {http://jmlr.csail.mit.edu/papers/volume12/decampos11a/decampos11a.pdf}
}
Download
de Campos, C., Ji, Q. (2011). Bayesian networks and the imprecise Dirichlet model applied to recognition problems. In Liu, W. (Ed), Symbolic and Quantitative Approaches to Reasoning With Uncertainty, Lecture Notes in Computer Science 6717, Springer, Berlin / Heidelberg, pp. 158–169.
Bayesian networks and the imprecise Dirichlet model applied to recognition problems
Authors: de Campos, C. and Ji, Q.
Year: 2011
Abstract: This paper describes an Imprecise Dirichlet Model and the maximum entropy criterion to learn Bayesian network parameters under insufficient and incomplete data. The method is applied to two distinct recognition problems, namely, a facial action unit recognition and an activity recognition in video surveillance sequences. The model treats a wide range of constraints that can be specified by experts, and deals with incomplete data using an ad-hoc expectation-maximization procedure. It is also described how the same idea can be used to learn dynamic Bayesian networks. With synthetic data, we show that our proposal and widely used methods, such as the Bayesian maximum a posteriori, achieve similar accuracy. However, when real data come in place, our method performs better than the others, because it does not rely on a single prior distribution, which might be far from the best one.
Published in Liu, W. (Ed), Symbolic and Quantitative Approaches to Reasoning With Uncertainty, Lecture Notes in Computer Science 6717, Springer, Berlin / Heidelberg, pp. 158–169.
Bayesian networks and the imprecise Dirichlet model applied to recognition problems
@INPROCEEDINGS{decampos2011f,
title = {Bayesian networks and the imprecise {D}irichlet model applied to recognition problems},
editor = {Liu, W.},
publisher = {Springer, Berlin / Heidelberg},
series = {Lecture Notes in Computer Science},
volume = {6717},
booktitle = {Symbolic and Quantitative Approaches to Reasoning With Uncertainty},
author = {de Campos, C. and Ji, Q.},
pages = {158--169},
year = {2011},
doi = {10.1007/978-3-642-22152-1_14}
}
Download
de Cooman, G., Miranda, E., Zaffalon, M. (2011). Independent natural extension. Artificial Intelligence 175, pp. 1911–1950.
Independent natural extension
Authors: de Cooman, G. and Miranda, E. and Zaffalon, M.
Year: 2011
Abstract: There is no unique extension of the standard notion of probabilistic independence to the case where probabilities are indeterminate or imprecisely specified. Epistemic independence is an extension that formalises the intuitive idea of mutual irrelevance between different sources of information. This gives epistemic independence very wide scope as well as appeal: this interpretation of independence is often taken as natural also in precise-probabilistic contexts. Nevertheless, epistemic independence has received little attention so far. This paper develops the foundations of this notion for variables assuming values in finite spaces. We define (epistemically) independent products of marginals (or possibly conditionals) and show that there always is a unique least-committal such independent product, which we call the independent natural extension. We supply an explicit formula for it, and study some of its properties, such as associativity, marginalisation and external additivity, which are basic tools to work with the independent natural extension. Additionally, we consider a number of ways in which the standard factorisation formula for independence can be generalised to an imprecise-probabilistic context. We show, under some mild conditions, that when the focus is on least-committal models, using the independent natural extension is equivalent to imposing a so-called strong factorisation property. This is an important outcome for applications as it gives a simple tool to make sure that inferences are consistent with epistemic independence judgements. We discuss the potential of our results for applications in Artificial Intelligence by recalling recent work by some of us, where the independent natural extension was applied to graphical models. It has allowed, for the first time, the development of an exact linear-time algorithm for the imprecise probability updating of credal trees.
Published in Artificial Intelligence 175, pp. 1911–1950.
Independent natural extension
@ARTICLE{zaffalon2011a,
title = {Independent natural extension},
journal = {Artificial Intelligence},
volume = {175},
author = {de Cooman, G. and Miranda, E. and Zaffalon, M.},
pages = {1911--1950},
year = {2011},
doi = {10.1016/j.artint.2011.06.001}
}
Download
de Lalla, C., Rinaldi, A., Montagna, D., Azzimonti, L., Bernardo, M.E., Sangalli, L.M., Paganoni, A.M., Maccario, R., Cesare-Merlone, A.D., Zecca, M., Locatelli, F., Dellabona, P., Casorati, G. (2011). Invariant Natural Killer T-cell reconstitution in pediatric leukemia patients given HLA-haploidentical stem cell transplantation defines distinct CD4+ and CD4- subset dynamics and associates with the remission state. The Journal of Immunology 186(7), pp. 4490–4499.
Invariant Natural Killer T-cell reconstitution in pediatric leukemia patients given HLA-haploidentical stem cell transplantation defines distinct CD4+ and CD4- subset dynamics and associates with the remission state
Authors: de Lalla, C. and Rinaldi, A. and Montagna, D. and Azzimonti, L. and Bernardo, M.E. and Sangalli, L.M. and Paganoni, A.M. and Maccario, R. and Cesare-Merlone, A.D. and Zecca, M. and Locatelli, F. and Dellabona, P. and Casorati, G.
Year: 2011
Abstract: Immune reconstitution plays a crucial role on the outcome of patients given T cell-depleted HLA-haploidentical hematopoietic stem cell transplantation (hHSCT) for hematological malignancies. CD1d-restricted invariant NKT (iNKT) cells are innate-like, lipid-reactive T lymphocytes controlling infections, cancer, and autoimmunity. Adult mature iNKT cells are divided in two functionally distinct CD4+ and CD4- subsets that express the NK receptor CD161 and derive from thymic CD4+CD161- precursors. We investigated iNKT cell reconstitution dynamics in 33 pediatric patients given hHSCT for hematological malignancies, with a follow-up reaching 6 y posttransplantation, and correlated their emergence with disease relapse. iNKT cells fully reconstitute and rapidly convert into IFN-γ{\textendash}expressing effectors in the 25 patients maintaining remission. CD4+ cells emerge earlier than the CD4- ones, both displaying CD161- immature phenotypes. CD4- cells expand more slowly than CD4+ cells, though they mature with significantly faster kinetics, reaching full maturation by 18 mo post-hHSCT. Between 4 and 6 y post-hHSCT, mature CD4- iNKT cells undergo a substantial expansion burst, resulting in a CD4+\
Published in The Journal of Immunology 186(7), pp. 4490–4499.
Invariant Natural Killer T-cell reconstitution in pediatric leukemia patients given HLA-haploidentical stem cell transplantation defines distinct CD4+ and CD4- subset dynamics and associates with the remission state
@ARTICLE{azzimonti2011a,
title = {Invariant {N}atural {K}iller {T}-cell reconstitution in pediatric leukemia patients given {HLA}-haploidentical stem cell transplantation defines distinct {CD4+} and {CD4}- subset dynamics and associates with the remission state},
journal = {The Journal of Immunology},
volume = {186},
author = {de Lalla, C. and Rinaldi, A. and Montagna, D. and Azzimonti, L. and Bernardo, M.E. and Sangalli, L.M. and Paganoni, A.M. and Maccario, R. and Cesare-Merlone, A.D. and Zecca, M. and Locatelli, F. and Dellabona, P. and Casorati, G.},
number = {7},
pages = {4490--4499},
year = {2011},
doi = {10.4049/jimmunol.1003748}
}
Download
Mauá, D.D., de Campos, C.P. (2011). Solving decision problems with limited information. In Shawe-Taylor, J., Zemel, R.S., Bartlett, P., Pereira, F.C.N., Weinberger, K.Q. (Eds), Advances in Neural Information Processing Systems 24 (NIPS 2011), pp. 603–611.
Solving decision problems with limited information
Authors: Mauá, D.D. and de Campos, C.P.
Year: 2011
Abstract: We present a new algorithm for exactly solving decision-making problems represented as an influence diagram. We do not require the usual assumptions of no forgetting and regularity, which allows us to solve problems with limited information. The algorithm, which implements a sophisticated variable elimination procedure, is empirically shown to outperform a state-of-the-art algorithm in randomly generated problems of up to 150 variables and 1064 strategies.
Published in Shawe-Taylor, J., Zemel, R.S., Bartlett, P., Pereira, F.C.N., Weinberger, K.Q. (Eds), Advances in Neural Information Processing Systems 24 (NIPS 2011), pp. 603–611.
Solving decision problems with limited information
@INCOLLECTION{maua2011a,
title = {Solving decision problems with limited information},
editor = {Shawe-Taylor, J. and Zemel, R.S. and Bartlett, P. and Pereira, F.C.N. and Weinberger, K.Q.},
booktitle = {Advances in Neural Information Processing Systems 24 ({NIPS} 2011)},
author = {Mau\'a, D.D. and de Campos, C.P.},
pages = {603--611},
year = {2011},
url = {http://books.nips.cc/papers/files/nips24/NIPS2011_0422.pdf}
}
Download
Mauá, D.D., de Campos, C.P., Zaffalon, M. (2011). A fully polynomial time approximation scheme for updating credal networks of bounded treewidth and number of variable states. In Coolen, F., de Cooman, G., Fetz, T., Oberguggenberger, M. (Eds), ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, Innsbruck, Austria, pp. 277–286.
A fully polynomial time approximation scheme for updating credal networks of bounded treewidth and number of variable states
Authors: Mauá, D.D. and de Campos, C.P. and Zaffalon, M.
Year: 2011
Abstract: Credal networks lift the precise probability assumption of Bayesian networks, enabling a richer representation of uncertainty in the form of closed convex sets of probability measures. The increase in expressiveness comes at the expense of higher computational costs. In this paper we present a new algorithm which is an extension of the well-known variable elimination algorithm for computing posterior inferences in extensively specified credal networks. The algorithm efficiency is empirically shown to outperform a state-of-the-art algorithm. We then provide the first fully polynomial time approximation scheme for inference in credal networks with bounded treewidth and number of states per variable.
Published in Coolen, F., de Cooman, G., Fetz, T., Oberguggenberger, M. (Eds), ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, Innsbruck, Austria, pp. 277–286.
A fully polynomial time approximation scheme for updating credal networks of bounded treewidth and number of variable states
@INPROCEEDINGS{maua2011b,
title = {A fully polynomial time approximation scheme for updating credal networks of bounded treewidth and number of variable states},
editor = {Coolen, F. and de Cooman, G. and Fetz, T. and Oberguggenberger, M.},
publisher = {SIPTA},
address = {Innsbruck, Austria},
booktitle = {{ISIPTA} '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications},
author = {Mau\'a, D.D. and de Campos, C.P. and Zaffalon, M.},
pages = {277--286},
year = {2011},
url = {http://leo.ugr.es/sipta/isipta11/proceedings/papers/s035.pdf}
}
Download
Miranda, E., Zaffalon, M., de Cooman, G. (2011). Conglomerable natural extension. In Coolen, F., de Cooman, G., Fetz, T., Oberguggenberger, M. (Eds), ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 287–296.
Conglomerable natural extension
Authors: Miranda, E. and Zaffalon, M. and de Cooman, G.
Year: 2011
Abstract: We study the weakest conglomerable model that is implied by desirability or probability assessments: the conglomerable natural extension. We show that taking the natural extension of the assessments while imposing conglomerability—the procedure adopted in Walley's theory—does not yield, in general, the conglomerable natural extension (but it does so in the case of the marginal extension). Iterating this process produces a sequence of models that approach the conglomerable natural extension, although it is not known, at this point, whether it is attained in the limit. We give sufficient conditions for this to happen in some special cases, and study the differences between working with coherent sets of desirable gambles and coherent lower previsions. Our results indicate that it might be necessary to re-think the foundations of Walley's theory of coherent conditional lower previsions for infinite partitions of conditioning events.
Published in Coolen, F., de Cooman, G., Fetz, T., Oberguggenberger, M. (Eds), ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 287–296.
Conglomerable natural extension
@INPROCEEDINGS{zaffalon2011c,
title = {Conglomerable natural extension},
editor = {Coolen, F. and de Cooman, G. and Fetz, T. and Oberguggenberger, M.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications},
author = {Miranda, E. and Zaffalon, M. and de Cooman, G.},
pages = {287--296},
year = {2011},
url = {http://www.sipta.org/isipta11/proceedings/papers/s030.pdf}
}
Download
Rinaldi, A., Mian, M., Chigrinova, E., Arcaini, L., Bhagat, G., Novak, U., Rancoita, P.M.V., Campos, C.P.D., Forconi, F., Gascoyne, R.D., Facchetti, F., Ponzoni, M., Govi, S., Ferreri, A.J.M., Mollejo, M., Piris, M.A., Baldini, L., Soulier, J., Thieblemont, C., Canzonieri, V., Gattei, V., Marasca, R., Franceschetti, S., Gaidano, G., Tucci, A., Uccella, S., Tibiletti, M.G., Dirnhofer, S., Tripodo, C., Doglioni, C., Favera, R.D., Cavalli, F., Zucca, E., Kwee, I., Bertoni, F. (2011). Genome-wide DNA profiling of marginal zone lymphomas identifies subtype-specific lesions with an impact on the clinical outcome. Blood 117(5), pp. 1595–1604.
Genome-wide DNA profiling of marginal zone lymphomas identifies subtype-specific lesions with an impact on the clinical outcome
Authors: Rinaldi, A. and Mian, M. and Chigrinova, E. and Arcaini, L. and Bhagat, G. and Novak, U. and Rancoita, P.M.V. and Campos, C.P.D. and Forconi, F. and Gascoyne, R.D. and Facchetti, F. and Ponzoni, M. and Govi, S. and Ferreri, A.J.M. and Mollejo, M. and Piris, M.A. and Baldini, L. and Soulier, J. and Thieblemont, C. and Canzonieri, V. and Gattei, V. and Marasca, R. and Franceschetti, S. and Gaidano, G. and Tucci, A. and Uccella, S. and Tibiletti, M.G. and Dirnhofer, S. and Tripodo, C. and Doglioni, C. and Favera, R.D. and Cavalli, F. and Zucca, E. and Kwee, I. and Bertoni, F.
Year: 2011
Abstract: Marginal zone B-cell lymphomas (MZLs) have been divided into 3 distinct subtypes (extranodal MZLs of mucosa-associated lymphoid tissue [MALT] type, nodal MZLs, and splenic MZLs). Nevertheless, the relationship between the subtypes is still unclear. We performed a comprehensive analysis of genomic DNA copy number changes in a very large series of MZL cases with the aim of addressing this question. Samples from 218 MZL patients (25 nodal, 57 MALT, 134 splenic, and 2 not better specified MZLs) were analyzed with the Affymetrix Human Mapping 250K SNP arrays, and the data combined with matched gene expression in 33 of 218 cases. MALT lymphoma presented significantly more frequently gains at 3p, 6p, 18p, and del(6q23) (TNFAIP3/A20), whereas splenic MZLs was associated with del(7q31), del(8p). Nodal MZLs did not show statistically significant differences compared with MALT lymphoma while lacking the splenic MZLs-related 7q losses. Gains of 3q and 18q were common to all 3 subtypes. del(8p) was often present together with del(17p) (TP53). Although del(17p) did not determine a worse outcome and del(8p) was only of borderline significance, the presence of both deletions had a highly significant negative impact on the outcome of splenic MZLs.
Published in Blood 117(5), The American Society of Hematology, pp. 1595–1604.
Genome-wide DNA profiling of marginal zone lymphomas identifies subtype-specific lesions with an impact on the clinical outcome
@ARTICLE{decampos2011b,
title = {Genome-wide {DNA} profiling of marginal zone lymphomas identifies subtype-specific lesions with an impact on the clinical outcome},
journal = {Blood},
publisher = {The American Society of Hematology},
volume = {117},
author = {Rinaldi, A. and Mian, M. and Chigrinova, E. and Arcaini, L. and Bhagat, G. and Novak, U. and Rancoita, P.M.V. and Campos, C.P.D. and Forconi, F. and Gascoyne, R.D. and Facchetti, F. and Ponzoni, M. and Govi, S. and Ferreri, A.J.M. and Mollejo, M. and Piris, M.A. and Baldini, L. and Soulier, J. and Thieblemont, C. and Canzonieri, V. and Gattei, V. and Marasca, R. and Franceschetti, S. and Gaidano, G. and Tucci, A. and Uccella, S. and Tibiletti, M.G. and Dirnhofer, S. and Tripodo, C. and Doglioni, C. and Favera, R.D. and Cavalli, F. and Zucca, E. and Kwee, I. and Bertoni, F.},
number = {5},
pages = {1595--1604},
year = {2011},
doi = {10.1182/blood-2010-01-264275},
url = {http://bloodjournal.hematologylibrary.org/content/117/5/1595.full.pdf+html}
}
Download
Zaffalon, M., Corani, G., Mauá, D.D. (2011). Utility-based accuracy measures to empirically evaluate credal classifiers. In Coolen, F., de Cooman, G., Fetz, T., Oberguggenberger, M. (Eds), ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 401–410.
Utility-based accuracy measures to empirically evaluate credal classifiers
Authors: Zaffalon, M. and Corani, G. and Mauá, D.D.
Year: 2011
Abstract: Predictions made by imprecise-probability models are often indeterminate (that is, set-valued). Measuring the quality of an indeterminate prediction by a single number is important to fairly compare different models, but a principled approach to this problem is currently missing. In this paper we derive a measure to evaluate the predictions of credal classifiers from a set of assumptions. The measure turns out to be made of an objective component, and another that is related to the decision-maker's degree of risk-aversion. We discuss when the measure can be rendered independent of such a degree, and provide insights as to how the comparison of classifiers based on the new measure changes with the number of predictions to be made. Finally, we empirically study the behavior of the proposed measure.
Published in Coolen, F., de Cooman, G., Fetz, T., Oberguggenberger, M. (Eds), ISIPTA '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 401–410.
Utility-based accuracy measures to empirically evaluate credal classifiers
@INPROCEEDINGS{zaffalon2011b,
title = {Utility-based accuracy measures to empirically evaluate credal classifiers},
editor = {Coolen, F. and de Cooman, G. and Fetz, T. and Oberguggenberger, M.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '11: Proceedings of the Seventh International Symposium on Imprecise Probability: Theories and Applications},
author = {Zaffalon, M. and Corani, G. and Mau\'a, D.D.},
pages = {401--410},
year = {2011},
url = {http://www.sipta.org/isipta11/proceedings/papers/s016.pdf}
}
Download top2010
Antonucci, A., Cuzzolin, F. (2010). Credal sets approximation by lower probabilities: application to credal networks. In Hüllermeier, E., Kruse, R., Hoffmann, F. (Eds), Computational Intelligence for Knowledge-based Systems Design, 13th International Conference on Information Processing and Management of Uncertainty, IPMU 2010, Dortmund, Germany, June 28 - July 2, 2010. Proceedings, Lecture Notes in Computer Science 6178, Springer, pp. 716–725.
Credal sets approximation by lower probabilities: application to credal networks
Authors: Antonucci, A. and Cuzzolin, F.
Year: 2010
Abstract: Credal sets are closed convex sets of probability mass functions. The lower probabilities specified by a credal set for each element of the power set can be used as constraints defining a second credal set. This simple procedure produces an outer approximation, with a bounded number of extreme points, for general credal sets. The approximation is optimal in the sense that no other lower probabilities can specify smaller supersets of the original credal set. Notably, in order to be computed, the approximation does not need the extreme points of the credal set, but only its lower probabilities. This makes the approximation particularly suited for credal networks, which are a generalization of Bayesian networks based on credal sets. Although most of the algorithms for credal networks updating only return lower posterior probabilities, the suggested approximation can be used to evaluate (as an outer approximation of) the posterior credal set. This makes it possible to adopt more sophisticated decision making criteria, without having to replace existing algorithms. The quality of the approximation is investigated by numerical tests.
Published in Hüllermeier, E., Kruse, R., Hoffmann, F. (Eds), Computational Intelligence for Knowledge-based Systems Design, 13th International Conference on Information Processing and Management of Uncertainty, IPMU 2010, Dortmund, Germany, June 28 - July 2, 2010. Proceedings, Lecture Notes in Computer Science 6178, Springer, pp. 716–725.
Credal sets approximation by lower probabilities: application to credal networks
@INPROCEEDINGS{antonucci2010a,
title = {Credal sets approximation by lower probabilities: application to credal networks},
editor = {H\"ullermeier, E. and Kruse, R. and Hoffmann, F.},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
volume = {6178},
booktitle = {Computational Intelligence for Knowledge-{b}ased Systems Design, 13th International Conference on Information Processing and Management of Uncertainty, {IPMU} 2010, Dortmund, Germany, June 28 - July 2, 2010. Proceedings},
author = {Antonucci, A. and Cuzzolin, F.},
pages = {716--725},
year = {2010},
doi = {10.1007/978-3-642-14049-5_73}
}
Download
Antonucci, A., Yi, S., de Campos, C.P., Zaffalon, M. (2010). Generalized loopy 2U: a new algorithm for approximate inference in credal networks. International Journal of Approximate Reasoning 55(5), pp. 474–484.
Generalized loopy 2U: a new algorithm for approximate inference in credal networks
Authors: Antonucci, A. and Yi, S. and de Campos, C.P. and Zaffalon, M.
Year: 2010
Abstract: Credal networks generalize Bayesian networks by relaxing the requirement of precision of probabilities. Credal networks are considerably more expressive than Bayesian networks, but this makes belief updating NP-hard even on polytrees. We develop a new efficient algorithm for approximate belief updating in credal networks. The algorithm is based on an important representation result we prove for general credal networks: that any credal network can be equivalently reformulated as a credal network with binary variables; moreover, the transformation, which is considerably more complex than in the Bayesian case, can be implemented in polynomial time. The equivalent binary credal network is then updated by L2U, a loopy approximate algorithm for binary credal networks. Overall, we generalize L2U to non-binary credal networks, obtaining a scalable algorithm for the general case, which is approximate only because of its loopy nature. The accuracy of the inferences with respect to other state-of-the-art algorithms is evaluated by extensive numerical tests.
Published in International Journal of Approximate Reasoning 55(5), pp. 474–484.
Generalized loopy 2U: a new algorithm for approximate inference in credal networks
@ARTICLE{antonucci2010c,
title = {Generalized loopy {2U}: a new algorithm for approximate inference in credal networks},
journal = {International Journal of Approximate Reasoning},
volume = {55},
author = {Antonucci, A. and Yi, S. and de Campos, C.P. and Zaffalon, M.},
number = {5},
pages = {474--484},
year = {2010},
doi = {10.1016/j.ijar.2010.01.007}
}
Download
Benavoli, A., Antonucci, A. (2010). Aggregating imprecise probabilistic knowledge: application to Zadeh's paradox and sensor networks. Int. Journal of Approximate Reasoning 51(9), pp. 1014–1028.
Aggregating imprecise probabilistic knowledge: application to Zadeh's paradox and sensor networks
Authors: Benavoli, A. and Antonucci, A.
Year: 2010
Abstract: The problem of aggregating two or more sources of information containing knowledge about a common domain is considered. We propose an aggregation framework for the case where the available information is modelled by coherent lower previsions, corresponding to convex sets of probability mass functions. The consistency between aggregated beliefs and sources of information is discussed. A closed formula, which specializes our rule to a particular class of models, is also derived. Two applications consisting in a possible explanation of Zadeh's paradox and an algorithm for estimation fusion in sensor networks are finally reported.
Published in Int. Journal of Approximate Reasoning 51(9), pp. 1014–1028.
Aggregating imprecise probabilistic knowledge: application to Zadeh's paradox and sensor networks
@ARTICLE{benavoli2010a,
title = {Aggregating imprecise probabilistic knowledge: application to {Z}adeh's paradox and sensor networks},
journal = {Int. Journal of Approximate Reasoning},
volume = {51},
author = {Benavoli, A. and Antonucci, A.},
number = {9},
pages = {1014--1028},
year = {2010},
doi = {10.1016/j.ijar.2010.08.010}
}
Download
de Campos, C.P., Ji, Q. (2010). Properties of Bayesian Dirichlet scores to learn Bayesian network structures. In AAAI Conference on Artificial Intelligence, AAAI Press, pp. 431–436.
Properties of Bayesian Dirichlet scores to learn Bayesian network structures
Authors: de Campos, C.P. and Ji, Q.
Year: 2010
Abstract: This paper addresses exact learning of Bayesian network structure from data based on the Bayesian Dirichlet score function and its derivations. We describe useful properties that strongly reduce the computational costs of many known methods without losing global optimality guarantees. We show empirically the advantages of the properties in terms of time and memory consumptions, demonstrating that state-of-the-art methods, with the use of such properties, might handle larger data sets than those currently possible.
Published in AAAI Conference on Artificial Intelligence, AAAI Press, pp. 431–436.
Properties of Bayesian Dirichlet scores to learn Bayesian network structures
@INPROCEEDINGS{decampos2010c,
title = {Properties of {B}ayesian {D}irichlet scores to learn {B}ayesian network structures},
publisher = {AAAI Press},
booktitle = {{AAAI} Conference on Artificial Intelligence},
author = {de Campos, C.P. and Ji, Q.},
pages = {431--436},
year = {2010},
url = {http://www.aaai.org/ocs/index.php/AAAI/AAAI10/paper/view/1704/2013}
}
Download
de Campos, C.P., Zeng, Z., Ji, Q. (2010). An improved structural EM to learn dynamic Bayesian nets. In 20th International Conference on Pattern Recognition (ICPR), pp. 601–604.
An improved structural EM to learn dynamic Bayesian nets
Authors: de Campos, C.P. and Zeng, Z. and Ji, Q.
Year: 2010
Abstract: This paper addresses the problem of learning structure of Bayesian and Dynamic Bayesian networks from incomplete data based on the Bayesian Information Criterion. We describe a procedure to map the problem of the dynamic case into a corresponding augmented Bayesian network through the use of structural constraints. Because the algorithm is exact and anytime, it is well suitable for a structural Expectation-Maximization (EM) method where the only source of approximation is due to the EM itself. We show empirically that the use a global maximizer inside the structural EM is computationally feasible and leads to more accurate models.
Published in 20th International Conference on Pattern Recognition (ICPR), pp. 601–604.
An improved structural EM to learn dynamic Bayesian nets
@INPROCEEDINGS{decampos2010d,
title = {An improved structural {EM} to learn dynamic {B}ayesian nets},
booktitle = {20th International Conference on Pattern Recognition ({ICPR})},
author = {de Campos, C.P. and Zeng, Z. and Ji, Q.},
pages = {601--604},
year = {2010},
doi = {10.1109/ICPR.2010.152}
}
Download
de Cooman, G., Hermans, F., Antonucci, A., Zaffalon, M. (2010). Epistemic irrelevance in credal nets: the case of imprecise markov trees. International Journal of Approximate Reasoning 51(9), pp. 1029–1052.
Epistemic irrelevance in credal nets: the case of imprecise markov trees
Authors: de Cooman, G. and Hermans, F. and Antonucci, A. and Zaffalon, M.
Year: 2010
Abstract: We focus on credal nets, which are graphical models that generalise Bayesian nets to imprecise probability. We replace the notion of strong independence commonly used in credal nets with the weaker notion of epistemic irrelevance, which is arguably more suited for a behavioural theory of probability. Focusing on directed trees, we show how to combine the given local uncertainty models in the nodes of the graph into a global model, and we use this to construct and justify an exact message-passing algorithm that computes updated beliefs for a variable in the tree. The algorithm, which is linear in the number of nodes, is formulated entirely in terms of coherent lower previsions, and is shown to satisfy a number of rationality requirements. We supply examples of the algorithm's operation, and report an application to on-line character recognition that illustrates the advantages of our approach for prediction. We comment on the perspectives, opened by the availability, for the first time, of a truly efficient algorithm based on epistemic irrelevance.
Published in International Journal of Approximate Reasoning 51(9), pp. 1029–1052.
Epistemic irrelevance in credal nets: the case of imprecise markov trees
@ARTICLE{antonucci2010b,
title = {Epistemic irrelevance in credal nets: the case of imprecise markov trees},
journal = {International Journal of Approximate Reasoning},
volume = {51},
author = {de Cooman, G. and Hermans, F. and Antonucci, A. and Zaffalon, M.},
number = {9},
pages = {1029--1052},
year = {2010},
doi = {10.1016/j.ijar.2010.08.011}
}
Download
de Cooman, G., Miranda, E., Zaffalon, M. (2010). Factorisation properties of the strong product. In Borgelt, C., González Rodrìguez, G., Trutschnig, W., Asunción Lubiano, M., Gil, M.A., Grzegorzewski, P., Hryniewicz, O. (Eds), Combining Soft Computing and Statistical Methods in Data Analysis, Advances in Intelligent and Soft Computing 77, Springer, pp. 139--147.
Factorisation properties of the strong product
Authors: de Cooman, G. and Miranda, E. and Zaffalon, M.
Year: 2010
Abstract: We investigate a number of factorisation conditions in the framework of sets of probability measures, or coherent lower previsions, with finite referential spaces. We show that the so-called strong product constitutes one way to combine a number of marginal coherent lower previsions into an independent joint lower prevision, and we prove that under some conditions it is the only independent product that satisfies the factorisation conditions.
Published in Borgelt, C., González Rodrìguez, G., Trutschnig, W., Asunción Lubiano, M., Gil, M.A., Grzegorzewski, P., Hryniewicz, O. (Eds), Combining Soft Computing and Statistical Methods in Data Analysis, Advances in Intelligent and Soft Computing 77, Springer, pp. 139--147.
Factorisation properties of the strong product
@INPROCEEDINGS{zaffalon2010c,
title = {Factorisation properties of the strong product},
editor = {Borgelt, C. and Gonz\'alez Rodr\`iguez, G. and Trutschnig, W. and Asunci\'on Lubiano, M. and Gil, M.A. and Grzegorzewski, P. and Hryniewicz, O.},
publisher = {Springer},
series = {Advances in Intelligent and Soft Computing},
volume = {77},
booktitle = {Combining Soft Computing and Statistical Methods in Data Analysis},
author = {de Cooman, G. and Miranda, E. and Zaffalon, M.},
pages = {139--147},
year = {2010},
doi = {10.1007/978-3-642-14746-3_18}
}
Download
de Cooman, G., Miranda, E., Zaffalon, M. (2010). Independent natural extension. In Hüllermeier, E., Kruse, R., Hoffmann, F. (Eds), Computational Intelligence for Knowledge-based Systems Design, Lecture Notes in Computer Science 6178, Springer, pp. 737–746.
Independent natural extension
Authors: de Cooman, G. and Miranda, E. and Zaffalon, M.
Year: 2010
Abstract: We introduce a general definition for the independence of a number of finite-valued variables, based on coherent lower previsions. Our definition has an epistemic flavour: it arises from personal judgements that a number of variables are irrelevant to one another. We show that a number of already existing notions, such as strong independence, satisfy our definition. Moreover, there always is a least-committal independent model, for which we provide an explicit formula: the independent natural extension. Our central result is that the independent natural extension satisfies so-called marginalisation, associativity and strong factorisation properties. These allow us to relate our research to more traditional ways of defining independence based on factorisation.
Published in Hüllermeier, E., Kruse, R., Hoffmann, F. (Eds), Computational Intelligence for Knowledge-based Systems Design, Lecture Notes in Computer Science 6178, Springer, pp. 737–746.
Independent natural extension
@INPROCEEDINGS{zaffalon2010b,
title = {Independent natural extension},
editor = {H\"ullermeier, E. and Kruse, R. and Hoffmann, F.},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
volume = {6178},
booktitle = {Computational Intelligence for Knowledge-{b}ased Systems Design},
author = {de Cooman, G. and Miranda, E. and Zaffalon, M.},
pages = {737--746},
year = {2010},
doi = {10.1007/978-3-642-14049-5_75}
}
Download
Corani, G., Benavoli, A. (2010). Restricting the IDM for classification. In Hullermeier, E., Kruse, R., Hoffmann, F. (Eds), Information Processing and Management of Uncertainty in Knowledge-based Systems. Theory and Methods, Communications in Computer and Information Science 80, Springer, Berlin / Heidelberg, pp. 328–337.
Restricting the IDM for classification
Authors: Corani, G. and Benavoli, A.
Year: 2010
Abstract: The naive credal classifier (NCC) extends naive Bayes classifier (NBC) to imprecise probabilities to robustly deal with the specification of the prior; NCC models a state of ignorance by using a set of priors, which is formalized by Walley's Imprecise Dirichlet Model (IDM). NCC has been shown to return more robust classification than NBC. However, there are particular situations (which we precisely characterize in the paper) under which the extreme densities included by the IDM force NCC to become very indeterminate, although NBC is able to issue accurately classifications. In this paper, we propose two approaches which overcome this issue, by restricting the set of priors of the IDM . We analyze both approaches theoretically and experimentally.
Published in Hullermeier, E., Kruse, R., Hoffmann, F. (Eds), Information Processing and Management of Uncertainty in Knowledge-based Systems. Theory and Methods, Communications in Computer and Information Science 80, Springer, Berlin / Heidelberg, pp. 328–337.
Note: 10.1007/978-3-642-14055-6_34
Restricting the IDM for classification
@INCOLLECTION{corani2010a,
title = {Restricting the {IDM} for classification},
editor = {Hullermeier, E. and Kruse, R. and Hoffmann, F.},
publisher = {Springer, Berlin / Heidelberg},
series = {Communications in Computer and Information Science},
volume = {80},
booktitle = {Information Processing and Management of Uncertainty in Knowledge-{b}ased Systems. Theory and Methods},
author = {Corani, G. and Benavoli, A.},
pages = {328--337},
year = {2010},
doi = {10.1007/978-3-642-14055-6_34}
}
Download
Corani, G., de Campos, C.P. (2010). A tree augmented classifier based on extreme imprecise Dirichlet model. International Journal of Approximate Reasoning 51(9), pp. 1053–1068.
A tree augmented classifier based on extreme imprecise Dirichlet model
Authors: Corani, G. and de Campos, C.P.
Year: 2010
Abstract: We present TANC, a TAN classifier (tree-augmented naive) based on imprecise probabilities. TANC models prior near-ignorance via the Extreme Imprecise Dirichlet Model (EDM). A first contribution of this paper is the experimental comparison between EDM and the global Imprecise Dirichlet Model using the naive credal classifier (NCC), with the aim of showing that EDM is a sensible approximation of the global IDM. TANC is able to deal with missing data in a conservative manner by considering all possible completions (without assuming them to be missing-at-random), but avoiding an exponential increase of the computational time. By experiments on real data sets, we show that TANC is more reliable than the Bayesian TAN and that it provides better performance compared to previous TANs based on imprecise probabilities. Yet, TANC is sometimes outperformed by NCC because the learned TAN structures are too complex; this calls for novel algorithms for learning the TAN structures
Published in International Journal of Approximate Reasoning 51(9), pp. 1053–1068.
A tree augmented classifier based on extreme imprecise Dirichlet model
@ARTICLE{Corani2010b,
title = {A tree augmented classifier based on extreme imprecise {D}irichlet model},
journal = {International Journal of Approximate Reasoning},
volume = {51},
author = {Corani, G. and de Campos, C.P.},
number = {9},
pages = {1053--1068},
year = {2010},
doi = {10.1016/j.ijar.2010.08.007}
}
Download
Corani, G., Giusti, A., Migliore, D., Schmidhuber, J. (2010). Robust texture recognition using credal classifiers. In Labrosse, F., Zwiggelaar, R., Liu, Y., Tiddeman, B. (Eds), Proceedings of the British Machine Vision Conference, BMVA Press, pp. 78.1–78.10.
Robust texture recognition using credal classifiers
Authors: Corani, G. and Giusti, A. and Migliore, D. and Schmidhuber, J.
Year: 2010
Abstract: Texture classification is used for many vision systems; in this paper we focus on improving the reliability of the classification through the so-called imprecise (or credal) classifiers, which suspend the judgment on the doubtful instances by returning a set of classes instead of a single class. Our view is that on critical instances it is more sensible to return a reliable set of classes rather than an unreliable single class. We compare the traditional naive Bayes classifier (NBC) against its imprecise counterpart, the naive credal classifier (NCC); we consider a standard classification dataset, when the problem is made progressively harder by introducing different image degradations or by providing smaller training sets. Experiments show that on the instances for which NCC returns more classes, NBC issues in fact unreliable classifications; the indeterminate classifications of NCC preserve reliability but at the same time also convey significant information
Published in Labrosse, F., Zwiggelaar, R., Liu, Y., Tiddeman, B. (Eds), Proceedings of the British Machine Vision Conference, BMVA Press, pp. 78.1–78.10.
Robust texture recognition using credal classifiers
@INPROCEEDINGS{corani2010c,
title = {Robust texture recognition using credal classifiers},
editor = {Labrosse, F. and Zwiggelaar, R. and Liu, Y. and Tiddeman, B.},
publisher = {BMVA Press},
booktitle = {Proceedings of the British Machine Vision Conference},
author = {Corani, G. and Giusti, A. and Migliore, D. and Schmidhuber, J.},
pages = {78.1--78.10},
year = {2010},
doi = {10.5244/C.24.78}
}
Download
Miranda, E., Zaffalon, M. (2010). Notes on desirability and conditional lower previsions. Annals of Mathematics and Artificial Intelligence 60(3–4), pp. 251–309.
Notes on desirability and conditional lower previsions
Authors: Miranda, E. and Zaffalon, M.
Year: 2010
Abstract: We detail the relationship between sets of desirable gambles and conditional lower previsions. The former is one the most general models of uncertainty. The latter corresponds to Walley's celebrated theory of imprecise probability. We consider two avenues: when a collection of conditional lower previsions is derived from a set of desirable gambles, and its converse. In either case, we relate the properties of the derived model with those of the originating one. Our results constitute basic tools to move from one formalism to the other, and thus to take advantage of work done in the two fronts.
Published in Annals of Mathematics and Artificial Intelligence 60(3–4), pp. 251–309.
Notes on desirability and conditional lower previsions
@ARTICLE{zaffalon2010e,
title = {Notes on desirability and conditional lower previsions},
journal = {Annals of Mathematics and Artificial Intelligence},
volume = {60},
author = {Miranda, E. and Zaffalon, M.},
number = {3--4},
pages = {251--309},
year = {2010},
doi = {10.1007/s10472-011-9231-4}
}
Download
Miranda, E., Zaffalon, M. (2010). Conditional models: coherence and inference through sequences of joint mass functions. Journal of Statistical Planning and Inference 140(7), pp. 1805–1833.
Conditional models: coherence and inference through sequences of joint mass functions
Authors: Miranda, E. and Zaffalon, M.
Year: 2010
Abstract: We call a conditional model any set of statements made of conditional probabilities or expectations. We take conditional models as primitive compared to unconditional probability, in the sense that conditional statements do not need to be derived from an unconditional probability. We focus on two problems: (coherence) giving conditions to guarantee that a conditional model is self-consistent; (inference) delivering methods to derive new probabilistic statements from a self-consistent conditional model. We address these problems in the case where the probabilistic statements can be specified imprecisely through sets of probabilities, while restricting the attention to finite spaces of possibilities. Using Walley's theory of coherent lower previsions, we fully characterise the question of coherence, and specialise it for the case of precisely specified probabilities, which is the most common case addressed in the literature. This shows that coherent conditional models are equivalent to sequences of (possibly sets of) unconditional mass functions. In turn, this implies that the inferences from a conditional model are the limits of the conditional inferences obtained by applying Bayes' rule, when possible, to the elements of the sequence. In doing so, we unveil the tight connection between conditional models and zero-probability events.
Published in Journal of Statistical Planning and Inference 140(7), pp. 1805–1833.
Conditional models: coherence and inference through sequences of joint mass functions
@ARTICLE{zaffalon2010a,
title = {Conditional models: coherence and inference through sequences of joint mass functions},
journal = {Journal of Statistical Planning and Inference},
volume = {140},
author = {Miranda, E. and Zaffalon, M.},
number = {7},
pages = {1805--1833},
year = {2010},
doi = {10.1016/j.jspi.2010.01.005}
}
Download
Pelessoni, R., Vicig, P., Zaffalon, M. (2010). Inference and risk measurement with the pari-mutuel model. International Journal of Approximate Reasoning 51(9), pp. 1145–1158.
Inference and risk measurement with the pari-mutuel model
Authors: Pelessoni, R. and Vicig, P. and Zaffalon, M.
Year: 2010
Abstract: We explore generalizations of the pari-mutuel model (PMM), a formalization of an intuitive way of assessing an upper probability from a precise one. We discuss a naive extension of the PMM considered in insurance, compare the PMM with a related model, the Total Variation Model, and generalize the natural extension of the PMM introduced by P. Walley and other pertained formulae. The results are subsequently given a risk measurement interpretation: in particular it is shown that a known risk measure, Tail Value at Risk (TVaR), is derived from the PMM, and a coherent risk measure more general than TVaR from its imprecise version. We analyze further the conditions for coherence of a related risk measure, Conditional Tail Expectation. Conditioning with the PMM is investigated too, computing its natural extension, characterising its dilation and studying the weaker concept of imprecision increase.
Published in International Journal of Approximate Reasoning 51(9), pp. 1145–1158.
Inference and risk measurement with the pari-mutuel model
@ARTICLE{zaffalon2010d,
title = {Inference and risk measurement with the pari-mutuel model},
journal = {International Journal of Approximate Reasoning},
volume = {51},
author = {Pelessoni, R. and Vicig, P. and Zaffalon, M.},
number = {9},
pages = {1145--1158},
year = {2010},
doi = {10.1016/j.ijar.2010.08.005}
}
Download
Piatti, A., Antonucci, A., Zaffalon, M. (2010). Building knowledge-based expert systems by credal networks: a tutorial. In Baswell, A.R. (Ed), Advances in Mathematics Research 11, Nova Science Publishers, New York.
Building knowledge-based expert systems by credal networks: a tutorial
Authors: Piatti, A. and Antonucci, A. and Zaffalon, M.
Year: 2010
Abstract: Knowledge-based systems are computer programs achieving expert-level competence in solving problems for specific task areas. This chapter is a tutorial on the construction of knowledge-based systems in the theoretical framework of credal networks. Credal networks are a generalization of Bayesian networks where credal sets, i.e., closed convex sets of probability measures, are used instead of precise probabilities. This allows for a more flexible process of elicitation than in the case of Bayesian networks. In fact, credal sets allow to represent ambiguity, contrast and contradiction in a natural and realistic way. The procedure we propose is based on a sharp distinction between the domain knowledge and the process linking this knowledge to the perceived evidence, which we call the observational process. This distinction leads to a very flexible representation of both domain knowledge and knowledge about the way the information is collected, together with a procedure of aggregation of the information coming from the different sources. The overall procedure is illustrated along the chapter by a simple knowledge-based system for the prediction of the result of a football match.
Published in Baswell, A.R. (Ed), Advances in Mathematics Research 11, Nova Science Publishers, New York.
Building knowledge-based expert systems by credal networks: a tutorial
@INBOOK{antonucci2010d,
title = {Building knowledge-based expert systems by credal networks: a tutorial},
editor = {Baswell, A.R.},
publisher = {Nova Science Publishers},
address = {New York},
volume = {11},
booktitle = {Advances in Mathematics Research},
author = {Piatti, A. and Antonucci, A. and Zaffalon, M.},
year = {2010},
chapter = {2}
}
Download
Scandurra, M., Mian, M., Greiner, T.C., Rancoita, P.M.V., De Campos, C.P., Chan, W.C., Vose, J.M., Chigrinova, E., Inghirami, G., Chiappella, A., Baldini, L., Ponzoni, M., Ferreri, A.J.M., Franceschetti, S., Gaidano, G., Montes-Moreno, S., Piris, M.A., Facchetti, F., Tucci, A., Nomdedeu, J.F., Lazure, T., Lambotte, O., Uccella, S., Pinotti, G., Pruneri, G., Martinelli, G., Young, K.H., Tibiletti, M.G., Rinaldi, A., Zucca, E., Kwee, I., Bertoni, F. (2010). Genomic lesions associated with a different clinical outcome in diffuse large B-Cell lymphoma treated with R-CHOP-21. British Journal of Haematology 151(3), pp. 221–231.
Genomic lesions associated with a different clinical outcome in diffuse large B-Cell lymphoma treated with R-CHOP-21
Authors: Scandurra, M. and Mian, M. and Greiner, T.C. and Rancoita, P.M.V. and De Campos, C.P. and Chan, W.C. and Vose, J.M. and Chigrinova, E. and Inghirami, G. and Chiappella, A. and Baldini, L. and Ponzoni, M. and Ferreri, A.J.M. and Franceschetti, S. and Gaidano, G. and Montes-Moreno, S. and Piris, M.A. and Facchetti, F. and Tucci, A. and Nomdedeu, J.F. and Lazure, T. and Lambotte, O. and Uccella, S. and Pinotti, G. and Pruneri, G. and Martinelli, G. and Young, K.H. and Tibiletti, M.G. and Rinaldi, A. and Zucca, E. and Kwee, I. and Bertoni, F.
Year: 2010
Abstract: Despite recent therapeutic improvements, the clinical course of diffuse large B-cell lymphoma (DLBCL) still differs considerably among patients. We conducted this retrospective multi-centre study to evaluate the impact of genomic aberrations detected using a high-density genome wide-single nucleotide polymorphism-based array on clinical outcome in a population of DLBCL patients treated with R-CHOP-21 (rituximab, cyclophosphamide, doxorubicine, vincristine and prednisone repeated every 21 d). 166 DNA samples were analysed using the GeneChip Human Mapping 250K NspI. Genomic anomalies were analysed regarding their impact on the clinical course of 124 patients treated with R-CHOP-21. Unsupervised clustering was performed to identify genetically related subgroups of patients with different clinical outcomes. Twenty recurrent genetic lesions showed an impact on the clinical course. Loss of genomic material at 8p23.1 showed the strongest statistical significance and was associated with additional aberrations, such as 17p- and 15q-. Unsupervised clustering identified five DLBCL clusters with distinct genetic profiles, clinical characteristics and outcomes. Genetic features and clusters, associated with a different outcome in patients treated with R-CHOP, have been identified by arrayCGH.
Published in British Journal of Haematology 151(3), Blackwell Publishing Ltd, pp. 221–231.
Genomic lesions associated with a different clinical outcome in diffuse large B-Cell lymphoma treated with R-CHOP-21
@ARTICLE{decampos2010a,
title = {Genomic lesions associated with a different clinical outcome in diffuse large {B}-{C}ell lymphoma treated with {R}-{CHOP}-21},
journal = {British Journal of Haematology},
publisher = {Blackwell Publishing Ltd},
volume = {151},
author = {Scandurra, M. and Mian, M. and Greiner, T.C. and Rancoita, P.M.V. and De Campos, C.P. and Chan, W.C. and Vose, J.M. and Chigrinova, E. and Inghirami, G. and Chiappella, A. and Baldini, L. and Ponzoni, M. and Ferreri, A.J.M. and Franceschetti, S. and Gaidano, G. and Montes-Moreno, S. and Piris, M.A. and Facchetti, F. and Tucci, A. and Nomdedeu, J.F. and Lazure, T. and Lambotte, O. and Uccella, S. and Pinotti, G. and Pruneri, G. and Martinelli, G. and Young, K.H. and Tibiletti, M.G. and Rinaldi, A. and Zucca, E. and Kwee, I. and Bertoni, F.},
number = {3},
pages = {221--231},
year = {2010},
doi = {10.1111/j.1365-2141.2010.08326.x},
url = {http://onlinelibrary.wiley.com/doi/10.1111/j.1365-2141.2010.08326.x/pdf}
}
Download top2009
Antonucci, A., Benavoli, A., Zaffalon, M., de Cooman, G., Hermans, F. (2009). Multiple model tracking by imprecise Markov trees. In FUSION 2009: Proceedings of the 12th International Conference on Information Fusion, IEEE.
Multiple model tracking by imprecise Markov trees
Authors: Antonucci, A. and Benavoli, A. and Zaffalon, M. and de Cooman, G. and Hermans, F.
Year: 2009
Abstract: We present a new procedure for tracking manoeuvring objects by hidden Markov chains. It leads to more reliable modelling of the transitions between hidden states compared to similar approaches proposed within the Bayesian framework: we adopt convex sets of probability mass functions rather than single 'precise probability' specifications, in order to provide a more realistic and cautious model of the manoeuvre dynamics. In general, the downside of such increased freedom in the modelling phase is a higher inferential complexity. However, the simple topology of hidden Markov chains allows for efficient tracking of the object through a recently developed belief propagation algorithm. Furthermore, the imprecise specification of the transitions can produce so-called indecision, meaning that more than one model may be suggested by our method as a possible explanation of the target kinematics. In summary, our approach leads to a multiple-model estimator whose performance, investigated through extensive numerical tests, turns out to be more accurate and robust than that of Bayesian ones.
Published in FUSION 2009: Proceedings of the 12th International Conference on Information Fusion, IEEE.
Multiple model tracking by imprecise Markov trees
@INPROCEEDINGS{antonucci2009e,
title = {Multiple model tracking by imprecise {M}arkov trees},
publisher = {IEEE},
booktitle = {{FUSION} 2009: Proceedings of the 12th International Conference on Information Fusion},
author = {Antonucci, A. and Benavoli, A. and Zaffalon, M. and de Cooman, G. and Hermans, F.},
year = {2009},
url = {http://isif.org/fusion/proceedings/fusion09CD/data/papers/0478.pdf}
}
Download
Antonucci, A., Brühlmann, R., Piatti, A., Zaffalon, M. (2009). Credal networks for military identification problems. International Journal of Approximate Reasoning 50(2), pp. 666–679.
Credal networks for military identification problems
Authors: Antonucci, A. and Brühlmann, R. and Piatti, A. and Zaffalon, M.
Year: 2009
Abstract: Credal networks are imprecise probabilistic graphical models generalizing Bayesian networks to convex sets of probability mass functions. This makes credal networks particularly suited to model expert knowledge under very general conditions, including states of qualitative and incomplete knowledge. In this paper, we present a credal network for risk evaluation in case of intrusion of civil aircrafts into a restricted flight area. The different factors relevant for this evaluation, together with an independence structure over them, are initially identified. These factors are observed by sensors, whose reliabilities can be affected by variable external factors, and even by the behaviour of the intruder. A model of these observation processes, and the necessary fusion scheme for the information returned by the sensors measuring the same factor, are both completely embedded into the structure of the credal network. A pool of experts, facilitated in their task by specific techniques to convert qualitative judgements into imprecise probabilistic assessments, has made possible the quantification of the network. We show the capabilities of the proposed model by means of some preliminary tests referred to simulated scenarios. Overall, we can regard this application as a useful tool to support military experts in their decision, but also as a quite general imprecise-probability paradigm for information fusion.
Published in International Journal of Approximate Reasoning 50(2), pp. 666–679.
Credal networks for military identification problems
@ARTICLE{antonucci2009a,
title = {Credal networks for military identification problems},
journal = {International Journal of Approximate Reasoning},
volume = {50},
author = {Antonucci, A. and Br\"uhlmann, R. and Piatti, A. and Zaffalon, M.},
number = {2},
pages = {666--679},
year = {2009},
doi = {10.1016/j.ijar.2009.01.005}
}
Download
Antonucci, A., Piatti, A. (2009). Modeling unreliable observations in Bayesian networks by credal networks. In Godo, L., Pugliese, A. (Eds), Scalable Uncertainty Management, Third International Conference, SUM 2009, Washington, DC, USA, September 28–30, 2009. Proceedings, Lecture Notes in Computer Science 5785, Springer, pp. 28–39.
Modeling unreliable observations in Bayesian networks by credal networks
Authors: Antonucci, A. and Piatti, A.
Year: 2009
Abstract: Bayesian networks are probabilistic graphical models widely employed in AI for the implementation of knowledge-based systems. Standard inference algorithms can update the beliefs about a variable of interest in the network after the observation of some other variables. This is usually achieved under the assumption that the observations could reveal the actual states of the variables in a fully reliable way. We propose a procedure for a more general modeling of the observations, which allows for updating beliefs in different situations, including various cases of unreliable, incomplete, uncertain and also missing observations. This is achieved by augmenting the original Bayesian network with a number of auxiliary variables corresponding to the observations. For a flexible modeling of the observational process, the quantification of the relations between these auxiliary variables and those of the original Bayesian network is done by credal sets, i.e., convex sets of probability mass functions. Without any lack of generality, we show how this can be done by simply estimating the bounds for the likelihoods of the observations. Overall, the Bayesian network is transformed into a credal network, for which a standard updating problem has to be solved. Finally, a number of transformations that might simplify the updating of the resulting credal network is provided.
Published in Godo, L., Pugliese, A. (Eds), Scalable Uncertainty Management, Third International Conference, SUM 2009, Washington, DC, USA, September 28–30, 2009. Proceedings, Lecture Notes in Computer Science 5785, Springer, pp. 28–39.
Modeling unreliable observations in Bayesian networks by credal networks
@INPROCEEDINGS{antonucci2009g,
title = {Modeling unreliable observations in {B}ayesian networks by credal networks},
editor = {Godo, L. and Pugliese, A.},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
volume = {5785},
booktitle = {Scalable Uncertainty Management, Third International Conference, {SUM} 2009, Washington, {DC}, {USA}, September 28–30, 2009. Proceedings},
author = {Antonucci, A. and Piatti, A.},
pages = {28--39},
year = {2009},
doi = {10.1007/978-3-642-04388-8_4}
}
Download
Benavoli, A., Antonucci, A. (2009). Aggregating imprecise probabilistic knowledge. In ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, Durham (UK), pp. 31–41.
Aggregating imprecise probabilistic knowledge
Authors: Benavoli, A. and Antonucci, A.
Year: 2009
Abstract: The problem of aggregating two or more sources of information containing knowledge about a same domain is considered. We propose an aggregation rule for the case where the available information is modeled by coherent lower previsions, corresponding to convex sets of probability mass functions. The consistency between aggregated beliefs and sources of information is discussed. A closed formula, which specializes our rule to a particular class of models, is also derived. Finally, an alternative explanation of Zadeh's paradox is provided.
Published in ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, Durham (UK), pp. 31–41.
Aggregating imprecise probabilistic knowledge
@INPROCEEDINGS{benavoli2009c,
title = {Aggregating imprecise probabilistic knowledge},
address = {Durham (UK)},
booktitle = {{ISIPTA} '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications},
author = {Benavoli, A. and Antonucci, A.},
pages = {31--41},
year = {2009},
url = {http://www.sipta.org/isipta09/proceedings/papers/s043.pdf}
}
Download
Benavoli, A., de Campos, C.P. (2009). Inference from multinomial data based on a MLE-dominance criterion. In Proc. on European Conf. on Symbolic and Quantitative Approaches to Reasoning and Uncertainty (ECSQARU), Springer, Berlin / Heidelberg, Verona (IT), pp. 22–33.
Inference from multinomial data based on a MLE-dominance criterion
Authors: Benavoli, A. and de Campos, C.P.
Year: 2009
Abstract: We consider the problem of inference from multinomial data with chances theta, subject to the a-priori information that the true parameter vector theta belongs to a known convex polytope Theta. The proposed estimator has the parametrized structure of the conditional-mean estimator with a prior Dirichlet distribution, whose parameters (s, t) are suitably designed via a dominance criterion so as to guarantee, for any theta in Theta, an improvement of the Mean Squared Error over the Maximum Likelihood Estimator (MLE). The solution of this MLE-dominance prob- lem allows us to give a different interpretation of: (1) the several Bayesian estimators proposed in the literature for the problem of inference from multinomial data; (2) the Imprecise Dirichlet Model (IDM) developed by Walley.
Published in Proc. on European Conf. on Symbolic and Quantitative Approaches to Reasoning and Uncertainty (ECSQARU), Springer, Berlin / Heidelberg, Verona (IT), pp. 22–33.
Inference from multinomial data based on a MLE-dominance criterion
@INPROCEEDINGS{benavoli2009b,
title = {Inference from multinomial data based on a {MLE}-dominance criterion},
publisher = {Springer, Berlin / Heidelberg},
address = {Verona (IT)},
booktitle = {Proc. {o}n European Conf. {o}n Symbolic and Quantitative Approaches to Reasoning and Uncertainty ({ECSQARU})},
author = {Benavoli, A. and de Campos, C.P.},
pages = {22--33},
year = {2009},
doi = {10.1007/978-3-642-02906-6_4}
}
Download
Benavoli, A., Ristic, B., Farina, A., Oxenham, M., Chisci, L. (2009). An application of evidential networks to threat assessment. Aerospace and Electronic Systems, IEEE Transactions on 45(2), pp. 620–639.
An application of evidential networks to threat assessment
Authors: Benavoli, A. and Ristic, B. and Farina, A. and Oxenham, M. and Chisci, L.
Year: 2009
Abstract: Decision makers operating in modern defence theatres need to comprehend and reason with huge quantities of potentially uncertain and imprecise data in a timely fashion. In this paper, an automatic information fusion system is developed which aims at supporting a commander's decision making by providing a threat assessment, that is an estimate of the extent to which an enemy platform poses a threat based on evidence about its intent and capability. Threat is modelled by a network of entities and relationships between them, while the uncertainties in the relationships are represented by belief functions as defined in the theory of evidence. To support the implementation of the threat assessment functionality, an efficient valuation-based reasoning scheme, referred to as an evidential network, is developed. To reduce computational overheads, the scheme performs local computations in the network by applying an inward propagation algorithm to the underlying binary join tree. This allows the dynamic nature of the external evidence, which drives the evidential network, to be taken into account by recomputing only the affected paths in the binary join tree.
Published in Aerospace and Electronic Systems, IEEE Transactions on 45(2), pp. 620–639.
An application of evidential networks to threat assessment
@ARTICLE{benavoli2009d,
title = {An application of evidential networks to threat assessment},
journal = {Aerospace and Electronic Systems, {IEEE} Transactions on},
volume = {45},
author = {Benavoli, A. and Ristic, B. and Farina, A. and Oxenham, M. and Chisci, L.},
number = {2},
pages = {620--639},
year = {2009},
doi = {10.1109/TAES.2009.5089545}
}
Download
Benavoli, A., Zaffalon, M., Miranda, E. (2009). Reliable hidden Markov model filtering through coherent lower previsions. In Information Fusion, 2009. FUSION '09. 12th International Conference on, Seattle (USA), pp. 1743–1750.
Reliable hidden Markov model filtering through coherent lower previsions
Authors: Benavoli, A. and Zaffalon, M. and Miranda, E.
Year: 2009
Abstract: We extend Hidden Markov Models for continuous variables taking into account imprecision in our knowledge about the probabilistic relationships involved. To achieve that, we consider sets of probabilities, also called coherent lower previsions. In addition to the general formulation, we study in detail a particular case of interest: linear-vacuous mixtures. We also show, in a practical case, that our extension outperforms the Kalman filter when modelling errors are present in the system.
Published in Information Fusion, 2009. FUSION '09. 12th International Conference on, Seattle (USA), pp. 1743–1750.
Reliable hidden Markov model filtering through coherent lower previsions
@INPROCEEDINGS{benavoli2009a,
title = {Reliable hidden {M}arkov model filtering through coherent lower previsions},
address = {Seattle (USA)},
booktitle = {Information Fusion, 2009. {FUSION} '09. 12th International Conference on},
author = {Benavoli, A. and Zaffalon, M. and Miranda, E.},
pages = {1743--1750},
year = {2009},
url = {http://isif.org/fusion/proceedings/fusion09CD/data/papers/0345.pdf}
}
Download
de Campos, C.P., Cozman, F.G., Luna, J.E.O. (2009). Assembling a consistent set of sentences in relational probabilistic logic with stochastic independence. Journal of Applied Logic 7(2), pp. 137–154.
Assembling a consistent set of sentences in relational probabilistic logic with stochastic independence
Authors: de Campos, C.P. and Cozman, F.G. and Luna, J.E.O.
Year: 2009
Abstract: We examine the representation of judgements of stochastic independence in probabilistic logics. We focus on a relational logic where (i) judgements of stochastic independence are encoded by directed acyclic graphs, and (ii) probabilistic assessments are flexible in the sense that they are not required to specify a single probability measure. We discuss issues of knowledge representation and inference that arise from our particular combination of graphs, stochastic independence, logical formulas and probabilistic assessments.
Published in Journal of Applied Logic 7(2), pp. 137–154.
Assembling a consistent set of sentences in relational probabilistic logic with stochastic independence
@ARTICLE{decampos2009d,
title = {Assembling a consistent set of sentences in relational probabilistic logic with stochastic independence},
journal = {Journal of Applied Logic},
volume = {7},
author = {de Campos, C.P. and Cozman, F.G. and Luna, J.E.O.},
number = {2},
pages = {137--154},
year = {2009},
doi = {10.1016/j.jal.2007.11.002}
}
Download
de Campos, C.P., Zeng, Z., Ji, Q. (2009). Structure learning of Bayesian networks using constraints. In International Conference on Machine Learning (ICML) 382, ACM, pp. 113–120.
Structure learning of Bayesian networks using constraints
Authors: de Campos, C.P. and Zeng, Z. and Ji, Q.
Year: 2009
Abstract: This paper addresses exact learning of Bayesian network structure from data and expert's knowledge based on score functions that are decomposable. First, it describes useful properties that strongly reduce the time and memory costs of many known methods such as hill-climbing, dynamic programming and sampling variable orderings. Secondly, a branch and bound algorithm is presented that integrates parameter and structural constraints with data in a way to guarantee global optimality with respect to the score function. It is an any-time procedure because, if stopped, it provides the best current solution and an estimation about how far it is from the global solution. We show empirically the advantages of the properties and the constraints, and the applicability of the algorithm to large data sets (up to one hundred variables) that cannot be handled by other current methods (limited to around 30 variables).
Published in International Conference on Machine Learning (ICML) 382, ACM, pp. 113–120.
Structure learning of Bayesian networks using constraints
@INPROCEEDINGS{decampos2009e,
title = {Structure learning of {B}ayesian networks using constraints},
publisher = {ACM},
volume = {382},
booktitle = {International Conference on Machine Learning ({ICML})},
author = {de Campos, C.P. and Zeng, Z. and Ji, Q.},
pages = {113--120},
year = {2009},
doi = {10.1145/1553374.1553389}
}
Download
de Campos, C.P., Zhang, L., Tong, Y., Ji, Q. (2009). Semi-qualitative probabilistic networks in computer vision problems. Journal of Statistical Theory and Practice 3(1), pp. 197–210.
Semi-qualitative probabilistic networks in computer vision problems
Authors: de Campos, C.P. and Zhang, L. and Tong, Y. and Ji, Q.
Year: 2009
Abstract: This paper explores the application of semi-qualitative probabilistic networks (SQPNs) that combine numeric and qualitative information to computer vision problems. Our version of SQPN allows qualitative influences and imprecise probability measures using intervals. We describe an Imprecise Dirichlet model for parameter learning and an iterative algorithm for evaluating posterior probabilities, maximum a posteriori and most probable explanations. Experiments on facial expression recognition and image segmentation problems are performed using real data.
Published in Journal of Statistical Theory and Practice 3(1), Grace Scientific Publishing LLC, pp. 197–210.
Semi-qualitative probabilistic networks in computer vision problems
@ARTICLE{decampos2009c,
title = {Semi-qualitative probabilistic networks in computer vision problems},
journal = {Journal of Statistical Theory and Practice},
publisher = {Grace Scientific Publishing LLC},
volume = {3},
author = {de Campos, C.P. and Zhang, L. and Tong, Y. and Ji, Q.},
number = {1},
pages = {197--210},
year = {2009},
doi = {10.1080/15598608.2009.10411920}
}
Download
de Campos, C.P., Zhang, L., Tong, Y., Ji, Q. (2009). Semi-qualitative probabilistic networks in computer vision problems. In Coolen-Schrijner, P., Coolen, F., Troffaes, M.C.M., Augustin, T. (Eds), Imprecision in Statistical Theory and Practice., Grace Scientific Publishing LLC, Greensboro, North-Carolina, USA, pp. 207–220.
Semi-qualitative probabilistic networks in computer vision problems
Authors: de Campos, C.P. and Zhang, L. and Tong, Y. and Ji, Q.
Year: 2009
Abstract: This paper explores the application of semi-qualitative probabilistic networks (SQPNs) that combine numeric and qualitative information to computer vision problems. Our version of SQPN allows qualitative influences and imprecise probability measures using intervals. We describe an Imprecise Dirichlet model for parameter learning and an iterative algorithm for evaluating posterior probabilities, maximum a posteriori and most probable explanations. Experiments on facial expression recognition and image segmentation problems are performed using real data.
Published in Coolen-Schrijner, P., Coolen, F., Troffaes, M.C.M., Augustin, T. (Eds), Imprecision in Statistical Theory and Practice., Grace Scientific Publishing LLC, Greensboro, North-Carolina, USA, pp. 207–220.
Semi-qualitative probabilistic networks in computer vision problems
@INBOOK{decampos2009a,
title = {Semi-qualitative probabilistic networks in computer vision problems},
editor = {Coolen-Schrijner, P. and Coolen, F. and Troffaes, M.C.M. and Augustin, T.},
publisher = {Grace Scientific Publishing LLC},
address = {Greensboro, North-Carolina, USA},
booktitle = {Imprecision in Statistical Theory and Practice.},
author = {de Campos, C.P. and Zhang, L. and Tong, Y. and Ji, Q.},
pages = {207--220},
year = {2009},
url = {http://www.amazon.com/Imprecision-Statistical-Practice-Pauline-Coolen-Schrijner/dp/0982399804}
}
Download
de Cooman, G., Hermans, F., Antonucci, A., Zaffalon, M. (2009). Epistemic irrelevance in credal networks: the case of imprecise Markov trees. In Augustin, T., Coolen, F., Moral, S., Troffaes, M.C.M. (Eds), ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 149–158.
Epistemic irrelevance in credal networks: the case of imprecise Markov trees
Authors: de Cooman, G. and Hermans, F. and Antonucci, A. and Zaffalon, M.
Year: 2009
Abstract: We replace strong independence in credal networks with the weaker notion of epistemic irrelevance. Focusing on directed trees, we show how to combine local credal sets into a global model, and we use this to construct and justify an exact message-passing algorithm that computes updated beliefs for a variable in the tree. The algorithm, which is essentially linear in the number of nodes, is formulated entirely in terms of coherent lower previsions. We supply examples of the algorithm's operation, and report an application to on-line character recognition that illustrates the advantages of our model for prediction.
Published in Augustin, T., Coolen, F., Moral, S., Troffaes, M.C.M. (Eds), ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 149–158.
Epistemic irrelevance in credal networks: the case of imprecise Markov trees
@INPROCEEDINGS{antonucci2009c,
title = {Epistemic irrelevance in credal networks: the case of imprecise {M}arkov trees},
editor = {Augustin, T. and Coolen, F. and Moral, S. and Troffaes, M.C.M.},
publisher = {SIPTA},
booktitle = {{ISIPTA} '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications},
author = {de Cooman, G. and Hermans, F. and Antonucci, A. and Zaffalon, M.},
pages = {149--158},
year = {2009},
url = {http://www.sipta.org/isipta09/proceedings/papers/s053.pdf}
}
Download
Corani, G., Campos, C., Yi, S. (2009). A tree augmented classifier based on extreme imprecise Dirichlet model. In Augustin, T., Coolen, F.P.A., Moral, S., Troffaes, M.C.M. (Eds), ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, Durham, UK, pp. 89–98.
A tree augmented classifier based on extreme imprecise Dirichlet model
Authors: Corani, G. and Campos, C. and Yi, S.
Year: 2009
Abstract: In this paper we present TANC, i.e., a tree-augmented naive credal classifier based on imprecise probabilities; it models prior near-ignorance via the Extreme Imprecise Dirichlet Model (EDM) (Cano et al., 2007) and deals conservatively with missing data in the training set, without assuming them to be missing-at-random. The EDM is an approximation of the global Imprecise Dirichlet Model (IDM), which considerably simplifies the computation of upper and lower probabilities; yet, having been only recently introduced, the quality of the provided approximation needs still to be verified. As first contribution, we extensively compare the output of the naive credal classifier (one of the few cases in which the global IDM can be exactly implemented) when learned with the EDM and the global IDM; the output of the classifier appears to be identical in the vast majority of cases, thus supporting the adoption of the EDM in real classification problems. Then, by experiments we show that TANC is more reliable than the precise TAN (learned with uniform prior), and also that it provides better performance compared to a previous (Zaffalon, 2003) TAN model based on imprecise probabilities. TANC treats missing data by considering all possible completions of the training set, but avoiding an exponential increase of the computational times; eventually
Published in Augustin, T., Coolen, F.P.A., Moral, S., Troffaes, M.C.M. (Eds), ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, Durham, UK, pp. 89–98.
A tree augmented classifier based on extreme imprecise Dirichlet model
@INPROCEEDINGS{corani2009c,
title = {A tree augmented classifier based on extreme imprecise {D}irichlet model},
editor = {Augustin, T. and Coolen, F.P.A. and Moral, S. and Troffaes, M.C.M.},
publisher = {SIPTA},
address = {Durham, UK},
booktitle = {{ISIPTA} '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications},
author = {Corani, G. and Campos, C. and Yi, S.},
pages = {89--98},
year = {2009},
url = {http://www.sipta.org/isipta09/proceedings/papers/s060.pdf}
}
Download
Corani, G., Rizzoli, A.E., Salvetti, A., Zaffalon, M. (2009). Reproducing human decisions in reservoir management: the case of lake lugano. In Information Technologies in Environmental Engineering, Springer, Berlin / Heidelberg, pp. 252–263.
Reproducing human decisions in reservoir management: the case of lake lugano
Authors: Corani, G. and Rizzoli, A.E. and Salvetti, A. and Zaffalon, M.
Year: 2009
Published in Information Technologies in Environmental Engineering, Springer, Berlin / Heidelberg, pp. 252–263.
Note: 10.1007/978-3-540-88351-7_19
Reproducing human decisions in reservoir management: the case of lake lugano
@INCOLLECTION{corani2009a,
title = {Reproducing human decisions in reservoir management: the case of lake lugano},
publisher = {Springer, Berlin / Heidelberg},
booktitle = {Information Technologies in Environmental Engineering},
author = {Corani, G. and Rizzoli, A.E. and Salvetti, A. and Zaffalon, M.},
pages = {252--263},
year = {2009},
doi = {10.1007/978-3-540-88351-7_19}
}
Download
Corani, G., Zaffalon, M. (2009). Lazy naive credal classifier. In Proceedings of the 1st ACM SIGKDD Workshop on Knowledge Discovery From Uncertain Data, U '09, ACM, New York, NY, USA, pp. 30–37.
Lazy naive credal classifier
Authors: Corani, G. and Zaffalon, M.
Year: 2009
Abstract: We propose a local (or lazy) version of the naive credal classifier. The latter is an extension of naive Bayes to imprecise probability developed to issue reliable classifications despite small amounts of data, which may then be carrying highly uncertain information about a domain. Reliability is maintained because credal classifiers can issue set-valued classifications on instances that are particularly difficult to classify. We show by extensive experiments that the local classifier outperforms the original one, both in terms of accuracy of classification and because it leads to stronger conclusions (i.e., set-valued classifications made by fewer classes). By comparing the local credal classifier with a local version of naive Bayes, we also show that the former reliably deals with instances which are difficult to classify
Published in Proceedings of the 1st ACM SIGKDD Workshop on Knowledge Discovery From Uncertain Data, U '09, ACM, New York, NY, USA, pp. 30–37.
Lazy naive credal classifier
@INPROCEEDINGS{corani2009b,
title = {Lazy naive credal classifier},
publisher = {ACM},
address = {New York, NY, USA},
series = {U '09},
booktitle = {Proceedings of the 1st {ACM} {SIGKDD} Workshop on Knowledge Discovery From Uncertain Data},
author = {Corani, G. and Zaffalon, M.},
pages = {30--37},
year = {2009},
doi = {10.1145/1610555.1610560}
}
Download
Miranda, E., Zaffalon, M. (2009). Coherence graphs. Artificial Intelligence 173, pp. 104–144.
Coherence graphs
Authors: Miranda, E. and Zaffalon, M.
Year: 2009
Abstract: We study the consistency of a number of probability distributions, which are allowed to be imprecise. To make the treatment as general as possible, we represent those probabilistic assessments as a collection of conditional lower previsions. The problem then becomes proving Walley's (strong) coherence of the assessments. In order to maintain generality in the analysis, we assume to be given nearly no information about the numbers that make up the lower previsions in the collection. Under this condition, we investigate the extent to which the above global task can be decomposed into simpler and more local ones. This is done by introducing a graphical representation of the conditional lower previsions that we call the coherence graph: we show that the coherence graph allows one to isolate some subsets of the collection whose coherence is sufficient for the coherence of all the assessments; and we provide a polynomial-time algorithm that finds the subsets efficiently. We show some of the implications of our results by focusing on three models and problems: Bayesian and credal networks, of which we prove coherence; the compatibility problem, for which we provide an optimal graphical decomposition; probabilistic satisfiability, of which we show that some intractable instances can instead be solved efficiently by exploiting coherence graphs.
Published in Artificial Intelligence 173, pp. 104–144.
Coherence graphs
@ARTICLE{zaffalon2009b,
title = {Coherence graphs},
journal = {Artificial Intelligence},
volume = {173},
author = {Miranda, E. and Zaffalon, M.},
pages = {104--144},
year = {2009},
doi = {10.1016/j.artint.2008.09.001}
}
Download
Miranda, E., Zaffalon, M. (2009). Natural extension as a limit of regular extensions. In Augustin, T., Coolen, F., Troffaes, M.C.M., Moral, S. (Eds), ISIPTA '09: Proceedings of the Sixth International Symposium on Imprecise Probability: Theories and Applications, SIPTA, pp. 327–336.