Hostname: page-component-89b8bd64d-b5k59 Total loading time: 0 Render date: 2026-05-06T17:12:22.300Z Has data issue: false hasContentIssue false

A zero-inflated Poisson latent position cluster model

Published online by Cambridge University Press:  23 January 2026

Chaoyi Lu*
Affiliation:
School of Mathematics and Statistics, University College Dublin, Dublin, Ireland Insight Research Ireland Centre for Data Analytics, University College Dublin, Dublin, Ireland
Riccardo Rastelli
Affiliation:
School of Mathematics and Statistics, University College Dublin, Dublin, Ireland
Nial Friel
Affiliation:
School of Mathematics and Statistics, University College Dublin, Dublin, Ireland Insight Research Ireland Centre for Data Analytics, University College Dublin, Dublin, Ireland
*
Corresponding author: Chaoyi Lu; Email: chaoyi.lu.stat@gmail.com
Rights & Permissions [Opens in a new window]

Abstract

The Latent Position Model (LPM) is a popular approach for the statistical analysis of network data. A central aspect of this model is that it assigns nodes to random positions in a latent space, such that the probability of an interaction between each pair of individuals or nodes is determined by their distance in this latent space. A key feature of this model is that it allows one to visualize nuanced structures via the latent space representation. The LPM can be further extended to the Latent Position Cluster Model (LPCM), to accommodate the clustering of nodes by assuming that the latent positions are distributed following a finite mixture distribution. In this paper, we extend the LPCM to accommodate missing network data and apply this to non-negative discrete weighted social networks. By treating missing data as “unusual” zero interactions, we propose a combination of the LPCM with the zero-inflated Poisson distribution. Statistical inference is based on a novel partially collapsed Markov chain Monte Carlo algorithm, where a Mixture-of-Finite-Mixtures (MFM) model is adopted to automatically determine the number of clusters and optimal group partitioning. Our algorithm features a truncated absorb-eject move, which is a novel adaptation of an idea commonly used in collapsed samplers, within the context of MFMs. Another aspect of our work is that we illustrate our results on 3-dimensional latent spaces, maintaining clear visualizations while achieving more flexibility than 2-dimensional models. The performance of this approach is illustrated via three carefully designed simulation studies, as well as four different publicly available real networks, where some interesting new perspectives are uncovered.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2026. Published by Cambridge University Press
Figure 0

Algorithm 1: A partially collapsed Metropolis-within-Gibbs sampler for ZIP-LPCM.

Figure 1

Figure 1. Simulation study 1 synthetic networks. The 1st row plots correspond to the scenario 1 network, while the 2nd row plots correspond to the scenario 2 network. The 1st column plots show the 3-dimensional plots of the latent positions with different node colors denoting the corresponding true clustering. Node sizes are proportional to node betweenness, whereas edge widths and colors are proportional to edge weights. The 2nd column plots are the rotated plots of the 1st column latent position plots that are rotated for $90^{\circ }$ clockwise with respect to the vertical axis.

Figure 2

Figure 2. Simulation study 1. Synthetic networks’ adjacency matrix heatmap plots. Darker entries correspond to higher edge weights. The side-bars indicate the reference clustering ${\boldsymbol{z}}^*$. Left plot: scenario 1 network, generated from a ZIP-LPCM. Right plot: scenario 2 network, generated from a Pois-LPCM.

Figure 3

Table 1. Simulation study 1. Performance of eight different implementations where (i) $\hat {K}$: the number of clusters in $\hat {{\boldsymbol{z}}}$; (ii) $\text{VI}(\hat {{\boldsymbol{z}}},{\boldsymbol{z}}^*)$: the VI distance between the point estimate $\hat {{\boldsymbol{z}}}$ and the true clustering ${\boldsymbol{z}}^*$; (iii) $\mathbb{E}_{{\boldsymbol{z}}}[\text{VI}(\hat {{\boldsymbol{z}}},{\boldsymbol{z}}) \mid {\boldsymbol{Y}}]$: the minimized expected posterior VI loss of the clustering with respect to $\hat {{\boldsymbol{z}}}$. This statistic measures the uncertainty of the posterior clustering around the $\hat {{\boldsymbol{z}}}$; (iv) $\mathbb{E}(\{|\hat {d}_{ij}-d^*_{ij}|\})$ [sd]: the mean of $\{|\hat {d}_{ij}-d^*_{ij}|\,:\, i,j=1,2,\dots ,N; i\gt j\}$ with the corresponding standard deviation (sd) shown in the square bracket; (v) $\hat {\beta }$: the posterior mean of $\beta$; (vi) $\mathbb{E}(\{|\hat {p}_{z_iz_j}-p^*_{z_iz_j}|\})$ [sd]: the mean of $\{|\hat {p}_{z_iz_j}-p^*_{z_iz_j}|\,:\, i,j=1,2,\dots ,N; i\gt j\}$ with sd in the square bracket. More details are included in Section 4.1. The best performance within each column are highlighted in bold font

Figure 4

Figure 3. Simulation study 1 scenario 1. Performance of the posterior mean $\hat {\boldsymbol{\nu }}$, which approximates the conditional probability in Eq. (18). The top-left plot is the heatmap of the reference values of Eq. (18), obtained by leveraging the reference model parameters used for simulating the network, whereby darker entry colors correspond to higher values. The other four heatmap plots describe $\hat {\boldsymbol{\nu }}$ as inferred by the corresponding priors indicated on top of the heatmap. The rows and columns of the matrices are rearranged and separated according to $\hat {{\boldsymbol{z}}}$ while the side-bars indicate the true clustering of each individual. The last plot shows the Receiver Operating Characteristic (ROC) curves for all the supervised ZIP-LPCM cases, where the reference $\boldsymbol{\nu }^*$ is the response variable.

Figure 5

Figure 4. Simulation study 2. Synthetic networks’ adjacency matrix heatmap plots. Darker entries correspond to higher edge weights. The side-bars indicate the reference clustering ${\boldsymbol{z}}^*$. Left plot: scenario 1 network, generated from a ZIP-SBM without hubs. Right plot: scenario 2 network, generated from a ZIP-SBM with hubs.

Figure 6

Figure 5. Simulation study 2. The 1st and the 2nd rows illustrate the inferred point estimate $\hat {{\boldsymbol{U}}}$ obtained by ZIP-LPCM Sup Beta(1,9) implementations for Scenario 1 and Scenario 2, respectively. The 2nd column plots are rotated version of the 1st column plots where each inferred latent position rotated for $90^{\circ }$ clockwise with respect to the vertical axis. Different node colors correspond to different inferred groups according to the corresponding $\hat {{\boldsymbol{z}}}$. Node sizes are proportional to node betweenness while edge widths and colors are proportional to edge weights.

Figure 7

Table 2. Simulation study 2. Performance of eight different implementations where (i) $\hat {K}$: the number of clusters in $\hat {{\boldsymbol{z}}}$; (ii) $\text{VI}(\hat {{\boldsymbol{z}}},{\boldsymbol{z}}^*)$: the VI distance between the point estimate $\hat {{\boldsymbol{z}}}$ and the true clustering ${\boldsymbol{z}}^*$; (iii) $\mathbb{E}_{{\boldsymbol{z}}}[\text{VI}(\hat {{\boldsymbol{z}}},{\boldsymbol{z}}) \mid {\boldsymbol{Y}}]$: the minimized expected posterior VI loss of the clustering with respect to $\hat {{\boldsymbol{z}}}$. This statistic measures the uncertainty of the posterior clustering around the $\hat {{\boldsymbol{z}}}$; (iv) $\hat {\beta }$: the posterior mean of $\beta$; (v) $\mathbb{E}(\{|\hat {p}_{z_iz_j}-p^*_{z_iz_j}|\})$[sd]: the mean of $\{|\hat {p}_{z_iz_j}-p^*_{z_iz_j}|\,:\, i,j=1,2,\dots ,N;\, i\gt j\}$ with the corresponding standard deviation (sd) shown in the square bracket; (vi) $\mathbb{E}(\{|\hat {\lambda }_{ij}-\lambda ^*_{ij}|\})$[sd]: the mean of $\{|\hat {\lambda }_{ij}-\lambda ^*_{ij}|\,:\, i,j=1,2,\dots ,N;\, i\gt j\}$ with the corresponding sd in the square bracket. More details are included in Section 4.2. The best performance within each column excluding the $\hat {\beta }$ column are highlighted in bold font

Figure 8

Table 3. Simulation study 3. Performance of the replication simulation study (top table) and the simulation studies for networks with larger network sizes (bottom table). The same set of summary statistics used in Table 1 is also leveraged here for performance explorations. The values in each row of the top table below correspond to the median, 10% and 90% quantiles of the corresponding summary statistics for all 50 replicated implementations. The values in the bottom table below correspond to the output values of the corresponding summary statistics for the two simulation studies of larger networks

Figure 9

Figure 6. Simulation study 2. Performance of $\hat {\boldsymbol{\nu }}$ based on the $\text{Beta}(1,9)$ and $\text{Beta}(1,19)$ prior settings. The 1st and 2nd rows, respectively, correspond to scenarios 1 and 2. Darker colors indicate higher (approximate) probability of unusual zero conditional on the fact that the corresponding observed interaction is a zero interaction. The rows and columns of the matrices are rearranged and separated according to $\hat {{\boldsymbol{z}}}$ while the side-bars indicate the true clustering of each individual.

Figure 10

Figure 7. Simulation study 3. A two-group synthetic network example of showing possible misclassifications of nodes’ clustering. The 2nd plot is the rotated plot of the 1st latent positions’ plot that is rotated for $45^{\circ }$ anti-clockwise with respect to the vertical axis. The latent positions are those used for generating the network, whereas the dark red (group 1) and dark blue (group 2) nodes correspond to the nodes which are inferred to have true clustering. The three light red nodes are the ones which should belong to group 1 according to the reference clustering ${\boldsymbol{z}}^*$ but they are instead inferred to be misclassified to group 2 based on their network structure. No misclassification observed for group 2 nodes’ clustering. Node sizes are proportional to node betweenness. Edge widths and colors are proportional to edge weights.

Figure 11

Figure 8. The heatmap plots for the Sampson monks real network where the grays are used for zero values in order to highlight other non-zero elements. Left plot: original observed adjacency matrix, ${\boldsymbol{Y}}$. Middle plot: plot of ${\boldsymbol{Y}}$ where the rows and columns of the matrices are rearranged and separated according to ${\boldsymbol{z}}^*$. The different colors in the side-bars of this and the last plot correspond to different reference clustering of each individual. Right plot: inferred $\hat {\text{P}}(x_{ij}\gt 0|y_{ij}=0,\dots )$.

Figure 12

Figure 9. The Sampson monks real network. Left plot: inferred latent positions, $\hat {{\boldsymbol{U}}}$. The three inferred groups in $\hat {{\boldsymbol{z}}}$ are distinguished by different colors, and perfectly agree with the reference clustering. Node sizes are proportional to node betweenness. Edge widths and colors are proportional to edge weights. Right plot: rotated version of the latent positions shown in the left plot where the whole latent space is rotated by $90^{\circ }$ clockwise with respect to the vertical axis.

Figure 13

Figure 10. Windsurfers real network. Left plot: inferred latent positions, $\hat {{\boldsymbol{U}}}$. The three inferred groups in $\hat {{\boldsymbol{z}}}$ are distinguished by different colors. Node sizes are proportional to node betweenness. Edge widths and colors are proportional to edge weights. Right plot: rotated version of the latent positions shown in the left plot where the whole latent space is rotated by $90^{\circ }$ clockwise with respect to the vertical axis.

Figure 14

Figure 11. The heatmap plots for the windsurfers real network where the grays are used for zero values in order to highlight other non-zero elements. Left plot: original observed adjacency matrix, ${\boldsymbol{Y}}$. Middle plot: plot of ${\boldsymbol{Y}}$ where the rows and columns of the matrices are rearranged and separated according to $\hat {{\boldsymbol{z}}}$. The different colors on the side-bars correspond to different inferred clustering of each individual. Right plot: inferred $\hat {\text{P}}(x_{ij}\gt 0|y_{ij}=0,\dots )$.

Figure 15

Figure 12. Train bombing real network. Left plot: inferred latent positions, $\hat {{\boldsymbol{U}}}$. The inferred clustering distinguished by different colors. Node sizes are proportional to node betweenness. Edge widths and colors are proportional to edge weights. Right plot: rotated version of the latent positions shown in the left plot where the whole latent space is rotated by $90^{\circ }$ clockwise with respect to the vertical axis.

Figure 16

Figure 13. The heatmap plots for the train bombing real network where the grays are used for zero values in order to highlight other non-zero elements. Left plot: original observed adjacency matrix, ${\boldsymbol{Y}}$. Middle plot: the plot of ${\boldsymbol{Y}}$ where the rows and columns of the matrices are rearranged and separated according to $\hat {{\boldsymbol{z}}}$. The different colors in the side-bars of this and the last plot correspond to different inferred clustering of each individual. Right plot: inferred $\hat {\text{P}}(x_{ij}\gt 0|y_{ij}=0,\dots )$.

Figure 17

Figure 14. Summit co-attendance criminality network. The 1st row plots illustrate the inferred latent positions, $\hat {{\boldsymbol{U}}}$, along with the different dark or light node colors indicating the reference clustering ${\boldsymbol{z}}^*$. Darker square nodes indicate a more central role in the organization. The same $\hat {{\boldsymbol{U}}}$ is shown in the 2nd row plots but the node colors denote instead the inferred partition $\hat {{\boldsymbol{z}}}$. The 2nd column plots are the rotated version of the latent positions shown in the 1st column plots where the whole latent space is rotated by $60^{\circ }$ clockwise with respect to the vertical axis. Node sizes are proportional to node betweenness. Edge widths and colors are proportional to edge weights.

Figure 18

Figure 15. The output heatmap plots for the summit co-attendance criminality network where the grays are used for zero values in order to highlight other non-zero elements. The different colors in the side-bars correspond to the role-affiliation information, ${\boldsymbol{z}}^*$. The rows and columns of the matrices are rearranged and separated according to $\hat {{\boldsymbol{z}}}$, where the inferred groups containing central nodes are placed at the bottom while the others are placed at the top. Left plot: adjacency matrix ${\boldsymbol{Y}}$. Middle plot: inferred $\hat {\boldsymbol{\nu }}$. Right plot: inferred $\hat {\text{P}}(x_{ij}\gt 0|y_{ij}=0,\dots )$.

Figure 19

Figure 16. The windsurfers, train bombing and summit co-attendance criminality network in 2-dimensional latent space. The plot settings regarding node sizes, colors, types and edge colors, widths are similar to those applied in the previous subsections. Top-left: inferred 2-d latent positions’ plot of the windsurfers network. Different node colors correspond to different inferred groups in $\hat {{\boldsymbol{z}}}$. Top-right: inferred 2-d latent positions’ plot of the train bombing network where the inferred clustering has only 1 group. Bottom plots: inferred 2-d latent positions of the summit co-attendance criminality network, where the different node colors in the left plot indicate the reference clustering ${\boldsymbol{z}}^*$, while those in the right plot correspond to different inferred groups in $\hat {{\boldsymbol{z}}}$.