|
1 | | -- abstract'@: >- |
2 | | - This study investigates the use of Variational |
3 | | - Auto-Encoders to build a simulator that approximates the law of |
4 | | - genuine observations. Using both simulated and real data in |
5 | | - scenarios involving counterfactuality, we discuss the general task |
6 | | - of evaluating a simulator’s quality, with a focus on comparisons of |
7 | | - statistical properties and predictive performance. While the |
8 | | - simulator built from simulated data shows minor discrepancies, the |
9 | | - results with real data reveal more substantial challenges. Beyond |
10 | | - the technical analysis, we reflect on the broader implications of |
11 | | - simulator design, and consider its role in modeling reality. |
12 | | - authors@: Sandrine Boulet and Antoine Chambaz |
13 | | - bibtex@: >+ |
14 | | - @article{boulet2025, |
15 | | - author = {Boulet, Sandrine and Chambaz, Antoine}, |
| 1 | +- abstract': >- |
| 2 | + In Bayesian statistics, the choice of the prior can have |
| 3 | + an important influence on the posterior and the parameter |
| 4 | + estimation, especially when few data samples are available. To limit |
| 5 | + the added subjectivity from a priori information, one can use the |
| 6 | + framework of objective priors, more particularly, we focus on |
| 7 | + reference priors in this work. However, computing such priors is a |
| 8 | + difficult task in general. Hence, we consider cases where the |
| 9 | + reference prior simplifies to the Jeffreys prior. We develop in this |
| 10 | + paper a flexible algorithm based on variational inference which |
| 11 | + computes approximations of priors from a set of parametric |
| 12 | + distributions using neural networks. We also show that our algorithm |
| 13 | + can retrieve modified Jeffreys priors when constraints are specified |
| 14 | + in the optimization problem to ensure the solution is proper. We |
| 15 | + propose a simple method to recover a relevant approximation of the |
| 16 | + parametric posterior distribution using Markov Chain Monte Carlo |
| 17 | + (MCMC) methods even if the density function of the parametric prior |
| 18 | + is not known in general. Numerical experiments on several |
| 19 | + statistical models of increasing complexity are presented. We show |
| 20 | + the usefulness of this approach by recovering the target |
| 21 | + distribution. The performance of the algorithm is evaluated on both |
| 22 | + prior and posterior distributions, jointly using variational |
| 23 | + inference and MCMC sampling. |
| 24 | + authors: Nils Baillie, Antoine Van Biesbroeck and Clément Gauchy |
| 25 | + bibtex: >+ |
| 26 | + @article{baillie2025, |
| 27 | + author = {Baillie, Nils and Van Biesbroeck, Antoine and Gauchy, |
| 28 | + Clément}, |
16 | 29 | publisher = {French Statistical Society}, |
17 | | - title = {Draw {Me} a {Simulator}}, |
| 30 | + title = {Variational Inference for Approximate Objective Priors Using |
| 31 | + Neural Networks}, |
18 | 32 | journal = {Computo}, |
19 | | - date = {2025-09-08}, |
20 | | - doi = {10.57750/w1hj-dw22}, |
| 33 | + date = {2025-12-01}, |
| 34 | + doi = {10.57750/76fh-t442}, |
21 | 35 | issn = {2824-7795}, |
22 | 36 | langid = {en}, |
23 | | - abstract = {This study investigates the use of Variational |
24 | | - Auto-Encoders to build a simulator that approximates the law of |
25 | | - genuine observations. Using both simulated and real data in |
26 | | - scenarios involving counterfactuality, we discuss the general task |
27 | | - of evaluating a simulator’s quality, with a focus on comparisons of |
28 | | - statistical properties and predictive performance. While the |
29 | | - simulator built from simulated data shows minor discrepancies, the |
30 | | - results with real data reveal more substantial challenges. Beyond |
31 | | - the technical analysis, we reflect on the broader implications of |
32 | | - simulator design, and consider its role in modeling reality.} |
| 37 | + abstract = {In Bayesian statistics, the choice of the prior can have |
| 38 | + an important influence on the posterior and the parameter |
| 39 | + estimation, especially when few data samples are available. To limit |
| 40 | + the added subjectivity from a priori information, one can use the |
| 41 | + framework of objective priors, more particularly, we focus on |
| 42 | + reference priors in this work. However, computing such priors is a |
| 43 | + difficult task in general. Hence, we consider cases where the |
| 44 | + reference prior simplifies to the Jeffreys prior. We develop in this |
| 45 | + paper a flexible algorithm based on variational inference which |
| 46 | + computes approximations of priors from a set of parametric |
| 47 | + distributions using neural networks. We also show that our algorithm |
| 48 | + can retrieve modified Jeffreys priors when constraints are specified |
| 49 | + in the optimization problem to ensure the solution is proper. We |
| 50 | + propose a simple method to recover a relevant approximation of the |
| 51 | + parametric posterior distribution using Markov Chain Monte Carlo |
| 52 | + (MCMC) methods even if the density function of the parametric prior |
| 53 | + is not known in general. Numerical experiments on several |
| 54 | + statistical models of increasing complexity are presented. We show |
| 55 | + the usefulness of this approach by recovering the target |
| 56 | + distribution. The performance of the algorithm is evaluated on both |
| 57 | + prior and posterior distributions, jointly using variational |
| 58 | + inference and MCMC sampling.} |
33 | 59 | } |
34 | 60 |
|
35 | | - date@: 2025-09-08 |
36 | | - description@: '' |
37 | | - doi@: 10.57750/w1hj-dw22 |
38 | | - draft@: false |
39 | | - journal@: Computo |
40 | | - pdf@: '' |
41 | | - repo@: published-202509-boulet-simulator |
42 | | - title@: Draw Me a Simulator |
43 | | - url@: '' |
44 | | - year@: 2025 |
45 | | - abstract': >- |
| 61 | + date: 2025-12-01 |
| 62 | + description: '' |
| 63 | + doi: 10.57750/76fh-t442 |
| 64 | + draft: false |
| 65 | + journal: Computo |
| 66 | + pdf: '' |
| 67 | + repo: published-202512-baillie-varp |
| 68 | + title: Variational inference for approximate objective priors using neural networks |
| 69 | + url: '' |
| 70 | + year: 2025 |
| 71 | +- abstract': >- |
| 72 | + The Maximum Mean Discrepancy (MMD) is a kernel-based |
| 73 | + metric widely used for nonparametric tests and estimation. Recently, |
| 74 | + it has also been studied as an objective function for parametric |
| 75 | + estimation, as it has been shown to yield robust estimators. We have |
| 76 | + implemented MMD minimization for parameter inference in a wide range |
| 77 | + of statistical models, including various regression models, within |
| 78 | + an `R` package called `regMMD`. This paper provides an introduction |
| 79 | + to the `regMMD` package. We describe the available kernels and |
| 80 | + optimization procedures, as well as the default settings. Detailed |
| 81 | + applications to simulated and real data are provided. |
| 82 | + authors: Pierre Alquier and Mathieu Gerber |
| 83 | + bibtex: >+ |
| 84 | + @article{alquier2025, |
| 85 | + author = {Alquier, Pierre and Gerber, Mathieu}, |
| 86 | + publisher = {French Statistical Society}, |
| 87 | + title = {`regMMD`: An {`R`} Package for Parametric Estimation and |
| 88 | + Regression with Maximum Mean Discrepancy}, |
| 89 | + journal = {Computo}, |
| 90 | + date = {2025-11-18}, |
| 91 | + doi = {10.57750/d6d1-gb09}, |
| 92 | + issn = {2824-7795}, |
| 93 | + langid = {en}, |
| 94 | + abstract = {The Maximum Mean Discrepancy (MMD) is a kernel-based |
| 95 | + metric widely used for nonparametric tests and estimation. Recently, |
| 96 | + it has also been studied as an objective function for parametric |
| 97 | + estimation, as it has been shown to yield robust estimators. We have |
| 98 | + implemented MMD minimization for parameter inference in a wide range |
| 99 | + of statistical models, including various regression models, within |
| 100 | + an `R` package called `regMMD`. This paper provides an introduction |
| 101 | + to the `regMMD` package. We describe the available kernels and |
| 102 | + optimization procedures, as well as the default settings. Detailed |
| 103 | + applications to simulated and real data are provided.} |
| 104 | + } |
| 105 | +
|
| 106 | + date: 2025-11-18 |
| 107 | + description: This document provides a complete introduction to the template based on the `regMMD` package for `R`, that implements minimum distance estimation in various parametric and regression models using the maximum mean discrepancy (MMD) metric. |
| 108 | + doi: 10.57750/d6d1-gb09 |
| 109 | + draft: false |
| 110 | + journal: Computo |
| 111 | + pdf: '' |
| 112 | + repo: published-202511-alquier-regmmd |
| 113 | + title: '`regMMD`: an `R` package for parametric estimation and regression with maximum mean discrepancy' |
| 114 | + url: '' |
| 115 | + year: 2025 |
| 116 | +- abstract': >- |
| 117 | + This paper presents a new algorithm (and an additional |
| 118 | + trick) that allows to compute fastly an entire curve of post hoc |
| 119 | + bounds for the False Discovery Proportion when the underlying bound |
| 120 | + \$V\^{}*\_\{\textbackslash mathfrak\{R\}\}\$ construction is based |
| 121 | + on a reference family \$\textbackslash mathfrak\{R\}\$ with a forest |
| 122 | + structure à la @MR4178188. By an entire curve, we mean the values |
| 123 | + \$V\^{}*\_\{\textbackslash mathfrak\{R\}\}(S\_1),\textbackslash |
| 124 | + dotsc,V\^{}*\_\{\textbackslash mathfrak\{R\}\}(S\_m)\$ computed on a |
| 125 | + path of increasing selection sets \$S\_1\textbackslash |
| 126 | + subsetneq\textbackslash dotsb\textbackslash subsetneq S\_m\$, |
| 127 | + \$\textbar S\_t\textbar=t\$. The new algorithm leverages the fact |
| 128 | + that going from \$S\_t\$ to \$S\_\{t+1\}\$ is done by adding only |
| 129 | + one hypothesis. Compared to a more naive approach, the new algorithm |
| 130 | + has a complexity in \$O(\textbar\textbackslash mathcal K\textbar |
| 131 | + m)\$ instead of \$O(\textbar\textbackslash mathcal K\textbar |
| 132 | + m\^{}2)\$, where \$\textbar\textbackslash mathcal K\textbar\$ is the |
| 133 | + cardinality of the family. |
| 134 | + authors: Guillermo Durand |
| 135 | + bibtex: >+ |
| 136 | + @article{durand2025, |
| 137 | + author = {Durand, Guillermo}, |
| 138 | + publisher = {French Statistical Society}, |
| 139 | + title = {Fast Confidence Bounds for the False Discovery Proportion |
| 140 | + over a Path of Hypotheses}, |
| 141 | + journal = {Computo}, |
| 142 | + date = {2025-10-09}, |
| 143 | + doi = {10.57750/efbs-ef14}, |
| 144 | + issn = {2824-7795}, |
| 145 | + langid = {en}, |
| 146 | + abstract = {This paper presents a new algorithm (and an additional |
| 147 | + trick) that allows to compute fastly an entire curve of post hoc |
| 148 | + bounds for the False Discovery Proportion when the underlying bound |
| 149 | + \$V\^{}*\_\{\textbackslash mathfrak\{R\}\}\$ construction is based |
| 150 | + on a reference family \$\textbackslash mathfrak\{R\}\$ with a forest |
| 151 | + structure à la @MR4178188. By an entire curve, we mean the values |
| 152 | + \$V\^{}*\_\{\textbackslash mathfrak\{R\}\}(S\_1),\textbackslash |
| 153 | + dotsc,V\^{}*\_\{\textbackslash mathfrak\{R\}\}(S\_m)\$ computed on a |
| 154 | + path of increasing selection sets \$S\_1\textbackslash |
| 155 | + subsetneq\textbackslash dotsb\textbackslash subsetneq S\_m\$, |
| 156 | + \$\textbar S\_t\textbar=t\$. The new algorithm leverages the fact |
| 157 | + that going from \$S\_t\$ to \$S\_\{t+1\}\$ is done by adding only |
| 158 | + one hypothesis. Compared to a more naive approach, the new algorithm |
| 159 | + has a complexity in \$O(\textbar\textbackslash mathcal K\textbar |
| 160 | + m)\$ instead of \$O(\textbar\textbackslash mathcal K\textbar |
| 161 | + m\^{}2)\$, where \$\textbar\textbackslash mathcal K\textbar\$ is the |
| 162 | + cardinality of the family.} |
| 163 | + } |
| 164 | +
|
| 165 | + date: 2025-10-09 |
| 166 | + description: '' |
| 167 | + doi: 10.57750/efbs-ef14 |
| 168 | + draft: false |
| 169 | + journal: Computo |
| 170 | + pdf: '' |
| 171 | + repo: published-202510-durand-fast |
| 172 | + title: Fast confidence bounds for the false discovery proportion over a path of hypotheses |
| 173 | + url: '' |
| 174 | + year: 2025 |
| 175 | +- abstract': >- |
46 | 176 | This study investigates the use of Variational |
47 | 177 | Auto-Encoders to build a simulator that approximates the law of |
48 | 178 | genuine observations. Using both simulated and real data in |
|
86 | 216 | title: Draw Me a Simulator |
87 | 217 | url: '' |
88 | 218 | year: 2025 |
89 | | -- abstract'@: >- |
90 | | - Model-based clustering provides a principled way of |
91 | | - developing clustering methods. We develop a new model-based |
92 | | - clustering methods for count data. The method combines clustering |
93 | | - and variable selection for improved clustering. The method is based |
94 | | - on conditionally independent Poisson mixture models and Poisson |
95 | | - generalized linear models. The method is demonstrated on simulated |
96 | | - data and data from an ultra running race, where the method yields |
97 | | - excellent clustering and variable selection performance. |
98 | | - authors@: Julien Jacques and Thomas Brendan Murphy |
99 | | - bibtex@: >+ |
100 | | - @article{jacques2025, |
101 | | - author = {Jacques, Julien and Brendan Murphy, Thomas}, |
102 | | - publisher = {French Statistical Society}, |
103 | | - title = {Model-Based {Clustering} and {Variable} {Selection} for |
104 | | - {Multivariate} {Count} {Data}}, |
105 | | - journal = {Computo}, |
106 | | - date = {2025-07-01}, |
107 | | - doi = {10.57750/6v7b-8483}, |
108 | | - issn = {2824-7795}, |
109 | | - langid = {en}, |
110 | | - abstract = {Model-based clustering provides a principled way of |
111 | | - developing clustering methods. We develop a new model-based |
112 | | - clustering methods for count data. The method combines clustering |
113 | | - and variable selection for improved clustering. The method is based |
114 | | - on conditionally independent Poisson mixture models and Poisson |
115 | | - generalized linear models. The method is demonstrated on simulated |
116 | | - data and data from an ultra running race, where the method yields |
117 | | - excellent clustering and variable selection performance.} |
118 | | - } |
119 | | -
|
120 | | - date@: 2025-07-01 |
121 | | - description@: '' |
122 | | - doi@: 10.57750/6v7b-8483 |
123 | | - draft@: false |
124 | | - journal@: Computo |
125 | | - pdf@: '' |
126 | | - repo@: published-202507-jacques-count-data |
127 | | - title@: Model-Based Clustering and Variable Selection for Multivariate Count Data |
128 | | - url@: '' |
129 | | - year@: 2025 |
130 | | - abstract': >- |
| 219 | +- abstract': >- |
131 | 220 | Model-based clustering provides a principled way of |
132 | 221 | developing clustering methods. We develop a new model-based |
133 | 222 | clustering methods for count data. The method combines clustering |
|
0 commit comments