Loading...
Loading...
Browse, search and filter the latest cybersecurity research papers from arXiv
This paper proposes a new test for inequalities that are linear in possibly partially identified nuisance parameters. This type of hypothesis arises in a broad set of problems, including subvector inference for linear unconditional moment (in)equality models, specification testing of such models, and inference for parameters bounded by linear programs. The new test uses a two-step test statistic and a chi-squared critical value with data-dependent degrees of freedom that can be calculated by an elementary formula. Its simple structure and tuning-parameter-free implementation make it attractive for practical use. We establish uniform asymptotic validity of the test, demonstrate its finite-sample size and power in simulations, and illustrate its use in an empirical application that analyzes women's labor supply in response to a welfare policy reform.
Understanding the sources that contribute to fine particulate matter (PM$_{2.5}$) is of crucial importance for designing and implementing targeted air pollution mitigation strategies. Determining what factors contribute to a pollutant's concentration goes under the name of source apportionment and it is a problem long studied by atmospheric scientists and statisticians alike. In this paper, we propose a Bayesian model for source apportionment, that advances the literature on source apportionment by allowing estimation of the number of sources and accounting for spatial and temporal dependence in the observed pollutants' concentrations. Taking as example observations of six species of fine particulate matter observed over the course of a year, we present a latent functional factor model that expresses the space-time varying observations of log concentrations of the six pollutant as a linear combination of space-time varying emissions produced by an unknown number of sources each multiplied by the corresponding source's relative contribution to the pollutant. Estimation of the number of sources is achieved by introducing source-specific shrinkage parameters. Application of the model to simulated data showcases its ability to retrieve the true number of sources and to reliably estimate the functional latent factors, whereas application to PM$_{2.5}$ speciation data in California identifies 3 major sources for the six PM$_{2.5}$ species.
Satellite and reanalysis rainfall products (SREs) can serve as valuable complements or alternatives in data-sparse regions, but their significant biases necessitate correction. This study rigorously evaluates a suite of bias correction (BC) methods, including statistical approaches (LOCI, QM), machine learning (SVR, GPR), and hybrid techniques (LOCI-GPR, QM-GPR), applied to seven SREs across 38 stations in Ghana and Zambia, aimed at assessing their performance in rainfall detection and intensity estimation. Results indicate that the ENACTS product, which uniquely integrates a large number of station records, was the most corrigible SRE; in Zambia, nearly all BC methods successfully reduced the mean error on daily rainfall amounts at over 70% of stations. However, this performance requires further validation at independent stations not incorporated into the ENACTS product. Overall, the statistical methods (QM and LOCI) generally outperformed other techniques, although QM exhibited a tendency to inflate rainfall values. All corrected SREs demonstrated a high capability for detecting dry days (POD $\ge$ 0.80), suggesting their potential utility for drought applications. A critical limitation persisted, however, as most SREs and BC methods consistently failed to improve the detection of heavy and violent rainfall events (POD $\leq$ 0.2), highlighting a crucial area for future research.
Count regression models are necessary for examining discrete dependent variables alongside covariates. Nonetheless, when data display outliers, overdispersion, and an abundance of zeros, traditional methods like the zero-inflated negative binomial (ZINB) model sometimes do not yield a satisfactory fit, especially in the tail regions. This research presents a versatile, heavy-tailed discrete model as a resilient substitute for the ZINB model. The suggested framework is built by extending the generalized Pareto distribution and its zero-inflated version to the discrete domain. This formulation efficiently addresses both overdispersion and zero inflation, providing increased flexibility for heavy-tailed count data. Through intensive simulation studies and real-world implementations, the proposed models are thoroughly tested to see how well they work. The results show that our models always do better than classic negative binomial and zero-inflated negative binomial regressions when it comes to goodness-of-fit. This is especially true for datasets with a lot of zeros and outliers. These results highlight the proposed framework's potential as a strong and flexible option for modeling complicated count data.
We analyze loss development in NAIC Schedule P loss triangles using functional data analysis methods. Adopting the functional viewpoint, our dataset comprises 3300+ curves of incremental loss ratios (ILR) of workers' compensation lines over 24 accident years. Relying on functional data depth, we first study similarities and differences in development patterns based on company-specific covariates, as well as identify anomalous ILR curves. The exploratory findings motivate the probabilistic forecasting framework developed in the second half of the paper. We propose a functional model to complete partially developed ILR curves based on partial least squares regression of PCA scores. Coupling the above with functional bootstrapping allows us to quantify future ILR uncertainty jointly across all future lags. We demonstrate that our method has much better probabilistic scores relative to Chain Ladder and in particular can provide accurate functional predictive intervals.
In this work we propose a generalized additive functional regression model for partially observed functional data. Our approach accommodates functional predictors of varying dimensions without requiring imputation of missing observations. Both the functional coefficients and covariates are represented using basis function expansions, with B-splines used in this study, though the method is not restricted to any specific basis choice. Model coefficients are estimated via penalized likelihood, leveraging the mixed model representation of penalized splines for efficient computation and smoothing parameter estimation.The performance of the proposed approach is assessed through two simulation studies: one involving two one-dimensional functional covariates, and another using a two-dimensional functional covariate. Finally, we demonstrate the practical utility of our method in an application to air-pollution classification in Dimapur, India, where images are treated as observations of a two-dimensional functional variable. This case study highlights the models ability to effectively handle incomplete functional data and to accurately discriminate between pollution levels.
Heavy-tailed distributions, prevalent in a lot of real-world applications such as finance, telecommunications, queuing theory, and natural language processing, are challenging to model accurately owing to their slow tail decay. Bernstein phase-type (BPH) distributions, through their analytical tractability and good approximations in the non-tail region, can present a good solution, but they suffer from an inability to reproduce these heavy-tailed behaviors exactly, thus leading to inadequate performance in important tail areas. On the contrary, while highly adaptable to heavy-tailed distributions, hyperexponential (HE) models struggle in the body part of the distribution. Additionally, they are highly sensitive to initial parameter selection, significantly affecting their precision. To solve these issues, we propose a novel hybrid model of BPH and HE distributions, borrowing the most desirable features from each for enhanced approximation quality. Specifically, we leverage an optimization to set initial parameters for the HE component, significantly enhancing its robustness and reducing the possibility that the associated procedure results in an invalid HE model. Experimental validation demonstrates that the novel hybrid approach is more performant than individual application of BPH or HE models. More precisely, it can capture both the body and the tail of heavy-tailed distributions, with a considerable enhancement in matching parameters such as mean and coefficient of variation. Additional validation through experiments utilizing queuing theory proves the practical usefulness, accuracy, and precision of our hybrid approach.
Variational system identification is a new formulation of maximum likelihood for estimation of parameters of dynamical systems subject to process and measurement noise, such as aircraft flying in turbulence. This formulation is an alternative to the filter-error method that circumvents the solution of a Riccati equation and does not have problems with unstable predictors. In this paper, variational system identification is demonstrated for estimating aircraft parameters from real flight-test data. The results show that, in real applications of practical interest, it has better convergence properties than the filter-error method, reaching the optimum even when null initial guesses are used for all parameters and decision variables. This paper also presents the theory behind the method and practical recommendations for its use.
Automated monitoring of marine mammals in the St. Lawrence Estuary faces extreme challenges: calls span low-frequency moans to ultrasonic clicks, often overlap, and are embedded in variable anthropogenic and environmental noise. We introduce a multi-step, attention-guided framework that first segments spectrograms to generate soft masks of biologically relevant energy and then fuses these masks with the raw inputs for multi-band, denoised classification. Image and mask embeddings are integrated via mid-level fusion, enabling the model to focus on salient spectrogram regions while preserving global context. Using real-world recordings from the Saguenay St. Lawrence Marine Park Research Station in Canada, we demonstrate that segmentation-driven attention and mid-level fusion improve signal discrimination, reduce false positive detections, and produce reliable representations for operational marine mammal monitoring across diverse environmental conditions and signal-to-noise ratios. Beyond in-distribution evaluation, we further assess the generalization of Mask-Guided Classification (MGC) under distributional shifts by testing on spectrograms generated with alternative acoustic transformations. While high-capacity baseline models lose accuracy in this Out-of-distribution (OOD) setting, MGC maintains stable performance, with even simple fusion mechanisms (gated, concat) achieving comparable results across distributions. This robustness highlights the capacity of MGC to learn transferable representations rather than overfitting to a specific transformation, thereby reinforcing its suitability for large-scale, real-world biodiversity monitoring. We show that in all experimental settings, the MGC framework consistently outperforms baseline architectures, yielding substantial gains in accuracy on both in-distribution and OOD data.
We present tractable methods for detecting changes in player performance metrics and apply these methods to Major League Baseball (MLB) batting and pitching data from the 2023 and 2024 seasons. First, we derive principled benchmarks for when performance metrics can be considered statistically reliable, assuming no underlying change, using distributional assumptions and standard concentration inequalities. We then propose a changepoint detection algorithm that combines a likelihood-based approach with split-sample inference to control false positives, using either nonparametric tests or tests appropriate to the underlying data distribution. These tests incorporate a shift parameter, allowing users to specify the minimum magnitude of change to detect. We demonstrate the utility of this approach across several baseball applications: detecting changes in batter plate discipline metrics (e.g., chase and whiff rate), identifying velocity changes in pitcher fastballs, and validating velocity changepoints against a curated ground-truth dataset of pitchers who transitioned from relief to starting roles. Our method flags meaningful changes in 91% of these `ground-truth' cases and reveals that, for some metrics, more than 60% of detected changes occur in-season. While developed for baseball, the proposed framework is broadly applicable to any setting involving monitoring of individual performance over time.
In modern industrial settings, advanced acquisition systems allow for the collection of data in the form of profiles, that is, as functional relationships linking responses to explanatory variables. In this context, statistical process monitoring (SPM) aims to assess the stability of profiles over time in order to detect unexpected behavior. This review focuses on SPM methods that model profiles as functional data, i.e., smooth functions defined over a continuous domain, and apply functional data analysis (FDA) tools to address limitations of traditional monitoring techniques. A reference framework for monitoring multivariate functional data is first presented. This review then offers a focused survey of several recent FDA-based profile monitoring methods that extend this framework to address common challenges encountered in real-world applications. These include approaches that integrate additional functional covariates to enhance detection power, a robust method designed to accommodate outlying observations, a real-time monitoring technique for partially observed profiles, and two adaptive strategies that target the characteristics of the out-of-control distribution. These methods are all implemented in the R package funcharts, available on CRAN. Finally, a review of additional existing FDA-based profile monitoring methods is also presented, along with suggestions for future research.
Weather predictions are often provided as ensembles generated by repeated runs of numerical weather prediction models. These forecasts typically exhibit bias and inaccurate dependence structures due to numerical and dispersion errors, requiring statistical postprocessing for improved precision. A common correction strategy is the two-step approach: first adjusting the univariate forecasts, then reconstructing the multivariate dependence. The second step is usually handled with nonparametric methods, which can underperform when historical data are limited. Parametric alternatives, such as the Gaussian Copula Approach (GCA), offer theoretical advantages but often produce poorly calibrated multivariate forecasts due to random sampling of the corrected univariate margins. In this work, we introduce COBASE, a novel copula-based postprocessing framework that preserves the flexibility of parametric modeling while mimicking the nonparametric techniques through a rank-shuffling mechanism. This design ensures calibrated margins and realistic dependence reconstruction. We evaluate COBASE on multi-site 2-meter temperature forecasts from the ALADIN-LAEF ensemble over Austria and on joint forecasts of temperature and dew point temperature from the ECMWF system in the Netherlands. Across all regions, COBASE variants consistently outperform traditional copula-based approaches, such as GCA, and achieve performance on par with state-of-the-art nonparametric methods like SimSchaake and ECC, with only minimal differences across settings. These results position COBASE as a competitive and robust alternative for multivariate ensemble postprocessing, offering a principled bridge between parametric and nonparametric dependence reconstruction.
Recent advancements in technology have established terrestrial laser scanners (TLS) as a powerful instrument in geodetic deformation analysis. As TLS becomes increasingly integrated into this field, it is essential to develop a comprehensive stochastic model that accurately captures the measurement uncertainties. A key component of this model is the construction of a complete and valid variance-covariance matrix (VCM) for TLS polar measurements, which requires the estimation of variances for range, vertical, and horizontal angles, as well as their correlations. While angular variances can be obtained from manufacturer specifications, the range variance varies with different intensity measurements. As a primary contribution, this study presents an effective methodology for measuring and estimating TLS range variances using both raw and scaled intensity values. A two-dimensional scanning approach is applied to both controlled targets and arbitrary objects using TLS instruments that provide raw intensity values (e.g., Z+F~Imager~5016A) and those that output scaled intensities (e.g., Leica~ScanStation~P50). The methodology is further evaluated using field observations on a water dam surface. Overall, this work introduces a comprehensive workflow for modeling range uncertainties in high-end TLS systems.
The COVID-19 pandemic reshaped human mobility through policy interventions and voluntary behavioral changes. Mobility adaptions helped mitigate pandemic spread, however our knowledge which environmental, social, and demographic factors helped mobility reduction and pandemic mitigation is patchy. We introduce a Bayesian hierarchical model to quantify heterogeneity in mobility responses across time and space in Germany's 400 districts using anonymized mobile phone data. Decomposing mobility into a disease-responsive component and disease-independent factors (temperature, school vacations, public holidays) allows us to quantify the impact of each factor. We find significant differences in reaction to disease spread along the urban-rural gradient, with large cities reducing mobility most strongly. Employment sectors further help explain variance in reaction strength during the first wave, while political variables gain significance during the second wave. However, reduced mobility only partially translates to lower peak incidence, indicating the influence of other hidden factors. Our results identify key drivers of mobility reductions and demonstrate that mobility behavior can serve as an operational proxy for population response.
Examining emotion interactions as an emotion network in social media offers key insights into human psychology, yet few studies have explored how fluctuations in such emotion network evolve during crises and normal times. This study proposes a novel computational approach grounded in network theory, leveraging large-scale Japanese social media data spanning varied crisis events (earthquakes and COVID-19 vaccination) and non-crisis periods over the past decade. Our analysis identifies and evaluates links between emotions through the co-occurrence of emotion-related concepts (words), revealing a stable structure of emotion network across situations and over time at the population level. We find that some emotion links (represented as link strength) such as emotion links associated with Tension are significantly strengthened during earthquake and pre-vaccination periods. However, the rank of emotion links remains highly intact. These findings challenge the assumption that emotion co-occurrence is context-based and offer a deeper understanding of emotions' intrinsic structure. Moreover, our network-based framework offers a systematic, scalable method for analyzing emotion co-occurrence dynamics, opening new avenues for psychological research using large-scale textual data.
Electricity demand and generation have become increasingly unpredictable with the growing share of variable renewable energy sources in the power system. Forecasting electricity supply by fuel mix is crucial for market operation, ensuring grid stability, optimizing costs, integrating renewable energy sources, and supporting sustainable energy planning. We introduce two statistical methods, centering on forecast reconciliation and compositional data analysis, to forecast short-term electricity supply by different types of fuel mix. Using data for five electricity markets in Australia, we study the forecast accuracy of these techniques. The bottom-up hierarchical forecasting method consistently outperforms the other approaches. Moreover, fuel mix forecasting is most accurate in power systems with a higher share of stable fossil fuel generation.
Cluster randomized trials (CRTs) offer a practical alternative for addressing logistical challenges and ensuring feasibility in community health, education, and prevention studies, even though randomized controlled trials are considered the gold standard in evaluating therapeutic interventions. Despite their utility, CRTs are often criticized for limited precision and complex modeling requirements. Advances in robust Bayesian methods and the incorporation of spatial correlation into CRT design and analysis remain relatively underdeveloped. This paper introduces a Bayesian spatial point process framework that models individuals nested within geographic clusters while explicitly accounting for spatial dependence. We demonstrate that conventional non-spatial models consistently underestimate uncertainty and lead to misleading inferences, whereas our spatial approach improves estimation stability, controls type I error, and enhances statistical power. Our results underscore the value and need for wider adoption of spatial methods in CRT.
In multi-temporal InSAR, phase linking refers to the estimation of a single-reference interferometric phase history from the information contained in the coherence matrix of a distributed scatterer. Since the phase information in the coherence matrix is typically inconsistent, the extent to which the estimated phase history captures it must be assessed to exclude unreliable pixels from further processing. We introduce three quality criteria in the form of coefficients, for threshold-based pixel selection: a coefficient based on closure phase that quantifies the internal consistency of the phase information in the coherence matrix; a goodness-of-fit coefficient that quantifies how well a resulting phase history estimate approximates the phase information according to the characteristic optimization model of a given phase linking method; and an ambiguity coefficient that compares the goodness of fit of the original estimate with that of an orthogonal alternative. We formulate the phase linking methods and these criteria within a unified mathematical framework and discuss computational and algorithmic aspects. Unlike existing goodness-of-fit indicators, the proposed coefficients are normalized to the unit interval with explicit noise-floor correction, improving interpretability across stacks of different size. Experiments on TerraSAR-X data over Visp, Switzerland, indicate that the closure phase coefficient effectively pre-screens stable areas, the goodness-of-fit coefficient aligns with and systematically generalizes established quality indicators, and the ambiguity coefficient flags solutions that fit well but are unstable. Together, the coefficients enable systematic pixel selection and quality control in the interferometric processing of distributed scatterers.