Loading...
Loading...
Browse, search and filter the latest cybersecurity research papers from arXiv
Remote driving, or teleoperating Autonomous Vehicles (AVs), is a key application that emerging 5G networks aim to support. In this paper, we conduct a systematic feasibility study of AV teleoperation over commercial 5G networks from both cross-layer and end-to-end (E2E) perspectives. Given the critical importance of timely delivery of sensor data, such as camera and LiDAR data, for AV teleoperation, we focus in particular on the performance of uplink sensor data delivery. We analyze the impacts of Physical Layer (PHY layer) 5G radio network factors, including channel conditions, radio resource allocation, and Handovers (HOs), on E2E latency performance. We also examine the impacts of 5G networks on the performance of upper-layer protocols and E2E application Quality-of-Experience (QoE) adaptation mechanisms used for real-time sensor data delivery, such as Real-Time Streaming Protocol (RTSP) and Web Real Time Communication (WebRTC). Our study reveals the challenges posed by today's 5G networks and the limitations of existing sensor data streaming mechanisms. The insights gained will help inform the co-design of future-generation wireless networks, edge cloud systems, and applications to overcome the low-latency barriers in AV teleoperation.
Single-board computers, with their wide range of external interfaces, provide a cost-effective solution for studying animals and plants in their natural habitat. With the introduction of the Raspberry Pi Zero 2 W, which provides hardware-based image and video encoders, it is now possible to extend this application area to include video surveillance capabilities. This paper demonstrates a solution that offloads video stream generation from the Central Processing Unit (CPU) to hardware-based encoders. The flow of data through an encoding application is described, followed by a method of accelerating image processing by reducing the number of memory copies. The paper concludes with an example use case demonstrating the application of this new feature in an underwater camera.
Many scientific disciplines have traditionally advanced by iterating over hypotheses using labor-intensive trial-and-error, which is a slow and expensive process. Recent advances in computing, digitalization, and machine learning have introduced tools that promise to make scientific research faster by assisting in this iterative process. However, these advances are scattered across disciplines and only loosely connected, with specific computational methods being primarily developed for narrow domain-specific applications. Virtual Laboratories are being proposed as a unified formulation to help researchers navigate this increasingly digital landscape using common AI technologies. While conceptually promising, VLs are not yet widely adopted in practice, and concrete implementations remain limited.This paper explains how the Virtual Laboratory concept can be implemented in practice by introducing the modular software library VAILabs, designed to support scientific discovery. VAILabs provides a flexible workbench and toolbox for a broad range of scientific domains. We outline the design principles and demonstrate a proof-of-concept by mapping three concrete research tasks from differing fields as virtual laboratory workflows.
Machine learning (ML) frameworks rely heavily on pseudorandom number generators (PRNGs) for tasks such as data shuffling, weight initialization, dropout, and optimization. Yet, the statistical quality and reproducibility of these generators-particularly when integrated into frameworks like PyTorch, TensorFlow, and NumPy-are underexplored. In this paper, we compare the statistical quality of PRNGs used in ML frameworks (Mersenne Twister, PCG, and Philox) against their original C implementations. Using the rigorous TestU01 BigCrush test suite, we evaluate 896 independent random streams for each generator. Our findings challenge claims of statistical robustness, revealing that even generators labeled ''crush-resistant'' (e.g., PCG, Philox) may fail certain statistical tests. Surprisingly, we can observe some differences in failure profiles between the native and framework-integrated versions of the same algorithm, highlighting some implementation differences that may exist.
One classic idea from the cybernetics literature is the Every Good Regulator Theorem (EGRT). The EGRT provides a means to identify good regulation, or the conditions under which an agent (regulator) can match the dynamical behavior of a system. We reevaluate and recast the EGRT in a modern context to provide insight into how intelligent autonomous learning systems might utilize a compressed global representation (world model). One-to-one mappings between a regulator (R) and the corresponding system (S) provide a reduced representation that preserves useful variety to match all possible outcomes of a system. Secondarily, we question the role of purpose or autonomy in this process, demonstrating how physical paradigms such as temporal criticality, non-normal denoising, and alternating procedural acquisition can recast behavior as statistical mechanics and yield regulatory relationships. These diverse physical systems challenge the notion of tightly-coupled good regulation when applied to non-uniform and out-of-distribution phenomena. Modern definitions of intelligence are found to be inadequate, and can be improved upon by viewing intelligence as embodied non-purposeful good regulation. Overall, we aim to recast the EGRT as a tool for contemporary Artificial Intelligence (AI) architectures by considering the role of good regulation in the implementation of world models.
This article critically examines the foundational principles of contemporary AI methods, exploring the limitations that hinder its potential. We draw parallels between the modern AI landscape and the 20th-century Modern Synthesis in evolutionary biology, and highlight how advancements in evolutionary theory that augmented the Modern Synthesis, particularly those of Evolutionary Developmental Biology, offer insights that can inform a new design paradigm for AI. By synthesizing findings across AI and evolutionary theory, we propose a pathway to overcome existing limitations, enabling AI to achieve its aspirational goals. We also examine how this perspective transforms the idea of an AI-driven technological singularity from speculative futurism into a grounded prospect.
The fast pace of technological growth has created a heightened need for intelligent, autonomous monitoring systems in a variety of fields, especially in environmental applications. This project shows the design process and implementation of a proper dual node (master-slave) IoT-based monitoring system using STM32F103C8T6 microcontrollers. The structure of the wireless monitoring system studies the environmental conditions in real-time and can measure parameters like temperature, humidity, soil moisture, raindrop detection and obstacle distance. The relay of information occurs between the primary master node (designated as the Green House) to the slave node (the Red House) employing the HC-05 Bluetooth module for information transmission. Each node displays the sensor data on OLED screens and a visual or auditory alert is triggered based on predetermined thresholds. A comparative analysis of STM32 (ARM Cortex-M3) and Arduino (AVR) is presented to justify the STM32 used in this work for greater processing power, less energy use, and better peripherals. Practical challenges in this project arise from power distribution and Bluetooth configuration limits. Future work will explore the transition of a Wi-Fi communication protocol and develop a mobile monitoring robot to enhance scalability of the system. Finally, this research shows that ARM based embedded systems can provide real-time environmental monitoring systems that are reliable and consume low power.
The theory that all processes in the universe are computational is attractive in its promise to provide an understandable theory of everything. I want to suggest here that this pancomputationalism is not sufficiently clear on which problem it is trying to solve, and how. I propose two interpretations of pancomputationalism as a theory: I) the world is a computer and II) the world can be described as a computer. The first implies a thesis of supervenience of the physical over computation and is thus reduced ad absurdum. The second is underdetermined by the world, and thus equally unsuccessful as theory. Finally, I suggest that pancomputationalism as metaphor can be useful.
Simulation was launched in the 1950s, nicknamed a tool of "last resort." Over the years, this Operations Research (OR) method has made significant progress, and utilizing the accelerated advances in computer science (hardware and software, processing speed, and advanced information visualization capabilities) to improve simulation usability in research and practice. After overcoming the initial obstacles and the scare of outliving its usefulness in the 2000s, computer simulation has remained a popular OR tool applied in diverse industries and sectors, earning its popularity leading to the term "simulation everywhere." This study uses bibliographic data from research and practice literature to evaluate the evolutionary expansion in simulation, focusing on discrete-event simulation (DES). The results show asymmetrical but positive yearly literature out-put, broadened DES adoption in diverse fields, and sustained relevance as a scientific method for tackling old, new, and emerging issues. Also, DES is an essential tool in Industry 4.0 and plays a central role in digital transformation that has swept the industrial space, from manufacturing to healthcare and other sectors. With the emergence, ongoing adoption, and deployment of generative artificial intelligence (GenAI), future studies seek ways to integrate GenAI in DES to remain relevant and improve the modeling and simulation processes.
With growing consumer health awareness, ensuring food safety and quality throughout the supply chain is crucial, particularly for perishable goods. Contamination can occur during production, processing, or distribution, making real-time monitoring essential. This study proposes an affordable Smartphone-based food traceability system (FTS) that utilizes RFID technology and smartphone sensors. A smartphone-based RFID reader tracks products, while integrated sensors monitor temperature, humidity, and location during storage and transport. The system is assessed in the kimchi supply chain in Korea, providing real-time data to both managers and consumers. It offered comprehensive product tracking, including temperature and humidity records, ensuring transparency and safety. Compared to traditional methods, the proposed system demonstrated improved efficiency in handling large volumes of data while maintaining accurate traceability. The results highlight its potential for enhancing food safety and quality across supply chains.
Governments around the world have increasingly adopted digital transformation (DT) initiatives to increase their strategic competitiveness in the global market. To support successful DT, governments have to introduce new governance logics and revise IT strategies to facilitate DT initiatives. In this study, we report a case study of how Enterprise Architecture (EA) concepts were introduced and translated into practices in Vietnamese government agencies over a span of 15 years. This translation process has enabled EA concepts to facilitate various DT initiatives such as e-government, digitalization, to name a few. Our findings suggest two mechanisms in the translation process: a theorization mechanism to generalize local practices into field-level abstract concepts, making them easier to spread, while a contextualization mechanism unpacks these concepts into practical, adaptable approaches, aligning EA with adopters' priorities and increasing its chances of dissemination. Furthermore, our findings illustrate how translation happened when the initial concepts are ambiguous and not-well-understood by adopters. In this situation, there is a need for widespread experiments and sense-making among pioneers before field- and organizational-level translation can occur.
Commuting Origin-Destination (OD) flows capture movements of people from residences to workplaces, representing the predominant form of intra-city mobility and serving as a critical reference for understanding urban dynamics and supporting sustainable policies. However, acquiring such data requires costly, time-consuming censuses. In this study, we introduce a commuting OD flow dataset for cities around the world, spanning 6 continents, 179 countries, and 1,625 cities, providing unprecedented coverage of dynamics under diverse urban environments. Specifically, we collected fine-grained demographic data, satellite imagery, and points of interest~(POIs) for each city as foundational inputs to characterize the functional roles of urban regions. Leveraging these, a deep generative model is employed to capture the complex relationships between urban geospatial features and human mobility, enabling the generation of commuting OD flows between urban regions. Comprehensively, validation shows that the spatial distributions of the generated flows closely align with real-world observations. We believe this dataset offers a valuable resource for advancing sustainable urban development research in urban science, data science, transportation engineering, and related fields.
The proliferation of IoT in cities, combined with Digital Twins, creates a rich data foundation for Smart Cities aimed at improving urban life and operations. Generative AI (GenAI) significantly enhances this potential, moving beyond traditional AI analytics and predictions by processing multimodal content and generating novel outputs like text and simulations. Using specialized or foundational models, GenAI's natural language abilities such as Natural Language Understanding (NLU) and Natural Language Generation (NLG) can power tailored applications and unified interfaces, dramatically lowering barriers for users interacting with complex smart city systems. In this paper, we focus on GenAI applications based on conversational interfaces within the context of three critical user archetypes in a Smart City - Citizens, Operators and Planners. We identify and review GenAI models and techniques that have been proposed or deployed for various urban subsystems in the contexts of these user archetypes. We also consider how GenAI can be built on the existing data foundation of official city records, IoT data streams and Urban Digital Twins. We believe this work represents the first comprehensive summarization of GenAI techniques for Smart Cities from the lens of the critical users in a Smart City.
Cooperative perception extends the perception capabilities of autonomous vehicles by enabling multi-agent information sharing via Vehicle-to-Everything (V2X) communication. Unlike traditional onboard sensors, V2X acts as a dynamic "information sensor" characterized by limited communication, heterogeneity, mobility, and scalability. This survey provides a comprehensive review of recent advancements from the perspective of information-centric cooperative perception, focusing on three key dimensions: information representation, information fusion, and large-scale deployment. We categorize information representation into data-level, feature-level, and object-level schemes, and highlight emerging methods for reducing data volume and compressing messages under communication constraints. In information fusion, we explore techniques under both ideal and non-ideal conditions, including those addressing heterogeneity, localization errors, latency, and packet loss. Finally, we summarize system-level approaches to support scalability in dense traffic scenarios. Compared with existing surveys, this paper introduces a new perspective by treating V2X communication as an information sensor and emphasizing the challenges of deploying cooperative perception in real-world intelligent transportation systems.
This paper presents a digital twin-empowered real-time optimal delivery system specifically validated through a proof-of-concept (PoC) demonstration of a real-world autonomous car-sharing service. This study integrates real-time data from roadside units (RSUs) and connected and autonomous vehicles (CAVs) within a digital twin of a campus environment to address the dynamic challenges of urban traffic. The proposed system leverages the Age of Information (AoI) metric to optimize vehicle routing by maintaining data freshness and dynamically adapting to real-time traffic conditions. Experimental results from the PoC demonstrate a 22% improvement in delivery efficiency compared to conventional shortest-path methods that do not consider information freshness. Furthermore, digital twin-based simulation results demonstrate that this proposed system improves overall delivery efficiency by 12% and effectively reduces the peak average AoI by 23% compared to the conventional method, where each vehicle selects the shortest route without considering information freshness. This study confirms the practical feasibility of cooperative driving systems, highlighting their potential to enhance smart mobility solutions through scalable digital twin deployments in complex urban environments.
Movement data is prevalent across various applications and scientific fields, often characterized by its massive scale and complexity. Exploratory Data Analysis (EDA) plays a crucial role in summarizing and describing such data, enabling researchers to generate insights and support scientific hypotheses. Despite its importance, traditional EDA practices face limitations when applied to high-dimensional, unlabeled movement data. The complexity and multi-faceted nature of this type of data require more advanced methods that go beyond the capabilities of current EDA techniques. This study addresses the gap in current EDA practices by proposing a novel approach that leverages movement variable taxonomies and outlier detection. We hypothesize that organizing movement features into a taxonomy, and applying anomaly detection to combinations of taxonomic nodes, can reveal meaningful patterns and lead to more interpretable descriptions of the data. To test this hypothesis, we introduce TUMD, a new method that integrates movement taxonomies with outlier detection to enhance data analysis and interpretation. TUMD was evaluated across four diverse datasets of moving objects using fixed parameter values. Its effectiveness was assessed through two passes: the first pass categorized the majority of movement patterns as Kinematic, Geometric, or Hybrid for all datasets, while the second pass refined these behaviors into more specific categories such as Speed, Acceleration, or Indentation. TUMD met the effectiveness criteria in three datasets, demonstrating its ability to describe and refine movement behaviors. The results confirmed our hypothesis, showing that the combination of movement taxonomies and anomaly detection successfully uncovers meaningful and interpretable patterns within high-dimensional, unlabeled movement data.
Symmetric alpha-stable (S alpha S) distributions with alpha<2 lack finite classical Fisher information. Building on Johnson's framework, we define Mixed Fractional Information (MFI) via the initial rate of relative entropy dissipation during interpolation between S alpha S laws with differing scales, v and s. We demonstrate two equivalent formulations for MFI in this specific S alpha S-to-S alpha S setting. The first involves the derivative D'(v) of the relative entropy between the two S alpha S densities. The second uses an integral expectation E_gv[u(x,0) (pF_v(x) - pF_s(x))] involving the difference between Fisher scores (pF_v, pF_s) and a specific MMSE-related score function u(x,0) derived from the interpolation dynamics. Our central contribution is a rigorous proof of the consistency identity: D'(v) = (1/(alpha v)) E_gv[X (pF_v(X) - pF_s(X))]. This identity mathematically validates the equivalence of the two MFI formulations for S alpha S inputs, establishing MFI's internal coherence and directly linking entropy dissipation rates to score function differences. We further establish MFI's non-negativity (zero if and only if v=s), derive its closed-form expression for the Cauchy case (alpha=1), and numerically validate the consistency identity. MFI provides a finite, coherent, and computable information-theoretic measure for comparing S alpha S distributions where classical Fisher information fails, connecting entropy dynamics to score functions and estimation concepts. This work lays a foundation for exploring potential fractional I-MMSE relations and new functional inequalities tailored to heavy-tailed systems.
Optimal statistical decisions should transcend the language used to describe them. Yet, how do we guarantee that the choice of coordinates - the parameterisation of an optimisation problem - does not subtly dictate the solution? This paper reveals a fundamental geometric invariance principle. We first analyse the optimal combination of two asymptotically normal estimators under a strictly convex trace-AMSE risk. While methods for finding optimal weights are known, we prove that the resulting optimal estimator is invariant under direct affine reparameterisations of the weighting scheme. This exemplifies a broader principle we term meta-equivariance: the unique minimiser of any strictly convex, differentiable scalar objective over a matrix space transforms covariantly under any invertible affine reparameterisation of that space. Distinct from classical statistical equivariance tied to data symmetries, meta-equivariance arises from the immutable geometry of convex optimisation itself. It guarantees that optimality, in these settings, is not an artefact of representation but an intrinsic, coordinate-free truth.