We present StratFormer, a transformer-based meta-agent that learns to simultaneously model and exploit opponents in imperfect-information games through a two-phase curriculum. The first phase trains an opponent modeling head to identify behavioral patterns from action histories while the agent plays a game-theoretic optimal (GTO) policy. The second phase progressively shifts the policy toward best-response (BR) exploitation, guided by a per-opponent regularization schedule tied to exploitability. Our architecture introduces dual-turn tokens -- feature vectors constructed at both agent and opponent decision points -- coupled with bucket-rate features that encode opponent tendencies across five strategic contexts. On Leduc Hold'em, a small poker variant with six cards and two betting rounds, we test against six opponent archetypes at two strength levels each, with exploitability ranging from 0.15 to 1.26 Big Blinds (BB) per hand. StratFormer achieves an average exploitation gain of +0.106 BB per hand over GTO, with peak gains of +0.821 against highly exploitable opponents, while maintaining near-equilibrium safety.
Sub-THz bands are promising high bandwidth and data rates, and in the recent years the device technologies made large progress and provided a multitude of transceiver, power amplifier (PA) and phased array devices supporting the frequency bands above 100 GHz. The more painful aspect of sub-THz transmission is the increased power consumption, caused by the large data rates and the related data conversion and processing effort, and on the analog side the low achievable PA efficiency and the reduced achievable output power. When planning a deployment of sub-THz communication systems, the target coverage and throughput can be achieved with a variety of scenarios, which will be different with respect to locations and number of base stations and system architectures. Although leading to similar performance, they will differ significantly in the overall power consumption. With an accurate power consumption model, including also baseband (BB) processing functionality, and system level simulations for different hybrid beamforming and MIMO schemes the related variations in power consumption in relation to a given performance are evaluated. This paper shows the critical design aspects for energy efficient sub-THz deployments by highlighting the sub- THz specific trade-offs between different number of BS with different transmit powers but also changing number of BB units and RF chains.
The joint design of analog beamforming and power allocation is investigated for a single radio-frequency chain multiuser time-division multiple access system under a max-min signal-to-noise ratio (SNR) criterion. A hardware-efficient phased-array architecture is considered, where the beamforming vector is shared by all users and is subject to constant-modulus constraints. For any fixed analog beamformer, the optimal power allocation is first derived in closed form, by which the original problem is reduced to phase-shift optimization only. Then, globally optimal branch-and-bound (BB) algorithms are developed for discrete and continuous phase shifts. Numerical results show that the proposed BB algorithms achieve the global optimum and provide reliable benchmarks for evaluating the performance gap of low-complexity alternating-optimization methods.
In what way could a data breach involving government-issued IDs such as passports, driver's licenses, etc., rival a random voluntary disclosure on a nondescript social-media platform? At first glance, the former appears more significant, and that is a valid assessment. The disclosed data could contain an individual's date of birth and address; for all intents and purposes, a leak of that data would be disastrous. Given the threat, the latter scenario involving an innocuous online post seems comparatively harmless--or does it? From that post and others like it, a forensic linguist could stylometrically uncover equivalent pieces of information, estimating an age range for the author (adolescent or adult) and narrowing down their geographical location (specific country). While not an exact science--the determinations are statistical--stylometry can reveal comparable, though noticeably diluted, information about an individual. To prevent an ID from being breached, simply sharing it as little as possible suffices. Preventing the leakage of personal information from written text requires a more complex solution: adversarial stylometry. In this paper, we explore how performing homoglyph substitution--the replacement of characters with visually similar alternatives (e.g., "h" $\texttt{[U+0068]}$ $\rightarrow$ "h" $\texttt{[U+04BB]}$)--on text can degrade stylometric systems.
Battery state-of-health (SOH) reported by on-board battery management systems (BMS) is the primary metric available to electric vehicle (EV) owners and regulators, yet no study has validated its reliability across manufacturers against independent measurements. Here we show, through an epidemiological study of 1,114 EVs spanning five manufacturers and 375 days, that battery health reporting is fundamentally unreliable: real capacity differences of 12-25% exist within every model, but BMS SOH fails to track them, with correlations ranging from \r{ho} = 0.10 (non-significant) to \r{ho} = 0.62 only under restrictive filtering, while 384 vehicles do not expose SOH at all. A manufacturer-independent electrochemical marker achieves 74-89% degradation classification accuracy across all platforms without requiring BMS data, and a controlled laboratory validation on cells identical to those in the fleet confirms that partial-voltage-window charge measurements track reference capacity with \r{ho} > 0.80 across all 60 voltage windows (p < 0.001). These findings reveal a structural information asymmetry with direct implications for the EU Battery Regulation's 2027 SOH transparency mandate, California's Advanced Clean Cars (ACC) II durability requirements, warranty enforcement, used-vehicle valuation, right-to-repair legislation, and second-life battery markets.
Atherosclerosis of the carotid artery increases stroke risk. Atherosclerosis assessment with MRI requires multimodal and multidimensional segmentation of the carotid artery, reproducible extraction of biomarkers, and the visualization of segmentations and biomarkers. We developed CaroTo, a tool that allows for standardized carotid atherosclerosis assessment. It combines the capabilities of MEVISFlow with specialized tools for carotid geometry and vessel wall assessment. It supports manual and automatic segmentation for 2D, 2D+time, and 3D images, facilitating precise and consistent evaluations of carotid artery stenosis.
Reconfigurable antennas (RAs) utilize the electromagnetic (EM) domain to provide dynamic control over antenna radiation patterns, which offers an effective way to enhance power efficiency in wireless links. Unlike conventional arrays with fixed element patterns, RAs enable on-demand beam-pattern synthesis by directly controlling each antenna's EM characteristics. While existing research on RAs has primarily focused on improving spectral efficiency, this paper explores their application for downlink localization. Moreover, the majority of existing works focus on far-field scenarios with little attention on near-field (NF). Motivated by these gaps, we consider a synthesis model in which each antenna generates desired beampatterns from a finite set of EM basis functions. We then formulate a joint optimization problem for the baseband (BB) and EM precoders with the objective of minimizing the user equipment (UE) position error bound (PEB) in NF conditions. Our analytical derivations and extensive simulation results demonstrate that the proposed hybrid precoder design for RAs significantly improves UE positioning accuracy compared to traditional non-reconfigurable arrays.
The celebrated Myerson--Satterthwaite theorem shows that in bilateral trade, no mechanism can be simultaneously fully efficient, Bayesian incentive compatible (BIC), and budget balanced (BB). This naturally raises the question of how closely the gains from trade (GFT) achievable by a BIC and BB mechanism can approximate the first-best (fully efficient) benchmark. The optimal BIC and BB mechanism is typically complex and highly distribution-dependent, making it difficult to characterize directly. Consequently, much of the literature analyzes simpler mechanisms such as the Random-Offerer (RO) mechanism and establishes constant-factor guarantees relative to the first-best GFT. An important open question concerns the worst-case performance of the RO mechanism relative to first-best (FB) efficiency. While it was originally hypothesized that the approximation ratio $\frac{\text{GFT}_{\text{FB}}}{\text{GFT}_{\text{RO}}}$ is bounded by $2$, recent work provided counterexamples to this conjecture: Cai et al. proved that the ratio can be strictly larger than $2$, and Babaioff et al. exhibited an explicit example with ratio approximately $2.02$. In this work, we employ AlphaEvolve, an AI-guided evolutionary search framework, to explore the space of value distributions. We identify a new worst-case instance that yields an improved lower bound of $\frac{\text{GFT}_{\text{FB}}}{\text{GFT}_{\text{RO}}} \ge \textbf{2.0749}$. This establishes a new lower bound on the worst-case performance of the Random-Offerer mechanism, demonstrating a wider efficiency gap than previously known.
As AI models progress beyond simple chatbots into more complex workflows, we draw ever closer to the event horizon beyond which AI systems will be utilized in autonomous, self-maintaining feedback loops. Any autonomous AI system will depend on automated, verifiable rewards and feedback; in settings where ground truth is sparse or non-deterministic, one practical source of such rewards is an LLM-as-a-Judge. Although LLM judges continue to improve, the literature has yet to introduce systems capable of enforcing standards with strong guarantees, particularly when bias vectors are unknown or adversarially discovered. To remedy this issue, we propose average bias-boundedness (A-BB), an algorithmic framework which formally guarantees reductions of harm/impact as a result of any measurable bias in an LLM judge. Evaluating on Arena-Hard-Auto with four LLM judges, we achieve (tau=0.5, delta=0.01) bias-bounded guarantees while retaining 61-99% correlation with original rankings across formatting and schematic bias settings, with most judge-bias combinations exceeding 80%. The code to reproduce our findings is available at https://github.com/penfever/bias-bounded-evaluation.
Sampling from unnormalized densities using diffusion models has emerged as a powerful paradigm. However, while recent approaches that use least-squares `matching' objectives have improved scalability, they often necessitate significant trade-offs, such as restricting prior distributions or relying on unstable optimization schemes. By generalizing these methods as special forms of fixed-point iterations rooted in Nelson's relation, we develop a new method that addresses these limitations, called Bridge Matching Sampler (BMS). Our approach enables learning a stochastic transport map between arbitrary prior and target distributions with a single, scalable, and stable objective. Furthermore, we introduce a damped variant of this iteration that incorporates a regularization term to mitigate mode collapse and further stabilize training. Empirically, we demonstrate that our method enables sampling at unprecedented scales while preserving mode diversity, achieving state-of-the-art results on complex synthetic densities and high-dimensional molecular benchmarks.