We consider a sensor that samples an $N$-state continuous-time Markov chain (CTMC)-based information source process, and transmits the observed state of the source, to a remote monitor tasked with timely tracking of the source process. The mismatch between the source and monitor processes is quantified by age of incorrect information (AoII), which penalizes the mismatch as it stays longer, and our objective is to minimize the average AoII under an average sampling rate constraint. We assume a perfect reverse channel and hence the sensor has information of the estimate while initiating a transmission or preempting an ongoing transmission. First, by modeling the problem as an average cost constrained semi-Markov decision process (CSMDP), we show that the structure of the problem gives rise to an optimum threshold policy for which the sensor initiates a transmission once the AoII exceeds a threshold depending on the instantaneous values of both the source and monitor processes. However, due to the high complexity of obtaining the optimum policy in this general setting, we consider a relaxed problem where the thresholds are allowed to be dependent only on the estimate. We show that this relaxed problem can be solved with a novel CSMDP formulation based on the theory of absorbing MCs, with a computational complexity of $\mathcal{O}(N^4)$, allowing one to obtain optimum policies for general CTMCs with over a hundred states.
Age of incorrect information (AoII) has recently been proposed as an alternative to existing information freshness metrics for real-time sampling and estimation problems involving information sources that are tracked by remote monitors. Different from existing metrics, AoII penalizes the incorrect information by increasing linearly with time as long as the source and the monitor are de-synchronized, and is reset when they are synchronized back. While AoII has generally been investigated for discrete time information sources, we develop a novel analytical model in this paper for push- and pull-based sampling and transmission of a continuous time Markov chain (CTMC) process. In the pull-based model, the sensor starts transmitting information on the observed CTMC only when a pull request from the monitor is received. On the other hand, in the push-based scenario, the sensor, being aware of the AoII process, samples and transmits when the AoII process exceeds a random threshold. The proposed analytical model for both scenarios is based on the construction of a discrete time MC (DTMC) making state transitions at the embedded epochs of synchronization points, using the theory of absorbing CTMCs, and in particular phase-type distributions. For a given sampling policy, analytical models to obtain the mean AoII and the average sampling rate are developed. Numerical results are presented to validate the analytical model as well as to provide insight on optimal sampling policies under sampling rate constraints.
A sensor observes a random phenomenon and transmits updates about the observed phenomenon to a remote monitor. The sensor may experience intermittent failures in which case the monitor will not receive any updates until the sensor has recovered. The monitor wants to keep a timely view of the observed process, as well as to detect any sensor failures, using the timings of the updates. We analyze this system model from a goal-oriented and semantic communication point of view, where the communication has multiple goals and multiple meanings/semantics. For the first goal, the performance is quantified by the age of information of the observed process at the monitor. For the second goal, the performance is quantified by the probability of error of the monitor's estimation of the sensor's failure status. Each arriving update packet brings both an information update and an indication about the sensor's status. The monitor estimates the failure status of the sensor by using the timings of the received updates. This estimation is subject to error, since a long period without any update receptions may be due to a low update rate or a failure of the sensor. We examine the trade-off between these two goals. We show that the probability of error of estimating a sensor failure decreases with increased update rate, however, the age of information is minimized with an intermediate update rate (not too low or high).