adds saving and loading history to and from disk

This commit is contained in:
2025-01-01 10:08:50 -05:00
parent dfd524ba19
commit 9136c5fd71
15 changed files with 602 additions and 146 deletions

View File

@@ -1,22 +1,56 @@
use crate::telemetry::data_item::TelemetryDataItem;
use crate::telemetry::data_value::TelemetryDataValue;
use chrono::{DateTime, SecondsFormat, TimeDelta, Utc};
use log::error;
use chrono::{DateTime, DurationRound, SecondsFormat, TimeDelta, Utc};
use log::{error, info};
use std::collections::VecDeque;
use std::sync::RwLock;
use std::{fs, path};
use std::cmp::min;
use std::io::SeekFrom;
use std::path::PathBuf;
use std::sync::{Arc, RwLock};
use anyhow::{ensure, Context};
use tokio::fs::File;
use tokio::io::{AsyncReadExt, AsyncSeekExt, AsyncWriteExt};
use tokio::task::JoinHandle;
use crate::core::TelemetryDataType;
use crate::telemetry::data::TelemetryData;
use crate::telemetry::definition::TelemetryDefinition;
const FOLDER_DURATION: TimeDelta = TimeDelta::hours(1);
fn update_next_from(time_since_next_from: TimeDelta, maximum_resolution: TimeDelta, t: DateTime<Utc>, next_from: DateTime<Utc>) -> DateTime<Utc> {
match (
time_since_next_from.num_nanoseconds(),
maximum_resolution.num_nanoseconds(),
) {
(_, Some(0)) => t,
(Some(nanos_since_next_from), Some(maximum_resolution_nanos)) => {
let nanos_since_next_from = nanos_since_next_from as u64;
let maximum_resolution_nanos = maximum_resolution_nanos as u64;
let num_steps =
nanos_since_next_from.div_ceil(maximum_resolution_nanos);
if num_steps > i32::MAX as u64 {
t + maximum_resolution
} else {
next_from + maximum_resolution * num_steps as i32
}
}
_ => t + maximum_resolution, // If there is a gap so big it can't be represented in 2^63 nanoseconds (over 200 years) just skip forward
}
}
struct SegmentData {
values: Vec<TelemetryDataValue>,
timestamps: Vec<DateTime<Utc>>,
}
struct HistorySegment {
struct HistorySegmentRam {
start: DateTime<Utc>,
end: DateTime<Utc>,
data: RwLock<SegmentData>,
}
impl HistorySegment {
impl HistorySegmentRam {
fn new(start: DateTime<Utc>, end: DateTime<Utc>) -> Self {
Self {
start,
@@ -34,7 +68,7 @@ impl HistorySegment {
}
let mut data = self.data.write().unwrap_or_else(|err| {
error!("HistorySegment::insert - data was poisoned: {}", err);
error!("HistorySegmentRam::insert - data was poisoned: {}", err);
let lock = err.into_inner();
self.data.clear_poison();
lock
@@ -60,38 +94,23 @@ impl HistorySegment {
if from < self.end && self.start < to {
// If there is overlap with the range
let data = self.data.read().unwrap_or_else(|err| {
error!("HistorySegment::get - data was poisoned: {}", err);
error!("HistorySegmentRam::get - data was poisoned: {}", err);
let lock = err.into_inner();
self.data.clear_poison();
lock
});
let start = data.timestamps.partition_point(|x| x < &from);
let end = data.timestamps.partition_point(|x| x < &to);
if start < data.timestamps.len() {
for i in start..end {
for i in start..data.timestamps.len() {
let t = data.timestamps[i];
if t >= self.end {
break;
}
if t >= next_from {
let time_since_next_from = t - next_from;
next_from = match (
time_since_next_from.num_nanoseconds(),
maximum_resolution.num_nanoseconds(),
) {
(_, Some(0)) => t,
(Some(nanos_since_next_from), Some(maximum_resolution_nanos)) => {
let nanos_since_next_from = nanos_since_next_from as u64;
let maximum_resolution_nanos = maximum_resolution_nanos as u64;
let num_steps =
nanos_since_next_from.div_ceil(maximum_resolution_nanos);
if num_steps > i32::MAX as u64 {
t + maximum_resolution
} else {
next_from + maximum_resolution * num_steps as i32
}
}
_ => t + maximum_resolution, // If there is a gap so big it can't be represented in 2^63 nanoseconds (over 200 years) just skip forward
};
next_from = update_next_from(time_since_next_from, maximum_resolution, t, next_from);
result.push(TelemetryDataItem {
value: data.values[i].clone(),
timestamp: t.to_rfc3339_opts(SecondsFormat::Millis, true),
@@ -105,65 +124,298 @@ impl HistorySegment {
}
}
pub struct TelemetryHistory {
segments: RwLock<VecDeque<HistorySegment>>,
struct HistorySegmentDisk {
start: DateTime<Utc>,
end: DateTime<Utc>,
length: u64,
file: File
}
impl TelemetryHistory {
pub fn new() -> Self {
Self {
segments: RwLock::new(VecDeque::new()),
impl HistorySegmentDisk {
const TIMESTAMP_LENGTH: u64 = 8 + 4;
const HEADER_LENGTH: u64 = Self::TIMESTAMP_LENGTH + Self::TIMESTAMP_LENGTH + 8;
async fn save_to_disk(mut folder: PathBuf, mut segment: HistorySegmentRam) -> anyhow::Result<Self> {
// Get the path for the specific timestamp we want to save to disk
let folder_time = segment.start.duration_trunc(FOLDER_DURATION)?;
folder.push(folder_time.to_rfc3339_opts(SecondsFormat::Secs, true));
// Create the necessary folders
fs::create_dir_all(&folder)?;
let mut file = folder;
file.push(format!("{}.dat", segment.start.to_rfc3339_opts(SecondsFormat::Secs, true)));
let file = File::create(file).await?;
let mut result = Self {
start: segment.start,
end: segment.end,
length: 0,
file
};
let utc_offset_start = result.start - DateTime::UNIX_EPOCH;
let utc_offset_end = result.end - DateTime::UNIX_EPOCH;
// Write the segment bounds
result.file.write_i64(utc_offset_start.num_seconds()).await?;
result.file.write_i32(utc_offset_start.subsec_nanos()).await?;
result.file.write_i64(utc_offset_end.num_seconds()).await?;
result.file.write_i32(utc_offset_end.subsec_nanos()).await?;
let data = segment.data.get_mut().unwrap_or_else(|err| {
error!("HistorySegmentDisk::save_to_disk - data was poisoned: {}", err);
let lock = err.into_inner();
lock
});
ensure!(data.timestamps.len() == data.values.len(), "Invalid Segment Cannot Be Saved to Disk");
result.length = data.timestamps.len() as u64;
result.file.write_u64(result.length).await?;
// Write all the timestamps
for timestamp in &data.timestamps {
let utc_offset = *timestamp - DateTime::UNIX_EPOCH;
result.file.write_i64(utc_offset.num_seconds()).await?;
result.file.write_i32(utc_offset.subsec_nanos()).await?;
}
// Write all the values
for value in &data.values {
match value {
TelemetryDataValue::Float32(value) => result.file.write_f32(*value).await?,
TelemetryDataValue::Float64(value) => result.file.write_f64(*value).await?,
}
}
result.file.flush().await?;
Ok(result)
}
async fn load_to_ram(mut self, telemetry_data_type: TelemetryDataType) -> anyhow::Result<HistorySegmentRam> {
let mut segment_data = SegmentData {
values: Vec::with_capacity(self.length as usize),
timestamps: Vec::with_capacity(self.length as usize),
};
self.file.seek(SeekFrom::Start(Self::HEADER_LENGTH)).await?;
for _ in 0..self.length {
segment_data.timestamps.push(self.read_date_time().await?);
}
for _ in 0..self.length {
segment_data.values.push(self.read_telemetry_item(telemetry_data_type).await?);
}
Ok(HistorySegmentRam {
start: self.start,
end: self.end,
data: RwLock::new(segment_data),
})
}
async fn open(folder: PathBuf, start: DateTime<Utc>) -> anyhow::Result<Self> {
// Get the path for the specific timestamp we want to save to disk
let folder_time = start.duration_trunc(FOLDER_DURATION)?;
let mut file = folder;
file.push(folder_time.to_rfc3339_opts(SecondsFormat::Secs, true));
file.push(format!("{}.dat", start.to_rfc3339_opts(SecondsFormat::Secs, true)));
let mut file = File::open(file).await?;
// Write the segment bounds
let start_seconds = file.read_i64().await?;
let start_nanos = file.read_i32().await?;
let end_seconds = file.read_i64().await?;
let end_nanos = file.read_i32().await?;
let start = TimeDelta::new(start_seconds, start_nanos as u32).context("Failed to reconstruct start TimeDelta")?;
let end = TimeDelta::new(end_seconds, end_nanos as u32).context("Failed to reconstruct end TimeDelta")?;
let length = file.read_u64().await?;
Ok(HistorySegmentDisk {
start: DateTime::UNIX_EPOCH + start,
end: DateTime::UNIX_EPOCH + end,
length,
file,
})
}
async fn get(
&mut self,
from: DateTime<Utc>,
to: DateTime<Utc>,
maximum_resolution: TimeDelta,
telemetry_data_type: TelemetryDataType,
) -> anyhow::Result<(DateTime<Utc>, Vec<TelemetryDataItem>)> {
let mut result = vec![];
let mut next_from = from;
if from < self.end && self.start < to {
let start = self.partition_point(from).await?;
if start < self.length {
for i in start..self.length {
let t = self.get_date_time(i).await?;
if t >= self.end {
break;
}
if t >= next_from {
let time_since_next_from = t - next_from;
next_from = update_next_from(time_since_next_from, maximum_resolution, t, next_from);
result.push(TelemetryDataItem {
value: self.get_telemetry_item(i, telemetry_data_type).await?,
timestamp: t.to_rfc3339_opts(SecondsFormat::Millis, true),
});
}
}
}
}
Ok((next_from, result))
}
async fn read_date_time(&mut self) -> anyhow::Result<DateTime<Utc>> {
let seconds = self.file.read_i64().await?;
let nanos = self.file.read_i32().await?;
let start = TimeDelta::new(seconds, nanos as u32).context("Failed to reconstruct TimeDelta")?;
Ok(DateTime::UNIX_EPOCH + start)
}
async fn get_date_time(&mut self, index: u64) -> anyhow::Result<DateTime<Utc>> {
self.file.seek(SeekFrom::Start(Self::HEADER_LENGTH + index * Self::TIMESTAMP_LENGTH)).await?;
self.read_date_time().await
}
async fn read_telemetry_item(&mut self, telemetry_data_type: TelemetryDataType) -> anyhow::Result<TelemetryDataValue> {
match telemetry_data_type {
TelemetryDataType::Float32 => Ok(TelemetryDataValue::Float32(self.file.read_f32().await?)),
TelemetryDataType::Float64 => Ok(TelemetryDataValue::Float64(self.file.read_f64().await?)),
}
}
pub fn insert(
async fn get_telemetry_item(&mut self, index: u64, telemetry_data_type: TelemetryDataType) -> anyhow::Result<TelemetryDataValue> {
let item_length = match telemetry_data_type {
TelemetryDataType::Float32 => 4,
TelemetryDataType::Float64 => 8,
};
self.file.seek(SeekFrom::Start(Self::HEADER_LENGTH + self.length * Self::TIMESTAMP_LENGTH + index * item_length)).await?;
self.read_telemetry_item(telemetry_data_type).await
}
async fn partition_point(&mut self, date_time: DateTime<Utc>) -> anyhow::Result<u64> {
if self.length == 0 {
return Ok(0);
}
// left should be too early to insert
// right should be too late to insert
let mut left = 0;
let mut size = self.length;
while size > 1 {
let half = size / 2;
let mid = left + half;
let is_less = self.get_date_time(mid).await? < date_time;
if is_less {
left = mid;
}
size -= half;
}
Ok(left + if self.get_date_time(left).await? < date_time { 1 } else { 0 })
}
}
pub struct TelemetryHistory {
pub data: TelemetryData,
segments: tokio::sync::RwLock<VecDeque<HistorySegmentRam>>,
}
impl From<TelemetryData> for TelemetryHistory {
fn from(value: TelemetryData) -> Self {
Self {
data: value,
segments: tokio::sync::RwLock::new(VecDeque::new()),
}
}
}
impl From<TelemetryDefinition> for TelemetryHistory {
fn from(value: TelemetryDefinition) -> Self {
<TelemetryDefinition as Into<TelemetryData>>::into(value).into()
}
}
impl TelemetryHistory {
fn cleanup_segment(&self, service: &TelemetryHistoryService, history_segment_ram: HistorySegmentRam) -> JoinHandle<()> {
let mut path = service.data_root_folder.clone();
path.push(&self.data.definition.uuid);
tokio::spawn(async move {
match HistorySegmentDisk::save_to_disk(path, history_segment_ram).await {
// Immediately drop the segment - now that we've saved it to disk we don't need to keep it in memory
Ok(segment) => drop(segment),
Err(err) => {
error!("An error occurred saving telemetry history to disk: {}", err);
}
}
})
}
async fn get_disk_segment(&self, service: &TelemetryHistoryService, start: DateTime<Utc>) -> anyhow::Result<HistorySegmentDisk> {
let mut path = service.data_root_folder.clone();
path.push(&self.data.definition.uuid);
HistorySegmentDisk::open(path, start).await
}
async fn create_ram_segment(&self, start: DateTime<Utc>, service: &TelemetryHistoryService, telemetry_data_type: TelemetryDataType) -> HistorySegmentRam {
let ram = match self.get_disk_segment(service, start).await {
Ok(disk) => disk.load_to_ram(telemetry_data_type).await,
Err(e) => Err(e),
};
match ram {
Ok(ram) => ram,
Err(_) => HistorySegmentRam::new(
start,
start + service.segment_width,
),
}
}
pub async fn insert(
&self,
service: &TelemetryHistoryService,
value: TelemetryDataValue,
timestamp: DateTime<Utc>,
) {
let segments = self.segments.read().unwrap_or_else(|err| {
error!("TelemetryHistory::insert - segments was poisoned: {}", err);
let lock = err.into_inner();
self.segments.clear_poison();
lock
});
let segments = self.segments.read().await;
let segments = if segments.is_empty() || segments[segments.len() - 1].end < timestamp {
// We want to insert something that doesn't fit into our history
drop(segments);
let mut segments = self.segments.write().unwrap_or_else(|err| {
error!("TelemetryHistory::insert - segments was poisoned: {}", err);
let lock = err.into_inner();
self.segments.clear_poison();
lock
});
let mut segments = self.segments.write().await;
if segments.len() == 0 {
segments.push_back(HistorySegment::new(
timestamp,
timestamp + service.segment_width,
));
let start_time = timestamp.duration_trunc(service.segment_width).unwrap();
segments.push_back(self.create_ram_segment(start_time, service, self.data.definition.data_type).await);
} else {
while segments[segments.len() - 1].end < timestamp {
if segments.len() == service.max_segments {
let _ = segments.pop_front();
if let Some(segment) = segments.pop_front() {
let _ = self.cleanup_segment(service, segment);
}
}
let start_time = segments[segments.len() - 1].end;
segments.push_back(HistorySegment::new(
start_time,
start_time + service.segment_width,
));
segments.push_back(self.create_ram_segment(start_time, service, self.data.definition.data_type).await);
}
}
drop(segments);
self.segments.read().unwrap_or_else(|err| {
error!("TelemetryHistory::insert - segments was poisoned: {}", err);
let lock = err.into_inner();
self.segments.clear_poison();
lock
})
self.segments.read().await
} else {
segments
};
@@ -174,23 +426,60 @@ impl TelemetryHistory {
segments[segment_index].insert(value, timestamp);
}
pub fn get(
pub fn insert_sync(
history: Arc<Self>,
service: Arc<TelemetryHistoryService>,
value: TelemetryDataValue,
timestamp: DateTime<Utc>,
) {
tokio::spawn(async move {
history.insert(&service, value, timestamp).await;
});
}
pub async fn get(
&self,
from: DateTime<Utc>,
to: DateTime<Utc>,
maximum_resolution: TimeDelta,
telemetry_history_service: &TelemetryHistoryService
) -> Vec<TelemetryDataItem> {
let mut result = vec![];
let segments = self.segments.read().unwrap_or_else(|err| {
error!("TelemetryHistory::get - segments was poisoned: {}", err);
let lock = err.into_inner();
self.segments.clear_poison();
lock
});
let segments = self.segments.read().await;
let mut from = from;
{
let first_ram_segment = segments[0].start;
let start = from.duration_trunc(telemetry_history_service.segment_width).unwrap();
let end = (to + telemetry_history_service.segment_width).duration_trunc(telemetry_history_service.segment_width).unwrap();
let end = min(end, first_ram_segment);
let mut path = telemetry_history_service.data_root_folder.clone();
path.push(&self.data.definition.uuid);
let mut start = start;
while start < end {
match self.get_disk_segment(telemetry_history_service, start).await {
Ok(mut disk) => {
match disk.get(from, to, maximum_resolution, self.data.definition.data_type).await {
Ok((new_from, new_data)) => {
from = new_from;
result.extend(new_data);
},
Err(err) => {
error!("Failed to get from disk segment: {err}");
}
}
},
Err(_) => {}, // Ignore errors
}
start += telemetry_history_service.segment_width;
}
}
for i in 0..segments.len() {
let (new_from, new_data) = segments[i].get(from, to, maximum_resolution);
from = new_from;
@@ -199,18 +488,42 @@ impl TelemetryHistory {
result
}
pub async fn cleanup(&self, service: &TelemetryHistoryService) -> anyhow::Result<()> {
let mut segments = self.segments.write().await;
for segment in segments.drain(..) {
self.cleanup_segment(service, segment).await?;
}
Ok(())
}
}
pub struct TelemetryHistoryService {
segment_width: TimeDelta,
max_segments: usize,
data_root_folder: PathBuf,
}
impl TelemetryHistoryService {
pub fn new() -> Self {
Self {
pub fn new() -> anyhow::Result<Self> {
let result = Self {
segment_width: TimeDelta::minutes(1),
max_segments: 5,
}
data_root_folder: path::absolute("telemetry")?,
};
fs::create_dir_all(&result.data_root_folder)?;
info!("Recording Telemetry Data to {}", result.data_root_folder.to_string_lossy());
Ok(result)
}
pub fn get_metadata_file(&self) -> PathBuf {
let mut result = self.data_root_folder.clone();
result.push("metadata.json");
result
}
}