**Rationale:** Integral Types were missing and are needed for Project Nautilus. **Changes:** - Implements Integral Data Types - u64 and i64 implemented through bigint Reviewed-on: #13 Co-authored-by: Sergey Savelyev <sergeysav.nn@gmail.com> Co-committed-by: Sergey Savelyev <sergeysav.nn@gmail.com>
664 lines
23 KiB
Rust
664 lines
23 KiB
Rust
use crate::serialization::file_ext::{ReadExt, WriteExt};
|
|
use crate::telemetry::data::TelemetryData;
|
|
use crate::telemetry::data_item::TelemetryDataItem;
|
|
use crate::telemetry::definition::TelemetryDefinition;
|
|
use anyhow::{anyhow, ensure, Context};
|
|
use api::data_type::DataType;
|
|
use api::data_value::DataValue;
|
|
use chrono::{DateTime, DurationRound, SecondsFormat, TimeDelta, Utc};
|
|
use log::{error, info};
|
|
use std::cmp::min;
|
|
use std::collections::VecDeque;
|
|
use std::fs;
|
|
use std::fs::File;
|
|
use std::io::{BufReader, BufWriter, Seek, SeekFrom, Write};
|
|
use std::path::PathBuf;
|
|
use std::sync::{Arc, RwLock};
|
|
use tokio::task::{spawn_blocking, JoinHandle};
|
|
|
|
const FOLDER_DURATION: TimeDelta = TimeDelta::hours(1);
|
|
|
|
fn update_next_from(
|
|
time_since_next_from: TimeDelta,
|
|
maximum_resolution: TimeDelta,
|
|
t: DateTime<Utc>,
|
|
next_from: DateTime<Utc>,
|
|
) -> DateTime<Utc> {
|
|
match (
|
|
time_since_next_from.num_nanoseconds(),
|
|
maximum_resolution.num_nanoseconds(),
|
|
) {
|
|
(_, Some(0)) => t,
|
|
(Some(nanos_since_next_from), Some(maximum_resolution_nanos)) => {
|
|
let nanos_since_next_from = nanos_since_next_from as u64;
|
|
let maximum_resolution_nanos = maximum_resolution_nanos as u64;
|
|
let num_steps = nanos_since_next_from.div_ceil(maximum_resolution_nanos);
|
|
if num_steps > i32::MAX as u64 {
|
|
t + maximum_resolution
|
|
} else {
|
|
next_from + maximum_resolution * num_steps as i32
|
|
}
|
|
}
|
|
_ => t + maximum_resolution, // If there is a gap so big it can't be represented in 2^63 nanoseconds (over 200 years) just skip forward
|
|
}
|
|
}
|
|
|
|
struct SegmentData {
|
|
values: Vec<DataValue>,
|
|
timestamps: Vec<DateTime<Utc>>,
|
|
}
|
|
|
|
struct HistorySegmentRam {
|
|
start: DateTime<Utc>,
|
|
end: DateTime<Utc>,
|
|
data: RwLock<SegmentData>,
|
|
}
|
|
|
|
impl HistorySegmentRam {
|
|
fn new(start: DateTime<Utc>, end: DateTime<Utc>) -> Self {
|
|
Self {
|
|
start,
|
|
end,
|
|
data: RwLock::new(SegmentData {
|
|
values: vec![],
|
|
timestamps: vec![],
|
|
}),
|
|
}
|
|
}
|
|
|
|
fn insert(&self, value: DataValue, timestamp: DateTime<Utc>) {
|
|
if timestamp < self.start || timestamp >= self.end {
|
|
return;
|
|
}
|
|
|
|
let mut data = self.data.write().unwrap_or_else(|err| {
|
|
error!("HistorySegmentRam::insert - data was poisoned: {}", err);
|
|
let lock = err.into_inner();
|
|
self.data.clear_poison();
|
|
lock
|
|
});
|
|
|
|
// Find the point where we should insert this item
|
|
let index = data.timestamps.partition_point(|item| item <= ×tamp);
|
|
// Insert the item
|
|
data.timestamps.insert(index, timestamp);
|
|
data.values.insert(index, value);
|
|
}
|
|
|
|
fn get(
|
|
&self,
|
|
from: DateTime<Utc>,
|
|
to: DateTime<Utc>,
|
|
maximum_resolution: TimeDelta,
|
|
) -> (DateTime<Utc>, Vec<TelemetryDataItem>) {
|
|
let mut result = vec![];
|
|
|
|
let mut next_from = from;
|
|
|
|
if from < self.end && self.start < to {
|
|
// If there is overlap with the range
|
|
let data = self.data.read().unwrap_or_else(|err| {
|
|
error!("HistorySegmentRam::get - data was poisoned: {}", err);
|
|
let lock = err.into_inner();
|
|
self.data.clear_poison();
|
|
lock
|
|
});
|
|
|
|
let start = data.timestamps.partition_point(|x| x < &from);
|
|
|
|
if start < data.timestamps.len() {
|
|
for i in start..data.timestamps.len() {
|
|
let t = data.timestamps[i];
|
|
if t >= self.end {
|
|
break;
|
|
}
|
|
if t >= next_from {
|
|
let time_since_next_from = t - next_from;
|
|
next_from = update_next_from(
|
|
time_since_next_from,
|
|
maximum_resolution,
|
|
t,
|
|
next_from,
|
|
);
|
|
result.push(TelemetryDataItem {
|
|
value: data.values[i],
|
|
timestamp: t.to_rfc3339_opts(SecondsFormat::Millis, true),
|
|
});
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
(next_from, result)
|
|
}
|
|
}
|
|
|
|
struct HistorySegmentFile {
|
|
start: DateTime<Utc>,
|
|
end: DateTime<Utc>,
|
|
length: u64,
|
|
file: BufReader<File>,
|
|
file_position: i64,
|
|
}
|
|
|
|
impl HistorySegmentFile {
|
|
const TIMESTAMP_LENGTH: u64 = 8 + 4;
|
|
const HEADER_LENGTH: u64 = Self::TIMESTAMP_LENGTH + Self::TIMESTAMP_LENGTH + 8;
|
|
|
|
fn save_to_disk(mut folder: PathBuf, mut segment: HistorySegmentRam) -> anyhow::Result<Self> {
|
|
// Get the path for the specific timestamp we want to save to disk
|
|
let folder_time = segment.start.duration_trunc(FOLDER_DURATION)?;
|
|
folder.push(folder_time.to_rfc3339_opts(SecondsFormat::Secs, true));
|
|
|
|
// Create the necessary folders
|
|
fs::create_dir_all(&folder)?;
|
|
|
|
let mut file = folder;
|
|
file.push(format!(
|
|
"{}.dat",
|
|
segment.start.to_rfc3339_opts(SecondsFormat::Secs, true)
|
|
));
|
|
|
|
let mut file = BufWriter::new(File::create(file)?);
|
|
|
|
let utc_offset_start = segment.start - DateTime::UNIX_EPOCH;
|
|
let utc_offset_end = segment.end - DateTime::UNIX_EPOCH;
|
|
|
|
// Write the segment bounds
|
|
file.write_data::<i64>(utc_offset_start.num_seconds())?;
|
|
file.write_data::<i32>(utc_offset_start.subsec_nanos())?;
|
|
file.write_data::<i64>(utc_offset_end.num_seconds())?;
|
|
file.write_data::<i32>(utc_offset_end.subsec_nanos())?;
|
|
|
|
let data = segment.data.get_mut().unwrap_or_else(|err| {
|
|
error!(
|
|
"HistorySegmentDisk::save_to_disk - data was poisoned: {}",
|
|
err
|
|
);
|
|
err.into_inner()
|
|
});
|
|
|
|
ensure!(
|
|
data.timestamps.len() == data.values.len(),
|
|
"Invalid Segment Cannot Be Saved to Disk"
|
|
);
|
|
|
|
let length = data.timestamps.len() as u64;
|
|
file.write_data::<u64>(length)?;
|
|
|
|
// Write all the timestamps
|
|
for timestamp in &data.timestamps {
|
|
let utc_offset = *timestamp - DateTime::UNIX_EPOCH;
|
|
file.write_data::<i64>(utc_offset.num_seconds())?;
|
|
file.write_data::<i32>(utc_offset.subsec_nanos())?;
|
|
}
|
|
|
|
// Write all the values
|
|
for value in &data.values {
|
|
match value {
|
|
DataValue::Float32(value) => file.write_data(*value)?,
|
|
DataValue::Float64(value) => file.write_data(*value)?,
|
|
DataValue::Boolean(value) => file.write_data(*value)?,
|
|
DataValue::Int8(value) => file.write_data(*value)?,
|
|
DataValue::Int16(value) => file.write_data(*value)?,
|
|
DataValue::Int32(value) => file.write_data(*value)?,
|
|
DataValue::Int64(value) => file.write_data(*value)?,
|
|
DataValue::Unsigned8(value) => file.write_data(*value)?,
|
|
DataValue::Unsigned16(value) => file.write_data(*value)?,
|
|
DataValue::Unsigned32(value) => file.write_data(*value)?,
|
|
DataValue::Unsigned64(value) => file.write_data(*value)?,
|
|
}
|
|
}
|
|
|
|
file.flush()?;
|
|
|
|
let mut file = BufReader::new(file.into_inner()?);
|
|
file.seek(SeekFrom::Start(0))?;
|
|
Ok(Self {
|
|
start: segment.start,
|
|
end: segment.end,
|
|
length: 0,
|
|
file,
|
|
file_position: 0,
|
|
})
|
|
}
|
|
|
|
fn load_to_ram(mut self, telemetry_data_type: DataType) -> anyhow::Result<HistorySegmentRam> {
|
|
let mut segment_data = SegmentData {
|
|
values: Vec::with_capacity(self.length as usize),
|
|
timestamps: Vec::with_capacity(self.length as usize),
|
|
};
|
|
|
|
self.file.seek(SeekFrom::Start(Self::HEADER_LENGTH))?;
|
|
self.file_position = Self::HEADER_LENGTH as i64;
|
|
for _ in 0..self.length {
|
|
segment_data.timestamps.push(self.read_date_time()?);
|
|
}
|
|
for _ in 0..self.length {
|
|
segment_data
|
|
.values
|
|
.push(self.read_telemetry_item(telemetry_data_type)?);
|
|
}
|
|
|
|
Ok(HistorySegmentRam {
|
|
start: self.start,
|
|
end: self.end,
|
|
data: RwLock::new(segment_data),
|
|
})
|
|
}
|
|
|
|
fn open(folder: PathBuf, start: DateTime<Utc>) -> anyhow::Result<Self> {
|
|
// Get the path for the specific timestamp we want to save to disk
|
|
let folder_time = start.duration_trunc(FOLDER_DURATION)?;
|
|
let mut file = folder;
|
|
file.push(folder_time.to_rfc3339_opts(SecondsFormat::Secs, true));
|
|
file.push(format!(
|
|
"{}.dat",
|
|
start.to_rfc3339_opts(SecondsFormat::Secs, true)
|
|
));
|
|
|
|
let mut file = BufReader::new(File::open(file)?);
|
|
|
|
// Write the segment bounds
|
|
let start_seconds = file.read_data::<i64>()?;
|
|
let start_nanos = file.read_data::<i32>()?;
|
|
let end_seconds = file.read_data::<i64>()?;
|
|
let end_nanos = file.read_data::<i32>()?;
|
|
let start = TimeDelta::new(start_seconds, start_nanos as u32)
|
|
.context("Failed to reconstruct start TimeDelta")?;
|
|
let end = TimeDelta::new(end_seconds, end_nanos as u32)
|
|
.context("Failed to reconstruct end TimeDelta")?;
|
|
|
|
let length = file.read_data::<u64>()?;
|
|
|
|
file.seek(SeekFrom::Start(0))?;
|
|
Ok(HistorySegmentFile {
|
|
start: DateTime::UNIX_EPOCH + start,
|
|
end: DateTime::UNIX_EPOCH + end,
|
|
length,
|
|
file,
|
|
file_position: 0,
|
|
})
|
|
}
|
|
|
|
fn get(
|
|
&mut self,
|
|
from: DateTime<Utc>,
|
|
to: DateTime<Utc>,
|
|
maximum_resolution: TimeDelta,
|
|
telemetry_data_type: DataType,
|
|
) -> anyhow::Result<(DateTime<Utc>, Vec<TelemetryDataItem>)> {
|
|
self.file_position = 0;
|
|
self.file.seek(SeekFrom::Start(0))?;
|
|
let mut result = vec![];
|
|
|
|
let mut next_from = from;
|
|
|
|
if from < self.end && self.start < to {
|
|
let start = self.partition_point(from)?;
|
|
if start < self.length {
|
|
for i in start..self.length {
|
|
let t = self.get_date_time(i)?;
|
|
if t >= self.end {
|
|
break;
|
|
}
|
|
if t >= next_from {
|
|
let time_since_next_from = t - next_from;
|
|
next_from = update_next_from(
|
|
time_since_next_from,
|
|
maximum_resolution,
|
|
t,
|
|
next_from,
|
|
);
|
|
result.push(TelemetryDataItem {
|
|
value: self.get_telemetry_item(i, telemetry_data_type)?,
|
|
timestamp: t.to_rfc3339_opts(SecondsFormat::Millis, true),
|
|
});
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok((next_from, result))
|
|
}
|
|
|
|
fn read_date_time(&mut self) -> anyhow::Result<DateTime<Utc>> {
|
|
let seconds = self.file.read_data::<i64>()?;
|
|
let nanos = self.file.read_data::<i32>()?;
|
|
self.file_position += 8 + 4;
|
|
let start =
|
|
TimeDelta::new(seconds, nanos as u32).context("Failed to reconstruct TimeDelta")?;
|
|
Ok(DateTime::UNIX_EPOCH + start)
|
|
}
|
|
|
|
fn get_date_time(&mut self, index: u64) -> anyhow::Result<DateTime<Utc>> {
|
|
let desired_position = Self::HEADER_LENGTH + index * Self::TIMESTAMP_LENGTH;
|
|
let seek_amount = desired_position as i64 - self.file_position;
|
|
self.file_position += seek_amount;
|
|
self.file.seek_relative(seek_amount)?;
|
|
self.read_date_time()
|
|
}
|
|
|
|
fn read_telemetry_item(&mut self, telemetry_data_type: DataType) -> anyhow::Result<DataValue> {
|
|
self.file_position += telemetry_data_type.get_data_length() as i64;
|
|
Ok(match telemetry_data_type {
|
|
DataType::Float32 => self.file.read_data::<f32>()?.into(),
|
|
DataType::Float64 => self.file.read_data::<f64>()?.into(),
|
|
DataType::Boolean => self.file.read_data::<bool>()?.into(),
|
|
DataType::Int8 => self.file.read_data::<i8>()?.into(),
|
|
DataType::Int16 => self.file.read_data::<i16>()?.into(),
|
|
DataType::Int32 => self.file.read_data::<i32>()?.into(),
|
|
DataType::Int64 => self.file.read_data::<i64>()?.into(),
|
|
DataType::Unsigned8 => self.file.read_data::<u8>()?.into(),
|
|
DataType::Unsigned16 => self.file.read_data::<u16>()?.into(),
|
|
DataType::Unsigned32 => self.file.read_data::<u32>()?.into(),
|
|
DataType::Unsigned64 => self.file.read_data::<u64>()?.into(),
|
|
})
|
|
}
|
|
|
|
fn get_telemetry_item(
|
|
&mut self,
|
|
index: u64,
|
|
telemetry_data_type: DataType,
|
|
) -> anyhow::Result<DataValue> {
|
|
let item_length = telemetry_data_type.get_data_length();
|
|
let desired_position =
|
|
Self::HEADER_LENGTH + self.length * Self::TIMESTAMP_LENGTH + index * item_length;
|
|
let seek_amount = desired_position as i64 - self.file_position;
|
|
self.file_position += seek_amount;
|
|
self.file.seek_relative(seek_amount)?;
|
|
self.read_telemetry_item(telemetry_data_type)
|
|
}
|
|
|
|
fn partition_point(&mut self, date_time: DateTime<Utc>) -> anyhow::Result<u64> {
|
|
if self.length == 0 {
|
|
return Ok(0);
|
|
}
|
|
|
|
// left should be too early to insert
|
|
// right should be too late to insert
|
|
let mut left = 0;
|
|
let mut size = self.length;
|
|
|
|
while size > 1 {
|
|
let half = size / 2;
|
|
let mid = left + half;
|
|
|
|
let is_less = self.get_date_time(mid)? < date_time;
|
|
if is_less {
|
|
left = mid;
|
|
}
|
|
size -= half;
|
|
}
|
|
|
|
Ok(left
|
|
+ if self.get_date_time(left)? < date_time {
|
|
1
|
|
} else {
|
|
0
|
|
})
|
|
}
|
|
}
|
|
|
|
pub struct TelemetryHistory {
|
|
pub data: TelemetryData,
|
|
segments: tokio::sync::RwLock<VecDeque<HistorySegmentRam>>,
|
|
}
|
|
|
|
impl From<TelemetryData> for TelemetryHistory {
|
|
fn from(value: TelemetryData) -> Self {
|
|
Self {
|
|
data: value,
|
|
segments: tokio::sync::RwLock::new(VecDeque::new()),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl From<TelemetryDefinition> for TelemetryHistory {
|
|
fn from(value: TelemetryDefinition) -> Self {
|
|
<TelemetryDefinition as Into<TelemetryData>>::into(value).into()
|
|
}
|
|
}
|
|
|
|
impl TelemetryHistory {
|
|
fn cleanup_segment(
|
|
&self,
|
|
service: &TelemetryHistoryService,
|
|
history_segment_ram: HistorySegmentRam,
|
|
) -> JoinHandle<()> {
|
|
let mut path = service.data_root_folder.clone();
|
|
path.push(self.data.definition.uuid.as_hyphenated().to_string());
|
|
spawn_blocking(move || {
|
|
match HistorySegmentFile::save_to_disk(path, history_segment_ram) {
|
|
// Immediately drop the segment - now that we've saved it to disk we don't need to keep it in memory
|
|
Ok(segment) => drop(segment),
|
|
Err(err) => {
|
|
error!(
|
|
"An error occurred saving telemetry history to disk: {}",
|
|
err
|
|
);
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
fn get_disk_segment(
|
|
&self,
|
|
service: &TelemetryHistoryService,
|
|
start: DateTime<Utc>,
|
|
) -> JoinHandle<anyhow::Result<HistorySegmentFile>> {
|
|
let mut path = service.data_root_folder.clone();
|
|
path.push(self.data.definition.uuid.as_hyphenated().to_string());
|
|
spawn_blocking(move || HistorySegmentFile::open(path, start))
|
|
}
|
|
|
|
async fn create_ram_segment(
|
|
&self,
|
|
start: DateTime<Utc>,
|
|
service: &TelemetryHistoryService,
|
|
telemetry_data_type: DataType,
|
|
) -> HistorySegmentRam {
|
|
let ram = self
|
|
.get_disk_segment(service, start)
|
|
.await
|
|
.unwrap_or_else(|e| Err(anyhow!("Join Error {e}")))
|
|
.map(|disk| spawn_blocking(move || disk.load_to_ram(telemetry_data_type)));
|
|
|
|
let ram = match ram {
|
|
Ok(ram) => ram.await.unwrap_or_else(|e| Err(anyhow!("Join Error {e}"))),
|
|
Err(e) => Err(e),
|
|
};
|
|
|
|
match ram {
|
|
Ok(ram) => ram,
|
|
_ => HistorySegmentRam::new(start, start + service.segment_width),
|
|
}
|
|
}
|
|
|
|
pub async fn insert(
|
|
&self,
|
|
service: &TelemetryHistoryService,
|
|
value: DataValue,
|
|
timestamp: DateTime<Utc>,
|
|
) {
|
|
let segments = self.segments.read().await;
|
|
|
|
let segments = if segments.is_empty() || segments[segments.len() - 1].end < timestamp {
|
|
// We want to insert something that doesn't fit into our history
|
|
drop(segments);
|
|
let mut segments = self.segments.write().await;
|
|
|
|
if segments.is_empty() {
|
|
let start_time = timestamp.duration_trunc(service.segment_width).unwrap();
|
|
segments.push_back(
|
|
self.create_ram_segment(start_time, service, self.data.definition.data_type)
|
|
.await,
|
|
);
|
|
} else {
|
|
while segments[segments.len() - 1].end < timestamp {
|
|
if segments.len() == service.max_segments {
|
|
if let Some(segment) = segments.pop_front() {
|
|
// We don't care about this future
|
|
drop(self.cleanup_segment(service, segment));
|
|
}
|
|
}
|
|
let start_time = segments[segments.len() - 1].end;
|
|
segments.push_back(
|
|
self.create_ram_segment(
|
|
start_time,
|
|
service,
|
|
self.data.definition.data_type,
|
|
)
|
|
.await,
|
|
);
|
|
}
|
|
}
|
|
|
|
drop(segments);
|
|
self.segments.read().await
|
|
} else {
|
|
segments
|
|
};
|
|
|
|
// Get the index of the first segment which has an end time AFTER the above timestamp
|
|
let segment_index = segments.partition_point(|segment| segment.end < timestamp);
|
|
|
|
segments[segment_index].insert(value, timestamp);
|
|
}
|
|
|
|
pub fn insert_sync(
|
|
history: Arc<Self>,
|
|
service: Arc<TelemetryHistoryService>,
|
|
value: DataValue,
|
|
timestamp: DateTime<Utc>,
|
|
) {
|
|
tokio::spawn(async move {
|
|
history.insert(&service, value, timestamp).await;
|
|
});
|
|
}
|
|
|
|
pub async fn get(
|
|
&self,
|
|
from: DateTime<Utc>,
|
|
to: DateTime<Utc>,
|
|
maximum_resolution: TimeDelta,
|
|
telemetry_history_service: &TelemetryHistoryService,
|
|
) -> Vec<TelemetryDataItem> {
|
|
let mut disk_result = vec![];
|
|
let mut ram_result = vec![];
|
|
|
|
let mut from = from;
|
|
let mut to = to;
|
|
let initial_to = to;
|
|
let mut ram_from_result = from;
|
|
|
|
{
|
|
let segments = self.segments.read().await;
|
|
let first_ram_segment = segments.front().map(|x| x.start);
|
|
if let Some(first_ram_segment) = first_ram_segment {
|
|
let mut ram_from = first_ram_segment;
|
|
for i in 0..segments.len() {
|
|
let (new_from, new_data) = segments[i].get(ram_from, to, maximum_resolution);
|
|
ram_from = new_from;
|
|
ram_result.extend(new_data);
|
|
}
|
|
from = min(from, first_ram_segment);
|
|
to = min(to, first_ram_segment);
|
|
ram_from_result = ram_from;
|
|
}
|
|
}
|
|
|
|
{
|
|
let start = from
|
|
.duration_trunc(telemetry_history_service.segment_width)
|
|
.unwrap();
|
|
let end = (to + telemetry_history_service.segment_width)
|
|
.duration_trunc(telemetry_history_service.segment_width)
|
|
.unwrap();
|
|
|
|
let mut path = telemetry_history_service.data_root_folder.clone();
|
|
path.push(self.data.definition.uuid.as_hyphenated().to_string());
|
|
|
|
let mut start = start;
|
|
while start < end {
|
|
// We're going to ignore errors with getting the disk segment
|
|
if let Ok(Ok(mut disk)) = self
|
|
.get_disk_segment(telemetry_history_service, start)
|
|
.await
|
|
{
|
|
match disk.get(from, to, maximum_resolution, self.data.definition.data_type) {
|
|
Ok((new_from, new_data)) => {
|
|
from = new_from;
|
|
disk_result.extend(new_data);
|
|
}
|
|
Err(err) => {
|
|
error!("Failed to get from disk segment: {err}");
|
|
}
|
|
}
|
|
}
|
|
start += telemetry_history_service.segment_width;
|
|
}
|
|
}
|
|
|
|
{
|
|
// Go through the ram segments a second time to capture any data added since we dealt
|
|
// with the disk data
|
|
from = ram_from_result;
|
|
to = initial_to;
|
|
let segments = self.segments.read().await;
|
|
for i in 0..segments.len() {
|
|
let (new_from, new_data) = segments[i].get(from, to, maximum_resolution);
|
|
from = new_from;
|
|
ram_result.extend(new_data);
|
|
}
|
|
}
|
|
|
|
disk_result.extend(ram_result);
|
|
disk_result
|
|
}
|
|
|
|
pub async fn cleanup(&self, service: &TelemetryHistoryService) -> anyhow::Result<()> {
|
|
let mut segments = self.segments.write().await;
|
|
|
|
let segments = segments
|
|
.drain(..)
|
|
.map(|segment| self.cleanup_segment(service, segment))
|
|
.collect::<Vec<_>>();
|
|
|
|
for segment in segments {
|
|
segment.await?;
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
pub struct TelemetryHistoryService {
|
|
segment_width: TimeDelta,
|
|
max_segments: usize,
|
|
data_root_folder: PathBuf,
|
|
}
|
|
|
|
impl TelemetryHistoryService {
|
|
pub fn new(data_folder: PathBuf) -> anyhow::Result<Self> {
|
|
let result = Self {
|
|
segment_width: TimeDelta::minutes(1),
|
|
max_segments: 5,
|
|
data_root_folder: data_folder,
|
|
};
|
|
|
|
fs::create_dir_all(&result.data_root_folder)?;
|
|
|
|
info!(
|
|
"Recording Telemetry Data to {}",
|
|
result.data_root_folder.to_string_lossy()
|
|
);
|
|
|
|
Ok(result)
|
|
}
|
|
|
|
pub fn get_metadata_file(&self) -> PathBuf {
|
|
self.data_root_folder.join("metadata.json")
|
|
}
|
|
}
|