Remove unnecessary thread structure
This commit is contained in:
parent
36b59240d9
commit
962d90fe1d
2 changed files with 92 additions and 208 deletions
218
src/bot.rs
218
src/bot.rs
|
@ -1,15 +1,12 @@
|
|||
use crate::config::{Config, PostBody, SeriesConfig};
|
||||
use crate::{config::{Config, PostBody, SeriesConfig}, fetchers::{jnovel}, lemmy};
|
||||
use crate::fetchers::jnovel::JPostInfo;
|
||||
use crate::lemmy::{Lemmy, PostInfo};
|
||||
use crate::post_history::SeriesHistory;
|
||||
use crate::{fetchers::{jnovel}, lemmy, write_error, write_info, write_warn, SharedData};
|
||||
use chrono::{DateTime, Duration, Utc};
|
||||
use chrono::{DateTime, Duration, Timelike, Utc};
|
||||
use lemmy_api_common::post::CreatePost;
|
||||
use lemmy_db_schema::newtypes::{CommunityId, LanguageId};
|
||||
use lemmy_db_schema::PostFeatureType;
|
||||
use std::collections::HashMap;
|
||||
use std::sync::Arc;
|
||||
use tokio::sync::RwLock;
|
||||
use tokio::time::sleep;
|
||||
use crate::fetchers::Fetcher;
|
||||
use systemd_journal_logger::connected_to_journal;
|
||||
|
@ -40,115 +37,91 @@ macro_rules! error {
|
|||
}
|
||||
};
|
||||
}
|
||||
|
||||
pub(crate) async fn run() {
|
||||
let mut last_reload: DateTime<Utc>;
|
||||
let mut lemmy: Lemmy;
|
||||
let mut login_error: bool;
|
||||
let mut communities;
|
||||
{
|
||||
let mut write = data.write().await;
|
||||
let mut communities: HashMap<String, CommunityId>;
|
||||
let mut post_history: SeriesHistory;
|
||||
let mut start: DateTime<Utc>;
|
||||
let mut config: Config = Config::load();
|
||||
last_reload = Utc::now();
|
||||
|
||||
// Errors during bot init are likely unrecoverable and therefore should panic the bot
|
||||
// Does not really matter since the bot will get restarted anyway but this way the uptime url logs a downtime
|
||||
write.config = Config::load();
|
||||
last_reload = Utc::now();
|
||||
}
|
||||
lemmy = match lemmy::login(&config).await {
|
||||
Ok(data) => data,
|
||||
Err(_) => panic!(),
|
||||
};
|
||||
login_error = false;
|
||||
|
||||
{
|
||||
let read = data.read().await;
|
||||
lemmy = match lemmy::login(&read.config).await {
|
||||
Ok(data) => data,
|
||||
Err(_) => panic!(),
|
||||
};
|
||||
login_error = false;
|
||||
communities = match lemmy.get_communities().await {
|
||||
Ok(data) => data,
|
||||
Err(_) => panic!(),
|
||||
};
|
||||
|
||||
communities = match lemmy.get_communities().await {
|
||||
Ok(data) => data,
|
||||
Err(_) => panic!(),
|
||||
};
|
||||
}
|
||||
|
||||
{
|
||||
let mut write = data.write().await;
|
||||
write.start = Utc::now();
|
||||
}
|
||||
start = Utc::now();
|
||||
|
||||
let info_msg = "Bot init successful, starting normal operations".to_owned();
|
||||
info!(info_msg);
|
||||
|
||||
loop {
|
||||
idle(&data).await;
|
||||
idle(&start, &config).await;
|
||||
start = Utc::now();
|
||||
|
||||
{
|
||||
let mut write = data.write().await;
|
||||
|
||||
write.start = Utc::now();
|
||||
// replace with watcher
|
||||
if start - last_reload >= Duration::seconds(config.config_reload_seconds as i64) {
|
||||
config = Config::load();
|
||||
let message = "Config reloaded".to_owned();
|
||||
info!(message);
|
||||
write.config = Config::load();
|
||||
let message = "Config reloaded".to_owned();
|
||||
write_info(message);
|
||||
}
|
||||
}
|
||||
|
||||
{
|
||||
if login_error {
|
||||
let info_msg = "Login invalid, refreshing session";
|
||||
info!(info_msg);
|
||||
lemmy = match lemmy::login(&config).await {
|
||||
Ok(data) => data,
|
||||
Err(_) => continue,
|
||||
};
|
||||
login_error = false;
|
||||
}
|
||||
Ok(data) => data,
|
||||
Err(_) => continue,
|
||||
};
|
||||
login_error = false;
|
||||
}
|
||||
|
||||
{
|
||||
let read = data.read().await;
|
||||
if read.start - last_reload >= Duration::seconds(read.config.config_reload_seconds as i64) {
|
||||
communities = match lemmy.get_communities().await {
|
||||
Ok(data) => data,
|
||||
Err(_) => {
|
||||
login_error = true;
|
||||
continue;
|
||||
if start - last_reload >= Duration::seconds(config.config_reload_seconds as i64) {
|
||||
communities = match lemmy.get_communities().await {
|
||||
Ok(data) => data,
|
||||
Err(_) => {
|
||||
login_error = true;
|
||||
continue;
|
||||
}
|
||||
};
|
||||
let message = "Communities reloaded".to_owned();
|
||||
info!(message);
|
||||
last_reload = Utc::now();
|
||||
write_info(message);
|
||||
last_reload = Utc::now();
|
||||
}
|
||||
}
|
||||
|
||||
{
|
||||
let mut write = data.write().await;
|
||||
write.post_history = SeriesHistory::load_history();
|
||||
post_history = SeriesHistory::load_history();
|
||||
|
||||
let series = config.series.clone();
|
||||
for series in series {
|
||||
if handle_series(&series, &communities, &lemmy, &config, &mut post_history)
|
||||
.await
|
||||
.is_err()
|
||||
{
|
||||
login_error = true;
|
||||
continue;
|
||||
};
|
||||
}
|
||||
|
||||
{
|
||||
let read = data.read().await;
|
||||
let series = read.config.series.clone();
|
||||
drop(read);
|
||||
for series in series {
|
||||
if handle_series(&series, &communities, &lemmy, &data)
|
||||
.await
|
||||
.is_err()
|
||||
{
|
||||
login_error = true;
|
||||
continue;
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
idle(&data).await;
|
||||
idle(&start, &config).await;
|
||||
}
|
||||
}
|
||||
|
||||
async fn idle(data: &Arc<RwLock<SharedData>>) {
|
||||
let read = data.read().await;
|
||||
async fn idle(start: &DateTime<Utc>, config: &Config) {
|
||||
let mut sleep_duration = Duration::seconds(30);
|
||||
if Utc::now() - read.start > sleep_duration {
|
||||
if Utc::now() - start > sleep_duration {
|
||||
sleep_duration = Duration::seconds(60);
|
||||
}
|
||||
|
||||
if let Some(status_url) = read.config.status_post_url.clone() {
|
||||
if let Some(status_url) = config.status_post_url.clone() {
|
||||
match reqwest::get(status_url).await {
|
||||
Ok(_) => {}
|
||||
Err(e) => {
|
||||
|
@ -158,12 +131,12 @@ async fn idle(data: &Arc<RwLock<SharedData>>) {
|
|||
}
|
||||
};
|
||||
|
||||
while Utc::now() - read.start < sleep_duration {
|
||||
while Utc::now() - start < sleep_duration {
|
||||
sleep(Duration::milliseconds(100).to_std().unwrap()).await;
|
||||
}
|
||||
}
|
||||
|
||||
async fn handle_series(series: &SeriesConfig, communities: &HashMap<String, CommunityId>, lemmy: &Lemmy, data: &Arc<RwLock<SharedData>>) -> Result<(), ()> {
|
||||
async fn handle_series(series: &SeriesConfig, communities: &HashMap<String, CommunityId>, lemmy: &Lemmy, config: &Config, post_history: &mut SeriesHistory ) -> Result<(), ()> {
|
||||
let jnc = jnovel::JFetcherOptions::new(series.slug.clone(), series.parted);
|
||||
let post_list = match jnc.check_feed().await {
|
||||
Ok(data) => data,
|
||||
|
@ -176,19 +149,15 @@ async fn handle_series(series: &SeriesConfig, communities: &HashMap<String, Comm
|
|||
}
|
||||
|
||||
for post_info in post_list.clone().iter() {
|
||||
// todo .clone() likely not needed
|
||||
let post_part_info = post_info.get_part_info();
|
||||
let post_lemmy_info = post_info.get_info();
|
||||
|
||||
{
|
||||
let read = data.read().await;
|
||||
if read.post_history.check_for_post(
|
||||
series.slug.as_str(),
|
||||
post_part_info.as_string().as_str(),
|
||||
post_lemmy_info.title.as_str(),
|
||||
) {
|
||||
continue;
|
||||
}
|
||||
if post_history.check_for_post(
|
||||
series.slug.as_str(),
|
||||
post_part_info.as_string().as_str(),
|
||||
post_lemmy_info.title.as_str(),
|
||||
) {
|
||||
continue;
|
||||
}
|
||||
|
||||
let post_series_config = match post_info {
|
||||
|
@ -224,51 +193,43 @@ async fn handle_series(series: &SeriesConfig, communities: &HashMap<String, Comm
|
|||
info!(info);
|
||||
let post_id = lemmy.post(post_data).await?;
|
||||
|
||||
if post_series_config.pin_settings.pin_new_post_community
|
||||
&& config
|
||||
.protected_communities
|
||||
.contains(&post_series_config.name)
|
||||
{
|
||||
let read = data.read().await;
|
||||
if post_series_config.pin_settings.pin_new_post_community
|
||||
&& !read
|
||||
.config
|
||||
.protected_communities
|
||||
let info = format!(
|
||||
"Pinning '{}' to {}",
|
||||
post_lemmy_info.title,
|
||||
post_series_config.name.as_str()
|
||||
);
|
||||
info!(info);
|
||||
{
|
||||
let info = format!(
|
||||
"Pinning '{}' to {}",
|
||||
post_lemmy_info.title,
|
||||
post_series_config.name.as_str()
|
||||
);
|
||||
write_info(info);
|
||||
let pinned_posts = lemmy.get_community_pinned(community_id).await?;
|
||||
if !pinned_posts.is_empty() {
|
||||
let community_pinned_post = &pinned_posts[0];
|
||||
lemmy
|
||||
.unpin(community_pinned_post.post.id, PostFeatureType::Community)
|
||||
.await?;
|
||||
}
|
||||
lemmy.pin(post_id, PostFeatureType::Community).await?;
|
||||
} else if read
|
||||
.config
|
||||
.protected_communities
|
||||
.contains(&post_series_config.name)
|
||||
{
|
||||
let message = format!(
|
||||
"Community '{}' for Series '{}' is protected. Is this intended?",
|
||||
&post_series_config.name, series.slug
|
||||
);
|
||||
write_warn(message);
|
||||
let pinned_posts = lemmy.get_community_pinned(community_id).await?;
|
||||
if !pinned_posts.is_empty() {
|
||||
let community_pinned_post = &pinned_posts[0];
|
||||
lemmy
|
||||
.unpin(community_pinned_post.post.id, PostFeatureType::Community)
|
||||
.await?;
|
||||
}
|
||||
lemmy.pin(post_id, PostFeatureType::Community).await?;
|
||||
} else if config
|
||||
.protected_communities
|
||||
.contains(&post_series_config.name)
|
||||
{
|
||||
let message = format!(
|
||||
"Community '{}' for Series '{}' is protected. Is this intended?",
|
||||
&post_series_config.name, series.slug
|
||||
);
|
||||
warn!(message);
|
||||
}
|
||||
|
||||
let read = data.read().await;
|
||||
if post_series_config.pin_settings.pin_new_post_local {
|
||||
let info = format!("Pinning '{}' to Instance", post_lemmy_info.title);
|
||||
info!(info);
|
||||
let pinned_posts = lemmy.get_local_pinned().await?;
|
||||
if !pinned_posts.is_empty() {
|
||||
for pinned_post in pinned_posts {
|
||||
if read
|
||||
.config
|
||||
if config
|
||||
.protected_communities
|
||||
.contains(&pinned_post.community.name)
|
||||
{
|
||||
|
@ -285,9 +246,8 @@ async fn handle_series(series: &SeriesConfig, communities: &HashMap<String, Comm
|
|||
lemmy.pin(post_id, PostFeatureType::Local).await?;
|
||||
}
|
||||
|
||||
let mut series_history = read.post_history.get_series(series.slug.as_str());
|
||||
let mut series_history = post_history.get_series(series.slug.as_str());
|
||||
let mut part_history = series_history.get_part(post_part_info.as_string().as_str());
|
||||
drop(read);
|
||||
|
||||
match post_info {
|
||||
JPostInfo::Chapter { .. } => part_history.chapter = post_info.get_info().title,
|
||||
|
@ -295,11 +255,9 @@ async fn handle_series(series: &SeriesConfig, communities: &HashMap<String, Comm
|
|||
}
|
||||
|
||||
series_history.set_part(post_part_info.as_string().as_str(), part_history);
|
||||
let mut write = data.write().await;
|
||||
write
|
||||
.post_history
|
||||
post_history
|
||||
.set_series(series.slug.as_str(), series_history);
|
||||
write.post_history.save_history();
|
||||
post_history.save_history();
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
|
82
src/main.rs
82
src/main.rs
|
@ -1,15 +1,9 @@
|
|||
use crate::config::Config;
|
||||
use crate::post_history::SeriesHistory;
|
||||
use chrono::{DateTime, Duration, Utc};
|
||||
use log::{error, info, warn, LevelFilter};
|
||||
use chrono::{Duration};
|
||||
use log::{LevelFilter};
|
||||
use once_cell::sync::Lazy;
|
||||
use reqwest::Client;
|
||||
use std::collections::HashMap;
|
||||
use std::fmt::Debug;
|
||||
use std::sync::Arc;
|
||||
use systemd_journal_logger::{connected_to_journal, JournalLog};
|
||||
use tokio::sync::RwLock;
|
||||
use tokio::time::sleep;
|
||||
use systemd_journal_logger::{JournalLog};
|
||||
|
||||
mod bot;
|
||||
mod config;
|
||||
|
@ -17,27 +11,6 @@ mod lemmy;
|
|||
mod post_history;
|
||||
mod fetchers;
|
||||
|
||||
pub(crate) fn write_error(err_msg: String) {
|
||||
match connected_to_journal() {
|
||||
true => error!("[ERROR] {err_msg}"),
|
||||
false => println!("[ERROR] {err_msg}"),
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn write_warn(warn_msg: String) {
|
||||
match connected_to_journal() {
|
||||
true => warn!("[WARN] {warn_msg}"),
|
||||
false => println!("[WARN] {warn_msg}"),
|
||||
}
|
||||
}
|
||||
|
||||
pub(crate) fn write_info(info_msg: String) {
|
||||
match connected_to_journal() {
|
||||
true => info!("[INFO] {info_msg}"),
|
||||
false => println!("[INFO] {info_msg}"),
|
||||
}
|
||||
}
|
||||
|
||||
pub static HTTP_CLIENT: Lazy<Client> = Lazy::new(|| {
|
||||
Client::builder()
|
||||
.timeout(Duration::seconds(30).to_std().unwrap())
|
||||
|
@ -46,25 +19,6 @@ pub static HTTP_CLIENT: Lazy<Client> = Lazy::new(|| {
|
|||
.expect("build client")
|
||||
});
|
||||
|
||||
#[derive(Clone, Debug)]
|
||||
pub(crate) struct SharedData {
|
||||
config: Config,
|
||||
post_history: SeriesHistory,
|
||||
start: DateTime<Utc>,
|
||||
}
|
||||
|
||||
impl SharedData {
|
||||
pub(crate) fn new() -> Self {
|
||||
SharedData {
|
||||
config: Config::default(),
|
||||
post_history: SeriesHistory {
|
||||
series: HashMap::new(),
|
||||
},
|
||||
start: Utc::now(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[tokio::main]
|
||||
async fn main() {
|
||||
JournalLog::new()
|
||||
|
@ -72,33 +26,5 @@ async fn main() {
|
|||
.install()
|
||||
.expect("Systemd-Logger crate error");
|
||||
log::set_max_level(LevelFilter::Info);
|
||||
|
||||
let mut data = SharedData::new();
|
||||
|
||||
loop {
|
||||
let write_data = Arc::new(RwLock::new(data.clone()));
|
||||
//let read_data = write_data.clone();
|
||||
let persistent_data = write_data.clone();
|
||||
|
||||
let bot_thread = tokio::spawn(async move { bot::run(write_data).await });
|
||||
|
||||
let _ = bot_thread.await;
|
||||
|
||||
data = persistent_data.read().await.clone();
|
||||
|
||||
{
|
||||
let err_msg = "Bot crashed due to unknown Error, restarting thread after wait...";
|
||||
match connected_to_journal() {
|
||||
true => error!("[ERROR] {err_msg}"),
|
||||
false => println!("[ERROR] {err_msg}"),
|
||||
}
|
||||
}
|
||||
|
||||
sleep(
|
||||
Duration::seconds(5)
|
||||
.to_std()
|
||||
.expect("Conversion should always work since static"),
|
||||
)
|
||||
.await;
|
||||
}
|
||||
bot::run().await;
|
||||
}
|
||||
|
|
Loading…
Reference in a new issue