Config Module rewrite
This commit is contained in:
parent
dd2cde35e4
commit
e7bbe9cf03
1 changed files with 47 additions and 400 deletions
|
@ -1,421 +1,68 @@
|
||||||
use std::{
|
use std::{error::Error};
|
||||||
error::Error,
|
|
||||||
fs::{self, OpenOptions},
|
|
||||||
io::Write,
|
|
||||||
path::Path,
|
|
||||||
};
|
|
||||||
|
|
||||||
use chrono::{DateTime, Duration, Utc};
|
|
||||||
use lemmy_api_common::{
|
|
||||||
community::{ListCommunities, ListCommunitiesResponse},
|
|
||||||
post::CreatePost,
|
|
||||||
sensitive::Sensitive,
|
|
||||||
};
|
|
||||||
use lemmy_db_schema::{
|
|
||||||
newtypes::{CommunityId, LanguageId},
|
|
||||||
ListingType,
|
|
||||||
};
|
|
||||||
use serde_derive::{Deserialize, Serialize};
|
use serde_derive::{Deserialize, Serialize};
|
||||||
use tokio::time::sleep;
|
|
||||||
use url::Url;
|
|
||||||
|
|
||||||
use crate::{
|
#[derive(Serialize, Deserialize, Clone)]
|
||||||
feeds::{FeedChapterData, FeedSeriesData, FeedSetting, FeedVolumeData},
|
|
||||||
PostQueueMetadata, CLIENT,
|
|
||||||
};
|
|
||||||
|
|
||||||
macro_rules! api_url {
|
|
||||||
() => {
|
|
||||||
"https://labs.j-novel.club/app/v1/".to_string()
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
macro_rules! chapter_url {
|
|
||||||
() => {
|
|
||||||
"https://j-novel.club/read/".to_string()
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
macro_rules! volume_url_base {
|
|
||||||
() => {
|
|
||||||
"https://j-novel.club/series/".to_string()
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
macro_rules! pub_struct {
|
|
||||||
($name:ident {$($field:ident: $t:ty,)*}) => {
|
|
||||||
#[derive(Serialize, Deserialize, Clone, PartialEq)]
|
|
||||||
pub(crate) struct $name {
|
|
||||||
$(pub(crate) $field: $t), *
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Secrets structs
|
|
||||||
pub_struct!(Secrets {
|
|
||||||
lemmy: LemmyLogin,
|
|
||||||
reddit: RedditLogin,
|
|
||||||
});
|
|
||||||
|
|
||||||
impl Secrets {
|
|
||||||
pub(crate) fn init() -> Secrets {
|
|
||||||
let file_contents = match fs::read_to_string("secrets.json") {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: secrets.json could not be read:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
let config_parse: Secrets = match serde_json::from_str(&file_contents) {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: secrets.json could not be parsed:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
|
|
||||||
return config_parse;
|
|
||||||
}
|
|
||||||
|
|
||||||
pub(crate) fn load(&mut self) {
|
|
||||||
let file_contents = match fs::read_to_string("secrets.json") {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: secrets.json could not be read:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
let config_parse: Secrets = match serde_json::from_str(&file_contents) {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: secrets.json could not be parsed:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
|
|
||||||
*self = config_parse;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Serialize, Deserialize, Clone, PartialEq)]
|
|
||||||
pub(crate) struct LemmyLogin {
|
|
||||||
pub(crate) username: String,
|
|
||||||
password: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl LemmyLogin {
|
|
||||||
pub(crate) fn get_username(&self) -> Sensitive<String> {
|
|
||||||
return Sensitive::new(self.username.clone());
|
|
||||||
}
|
|
||||||
|
|
||||||
pub(crate) fn get_password(&self) -> Sensitive<String> {
|
|
||||||
return Sensitive::new(self.password.clone());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Serialize, Deserialize, Clone, PartialEq)]
|
|
||||||
pub(crate) struct RedditLogin {
|
|
||||||
pub(crate) app_id: String,
|
|
||||||
app_secret: String,
|
|
||||||
refresh_token: String,
|
|
||||||
redirect_uri: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
// Config structs
|
|
||||||
#[derive(Serialize, Deserialize, Clone, PartialEq)]
|
|
||||||
pub(crate) struct Config {
|
pub(crate) struct Config {
|
||||||
pub(crate) instance: String,
|
instance: String,
|
||||||
pub(crate) reddit_config: RedditConfig,
|
status_post_url: Option<String>,
|
||||||
pub(crate) feeds: Vec<FeedSetting>,
|
config_reload_seconds: u32,
|
||||||
pub(crate) uptime_kuma: Option<String>,
|
series: Vec<SeriesConfig>,
|
||||||
pub(crate) config_reload: Option<usize>,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Config {
|
impl Config {
|
||||||
pub(crate) fn init() -> Config {
|
pub(crate) fn new() -> Result<Self, Box<dyn Error>> {
|
||||||
let file_contents = match fs::read_to_string("config.json") {
|
todo!()
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: config.json could not be read:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
let config_parse: Config = match serde_json::from_str(&file_contents) {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: config.json could not be parsed:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
|
|
||||||
return config_parse;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(crate) fn load(&mut self) {
|
pub(crate) fn load() -> Result<Self, Box<dyn Error>> {
|
||||||
let file_contents = match fs::read_to_string("config.json") {
|
let cfg: Self = confy::load_path("./config.toml")?;
|
||||||
Ok(data) => data,
|
if cfg.instance.is_empty() {
|
||||||
Err(e) => panic!("ERROR: config.json could not be read:\n\n{:#?}", e),
|
panic!("config.toml not found!")
|
||||||
};
|
|
||||||
let config_parse: Config = match serde_json::from_str(&file_contents) {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: config.json could not be parsed:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
|
|
||||||
*self = config_parse;
|
|
||||||
}
|
|
||||||
|
|
||||||
#[warn(unused_results)]
|
|
||||||
pub(crate) async fn fetch_infos(&self) -> Result<(), Box<dyn Error>> {
|
|
||||||
// Get Series Data
|
|
||||||
for feed in &self.feeds {
|
|
||||||
let res = CLIENT
|
|
||||||
.get(api_url!() + "series/" + &feed.series_slug + "?format=json")
|
|
||||||
.send()
|
|
||||||
.await?
|
|
||||||
.text()
|
|
||||||
.await?;
|
|
||||||
|
|
||||||
let data: FeedSeriesData = serde_json::from_str(&res).unwrap();
|
|
||||||
}
|
}
|
||||||
return Ok(());
|
Ok(cfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[warn(unused_results)]
|
pub(crate) fn save() -> Result<(), Box<dyn Error>> {
|
||||||
pub(crate) async fn check_feeds(
|
todo!()
|
||||||
&self,
|
}
|
||||||
post_history: &Vec<PrevPost>,
|
}
|
||||||
community_ids: &CommunitiesVector,
|
|
||||||
auth: &Sensitive<String>,
|
|
||||||
) -> Result<Vec<(CreatePost, PostQueueMetadata)>, Box<dyn Error>> {
|
|
||||||
let mut post_queue: Vec<(CreatePost, PostQueueMetadata)> = vec![];
|
|
||||||
|
|
||||||
for feed in &self.feeds {
|
impl Default for Config {
|
||||||
let mut history_data: &PrevPost = &PrevPost {
|
fn default() -> Self {
|
||||||
id: feed.id,
|
Config {
|
||||||
last_volume_slug: None,
|
instance: "".to_string(),
|
||||||
last_volume_time: None,
|
status_post_url: None,
|
||||||
last_part_slug: None,
|
config_reload_seconds: 21600,
|
||||||
last_part_time: None,
|
series: vec![]
|
||||||
};
|
|
||||||
for post in post_history {
|
|
||||||
if post.id == feed.id {
|
|
||||||
history_data = post;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Check for Volume Release
|
|
||||||
let res = CLIENT
|
|
||||||
.get(api_url!() + "series/" + &feed.series_slug + "/volumes?format=json")
|
|
||||||
.send()
|
|
||||||
.await?
|
|
||||||
.text()
|
|
||||||
.await?;
|
|
||||||
|
|
||||||
#[derive(Serialize, Deserialize, Debug, Clone, PartialEq)]
|
|
||||||
struct FeedVolumeReturn {
|
|
||||||
volumes: Vec<FeedVolumeData>,
|
|
||||||
}
|
|
||||||
let mut data: FeedVolumeReturn = serde_json::from_str(&res).unwrap();
|
|
||||||
|
|
||||||
let current_volume_slug = data.volumes[&data.volumes.len() - 1].slug.clone();
|
|
||||||
|
|
||||||
let now = Utc::now();
|
|
||||||
data.volumes.reverse();
|
|
||||||
for volume in data.volumes {
|
|
||||||
let published = DateTime::parse_from_rfc3339(&volume.publishing).unwrap();
|
|
||||||
// Get First Volume that has valid Release Data
|
|
||||||
if now >= published {
|
|
||||||
if Some(volume.slug.clone()) != history_data.last_volume_slug {
|
|
||||||
if let Some(volume_community) = &feed.communities.volume {
|
|
||||||
let mut post_url = Url::parse(&(volume_url_base!() + &feed.series_slug))?;
|
|
||||||
post_url.set_fragment(Some(&("volume-".to_string() + &volume.number.to_string())));
|
|
||||||
|
|
||||||
let new_post = CreatePost {
|
|
||||||
name: volume.title.clone(),
|
|
||||||
community_id: community_ids.find(&volume_community),
|
|
||||||
url: Some(post_url),
|
|
||||||
body: Some(
|
|
||||||
"[Reddit](https://reddit.com/r/HonzukinoGekokujou)\n\n\
|
|
||||||
[Discord](https://discord.com/invite/fGefmzu)\n\n\
|
|
||||||
[#ascendanceofabookworm](https://mastodon.social/tags/ascendanceofabookworm)\n\n\
|
|
||||||
[#honzuki](https://mastodon.social/tags/honzuki)\n\n\
|
|
||||||
[#本好きの下剋上](https://mastodon.social/tags/%E6%9C%AC%E5%A5%BD%E3%81%8D%E3%81%AE%E4%B8%8B%E5%89%8B%E4%B8%8A)"
|
|
||||||
.into(),
|
|
||||||
),
|
|
||||||
honeypot: None,
|
|
||||||
nsfw: Some(false),
|
|
||||||
language_id: Some(LanguageId(37)), // TODO get this id once every few hours per API request, the ordering of IDs suggests that the EN Id might change in the future
|
|
||||||
auth: auth.clone(),
|
|
||||||
};
|
|
||||||
post_queue.push((
|
|
||||||
new_post,
|
|
||||||
PostQueueMetadata {
|
|
||||||
id: feed.id,
|
|
||||||
series: feed.series_slug.clone(),
|
|
||||||
part: None,
|
|
||||||
volume: Some(volume.slug),
|
|
||||||
},
|
|
||||||
));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if let Some(chapter_community) = &feed.communities.chapter {
|
|
||||||
// Check for Part Release
|
|
||||||
let res = CLIENT
|
|
||||||
.get(api_url!() + "volumes/" + ¤t_volume_slug + "/parts?format=json")
|
|
||||||
.send()
|
|
||||||
.await?
|
|
||||||
.text()
|
|
||||||
.await?;
|
|
||||||
|
|
||||||
#[derive(Serialize, Deserialize, Debug, Clone, PartialEq)]
|
|
||||||
struct FeedChapterReturn {
|
|
||||||
parts: Vec<FeedChapterData>,
|
|
||||||
}
|
|
||||||
let mut data: FeedChapterReturn = serde_json::from_str(&res).unwrap();
|
|
||||||
|
|
||||||
data.parts.reverse();
|
|
||||||
for part in data.parts {
|
|
||||||
if Some(part.slug.clone()) != history_data.last_part_slug {
|
|
||||||
let new_post = CreatePost {
|
|
||||||
name: part.title.clone(),
|
|
||||||
community_id: community_ids.find(&chapter_community),
|
|
||||||
url: Some(Url::parse(&(chapter_url!() + &part.slug)).unwrap()),
|
|
||||||
body: Some(
|
|
||||||
"[Reddit](https://reddit.com/r/HonzukinoGekokujou)\n\n\
|
|
||||||
[Discord](https://discord.com/invite/fGefmzu)\n\n\
|
|
||||||
[#ascendanceofabookworm](https://mastodon.social/tags/ascendanceofabookworm)\n\n\
|
|
||||||
[#honzuki](https://mastodon.social/tags/honzuki)\n\n\
|
|
||||||
[#本好きの下剋上](https://mastodon.social/tags/%E6%9C%AC%E5%A5%BD%E3%81%8D%E3%81%AE%E4%B8%8B%E5%89%8B%E4%B8%8A)"
|
|
||||||
.into(),
|
|
||||||
),
|
|
||||||
honeypot: None,
|
|
||||||
nsfw: Some(false),
|
|
||||||
language_id: Some(LanguageId(37)), // TODO get this id once every few hours per API request, the ordering of IDs suggests that the EN Id might change in the future
|
|
||||||
auth: auth.clone(),
|
|
||||||
};
|
|
||||||
post_queue.push((
|
|
||||||
new_post,
|
|
||||||
PostQueueMetadata {
|
|
||||||
id: feed.id,
|
|
||||||
series: feed.series_slug.clone(),
|
|
||||||
part: Some(part.slug),
|
|
||||||
volume: None,
|
|
||||||
},
|
|
||||||
));
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sleep(Duration::milliseconds(100).to_std().unwrap()).await; // Should prevent dos-ing J-Novel servers
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return Ok(post_queue);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub_struct!(RedditConfig {
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
user_agent: String,
|
pub(crate) struct SeriesConfig {
|
||||||
subreddit: String,
|
slug: String,
|
||||||
});
|
parted: bool,
|
||||||
|
prepub_community: PostConfig,
|
||||||
//noinspection ALL
|
volume_community: PostConfig,
|
||||||
#[derive(Serialize, Deserialize, Clone, PartialEq, strum_macros::Display)]
|
|
||||||
#[allow(non_camel_case_types)]
|
|
||||||
pub(crate) enum LemmyCommunities {
|
|
||||||
aobwebnovel,
|
|
||||||
aobprepub,
|
|
||||||
aoblightnovel,
|
|
||||||
aobmanga,
|
|
||||||
metadiscussions,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Posts structs
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
pub_struct!(PrevPost {
|
pub(crate) struct PostConfig {
|
||||||
id: usize,
|
community_name: String,
|
||||||
last_volume_slug: Option<String>,
|
pin_settings: PinConfig,
|
||||||
last_volume_time: Option<String>,
|
post_body: PostBody,
|
||||||
last_part_slug: Option<String>,
|
|
||||||
last_part_time: Option<String>,
|
|
||||||
});
|
|
||||||
|
|
||||||
impl PrevPost {
|
|
||||||
pub(crate) fn load() -> Vec<PrevPost> {
|
|
||||||
let mut history;
|
|
||||||
|
|
||||||
if Path::new("posts.json").exists() {
|
|
||||||
let file_contents = match fs::read_to_string("posts.json") {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: posts.json could not be read:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
|
|
||||||
if file_contents.len() > 0 {
|
|
||||||
let history_parse: Vec<PrevPost> = match serde_json::from_str(&file_contents) {
|
|
||||||
Ok(data) => data,
|
|
||||||
Err(e) => panic!("ERROR: posts.json could not be parsed:\n\n{:#?}", e),
|
|
||||||
};
|
|
||||||
history = history_parse;
|
|
||||||
} else {
|
|
||||||
history = [].to_vec()
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
let _ = fs::File::create("posts.json");
|
|
||||||
history = [].to_vec()
|
|
||||||
}
|
|
||||||
|
|
||||||
history.sort_by(|a, b| a.id.cmp(&b.id));
|
|
||||||
|
|
||||||
return history;
|
|
||||||
}
|
|
||||||
|
|
||||||
pub(crate) fn save(data: &Vec<PrevPost>) {
|
|
||||||
let mut file = OpenOptions::new()
|
|
||||||
.read(true)
|
|
||||||
.write(true)
|
|
||||||
.create(true)
|
|
||||||
.open("posts.json")
|
|
||||||
.unwrap();
|
|
||||||
|
|
||||||
let json_data = serde_json::to_string_pretty(&data).unwrap();
|
|
||||||
|
|
||||||
write!(&mut file, "{}", json_data).unwrap();
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Bot Helper Structs
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
pub_struct!(CommunitiesVector {
|
pub(crate) struct PinConfig {
|
||||||
ids: Vec<(CommunityId, String)>,
|
pin_new_post_local: bool,
|
||||||
});
|
pin_new_post_community: bool,
|
||||||
|
}
|
||||||
impl CommunitiesVector {
|
|
||||||
pub(crate) fn new() -> CommunitiesVector {
|
#[derive(Debug, Serialize, Deserialize, Clone)]
|
||||||
CommunitiesVector { ids: vec![] }
|
#[serde(tag = "body_type", content = "body_content")]
|
||||||
}
|
pub(crate) enum PostBody {
|
||||||
|
None,
|
||||||
#[warn(unused_results)]
|
Description,
|
||||||
pub(crate) async fn load(&mut self, auth: &Sensitive<String>, base: &String) -> Result<(), Box<dyn Error>> {
|
Custom(String),
|
||||||
let params = ListCommunities {
|
|
||||||
auth: Some(auth.clone()),
|
|
||||||
type_: Some(ListingType::Local),
|
|
||||||
..Default::default()
|
|
||||||
};
|
|
||||||
|
|
||||||
let res = CLIENT
|
|
||||||
.get(base.clone() + "/api/v3/community/list")
|
|
||||||
.query(¶ms)
|
|
||||||
.send()
|
|
||||||
.await?
|
|
||||||
.text()
|
|
||||||
.await?;
|
|
||||||
|
|
||||||
let site_data: ListCommunitiesResponse = serde_json::from_str(&res).unwrap();
|
|
||||||
|
|
||||||
let mut ids = [].to_vec();
|
|
||||||
|
|
||||||
site_data.communities.iter().for_each(|entry| {
|
|
||||||
let new_id = (entry.community.id, entry.community.name.clone());
|
|
||||||
ids.push(new_id);
|
|
||||||
});
|
|
||||||
|
|
||||||
self.ids = ids;
|
|
||||||
return Ok(());
|
|
||||||
}
|
|
||||||
|
|
||||||
pub(crate) fn find(&self, name: &LemmyCommunities) -> CommunityId {
|
|
||||||
let mut ret_id = CommunityId(0);
|
|
||||||
|
|
||||||
self.ids.iter().for_each(|id| {
|
|
||||||
let id_name = &id.1;
|
|
||||||
if &name.to_string() == id_name {
|
|
||||||
ret_id = id.0;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
return ret_id;
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue