Add support for multiple websites, contents, etc.

This commit is contained in:
Elnu 2022-07-14 17:05:45 -07:00
parent 0de8921306
commit 5d70793e2b
10 changed files with 1034 additions and 78 deletions

View file

@ -3,8 +3,8 @@ use chrono::{DateTime, Utc};
use serde::{Deserialize, Serialize, Serializer};
use validator::Validate;
// Master comment type that is stored in database
#[derive(Serialize, Deserialize, Validate)]
#[serde(rename_all = "camelCase")]
pub struct Comment {
#[serde(skip_serializing_if = "Option::is_none")]
pub author: Option<String>, // None is Anonymous
@ -19,6 +19,7 @@ pub struct Comment {
#[serde(with = "ts_seconds_option")]
#[serde(skip_serializing_if = "Option::is_none")]
pub timestamp: Option<DateTime<Utc>>,
pub content_id: String,
}
fn serialize_gravatar<S>(email: &Option<String>, s: S) -> Result<S::Ok, S::Error>

View file

@ -13,26 +13,28 @@ impl Database {
let conn = Connection::open_in_memory()?;
conn.execute(
"CREATE TABLE comment (
id INTEGER PRIMARY KEY,
email TEXT,
author TEXT,
text TEXT NOT NULL,
timestamp DATETIME DEFAULT CURRENT_TIMESTAMP
id INTEGER PRIMARY KEY,
email TEXT,
author TEXT,
text TEXT NOT NULL,
timestamp DATETIME DEFAULT CURRENT_TIMESTAMP,
content_id TEXT NOT NULL
)",
params![],
)?;
Ok(Self { conn })
}
pub fn get_comments(&self) -> Result<Vec<Comment>> {
pub fn get_comments(&self, content_id: &str) -> Result<Vec<Comment>> {
self.conn
.prepare("SELECT author, email, text, timestamp FROM comment ORDER BY timestamp DESC")?
.prepare(&format!("SELECT author, email, text, timestamp FROM comment WHERE content_id='{content_id}' ORDER BY timestamp DESC"))?
.query_map([], |row| {
Ok(Comment {
author: row.get(0)?,
email: row.get(1)?,
text: row.get(2)?,
timestamp: row.get(3)?,
content_id: content_id.to_owned(),
})
})?
.collect()
@ -40,8 +42,8 @@ impl Database {
pub fn create_comment(&self, comment: &Comment) -> Result<()> {
self.conn.execute(
"INSERT INTO comment (author, email, text) VALUES (?1, ?2, ?3)",
params![&comment.author, &comment.email, &comment.text],
"INSERT INTO comment (author, email, text, content_id) VALUES (?1, ?2, ?3, ?4)",
params![&comment.author, &comment.email, &comment.text, &comment.content_id],
)?;
Ok(())
}

View file

@ -1,63 +1,161 @@
mod comment;
use actix_cors::Cors;
mod comment; use actix_cors::Cors;
pub use comment::*;
mod database;
pub use database::Database;
use actix_web::{get, post, web, App, HttpResponse, HttpServer, Responder};
use std::{env, sync::Mutex};
use actix_web::{get, post, web, App, HttpResponse, HttpRequest, HttpServer, Responder};
use std::{env, sync::{Mutex, MutexGuard}};
use validator::Validate;
use scraper::{Html, Selector};
use std::collections::HashMap;
use serde::Deserialize;
struct AppState {
db: Mutex<Database>,
databases: HashMap<String, Mutex<Database>>,
}
#[get("/")]
async fn get_comments(data: web::Data<AppState>) -> impl Responder {
let db = match data.db.lock() {
Ok(db) => db,
Err(_) => return HttpResponse::InternalServerError().into(),
fn get_db<'a>(data: &'a web::Data<AppState>, request: &HttpRequest) -> Result<MutexGuard<'a, Database>, HttpResponse> {
// all the .into() are converting from HttpResponseBuilder to HttpResponse
let origin = match request.head().headers().get("Origin") {
Some(origin) => match origin.to_str() {
Ok(origin) => origin,
Err(_) => return Err(HttpResponse::BadRequest().into()),
}
None => return Err(HttpResponse::BadRequest().into()),
};
HttpResponse::Ok().json(&db.get_comments().unwrap())
match data.databases.get(origin) {
Some(database) => Ok(match database.lock() {
Ok(database) => database,
Err(_) => return Err(HttpResponse::InternalServerError().into()),
}),
None => return Err(HttpResponse::BadRequest().into()),
}
}
#[get("/{content_id}")]
async fn get_comments(data: web::Data<AppState>, request: HttpRequest, content_id: web::Path<String>) -> impl Responder {
let database = match get_db(&data, &request) {
Ok(database) => database,
Err(response) => return response,
};
HttpResponse::Ok().json(database.get_comments(&content_id).unwrap())
}
#[derive(Deserialize)]
struct PostCommentsRequest {
url: String,
comment: Comment,
}
#[post("/")]
async fn post_comment(data: web::Data<AppState>, bytes: web::Bytes) -> impl Responder {
async fn post_comment(data: web::Data<AppState>, request: HttpRequest, bytes: web::Bytes) -> impl Responder {
match String::from_utf8(bytes.to_vec()) {
Ok(text) => {
let db = match data.db.lock() {
Ok(db) => db,
Err(_) => return HttpResponse::InternalServerError(),
};
let comment: Comment = match serde_json::from_str(&text) {
Ok(comment) => comment,
Err(_) => return HttpResponse::BadRequest(),
let PostCommentsRequest { url, comment } = match serde_json::from_str(&text) {
Ok(req) => req,
Err(_) => return HttpResponse::BadRequest().into(),
};
if comment.validate().is_err() {
return HttpResponse::BadRequest();
return HttpResponse::BadRequest().into();
}
db.create_comment(&comment).unwrap();
HttpResponse::Ok()
let origin = match request.head().headers().get("Origin") {
Some(origin) => match origin.to_str() {
Ok(origin) => origin,
// If the Origin is not valid ASCII, it is a bad request not sent from a browser
Err(_) => return HttpResponse::BadRequest().into(),
},
// If there is no Origin header, it is a bad request not sent from a browser
None => return HttpResponse::BadRequest().into(),
};
// Check to see if provided URL is in scope.
// This is to prevent malicious requests that try to get server to fetch external websites.
// (requires loop because "labels on blocks are unstable")
// https://github.com/rust-lang/rust/issues/48594
'outer: loop {
for site_root in data.databases.keys() {
if site_root.starts_with(origin) && url.starts_with(site_root) {
break 'outer;
}
}
return HttpResponse::BadRequest().into();
}
match get_page_data(&url).await {
Ok(page_data_option) => match page_data_option {
Some(page_data) => if page_data.content_id != comment.content_id {
return HttpResponse::BadRequest().into();
},
None => return HttpResponse::BadRequest().into(),
},
Err(_) => return HttpResponse::InternalServerError().into(),
};
let database = match get_db(&data, &request) {
Ok(database) => database,
Err(response) => return response,
};
database.create_comment(&comment).unwrap();
HttpResponse::Ok().into()
}
Err(_) => HttpResponse::BadRequest().into(),
}
}
// Contains all page details stored in meta tags.
// Currently, only content_id, but this is wrapped in this struct
// to make adding other meta tags, such as locked comments, in the future
struct PageData {
content_id: String,
}
async fn get_page_data(url: &str) -> Result<Option<PageData>, reqwest::Error> {
let response = reqwest::get(url).await?;
if !response.status().is_success() {
return Ok(None);
}
let content = response.text_with_charset("utf-8").await?;
let document = Html::parse_document(&content);
let get_meta = |name: &str| -> Option<String> {
let selector = Selector::parse(&format!("meta[name=\"{}\"]", name)).unwrap();
match document.select(&selector).next() {
Some(element) => match element.value().attr("content") {
Some(value) => Some(value.to_owned()),
None => return None,
},
None => return None,
}
};
return Ok(Some(PageData {
content_id: match get_meta("soudan-content-id") {
Some(id) => id,
None => return Ok(None),
},
}))
}
#[actix_web::main]
async fn main() -> Result<(), std::io::Error> {
let mut domains = Vec::new();
let testing = {
let mut testing = false;
for argument in env::args() {
let mut args = env::args();
args.next(); // Skip first, will be executable name
for argument in args {
if argument == "--testing" || argument == "-t" {
testing = true;
break;
} else {
domains.push(argument);
}
}
testing
};
let db = Database::new(testing).unwrap();
let state = web::Data::new(AppState { db: Mutex::new(db) });
if domains.len() == 0 {
panic!("At least one domain is required!");
}
let mut databases = HashMap::new();
for domain in domains.iter() {
databases.insert(domain.to_owned(), Mutex::new(Database::new(testing).unwrap()));
}
let state = web::Data::new(AppState { databases });
HttpServer::new(move || {
App::new()
.service(get_comments)