18 Commits

Author SHA1 Message Date
eda704d2f1 test concurent get q from db/cache
All checks were successful
continuous-integration/drone/push Build is passing
2023-08-18 14:56:08 +03:00
da962a9f7a v0.2.0
All checks were successful
continuous-integration/drone/push Build is passing
2023-08-18 11:14:36 +03:00
Dmitry Belyaev
f86a3a9050 Merge pull request 'migrate/rocket/0.5' (#3) from migrate/rocket/0.5 into master
All checks were successful
continuous-integration/drone/push Build is passing
Reviewed-on: #3
2023-08-18 08:07:13 +00:00
971c79a111 use async db
All checks were successful
continuous-integration/drone/push Build is passing
continuous-integration/drone/pr Build is passing
2023-08-18 10:50:56 +03:00
a4dd460c37 use serde from rocket_dyn_templates
All checks were successful
continuous-integration/drone/push Build is passing
continuous-integration/drone/pr Build is passing
2023-08-05 19:42:19 +03:00
67087e6583 ci: add pipeline type
Some checks reported errors
continuous-integration/drone/push Build was killed
continuous-integration/drone/pr Build is passing
2023-08-05 19:31:35 +03:00
ca75f84ee2 ci: fix img
All checks were successful
continuous-integration/drone/push Build is passing
continuous-integration/drone/pr Build is passing
2023-08-05 19:25:50 +03:00
a951eff714 ci: use rust stable
Some checks reported errors
continuous-integration/drone/push Build encountered an error
2023-08-05 19:24:47 +03:00
18eb3ec51b min migrate to rocket 0.5
All checks were successful
continuous-integration/drone/push Build is passing
2023-08-05 19:21:00 +03:00
78d57ca2e6 update lockfile
All checks were successful
continuous-integration/drone/push Build is passing
2023-08-05 17:50:49 +03:00
103391c891 mini_moka
All checks were successful
continuous-integration/drone/push Build is passing
2023-04-04 11:39:52 +03:00
6d1d8eaea1 upd deps
All checks were successful
continuous-integration/drone/push Build is passing
2023-04-03 15:43:30 +03:00
d1cb2187c1 add cache
All checks were successful
continuous-integration/drone/push Build is passing
b4tman/qchgk_web#1
using moka
2023-04-02 17:01:51 +03:00
d9a060b1b7 add .drone.yml
All checks were successful
continuous-integration/drone/push Build is passing
continuous-integration/drone/tag Build is passing
2023-03-29 09:34:01 +03:00
d6571959db Merge branch 'rocket' 2023-03-28 16:37:03 +03:00
b281f85a70 share database reader instance 2023-03-28 16:36:04 +03:00
89735f98d3 upd chgk_ledb 2023-03-28 16:30:40 +03:00
f56ccf471f new db 2023-01-04 01:12:03 +03:00
5 changed files with 1680 additions and 1108 deletions

13
.drone.yml Normal file
View File

@@ -0,0 +1,13 @@
kind: pipeline
type: docker
name: default
steps:
- name: build
image: rust:1-alpine
commands:
- apk add --no-cache musl-dev
- cargo fetch
- cargo build --all
environment:
CARGO_REGISTRIES_CRATES_IO_PROTOCOL: sparse

4
.gitignore vendored
View File

@@ -1,4 +1,4 @@
/target
**/*.rs.bk
/db
db.dat
/db

2571
Cargo.lock generated

File diff suppressed because it is too large Load Diff

View File

@@ -1,6 +1,6 @@
[package]
name = "qchgk_web"
version = "0.1.1"
version = "0.2.0"
authors = ["Dmitry <b4tm4n@mail.ru>"]
edition = "2021"
license = "MIT"
@@ -11,21 +11,14 @@ readme = "README.md"
[dependencies]
rand="0.8"
serde="1.0"
serde_json="1.0"
ledb = {git = "https://github.com/b4tman/ledb.git", rev="a646b90e", package="ledb"}
rocket="0.4"
chgk_ledb_lib = {git = "https://gitea.b4tman.ru/b4tman/chgk_ledb.git", rev="e521e39f5e", package="chgk_ledb_lib"}
# lmdb-zero="0.4"
[dependencies.rocket_contrib]
version = "0.4"
default-features = false
features = ["serve", "tera_templates"]
rocket = { version = "=0.5.0-rc.3", features = ["json"] }
rocket_dyn_templates = { version = "=0.1.0-rc.3", features = ["tera"] }
chgk_ledb_lib = {git = "https://gitea.b4tman.ru/b4tman/chgk_ledb.git", rev="699478f85e", package="chgk_ledb_lib", features=["async"]}
mini-moka = "0.10.0"
[profile.release]
opt-level = 3
debug = false
lto = true
strip = true

View File

@@ -1,25 +1,27 @@
#![feature(proc_macro_hygiene, decl_macro)]
extern crate ledb;
extern crate serde;
extern crate serde_json;
#[macro_use]
extern crate rocket;
extern crate rocket_contrib;
use rocket::fs::FileServer;
use rocket::futures::FutureExt;
use rocket::response::Redirect;
use rocket::{Rocket, State};
use rocket_contrib::serve::StaticFiles;
use rocket_contrib::templates::Template;
use rocket::State;
use rocket_dyn_templates::tera;
use rocket_dyn_templates::Template;
use rand::distributions::Uniform;
use rand::Rng;
use ledb::{Options, Storage};
use std::ops::Deref;
use std::sync::Arc;
use chgk_ledb_lib::async_db;
use chgk_ledb_lib::questions::Question;
use mini_moka::sync::Cache;
use std::time::Duration;
const DB_FILENAME: &str = "db.dat";
trait ErrorEmpty {
type Output;
fn err_empty(self) -> Result<Self::Output, ()>;
@@ -32,60 +34,114 @@ impl<T, E> ErrorEmpty for Result<T, E> {
}
}
#[derive(Clone)]
struct ArcTemplateData {
value: Arc<tera::Value>,
}
impl ArcTemplateData {
fn new(value: tera::Value) -> ArcTemplateData {
ArcTemplateData {
value: Arc::new(value),
}
}
fn render(&self, name: &'static str) -> Template {
Template::render(name, self.value.deref())
}
}
type TemplateCache = mini_moka::sync::Cache<usize, ArcTemplateData>;
type DataBaseInner = async_db::Reader<Question>;
type DataBase = Arc<DataBaseInner>;
struct AppState {
storage: Storage,
database_distribution: Uniform<u32>,
db: DataBase,
database_distribution: Uniform<usize>,
}
fn get_database_distribution(storage: &Storage) -> Uniform<u32> {
let collection = storage
.collection("questions")
.expect("collection \"questions\"");
let last_id = collection.last_id().expect("\"questions\" last_id");
impl From<DataBaseInner> for AppState {
fn from(db: DataBaseInner) -> Self {
let last_id = db.len();
let database_distribution = rand::distributions::Uniform::new_inclusive(1usize, last_id);
let db = Arc::new(db);
rand::distributions::Uniform::new_inclusive(1u32, last_id)
Self {
db,
database_distribution,
}
}
}
fn random_question_id(database_distribution: &Uniform<u32>) -> u32 {
fn random_question_id(database_distribution: &Uniform<usize>) -> usize {
let mut rng = rand::thread_rng();
rng.sample(database_distribution)
}
fn get_question(storage: &Storage, id: u32) -> Result<Question, ()> {
storage
.collection("questions")
.err_empty()?
.get::<Question>(id)
.err_empty()?
.ok_or(())
async fn get_question(db: &DataBase, id: usize) -> Result<Question, ()> {
db.get(id - 1).await.err_empty()
}
fn show_question_details(template_name: &'static str, data: &AppState, id: u32) -> Template {
match get_question(&data.storage, id) {
Ok(question) => {
let mut context = serde_json::to_value(question).expect("question serialize");
if context.is_object() {
let next_id = random_question_id(&data.database_distribution);
context["next"] = serde_json::to_value(next_id).expect("question id serialize");
async fn show_question_details(
template_name: &'static str,
data: &AppState,
cache: &TemplateCache,
id: usize,
) -> Template {
let cache_cloned = cache.clone();
let from_cache = rocket::tokio::spawn(async move { cache_cloned.get(&id) }).fuse();
let from_db = async move {
match get_question(&data.db, id).await {
Ok(question) => {
let mut context = tera::to_value(question).expect("question serialize");
if context.is_object() {
let next_id = random_question_id(&data.database_distribution);
context["next"] = tera::to_value(next_id).expect("question id serialize");
}
let value = ArcTemplateData::new(context);
let result = value.render(template_name);
cache.insert(id, value);
result
}
Err(_) => {
use std::collections::HashMap;
let context: HashMap<String, String> = HashMap::new();
Template::render("404", context)
}
Template::render(template_name, &context)
}
Err(_) => {
use std::collections::HashMap;
let context: HashMap<String, String> = HashMap::new();
Template::render("404", &context)
}
.fuse();
loop {
rocket::tokio::select! {
biased;
Ok(Some(template)) = from_cache => {
println!("from cache");
break template.render(template_name)
},
template = from_db => {
println!("from db");
break template
}
}
}
}
#[get("/q/<id>")]
fn show_question(data: State<AppState>, id: u32) -> Template {
show_question_details("question", data.inner(), id)
async fn show_question(
data: &State<AppState>,
cache: &State<TemplateCache>,
id: usize,
) -> Template {
show_question_details("question", data.inner(), cache.inner(), id).await
}
#[get("/q/<id>/a")]
fn show_answer(data: State<AppState>, id: u32) -> Template {
show_question_details("answer", data.inner(), id)
async fn show_answer(data: &State<AppState>, cache: &State<TemplateCache>, id: usize) -> Template {
show_question_details("answer", data.inner(), cache.inner(), id).await
}
#[get("/q/0")]
@@ -99,7 +155,7 @@ fn answer0() -> Redirect {
}
#[get("/")]
fn index(data: State<AppState>) -> Redirect {
fn index(data: &State<AppState>) -> Redirect {
let id = random_question_id(&data.database_distribution);
Redirect::temporary(format!("/q/{}", id))
}
@@ -108,36 +164,29 @@ fn index(data: State<AppState>) -> Redirect {
fn not_found(_req: &rocket::Request) -> Template {
use std::collections::HashMap;
let context: HashMap<String, String> = HashMap::new();
Template::render("404", &context)
Template::render("404", context)
}
fn rocket() -> Rocket {
let options: Options = serde_json::from_value(serde_json::json!({
"read_only": true,
"no_lock": true,
}))
.unwrap();
#[launch]
async fn rocket() -> _ {
let state: AppState = async_db::Reader::new(DB_FILENAME)
.await
.expect("open db")
.into();
let cache: TemplateCache = Cache::builder()
.time_to_idle(Duration::from_secs(15 * 60))
.max_capacity(300)
.build();
let mut state = AppState {
storage: Storage::new("db", options).expect("open db"),
database_distribution: Uniform::new(1, 3),
};
state.database_distribution = get_database_distribution(&state.storage);
let state = state;
rocket::ignite()
rocket::build()
.manage(state)
.register(catchers![not_found])
.manage(cache)
.register("/", catchers![not_found])
.mount(
"/",
routes![index, show_question, show_answer, question0, answer0],
)
.mount("/q", routes![index])
.mount("/q/static", StaticFiles::from("static/"))
.mount("/q/static", FileServer::from("static/"))
.attach(Template::fairing())
}
fn main() {
rocket().launch();
}