1
0
mirror of https://github.com/actix/examples synced 2024-11-30 17:14:35 +01:00
examples/basics/state/src/main.rs

83 lines
2.9 KiB
Rust
Raw Normal View History

2019-03-17 04:23:09 +01:00
//! Application may have multiple data objects that are shared across
//! all handlers within same Application.
2018-05-02 15:20:43 +02:00
//!
2022-01-29 15:23:32 +01:00
//! For global shared state, we wrap our state in a [`actix_web::web::Data`] and move it into
//! the factory closure. The closure is called once-per-thread, and we clone our state
2022-01-29 15:23:32 +01:00
//! and attach to each instance of the [`App`] with `.app_data(state.clone())`.
//!
//! For thread-local state, we construct our state within the factory closure and attach to
2022-01-29 15:23:32 +01:00
//! the app with `.app_data(Data::new(state))`.
//!
//! We retrieve our app state within our handlers with a `state: Data<...>` argument.
//!
2022-02-06 09:13:24 +01:00
//! By default, Actix Web runs one [`App`] per logical cpu core.
2022-01-29 15:23:32 +01:00
//! When running on `<N>` cores, we see that the example will increment `counter_mutex` (global state via
//! Mutex) and `counter_atomic` (global state via Atomic variable) each time the endpoint is called,
//! but only appear to increment `counter_cell` every Nth time on average (thread-local state). This
//! is because the workload is being shared equally among cores.
2018-05-02 15:20:43 +02:00
//!
2019-11-14 21:16:48 +01:00
//! Check [user guide](https://actix.rs/docs/application/#state) for more info.
use std::cell::Cell;
2019-03-07 23:50:29 +01:00
use std::io;
use std::sync::atomic::{AtomicUsize, Ordering};
2019-06-06 12:47:40 +02:00
use std::sync::Mutex;
2022-01-29 15:23:32 +01:00
use actix_web::{
middleware,
web::{self, Data},
App, HttpRequest, HttpResponse, HttpServer,
};
/// simple handle
async fn index(
2022-01-29 15:23:32 +01:00
counter_mutex: Data<Mutex<usize>>,
counter_cell: Data<Cell<u32>>,
counter_atomic: Data<AtomicUsize>,
req: HttpRequest,
) -> HttpResponse {
println!("{:?}", req);
// Increment the counters
2022-01-29 15:23:32 +01:00
*counter_mutex.lock().unwrap() += 1;
counter_cell.set(counter_cell.get() + 1);
counter_atomic.fetch_add(1, Ordering::SeqCst);
let body = format!(
"global mutex counter: {}, local counter: {}, global atomic counter: {}",
2022-01-29 15:23:32 +01:00
*counter_mutex.lock().unwrap(),
counter_cell.get(),
counter_atomic.load(Ordering::SeqCst),
);
HttpResponse::Ok().body(body)
}
2020-09-12 17:49:45 +02:00
#[actix_web::main]
2019-12-07 18:59:24 +01:00
async fn main() -> io::Result<()> {
2019-03-07 23:50:29 +01:00
std::env::set_var("RUST_LOG", "actix_web=info");
env_logger::init();
// Create some global state prior to building the server
2020-04-03 09:14:30 +02:00
#[allow(clippy::mutex_atomic)] // it's intentional.
2022-01-29 15:23:32 +01:00
let counter_mutex = Data::new(Mutex::new(0usize));
let counter_atomic = Data::new(AtomicUsize::new(0usize));
2019-03-07 23:50:29 +01:00
// move is necessary to give closure below ownership of counter1
2019-03-07 23:50:29 +01:00
HttpServer::new(move || {
// Create some thread-local state
2022-01-29 15:23:32 +01:00
let counter_cell = Cell::new(0u32);
2019-03-07 23:50:29 +01:00
App::new()
2022-01-29 15:23:32 +01:00
.app_data(counter_mutex.clone()) // add shared state
.app_data(counter_atomic.clone()) // add shared state
.app_data(Data::new(counter_cell)) // add thread-local state
// enable logger
2019-03-26 04:29:00 +01:00
.wrap(middleware::Logger::default())
// register simple handler
2019-03-07 23:50:29 +01:00
.service(web::resource("/").to(index))
})
.bind("127.0.0.1:8080")?
2019-12-25 17:48:33 +01:00
.run()
2019-12-07 18:59:24 +01:00
.await
}