1
0
mirror of https://github.com/fafhrd91/actix-web synced 2024-11-24 16:32:59 +01:00
actix-web/src/ws/mask.rs

138 lines
3.7 KiB
Rust
Raw Normal View History

//! This is code from [Tungstenite project](https://github.com/snapview/tungstenite-rs)
2018-04-29 18:09:08 +02:00
#![cfg_attr(feature = "cargo-clippy", allow(cast_ptr_alignment))]
use std::cmp::min;
use std::mem::uninitialized;
use std::ptr::copy_nonoverlapping;
/// Mask/unmask a frame.
#[inline]
2018-03-09 05:39:05 +01:00
pub fn apply_mask(buf: &mut [u8], mask: u32) {
2018-06-19 21:27:41 +02:00
unsafe { apply_mask_fast32(buf, mask) }
}
/// A safe unoptimized mask application.
#[inline]
#[allow(dead_code)]
fn apply_mask_fallback(buf: &mut [u8], mask: &[u8; 4]) {
for (i, byte) in buf.iter_mut().enumerate() {
*byte ^= mask[i & 3];
}
}
2018-03-09 02:19:50 +01:00
/// Faster version of `apply_mask()` which operates on 8-byte blocks.
2018-06-19 21:27:41 +02:00
///
/// unsafe because uses pointer math and bit operations for performance
#[inline]
2018-04-14 01:02:01 +02:00
#[cfg_attr(feature = "cargo-clippy", allow(cast_lossless))]
2018-06-19 21:27:41 +02:00
unsafe fn apply_mask_fast32(buf: &mut [u8], mask_u32: u32) {
let mut ptr = buf.as_mut_ptr();
let mut len = buf.len();
// Possible first unaligned block.
2018-03-09 19:25:47 +01:00
let head = min(len, (8 - (ptr as usize & 0x7)) & 0x3);
let mask_u32 = if head > 0 {
2018-04-29 18:09:08 +02:00
let n = if head > 4 { head - 4 } else { head };
2018-03-09 02:19:50 +01:00
2018-03-09 19:25:47 +01:00
let mask_u32 = if n > 0 {
2018-06-19 21:27:41 +02:00
xor_mem(ptr, mask_u32, n);
ptr = ptr.offset(head as isize);
2018-03-09 19:25:47 +01:00
len -= n;
2018-03-09 02:19:50 +01:00
if cfg!(target_endian = "big") {
2018-03-09 19:25:47 +01:00
mask_u32.rotate_left(8 * n as u32)
2018-03-09 02:19:50 +01:00
} else {
2018-03-09 19:25:47 +01:00
mask_u32.rotate_right(8 * n as u32)
2018-03-09 02:19:50 +01:00
}
} else {
mask_u32
2018-03-09 19:25:47 +01:00
};
if head > 4 {
2018-06-19 21:27:41 +02:00
*(ptr as *mut u32) ^= mask_u32;
ptr = ptr.offset(4);
len -= 4;
2018-03-09 19:25:47 +01:00
}
mask_u32
} else {
mask_u32
};
if len > 0 {
debug_assert_eq!(ptr as usize % 4, 0);
}
// Properly aligned middle of the data.
2018-03-09 02:19:50 +01:00
if len >= 8 {
let mut mask_u64 = mask_u32 as u64;
mask_u64 = mask_u64 << 32 | mask_u32 as u64;
while len >= 8 {
2018-06-19 21:27:41 +02:00
*(ptr as *mut u64) ^= mask_u64;
ptr = ptr.offset(8);
len -= 8;
2018-03-09 02:19:50 +01:00
}
}
while len >= 4 {
2018-06-19 21:27:41 +02:00
*(ptr as *mut u32) ^= mask_u32;
ptr = ptr.offset(4);
len -= 4;
}
// Possible last block.
if len > 0 {
2018-06-19 21:27:41 +02:00
xor_mem(ptr, mask_u32, len);
}
}
#[inline]
2018-04-14 01:02:01 +02:00
// TODO: copy_nonoverlapping here compiles to call memcpy. While it is not so
// inefficient, it could be done better. The compiler does not see that len is
// limited to 3.
unsafe fn xor_mem(ptr: *mut u8, mask: u32, len: usize) {
let mut b: u32 = uninitialized();
#[allow(trivial_casts)]
copy_nonoverlapping(ptr, &mut b as *mut _ as *mut u8, len);
b ^= mask;
#[allow(trivial_casts)]
copy_nonoverlapping(&b as *const _ as *const u8, ptr, len);
}
#[cfg(test)]
mod tests {
2018-06-19 21:27:41 +02:00
use super::{apply_mask, apply_mask_fallback};
2018-03-09 05:39:05 +01:00
use std::ptr;
#[test]
fn test_apply_mask() {
2018-04-14 01:02:01 +02:00
let mask = [0x6d, 0xb6, 0xb2, 0x80];
let mask_u32: u32 = unsafe { ptr::read_unaligned(mask.as_ptr() as *const u32) };
2018-03-09 05:39:05 +01:00
let unmasked = vec![
2018-04-14 01:02:01 +02:00
0xf3, 0x00, 0x01, 0x02, 0x03, 0x80, 0x81, 0x82, 0xff, 0xfe, 0x00, 0x17,
0x74, 0xf9, 0x12, 0x03,
];
// Check masking with proper alignment.
{
let mut masked = unmasked.clone();
apply_mask_fallback(&mut masked, &mask);
let mut masked_fast = unmasked.clone();
2018-06-19 21:27:41 +02:00
apply_mask(&mut masked_fast, mask_u32);
assert_eq!(masked, masked_fast);
}
// Check masking without alignment.
{
let mut masked = unmasked.clone();
apply_mask_fallback(&mut masked[1..], &mask);
let mut masked_fast = unmasked.clone();
2018-06-19 21:27:41 +02:00
apply_mask(&mut masked_fast[1..], mask_u32);
assert_eq!(masked, masked_fast);
}
}
}