2018-01-21 01:47:34 +01:00
|
|
|
//! This is code from [Tungstenite project](https://github.com/snapview/tungstenite-rs)
|
|
|
|
use std::cmp::min;
|
|
|
|
use std::mem::uninitialized;
|
|
|
|
use std::ptr::copy_nonoverlapping;
|
|
|
|
|
|
|
|
/// Mask/unmask a frame.
|
|
|
|
#[inline]
|
2018-03-09 05:39:05 +01:00
|
|
|
pub fn apply_mask(buf: &mut [u8], mask: u32) {
|
2018-01-21 01:47:34 +01:00
|
|
|
apply_mask_fast32(buf, mask)
|
|
|
|
}
|
|
|
|
|
|
|
|
/// A safe unoptimized mask application.
|
|
|
|
#[inline]
|
|
|
|
#[allow(dead_code)]
|
|
|
|
fn apply_mask_fallback(buf: &mut [u8], mask: &[u8; 4]) {
|
|
|
|
for (i, byte) in buf.iter_mut().enumerate() {
|
|
|
|
*byte ^= mask[i & 3];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-09 02:19:50 +01:00
|
|
|
/// Faster version of `apply_mask()` which operates on 8-byte blocks.
|
2018-01-21 01:47:34 +01:00
|
|
|
#[inline]
|
2018-03-09 05:39:05 +01:00
|
|
|
fn apply_mask_fast32(buf: &mut [u8], mask_u32: u32) {
|
2018-01-21 01:47:34 +01:00
|
|
|
let mut ptr = buf.as_mut_ptr();
|
|
|
|
let mut len = buf.len();
|
|
|
|
|
|
|
|
// Possible first unaligned block.
|
|
|
|
let head = min(len, (4 - (ptr as usize & 3)) & 3);
|
|
|
|
let mask_u32 = if head > 0 {
|
|
|
|
unsafe {
|
|
|
|
xor_mem(ptr, mask_u32, head);
|
|
|
|
ptr = ptr.offset(head as isize);
|
|
|
|
}
|
|
|
|
len -= head;
|
2018-03-09 05:39:05 +01:00
|
|
|
//let mask_u32 =
|
|
|
|
if cfg!(target_endian = "big") {
|
2018-01-21 01:47:34 +01:00
|
|
|
mask_u32.rotate_left(8 * head as u32)
|
|
|
|
} else {
|
|
|
|
mask_u32.rotate_right(8 * head as u32)
|
2018-03-09 05:39:05 +01:00
|
|
|
}//;
|
2018-03-09 02:19:50 +01:00
|
|
|
|
2018-03-09 05:39:05 +01:00
|
|
|
/*
|
2018-03-09 02:19:50 +01:00
|
|
|
let head = min(len, (4 - (ptr as usize & 3)) & 3);
|
|
|
|
if head > 0 {
|
|
|
|
unsafe {
|
|
|
|
xor_mem(ptr, mask_u32, head);
|
|
|
|
ptr = ptr.offset(head as isize);
|
|
|
|
}
|
|
|
|
len -= head;
|
|
|
|
if cfg!(target_endian = "big") {
|
|
|
|
mask_u32.rotate_left(8 * head as u32)
|
|
|
|
} else {
|
|
|
|
mask_u32.rotate_right(8 * head as u32)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
mask_u32
|
2018-03-09 05:39:05 +01:00
|
|
|
}*/
|
2018-01-21 01:47:34 +01:00
|
|
|
} else {
|
|
|
|
mask_u32
|
|
|
|
};
|
|
|
|
|
|
|
|
if len > 0 {
|
|
|
|
debug_assert_eq!(ptr as usize % 4, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Properly aligned middle of the data.
|
2018-03-09 02:19:50 +01:00
|
|
|
if len >= 8 {
|
|
|
|
let mut mask_u64 = mask_u32 as u64;
|
|
|
|
mask_u64 = mask_u64 << 32 | mask_u32 as u64;
|
|
|
|
|
|
|
|
while len >= 8 {
|
|
|
|
unsafe {
|
|
|
|
*(ptr as *mut u64) ^= mask_u64;
|
|
|
|
ptr = ptr.offset(8);
|
|
|
|
len -= 8;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
while len >= 4 {
|
2018-01-21 01:47:34 +01:00
|
|
|
unsafe {
|
|
|
|
*(ptr as *mut u32) ^= mask_u32;
|
|
|
|
ptr = ptr.offset(4);
|
|
|
|
len -= 4;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Possible last block.
|
|
|
|
if len > 0 {
|
|
|
|
unsafe { xor_mem(ptr, mask_u32, len); }
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[inline]
|
|
|
|
// TODO: copy_nonoverlapping here compiles to call memcpy. While it is not so inefficient,
|
|
|
|
// it could be done better. The compiler does not see that len is limited to 3.
|
|
|
|
unsafe fn xor_mem(ptr: *mut u8, mask: u32, len: usize) {
|
|
|
|
let mut b: u32 = uninitialized();
|
|
|
|
#[allow(trivial_casts)]
|
|
|
|
copy_nonoverlapping(ptr, &mut b as *mut _ as *mut u8, len);
|
|
|
|
b ^= mask;
|
|
|
|
#[allow(trivial_casts)]
|
|
|
|
copy_nonoverlapping(&b as *const _ as *const u8, ptr, len);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
mod tests {
|
2018-03-09 05:39:05 +01:00
|
|
|
use std::ptr;
|
2018-01-21 01:47:34 +01:00
|
|
|
use super::{apply_mask_fallback, apply_mask_fast32};
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_apply_mask() {
|
|
|
|
let mask = [
|
|
|
|
0x6d, 0xb6, 0xb2, 0x80,
|
|
|
|
];
|
2018-03-09 05:39:05 +01:00
|
|
|
let mask_u32: u32 = unsafe {ptr::read_unaligned(mask.as_ptr() as *const u32)};
|
|
|
|
|
2018-01-21 01:47:34 +01:00
|
|
|
let unmasked = vec![
|
|
|
|
0xf3, 0x00, 0x01, 0x02, 0x03, 0x80, 0x81, 0x82,
|
|
|
|
0xff, 0xfe, 0x00, 0x17, 0x74, 0xf9, 0x12, 0x03,
|
|
|
|
];
|
|
|
|
|
|
|
|
// Check masking with proper alignment.
|
|
|
|
{
|
|
|
|
let mut masked = unmasked.clone();
|
|
|
|
apply_mask_fallback(&mut masked, &mask);
|
|
|
|
|
|
|
|
let mut masked_fast = unmasked.clone();
|
2018-03-09 05:39:05 +01:00
|
|
|
apply_mask_fast32(&mut masked_fast, mask_u32);
|
2018-01-21 01:47:34 +01:00
|
|
|
|
|
|
|
assert_eq!(masked, masked_fast);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check masking without alignment.
|
|
|
|
{
|
|
|
|
let mut masked = unmasked.clone();
|
|
|
|
apply_mask_fallback(&mut masked[1..], &mask);
|
|
|
|
|
|
|
|
let mut masked_fast = unmasked.clone();
|
2018-03-09 05:39:05 +01:00
|
|
|
apply_mask_fast32(&mut masked_fast[1..], mask_u32);
|
2018-01-21 01:47:34 +01:00
|
|
|
|
|
|
|
assert_eq!(masked, masked_fast);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|