More comprehensive unit tests for router

This commit is contained in:
Mathias Hall-Andersen
2020-02-20 13:21:37 +01:00
parent ead75828cd
commit db02609334
9 changed files with 667 additions and 605 deletions

View File

@@ -5,7 +5,7 @@ use std::sync::Arc;
use std::thread; use std::thread;
use std::time::Instant; use std::time::Instant;
use log::debug; use log;
use spin::{Mutex, RwLock}; use spin::{Mutex, RwLock};
use zerocopy::LayoutVerified; use zerocopy::LayoutVerified;
@@ -91,20 +91,17 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Drop
for DeviceHandle<E, C, T, B> for DeviceHandle<E, C, T, B>
{ {
fn drop(&mut self) { fn drop(&mut self) {
debug!("router: dropping device"); log::debug!("router: dropping device");
// close worker queue // close worker queue
self.state.work.close(); self.state.work.close();
// join all worker threads // join all worker threads
while match self.handles.pop() { while let Some(handle) = self.handles.pop() {
Some(handle) => {
handle.thread().unpark(); handle.thread().unpark();
handle.join().unwrap(); handle.join().unwrap();
true
} }
_ => false, log::debug!("router: joined with all workers from pool");
} {}
} }
} }
@@ -124,8 +121,13 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> DeviceHandle<
// start worker threads // start worker threads
let mut threads = Vec::with_capacity(num_workers); let mut threads = Vec::with_capacity(num_workers);
while let Some(rx) = consumers.pop() { while let Some(rx) = consumers.pop() {
threads.push(thread::spawn(move || worker(rx))); println!("spawn");
threads.push(thread::spawn(move || {
println!("spawned");
worker(rx);
}));
} }
debug_assert!(num_workers > 0, "zero worker threads");
debug_assert_eq!(threads.len(), num_workers); debug_assert_eq!(threads.len(), num_workers);
// return exported device handle // return exported device handle
@@ -142,7 +144,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> DeviceHandle<
return bind.write(msg, dst); return bind.write(msg, dst);
} }
} }
return Ok(()) return Ok(());
} }
/// Brings the router down. /// Brings the router down.

View File

@@ -1,5 +1,8 @@
use std::mem;
use byteorder::BigEndian; use byteorder::BigEndian;
use zerocopy::byteorder::U16; use zerocopy::byteorder::U16;
use zerocopy::LayoutVerified;
use zerocopy::{AsBytes, FromBytes}; use zerocopy::{AsBytes, FromBytes};
pub const VERSION_IP4: u8 = 4; pub const VERSION_IP4: u8 = 4;
@@ -24,3 +27,23 @@ pub struct IPv6Header {
pub f_source: [u8; 16], pub f_source: [u8; 16],
pub f_destination: [u8; 16], pub f_destination: [u8; 16],
} }
#[inline(always)]
pub fn inner_length(packet: &[u8]) -> Option<usize> {
match packet.get(0)? >> 4 {
VERSION_IP4 => {
let (header, _): (LayoutVerified<&[u8], IPv4Header>, _) =
LayoutVerified::new_from_prefix(packet)?;
Some(header.f_total_len.get() as usize)
}
VERSION_IP6 => {
// check length and cast to IPv6 header
let (header, _): (LayoutVerified<&[u8], IPv6Header>, _) =
LayoutVerified::new_from_prefix(packet)?;
Some(header.f_len.get() as usize + mem::size_of::<IPv6Header>())
}
_ => None,
}
}

View File

@@ -24,6 +24,7 @@ use super::types::*;
pub const SIZE_TAG: usize = 16; pub const SIZE_TAG: usize = 16;
pub const SIZE_MESSAGE_PREFIX: usize = mem::size_of::<TransportHeader>(); pub const SIZE_MESSAGE_PREFIX: usize = mem::size_of::<TransportHeader>();
pub const SIZE_KEEPALIVE: usize = mem::size_of::<TransportHeader>() + SIZE_TAG;
pub const CAPACITY_MESSAGE_POSTFIX: usize = SIZE_TAG; pub const CAPACITY_MESSAGE_POSTFIX: usize = SIZE_TAG;
pub const fn message_data_len(payload: usize) -> usize { pub const fn message_data_len(payload: usize) -> usize {

View File

@@ -22,7 +22,7 @@ use std::sync::atomic::AtomicBool;
use std::sync::Arc; use std::sync::Arc;
use arraydeque::{ArrayDeque, Wrapping}; use arraydeque::{ArrayDeque, Wrapping};
use log::debug; use log;
use spin::Mutex; use spin::Mutex;
pub struct KeyWheel { pub struct KeyWheel {
@@ -148,7 +148,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Drop for Peer
*peer.enc_key.lock() = None; *peer.enc_key.lock() = None;
*peer.endpoint.lock() = None; *peer.endpoint.lock() = None;
debug!("peer dropped & removed from device"); log::debug!("peer dropped & removed from device");
} }
} }
@@ -192,8 +192,6 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> PeerInner<E,
/// ///
/// Unit if packet was sent, or an error indicating why sending failed /// Unit if packet was sent, or an error indicating why sending failed
pub fn send_raw(&self, msg: &[u8]) -> Result<(), RouterError> { pub fn send_raw(&self, msg: &[u8]) -> Result<(), RouterError> {
debug!("peer.send");
// send to endpoint (if known) // send to endpoint (if known)
match self.endpoint.lock().as_mut() { match self.endpoint.lock().as_mut() {
Some(endpoint) => { Some(endpoint) => {
@@ -227,6 +225,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Peer<E, C, T,
let mut enc_key = self.enc_key.lock(); let mut enc_key = self.enc_key.lock();
match enc_key.as_mut() { match enc_key.as_mut() {
None => { None => {
log::debug!("no key encryption key available");
if stage { if stage {
self.staged_packets.lock().push_back(msg); self.staged_packets.lock().push_back(msg);
}; };
@@ -235,13 +234,14 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Peer<E, C, T,
Some(mut state) => { Some(mut state) => {
// avoid integer overflow in nonce // avoid integer overflow in nonce
if state.nonce >= REJECT_AFTER_MESSAGES - 1 { if state.nonce >= REJECT_AFTER_MESSAGES - 1 {
log::debug!("encryption key expired");
*enc_key = None; *enc_key = None;
if stage { if stage {
self.staged_packets.lock().push_back(msg); self.staged_packets.lock().push_back(msg);
} }
(None, true) (None, true)
} else { } else {
debug!("encryption state available, nonce = {}", state.nonce); log::debug!("encryption state available, nonce = {}", state.nonce);
let job = let job =
SendJob::new(msg, state.nonce, state.keypair.clone(), self.clone()); SendJob::new(msg, state.nonce, state.keypair.clone(), self.clone());
if self.outbound.push(job.clone()) { if self.outbound.push(job.clone()) {
@@ -256,18 +256,20 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Peer<E, C, T,
}; };
if need_key { if need_key {
log::debug!("request new key");
debug_assert!(job.is_none()); debug_assert!(job.is_none());
C::need_key(&self.opaque); C::need_key(&self.opaque);
}; };
if let Some(job) = job { if let Some(job) = job {
log::debug!("schedule outbound job");
self.device.work.send(JobUnion::Outbound(job)) self.device.work.send(JobUnion::Outbound(job))
} }
} }
// Transmit all staged packets // Transmit all staged packets
fn send_staged(&self) -> bool { fn send_staged(&self) -> bool {
debug!("peer.send_staged"); log::trace!("peer.send_staged");
let mut sent = false; let mut sent = false;
let mut staged = self.staged_packets.lock(); let mut staged = self.staged_packets.lock();
loop { loop {
@@ -282,7 +284,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Peer<E, C, T,
} }
pub(super) fn confirm_key(&self, keypair: &Arc<KeyPair>) { pub(super) fn confirm_key(&self, keypair: &Arc<KeyPair>) {
debug!("peer.confirm_key"); log::trace!("peer.confirm_key");
{ {
// take lock and check keypair = keys.next // take lock and check keypair = keys.next
let mut keys = self.keys.lock(); let mut keys = self.keys.lock();
@@ -329,7 +331,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> PeerHandle<E,
/// This API still permits support for the "sticky socket" behavior, /// This API still permits support for the "sticky socket" behavior,
/// as sockets should be "unsticked" when manually updating the endpoint /// as sockets should be "unsticked" when manually updating the endpoint
pub fn set_endpoint(&self, endpoint: E) { pub fn set_endpoint(&self, endpoint: E) {
debug!("peer.set_endpoint"); log::trace!("peer.set_endpoint");
*self.peer.endpoint.lock() = Some(endpoint); *self.peer.endpoint.lock() = Some(endpoint);
} }
@@ -339,13 +341,13 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> PeerHandle<E,
/// ///
/// Does not convey potential "sticky socket" information /// Does not convey potential "sticky socket" information
pub fn get_endpoint(&self) -> Option<SocketAddr> { pub fn get_endpoint(&self) -> Option<SocketAddr> {
debug!("peer.get_endpoint"); log::trace!("peer.get_endpoint");
self.peer.endpoint.lock().as_ref().map(|e| e.into_address()) self.peer.endpoint.lock().as_ref().map(|e| e.into_address())
} }
/// Zero all key-material related to the peer /// Zero all key-material related to the peer
pub fn zero_keys(&self) { pub fn zero_keys(&self) {
debug!("peer.zero_keys"); log::trace!("peer.zero_keys");
let mut release: Vec<u32> = Vec::with_capacity(3); let mut release: Vec<u32> = Vec::with_capacity(3);
let mut keys = self.peer.keys.lock(); let mut keys = self.peer.keys.lock();
@@ -416,7 +418,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> PeerHandle<E,
// update incoming packet id map // update incoming packet id map
{ {
debug!("peer.add_keypair: updating inbound id map"); log::trace!("peer.add_keypair: updating inbound id map");
let mut recv = self.peer.device.recv.write(); let mut recv = self.peer.device.recv.write();
// purge recv map of previous id // purge recv map of previous id
@@ -438,14 +440,14 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> PeerHandle<E,
// schedule confirmation // schedule confirmation
if initiator { if initiator {
debug_assert!(self.peer.enc_key.lock().is_some()); debug_assert!(self.peer.enc_key.lock().is_some());
debug!("peer.add_keypair: is initiator, must confirm the key"); log::trace!("peer.add_keypair: is initiator, must confirm the key");
// attempt to confirm using staged packets // attempt to confirm using staged packets
if !self.peer.send_staged() { if !self.peer.send_staged() {
// fall back to keepalive packet // fall back to keepalive packet
self.send_keepalive(); self.send_keepalive();
debug!("peer.add_keypair: keepalive for confirmation",); log::debug!("peer.add_keypair: keepalive for confirmation",);
} }
debug!("peer.add_keypair: key attempted confirmed"); log::trace!("peer.add_keypair: key attempted confirmed");
} }
debug_assert!( debug_assert!(
@@ -456,7 +458,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> PeerHandle<E,
} }
pub fn send_keepalive(&self) { pub fn send_keepalive(&self) {
debug!("peer.send_keepalive"); log::trace!("peer.send_keepalive");
self.peer.send(vec![0u8; SIZE_MESSAGE_PREFIX], false) self.peer.send(vec![0u8; SIZE_MESSAGE_PREFIX], false)
} }

View File

@@ -1,12 +1,12 @@
use super::device::DecryptionState; use super::device::DecryptionState;
use super::ip::inner_length;
use super::messages::TransportHeader; use super::messages::TransportHeader;
use super::queue::{ParallelJob, Queue, SequentialJob}; use super::queue::{ParallelJob, Queue, SequentialJob};
use super::types::Callbacks; use super::types::Callbacks;
use super::{REJECT_AFTER_MESSAGES, SIZE_TAG}; use super::{REJECT_AFTER_MESSAGES, SIZE_KEEPALIVE};
use super::super::{tun, udp, Endpoint}; use super::super::{tun, udp, Endpoint};
use std::mem;
use std::sync::atomic::{AtomicBool, Ordering}; use std::sync::atomic::{AtomicBool, Ordering};
use std::sync::Arc; use std::sync::Arc;
@@ -15,7 +15,7 @@ use spin::Mutex;
use zerocopy::{AsBytes, LayoutVerified}; use zerocopy::{AsBytes, LayoutVerified};
struct Inner<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> { struct Inner<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> {
ready: AtomicBool, ready: AtomicBool, // job status
buffer: Mutex<(Option<E>, Vec<u8>)>, // endpoint & ciphertext buffer buffer: Mutex<(Option<E>, Vec<u8>)>, // endpoint & ciphertext buffer
state: Arc<DecryptionState<E, C, T, B>>, // decryption state (keys and replay protector) state: Arc<DecryptionState<E, C, T, B>>, // decryption state (keys and replay protector)
} }
@@ -53,26 +53,41 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> ParallelJob
&self.0.state.peer.inbound &self.0.state.peer.inbound
} }
/* The parallel section of an incoming job:
*
* - Decryption.
* - Crypto-key routing lookup.
*
* Note: We truncate the message buffer to 0 bytes in case of authentication failure
* or crypto-key routing failure (attempted impersonation).
*
* Note: We cannot do replay protection in the parallel job,
* since this can cause dropping of packets (leaving the window) due to scheduling.
*/
fn parallel_work(&self) { fn parallel_work(&self) {
// TODO: refactor debug_assert_eq!(
self.is_ready(),
false,
"doing parallel work on completed job"
);
log::trace!("processing parallel receive job");
// decrypt // decrypt
{ {
// closure for locking
let job = &self.0; let job = &self.0;
let peer = &job.state.peer; let peer = &job.state.peer;
let mut msg = job.buffer.lock(); let mut msg = job.buffer.lock();
// process buffer
let ok = (|| {
// cast to header followed by payload // cast to header followed by payload
let (header, packet): (LayoutVerified<&mut [u8], TransportHeader>, &mut [u8]) = let (header, packet): (LayoutVerified<&mut [u8], TransportHeader>, &mut [u8]) =
match LayoutVerified::new_from_prefix(&mut msg.1[..]) { match LayoutVerified::new_from_prefix(&mut msg.1[..]) {
Some(v) => v, Some(v) => v,
None => { None => return false,
log::debug!("inbound worker: failed to parse message");
return;
}
}; };
// authenticate and decrypt payload
{
// create nonce object // create nonce object
let mut nonce = [0u8; 12]; let mut nonce = [0u8; 12];
debug_assert_eq!(nonce.len(), CHACHA20_POLY1305.nonce_len()); debug_assert_eq!(nonce.len(), CHACHA20_POLY1305.nonce_len());
@@ -87,48 +102,25 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> ParallelJob
// attempt to open (and authenticate) the body // attempt to open (and authenticate) the body
match key.open_in_place(nonce, Aad::empty(), packet) { match key.open_in_place(nonce, Aad::empty(), packet) {
Ok(_) => (), Ok(_) => (),
Err(_) => { Err(_) => return false,
// fault and return early
log::trace!("inbound worker: authentication failure");
msg.1.truncate(0);
return;
}
}
} }
// check that counter not after reject // check that counter not after reject
if header.f_counter.get() >= REJECT_AFTER_MESSAGES { if header.f_counter.get() >= REJECT_AFTER_MESSAGES {
msg.1.truncate(0); return false;
return;
} }
// cryptokey route and strip padding // check crypto-key router
let inner_len = { packet.len() == SIZE_KEEPALIVE || peer.device.table.check_route(&peer, &packet)
let length = packet.len() - SIZE_TAG; })();
if length > 0 {
peer.device.table.check_route(&peer, &packet[..length]) // remove message in case of failure:
} else { // to indicate failure and avoid later accidental use of unauthenticated data.
Some(0) if !ok {
msg.1.truncate(0);
} }
}; };
// truncate to remove tag
match inner_len {
None => {
log::trace!("inbound worker: cryptokey routing failed");
msg.1.truncate(0);
}
Some(len) => {
log::trace!(
"inbound worker: good route, length = {} {}",
len,
if len == 0 { "(keepalive)" } else { "" }
);
msg.1.truncate(mem::size_of::<TransportHeader>() + len);
}
}
}
// mark ready // mark ready
self.0.ready.store(true, Ordering::Release); self.0.ready.store(true, Ordering::Release);
} }
@@ -142,6 +134,13 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> SequentialJob
} }
fn sequential_work(self) { fn sequential_work(self) {
debug_assert_eq!(
self.is_ready(),
true,
"doing sequential work on an incomplete job"
);
log::trace!("processing sequential receive job");
let job = &self.0; let job = &self.0;
let peer = &job.state.peer; let peer = &job.state.peer;
let mut msg = job.buffer.lock(); let mut msg = job.buffer.lock();
@@ -152,7 +151,7 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> SequentialJob
match LayoutVerified::new_from_prefix(&msg.1[..]) { match LayoutVerified::new_from_prefix(&msg.1[..]) {
Some(v) => v, Some(v) => v,
None => { None => {
// also covers authentication failure // also covers authentication failure (will fail to parse header)
return; return;
} }
}; };
@@ -173,20 +172,16 @@ impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> SequentialJob
*peer.endpoint.lock() = endpoint; *peer.endpoint.lock() = endpoint;
// check if should be written to TUN // check if should be written to TUN
let mut sent = false; // (keep-alive and malformed packets will have no inner length)
if packet.len() > 0 { if let Some(inner) = inner_length(packet) {
sent = match peer.device.inbound.write(&packet[..]) { if inner >= packet.len() {
Err(e) => { let _ = peer.device.inbound.write(&packet[..inner]).map_err(|e| {
log::debug!("failed to write inbound packet to TUN: {:?}", e); log::debug!("failed to write inbound packet to TUN: {:?}", e);
false });
} }
Ok(_) => true,
}
} else {
log::debug!("inbound worker: received keepalive")
} }
// trigger callback // trigger callback
C::recv(&peer.opaque, msg.1.len(), sent, &job.state.keypair); C::recv(&peer.opaque, msg.1.len(), true, &job.state.keypair);
} }
} }

View File

@@ -1,13 +1,11 @@
use super::ip::*; use super::ip::*;
use zerocopy::LayoutVerified;
use std::mem;
use std::net::{IpAddr, Ipv4Addr, Ipv6Addr}; use std::net::{IpAddr, Ipv4Addr, Ipv6Addr};
use spin::RwLock; use spin::RwLock;
use treebitmap::address::Address; use treebitmap::address::Address;
use treebitmap::IpLookupTable; use treebitmap::IpLookupTable;
use zerocopy::LayoutVerified;
/* Functions for obtaining and validating "cryptokey" routes */ /* Functions for obtaining and validating "cryptokey" routes */
@@ -115,53 +113,26 @@ impl<T: Eq + Clone> RoutingTable<T> {
} }
#[inline(always)] #[inline(always)]
pub fn check_route(&self, peer: &T, packet: &[u8]) -> Option<usize> { pub fn check_route(&self, peer: &T, packet: &[u8]) -> bool {
match packet.get(0)? >> 4 { match packet.get(0).map(|v| v >> 4) {
VERSION_IP4 => { Some(VERSION_IP4) => LayoutVerified::new_from_prefix(packet)
// check length and cast to IPv4 header .and_then(|(header, _): (LayoutVerified<&[u8], IPv4Header>, _)| {
let (header, _): (LayoutVerified<&[u8], IPv4Header>, _) =
LayoutVerified::new_from_prefix(packet)?;
log::trace!(
"router, check route for IPv4 source: {:?}",
Ipv4Addr::from(header.f_source)
);
// check IPv4 source address
self.ipv4 self.ipv4
.read() .read()
.longest_match(Ipv4Addr::from(header.f_source)) .longest_match(Ipv4Addr::from(header.f_source))
.and_then(|(_, _, p)| { .map(|(_, _, p)| p == peer)
if p == peer {
Some(header.f_total_len.get() as usize)
} else {
None
}
}) })
} .is_some(),
VERSION_IP6 => {
// check length and cast to IPv6 header
let (header, _): (LayoutVerified<&[u8], IPv6Header>, _) =
LayoutVerified::new_from_prefix(packet)?;
log::trace!( Some(VERSION_IP6) => LayoutVerified::new_from_prefix(packet)
"router, check route for IPv6 source: {:?}", .and_then(|(header, _): (LayoutVerified<&[u8], IPv6Header>, _)| {
Ipv6Addr::from(header.f_source)
);
// check IPv6 source address
self.ipv6 self.ipv6
.read() .read()
.longest_match(Ipv6Addr::from(header.f_source)) .longest_match(Ipv6Addr::from(header.f_source))
.and_then(|(_, _, p)| { .map(|(_, _, p)| p == peer)
if p == peer {
Some(header.f_len.get() as usize + mem::size_of::<IPv6Header>())
} else {
None
}
}) })
} .is_some(),
_ => None, _ => false,
} }
} }
} }

View File

@@ -1,9 +1,9 @@
use super::queue::{SequentialJob, ParallelJob, Queue};
use super::KeyPair;
use super::types::Callbacks;
use super::peer::Peer;
use super::{REJECT_AFTER_MESSAGES, SIZE_TAG};
use super::messages::{TransportHeader, TYPE_TRANSPORT}; use super::messages::{TransportHeader, TYPE_TRANSPORT};
use super::peer::Peer;
use super::queue::{ParallelJob, Queue, SequentialJob};
use super::types::Callbacks;
use super::KeyPair;
use super::{REJECT_AFTER_MESSAGES, SIZE_TAG};
use super::super::{tun, udp, Endpoint}; use super::super::{tun, udp, Endpoint};
@@ -11,8 +11,8 @@ use std::sync::atomic::{AtomicBool, Ordering};
use std::sync::Arc; use std::sync::Arc;
use ring::aead::{Aad, LessSafeKey, Nonce, UnboundKey, CHACHA20_POLY1305}; use ring::aead::{Aad, LessSafeKey, Nonce, UnboundKey, CHACHA20_POLY1305};
use zerocopy::{AsBytes, LayoutVerified};
use spin::Mutex; use spin::Mutex;
use zerocopy::{AsBytes, LayoutVerified};
struct Inner<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> { struct Inner<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> {
ready: AtomicBool, ready: AtomicBool,
@@ -23,7 +23,7 @@ struct Inner<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> {
} }
pub struct SendJob<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>>( pub struct SendJob<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>>(
Arc<Inner<E, C, T, B>> Arc<Inner<E, C, T, B>>,
); );
impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Clone for SendJob<E, C, T, B> { impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> Clone for SendJob<E, C, T, B> {
@@ -37,52 +37,21 @@ impl <E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> SendJob<E, C
buffer: Vec<u8>, buffer: Vec<u8>,
counter: u64, counter: u64,
keypair: Arc<KeyPair>, keypair: Arc<KeyPair>,
peer: Peer<E, C, T, B> peer: Peer<E, C, T, B>,
) -> SendJob<E, C, T, B> { ) -> SendJob<E, C, T, B> {
SendJob(Arc::new(Inner { SendJob(Arc::new(Inner {
buffer: Mutex::new(buffer), buffer: Mutex::new(buffer),
counter, counter,
keypair, keypair,
peer, peer,
ready: AtomicBool::new(false) ready: AtomicBool::new(false),
})) }))
} }
} }
impl <E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> SequentialJob for SendJob<E, C, T, B> { impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> ParallelJob
for SendJob<E, C, T, B>
fn is_ready(&self) -> bool { {
self.0.ready.load(Ordering::Acquire)
}
fn sequential_work(self) {
debug_assert_eq!(
self.is_ready(),
true,
"doing sequential work
on an incomplete job"
);
log::trace!("processing sequential send job");
// send to peer
let job = &self.0;
let msg = job.buffer.lock();
let xmit = job.peer.send_raw(&msg[..]).is_ok();
// trigger callback (for timers)
C::send(
&job.peer.opaque,
msg.len(),
xmit,
&job.keypair,
job.counter,
);
}
}
impl <E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> ParallelJob for SendJob<E, C, T, B> {
fn queue(&self) -> &Queue<Self> { fn queue(&self) -> &Queue<Self> {
&self.0.peer.outbound &self.0.peer.outbound
} }
@@ -141,3 +110,29 @@ impl <E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> ParallelJob
self.0.ready.store(true, Ordering::Release); self.0.ready.store(true, Ordering::Release);
} }
} }
impl<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> SequentialJob
for SendJob<E, C, T, B>
{
fn is_ready(&self) -> bool {
self.0.ready.load(Ordering::Acquire)
}
fn sequential_work(self) {
debug_assert_eq!(
self.is_ready(),
true,
"doing sequential work
on an incomplete job"
);
log::trace!("processing sequential send job");
// send to peer
let job = &self.0;
let msg = job.buffer.lock();
let xmit = job.peer.send_raw(&msg[..]).is_ok();
// trigger callback (for timers)
C::send(&job.peer.opaque, msg.len(), xmit, &job.keypair, job.counter);
}
}

View File

@@ -1,113 +1,134 @@
use std::net::IpAddr;
use std::sync::atomic::Ordering;
use std::sync::Arc;
use std::sync::Mutex;
use std::thread;
use std::time::Duration;
use num_cpus;
use super::super::dummy;
use super::super::dummy_keypair;
use super::super::tests::make_packet;
use super::super::udp::*;
use super::KeyPair; use super::KeyPair;
use super::SIZE_MESSAGE_PREFIX; use super::SIZE_MESSAGE_PREFIX;
use super::{Callbacks, Device}; use super::{Callbacks, Device};
extern crate test; use super::SIZE_KEEPALIVE;
const SIZE_KEEPALIVE: usize = 32; use super::super::dummy;
use super::super::dummy_keypair;
use super::super::tests::make_packet;
#[cfg(test)] use crate::platform::udp::Reader;
mod tests {
use super::*; use std::net::IpAddr;
use env_logger; use std::ops::Deref;
use log::debug;
use std::sync::atomic::AtomicUsize; use std::sync::atomic::AtomicUsize;
use std::sync::atomic::Ordering;
use std::sync::mpsc::{channel, Receiver, RecvTimeoutError, Sender};
use std::sync::Arc;
use std::sync::Mutex;
use std::time::Duration;
use env_logger;
use num_cpus;
use test::Bencher; use test::Bencher;
extern crate test;
const SIZE_MSG: usize = 1024;
const TIMEOUT: Duration = Duration::from_millis(1000);
struct EventTracker<E> {
rx: Mutex<Receiver<E>>,
tx: Mutex<Sender<E>>,
}
impl<E> EventTracker<E> {
fn new() -> Self {
let (tx, rx) = channel();
EventTracker {
rx: Mutex::new(rx),
tx: Mutex::new(tx),
}
}
fn log(&self, e: E) {
self.tx.lock().unwrap().send(e).unwrap();
}
fn wait(&self, timeout: Duration) -> Option<E> {
match self.rx.lock().unwrap().recv_timeout(timeout) {
Ok(v) => Some(v),
Err(RecvTimeoutError::Timeout) => None,
Err(RecvTimeoutError::Disconnected) => panic!("Disconnect"),
}
}
fn now(&self) -> Option<E> {
self.wait(Duration::from_millis(0))
}
}
// type for tracking events inside the router module // type for tracking events inside the router module
struct Flags { struct Inner {
send: Mutex<Vec<(usize, bool)>>, send: EventTracker<(usize, bool)>,
recv: Mutex<Vec<(usize, bool)>>, recv: EventTracker<(usize, bool)>,
need_key: Mutex<Vec<()>>, need_key: EventTracker<()>,
key_confirmed: Mutex<Vec<()>>, key_confirmed: EventTracker<()>,
} }
#[derive(Clone)] #[derive(Clone)]
struct Opaque(Arc<Flags>); struct Opaque {
inner: Arc<Inner>,
}
impl Deref for Opaque {
type Target = Inner;
fn deref(&self) -> &Self::Target {
&self.inner
}
}
struct TestCallbacks(); struct TestCallbacks();
impl Opaque { impl Opaque {
fn new() -> Opaque { fn new() -> Opaque {
Opaque(Arc::new(Flags { Opaque {
send: Mutex::new(vec![]), inner: Arc::new(Inner {
recv: Mutex::new(vec![]), send: EventTracker::new(),
need_key: Mutex::new(vec![]), recv: EventTracker::new(),
key_confirmed: Mutex::new(vec![]), need_key: EventTracker::new(),
})) key_confirmed: EventTracker::new(),
}),
}
}
} }
fn reset(&self) { macro_rules! no_events {
self.0.send.lock().unwrap().clear(); ($opq:expr) => {
self.0.recv.lock().unwrap().clear(); assert_eq!($opq.send.now(), None, "unexpected send event");
self.0.need_key.lock().unwrap().clear(); assert_eq!($opq.recv.now(), None, "unexpected recv event");
self.0.key_confirmed.lock().unwrap().clear(); assert_eq!($opq.need_key.now(), None, "unexpected need_key event");
} assert_eq!(
$opq.key_confirmed.now(),
fn send(&self) -> Option<(usize, bool)> { None,
self.0.send.lock().unwrap().pop() "unexpected key_confirmed event"
} );
};
fn recv(&self) -> Option<(usize, bool)> {
self.0.recv.lock().unwrap().pop()
}
fn need_key(&self) -> Option<()> {
self.0.need_key.lock().unwrap().pop()
}
fn key_confirmed(&self) -> Option<()> {
self.0.key_confirmed.lock().unwrap().pop()
}
// has all events been accounted for by assertions?
fn is_empty(&self) -> bool {
let send = self.0.send.lock().unwrap();
let recv = self.0.recv.lock().unwrap();
let need_key = self.0.need_key.lock().unwrap();
let key_confirmed = self.0.key_confirmed.lock().unwrap();
send.is_empty() && recv.is_empty() && need_key.is_empty() & key_confirmed.is_empty()
}
} }
impl Callbacks for TestCallbacks { impl Callbacks for TestCallbacks {
type Opaque = Opaque; type Opaque = Opaque;
fn send(t: &Self::Opaque, size: usize, sent: bool, _keypair: &Arc<KeyPair>, _counter: u64) { fn send(t: &Self::Opaque, size: usize, sent: bool, _keypair: &Arc<KeyPair>, _counter: u64) {
t.0.send.lock().unwrap().push((size, sent)) t.send.log((size, sent))
} }
fn recv(t: &Self::Opaque, size: usize, sent: bool, _keypair: &Arc<KeyPair>) { fn recv(t: &Self::Opaque, size: usize, sent: bool, _keypair: &Arc<KeyPair>) {
t.0.recv.lock().unwrap().push((size, sent)) t.recv.log((size, sent))
} }
fn need_key(t: &Self::Opaque) { fn need_key(t: &Self::Opaque) {
t.0.need_key.lock().unwrap().push(()); t.need_key.log(());
} }
fn key_confirmed(t: &Self::Opaque) { fn key_confirmed(t: &Self::Opaque) {
t.0.key_confirmed.lock().unwrap().push(()); t.key_confirmed.log(());
} }
} }
// wait for scheduling
fn wait() {
thread::sleep(Duration::from_millis(15));
}
fn init() { fn init() {
let _ = env_logger::builder().is_test(true).try_init(); let _ = env_logger::builder().is_test(true).try_init();
} }
@@ -203,27 +224,41 @@ mod tests {
), ),
]; ];
for (num, (mask, len, dst, okay)) in tests.iter().enumerate() { for (mask, len, dst, okay) in tests.iter() {
let len = *len;
let okay = *okay;
println!( println!(
"Check: {} {} {}/{}", "Check: {} {} {}/{}",
dst, dst,
if *okay { "\\in" } else { "\\notin" }, if okay { "\\in" } else { "\\notin" },
mask, mask,
len len
); );
for set_key in vec![true, false] { for set_key in vec![true, false] {
debug!("index = {}, set_key = {}", num, set_key); for confirm_with_staged_packet in vec![true, false] {
let send_keepalive = (!confirm_with_staged_packet || !okay) && set_key;
let send_payload = okay && set_key;
let need_key = ((confirm_with_staged_packet && set_key) || !set_key) && okay;
println!(
" confirm_with_staged_packet = {}, send_keepalive = {}, set_key = {}",
confirm_with_staged_packet, send_keepalive, set_key
);
// add new peer // add new peer
let opaque = Opaque::new(); let opaque = Opaque::new();
let peer = router.new_peer(opaque.clone()); let peer = router.new_peer(opaque.clone());
let mask: IpAddr = mask.parse().unwrap(); let mask: IpAddr = mask.parse().unwrap();
if set_key {
// confirm using keepalive
if set_key && (!confirm_with_staged_packet) {
peer.add_keypair(dummy_keypair(true)); peer.add_keypair(dummy_keypair(true));
} }
// map subnet to peer // map subnet to peer
peer.add_allowed_ip(mask, *len); peer.add_allowed_ip(mask, len);
// create "IP packet" // create "IP packet"
let dst = dst.parse().unwrap(); let dst = dst.parse().unwrap();
@@ -231,56 +266,53 @@ mod tests {
IpAddr::V4(_) => "127.0.0.1".parse().unwrap(), IpAddr::V4(_) => "127.0.0.1".parse().unwrap(),
IpAddr::V6(_) => "::1".parse().unwrap(), IpAddr::V6(_) => "::1".parse().unwrap(),
}; };
let msg = make_packet_padded(1024, src, dst, 0); let msg = make_packet_padded(SIZE_MSG, src, dst, 0);
// cryptkey route the IP packet // crypto-key route the IP packet
let res = router.send(msg); let res = router.send(msg);
// allow some scheduling
wait();
if *okay {
// cryptkey routing succeeded
assert!(res.is_ok(), "crypt-key routing should succeed: {:?}", res);
assert_eq!( assert_eq!(
opaque.need_key().is_some(), res.is_ok(),
!set_key, okay,
"crypto-routing / destination lookup failure"
);
// confirm using staged packet
if set_key && confirm_with_staged_packet {
peer.add_keypair(dummy_keypair(true));
}
// check for key-material request
if need_key {
assert_eq!(
opaque.need_key.wait(TIMEOUT),
Some(()),
"should have requested a new key, if no encryption state was set" "should have requested a new key, if no encryption state was set"
); );
assert_eq!(
opaque.send().is_some(),
set_key,
"transmission should have been attempted"
);
assert!(
opaque.recv().is_none(),
"no messages should have been marked as received"
);
} else {
// no such cryptkey route
assert!(res.is_err(), "crypt-key routing should fail");
assert!(
opaque.need_key().is_none(),
"should not request a new-key if crypt-key routing failed"
);
assert_eq!(
opaque.send(),
if set_key {
Some((SIZE_KEEPALIVE, false))
} else {
None
},
"transmission should only happen if key was set (keepalive)",
);
assert!(
opaque.recv().is_none(),
"no messages should have been marked as received",
);
}
}
} }
println!("Test complete, drop device"); // check for keepalive
if send_keepalive {
assert_eq!(
opaque.send.wait(TIMEOUT),
Some((SIZE_KEEPALIVE, false)),
"keepalive should be sent before transport message"
);
}
// check for encryption of payload
if send_payload {
assert_eq!(
opaque.send.wait(TIMEOUT),
Some((SIZE_KEEPALIVE + SIZE_MSG, false)),
"message buffer should be encrypted"
)
}
// check that we handled all events
no_events!(opaque);
}
}
}
} }
#[test] #[test]
@@ -326,11 +358,22 @@ mod tests {
), ),
]; ];
for stage in vec![true, false] {
for (p1, p2) in tests.iter() { for (p1, p2) in tests.iter() {
for confirm_with_staged_packet in vec![true, false] {
println!(
"peer1 = {:?}, peer2 = {:?}, confirm_with_staged_packet = {}",
p1, p2, confirm_with_staged_packet
);
let ((bind_reader1, bind_writer1), (bind_reader2, bind_writer2)) = let ((bind_reader1, bind_writer1), (bind_reader2, bind_writer2)) =
dummy::PairBind::pair(); dummy::PairBind::pair();
let confirm_packet_size = if confirm_with_staged_packet {
SIZE_KEEPALIVE + SIZE_MSG
} else {
SIZE_KEEPALIVE
};
// create matching device // create matching device
let (_fake, _, tun_writer1, _) = dummy::TunTest::create(false); let (_fake, _, tun_writer1, _) = dummy::TunTest::create(false);
let (_fake, _, tun_writer2, _) = dummy::TunTest::create(false); let (_fake, _, tun_writer2, _) = dummy::TunTest::create(false);
@@ -365,14 +408,12 @@ mod tests {
peer2.set_endpoint(dummy::UnitEndpoint::new()); peer2.set_endpoint(dummy::UnitEndpoint::new());
} }
if stage { if confirm_with_staged_packet {
println!("confirm using staged packet");
// create IP packet // create IP packet
let (_mask, _len, ip1, _okay) = p1; let (_mask, _len, ip1, _okay) = p1;
let (_mask, _len, ip2, _okay) = p2; let (_mask, _len, ip2, _okay) = p2;
let msg = make_packet_padded( let msg = make_packet_padded(
1024, SIZE_MSG,
ip1.parse().unwrap(), // src ip1.parse().unwrap(), // src
ip2.parse().unwrap(), // dst ip2.parse().unwrap(), // dst
0, 0,
@@ -380,79 +421,107 @@ mod tests {
// stage packet for sending // stage packet for sending
router2.send(msg).expect("failed to sent staged packet"); router2.send(msg).expect("failed to sent staged packet");
wait();
// validate events // a new key should have been requested from the handshake machine
assert!(opaque2.recv().is_none()); assert_eq!(
assert!( opaque2.need_key.wait(TIMEOUT),
opaque2.send().is_none(), Some(()),
"sending should fail as not key is set"
);
assert!(
opaque2.need_key().is_some(),
"a new key should be requested since a packet was attempted transmitted" "a new key should be requested since a packet was attempted transmitted"
); );
assert!(opaque2.is_empty(), "callbacks should only run once");
no_events!(opaque1);
no_events!(opaque2);
} }
// this should cause a key-confirmation packet (keepalive or staged packet) // add a keypair
// this also causes peer1 to learn the "endpoint" for peer2 assert_eq!(peer1.get_endpoint(), None, "no endpoint has yet been set");
assert!(peer1.get_endpoint().is_none());
peer2.add_keypair(dummy_keypair(true)); peer2.add_keypair(dummy_keypair(true));
wait(); // this should cause a key-confirmation packet (keepalive or staged packet)
assert!(opaque2.send().is_some()); assert_eq!(
assert!(opaque2.is_empty(), "events on peer2 should be 'send'"); opaque2.send.wait(TIMEOUT),
assert!(opaque1.is_empty(), "nothing should happened on peer1"); Some((confirm_packet_size, true)),
"expected successful transmission of a confirmation packet"
);
// no other events should fire
no_events!(opaque1);
no_events!(opaque2);
// read confirming message received by the other end ("across the internet") // read confirming message received by the other end ("across the internet")
let mut buf = vec![0u8; 2048]; let mut buf = vec![0u8; SIZE_MSG * 2];
let (len, from) = bind_reader1.read(&mut buf).unwrap(); let (len, from) = bind_reader1.read(&mut buf).unwrap();
buf.truncate(len); buf.truncate(len);
router1.recv(from, buf).unwrap();
wait(); assert_eq!(
assert!(opaque1.recv().is_some()); len,
assert!(opaque1.key_confirmed().is_some()); if confirm_with_staged_packet {
assert!( SIZE_MSG + SIZE_KEEPALIVE
opaque1.is_empty(), } else {
"events on peer1 should be 'recv' and 'key_confirmed'" SIZE_KEEPALIVE
},
"unexpected size of confirmation message"
); );
assert!(peer1.get_endpoint().is_some());
assert!(opaque2.is_empty(), "nothing should happened on peer2");
// pass to the router for processing
router1
.recv(from, buf)
.expect("failed to receive confirmation message");
// check that a receive event is fired
assert_eq!(
opaque1.recv.wait(TIMEOUT),
Some((confirm_packet_size, true)),
"we expect processing to be successful"
);
// the key is confirmed
assert_eq!(
opaque1.key_confirmed.wait(TIMEOUT),
Some(()),
"confirmation message should confirm the key"
);
// peer1 learns the endpoint
assert!(
peer1.get_endpoint().is_some(),
"peer1 should learn the endpoint of peer2 from the confirmation message (roaming)"
);
// no other events should fire
no_events!(opaque1);
no_events!(opaque2);
// now that peer1 has an endpoint // now that peer1 has an endpoint
// route packets : peer1 -> peer2 // route packets in the other direction: peer1 -> peer2
for id in 1..11 { for id in 1..11 {
println!("round: {}", id); println!("packet: {}", id);
assert!(
opaque1.is_empty(), let message_size = 1024;
"we should have asserted a value for every callback on peer1"
);
assert!(
opaque2.is_empty(),
"we should have asserted a value for every callback on peer2"
);
// pass IP packet to router // pass IP packet to router
let (_mask, _len, ip1, _okay) = p1; let (_mask, _len, ip1, _okay) = p1;
let (_mask, _len, ip2, _okay) = p2; let (_mask, _len, ip2, _okay) = p2;
let msg = make_packet_padded( let msg = make_packet_padded(
1024, message_size,
ip2.parse().unwrap(), // src ip2.parse().unwrap(), // src
ip1.parse().unwrap(), // dst ip1.parse().unwrap(), // dst
id, id,
); );
router1.send(msg).unwrap();
wait(); router1
assert!(opaque1.send().is_some(), "encryption should succeed"); .send(msg)
assert!( .expect("we expect routing to be successful");
opaque1.recv().is_none(),
"receiving callback should not be called" // encryption succeeds and the correct size is logged
assert_eq!(
opaque1.send.wait(TIMEOUT),
Some((message_size + SIZE_KEEPALIVE, true)),
"expected send event for peer1 -> peer2 payload"
); );
assert!(opaque1.need_key().is_none());
// otherwise no events
no_events!(opaque1);
no_events!(opaque2);
// receive ("across the internet") on the other end // receive ("across the internet") on the other end
let mut buf = vec![0u8; 2048]; let mut buf = vec![0u8; 2048];
@@ -460,17 +529,16 @@ mod tests {
buf.truncate(len); buf.truncate(len);
router2.recv(from, buf).unwrap(); router2.recv(from, buf).unwrap();
wait(); // check that decryption succeeds
assert!( assert_eq!(
opaque2.send().is_none(), opaque2.recv.wait(TIMEOUT),
"sending callback should not be called" Some((message_size + SIZE_KEEPALIVE, true)),
);
assert!(
opaque2.recv().is_some(),
"decryption and routing should succeed" "decryption and routing should succeed"
); );
assert!(opaque2.need_key().is_none());
} // otherwise no events
no_events!(opaque1);
no_events!(opaque2);
} }
} }
} }

View File

@@ -6,6 +6,7 @@ use super::receive::ReceiveJob;
use super::send::SendJob; use super::send::SendJob;
use crossbeam_channel::Receiver; use crossbeam_channel::Receiver;
use log;
pub enum JobUnion<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> { pub enum JobUnion<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>> {
Outbound(SendJob<E, C, T, B>), Outbound(SendJob<E, C, T, B>),
@@ -16,8 +17,12 @@ pub fn worker<E: Endpoint, C: Callbacks, T: tun::Writer, B: udp::Writer<E>>(
receiver: Receiver<JobUnion<E, C, T, B>>, receiver: Receiver<JobUnion<E, C, T, B>>,
) { ) {
loop { loop {
log::trace!("pool worker awaiting job");
match receiver.recv() { match receiver.recv() {
Err(_) => break, Err(e) => {
log::debug!("worker stopped with {}", e);
break;
}
Ok(JobUnion::Inbound(job)) => { Ok(JobUnion::Inbound(job)) => {
job.parallel_work(); job.parallel_work();
job.queue().consume(); job.queue().consume();