Reset peer's receive timeout when a message is received

* Make advance_clock more realistic by waking timers in order,
  instead of all at once.
* Don't advance the clock when simulating random delays.

Co-Authored-By: Keith Simmons <keith@zed.dev>
Co-Authored-By: Nathan Sobo <nathan@zed.dev>
This commit is contained in:
Max Brunsfeld 2022-03-07 15:33:39 -08:00
parent 120f7bbc3d
commit 1f5eab39a9
5 changed files with 67 additions and 32 deletions

View File

@ -966,8 +966,6 @@ mod tests {
server.roll_access_token(); server.roll_access_token();
server.allow_connections(); server.allow_connections();
cx.foreground().advance_clock(Duration::from_secs(10)); cx.foreground().advance_clock(Duration::from_secs(10));
assert_eq!(server.auth_count(), 1);
cx.foreground().advance_clock(Duration::from_secs(10));
while !matches!(status.next().await, Some(Status::Connected { .. })) {} while !matches!(status.next().await, Some(Status::Connected { .. })) {}
assert_eq!(server.auth_count(), 2); // Client re-authenticated due to an invalid token assert_eq!(server.auth_count(), 2); // Client re-authenticated due to an invalid token
} }

View File

@ -330,14 +330,34 @@ impl Deterministic {
} }
pub fn advance_clock(&self, duration: Duration) { pub fn advance_clock(&self, duration: Duration) {
let new_now = self.state.lock().now + duration;
loop {
self.run_until_parked();
let mut state = self.state.lock(); let mut state = self.state.lock();
state.now += duration;
let now = state.now;
let mut pending_timers = mem::take(&mut state.pending_timers);
drop(state);
pending_timers.retain(|(_, wakeup, _)| *wakeup > now); if let Some((_, wakeup_time, _)) = state.pending_timers.first() {
self.state.lock().pending_timers.extend(pending_timers); let wakeup_time = *wakeup_time;
if wakeup_time < new_now {
let timer_count = state
.pending_timers
.iter()
.take_while(|(_, t, _)| *t == wakeup_time)
.count();
state.now = wakeup_time;
let timers_to_wake = state
.pending_timers
.drain(0..timer_count)
.collect::<Vec<_>>();
drop(state);
drop(timers_to_wake);
continue;
}
}
break;
}
self.state.lock().now = new_now;
} }
} }
@ -640,9 +660,6 @@ impl Background {
for _ in 0..yields { for _ in 0..yields {
yield_now().await; yield_now().await;
} }
let delay = Duration::from_millis(executor.state.lock().rng.gen_range(0..100));
executor.advance_clock(delay);
} }
} }
_ => panic!("this method can only be called on a deterministic executor"), _ => panic!("this method can only be called on a deterministic executor"),

View File

@ -88,13 +88,14 @@ pub struct Peer {
#[derive(Clone)] #[derive(Clone)]
pub struct ConnectionState { pub struct ConnectionState {
outgoing_tx: futures::channel::mpsc::UnboundedSender<proto::Envelope>, outgoing_tx: futures::channel::mpsc::UnboundedSender<proto::Message>,
next_message_id: Arc<AtomicU32>, next_message_id: Arc<AtomicU32>,
response_channels: response_channels:
Arc<Mutex<Option<HashMap<u32, oneshot::Sender<(proto::Envelope, barrier::Sender)>>>>>, Arc<Mutex<Option<HashMap<u32, oneshot::Sender<(proto::Envelope, barrier::Sender)>>>>>,
} }
const KEEPALIVE_INTERVAL: Duration = Duration::from_secs(1); const KEEPALIVE_INTERVAL: Duration = Duration::from_secs(1);
const WRITE_TIMEOUT: Duration = Duration::from_secs(2);
impl Peer { impl Peer {
pub fn new() -> Arc<Self> { pub fn new() -> Arc<Self> {
@ -142,19 +143,25 @@ impl Peer {
this.connections.write().remove(&connection_id); this.connections.write().remove(&connection_id);
}); });
// Send messages on this frequency so the connection isn't closed.
let keepalive_timer = create_timer(KEEPALIVE_INTERVAL).fuse();
futures::pin_mut!(keepalive_timer);
loop { loop {
let read_message = reader.read().fuse(); let read_message = reader.read().fuse();
futures::pin_mut!(read_message); futures::pin_mut!(read_message);
let read_timeout = create_timer(2 * KEEPALIVE_INTERVAL).fuse();
futures::pin_mut!(read_timeout); // Disconnect if we don't receive messages at least this frequently.
let receive_timeout = create_timer(3 * KEEPALIVE_INTERVAL).fuse();
futures::pin_mut!(receive_timeout);
loop { loop {
futures::select_biased! { futures::select_biased! {
outgoing = outgoing_rx.next().fuse() => match outgoing { outgoing = outgoing_rx.next().fuse() => match outgoing {
Some(outgoing) => { Some(outgoing) => {
let outgoing = proto::Message::Envelope(outgoing); if let Some(result) = writer.write(outgoing).timeout(WRITE_TIMEOUT).await {
if let Some(result) = writer.write(outgoing).timeout(2 * KEEPALIVE_INTERVAL).await {
result.context("failed to write RPC message")?; result.context("failed to write RPC message")?;
keepalive_timer.set(create_timer(KEEPALIVE_INTERVAL).fuse());
} else { } else {
Err(anyhow!("timed out writing message"))?; Err(anyhow!("timed out writing message"))?;
} }
@ -168,18 +175,18 @@ impl Peer {
return Ok(()); return Ok(());
} }
} }
break; break;
}, },
_ = create_timer(KEEPALIVE_INTERVAL).fuse() => { _ = keepalive_timer => {
if let Some(result) = writer.write(proto::Message::Ping).timeout(2 * KEEPALIVE_INTERVAL).await { if let Some(result) = writer.write(proto::Message::Ping).timeout(WRITE_TIMEOUT).await {
result.context("failed to send websocket ping")?; result.context("failed to send keepalive")?;
keepalive_timer.set(create_timer(KEEPALIVE_INTERVAL).fuse());
} else { } else {
Err(anyhow!("timed out sending websocket ping"))?; Err(anyhow!("timed out sending keepalive"))?;
} }
} }
_ = read_timeout => { _ = receive_timeout => {
Err(anyhow!("timed out reading message"))? Err(anyhow!("delay between messages too long"))?
} }
} }
} }
@ -278,11 +285,11 @@ impl Peer {
.insert(message_id, tx); .insert(message_id, tx);
connection connection
.outgoing_tx .outgoing_tx
.unbounded_send(request.into_envelope( .unbounded_send(proto::Message::Envelope(request.into_envelope(
message_id, message_id,
None, None,
original_sender_id.map(|id| id.0), original_sender_id.map(|id| id.0),
)) )))
.map_err(|_| anyhow!("connection was closed"))?; .map_err(|_| anyhow!("connection was closed"))?;
Ok(()) Ok(())
}); });
@ -305,7 +312,9 @@ impl Peer {
.fetch_add(1, atomic::Ordering::SeqCst); .fetch_add(1, atomic::Ordering::SeqCst);
connection connection
.outgoing_tx .outgoing_tx
.unbounded_send(message.into_envelope(message_id, None, None))?; .unbounded_send(proto::Message::Envelope(
message.into_envelope(message_id, None, None),
))?;
Ok(()) Ok(())
} }
@ -321,7 +330,11 @@ impl Peer {
.fetch_add(1, atomic::Ordering::SeqCst); .fetch_add(1, atomic::Ordering::SeqCst);
connection connection
.outgoing_tx .outgoing_tx
.unbounded_send(message.into_envelope(message_id, None, Some(sender_id.0)))?; .unbounded_send(proto::Message::Envelope(message.into_envelope(
message_id,
None,
Some(sender_id.0),
)))?;
Ok(()) Ok(())
} }
@ -336,7 +349,11 @@ impl Peer {
.fetch_add(1, atomic::Ordering::SeqCst); .fetch_add(1, atomic::Ordering::SeqCst);
connection connection
.outgoing_tx .outgoing_tx
.unbounded_send(response.into_envelope(message_id, Some(receipt.message_id), None))?; .unbounded_send(proto::Message::Envelope(response.into_envelope(
message_id,
Some(receipt.message_id),
None,
)))?;
Ok(()) Ok(())
} }
@ -351,7 +368,11 @@ impl Peer {
.fetch_add(1, atomic::Ordering::SeqCst); .fetch_add(1, atomic::Ordering::SeqCst);
connection connection
.outgoing_tx .outgoing_tx
.unbounded_send(response.into_envelope(message_id, Some(receipt.message_id), None))?; .unbounded_send(proto::Message::Envelope(response.into_envelope(
message_id,
Some(receipt.message_id),
None,
)))?;
Ok(()) Ok(())
} }

View File

@ -283,6 +283,7 @@ pub struct MessageStream<S> {
encoding_buffer: Vec<u8>, encoding_buffer: Vec<u8>,
} }
#[derive(Debug)]
pub enum Message { pub enum Message {
Envelope(Envelope), Envelope(Envelope),
Ping, Ping,

View File

@ -2732,8 +2732,6 @@ mod tests {
.read_with(cx_a, |tree, _| tree.as_local().unwrap().scan_complete()) .read_with(cx_a, |tree, _| tree.as_local().unwrap().scan_complete())
.await; .await;
eprintln!("sharing");
project_a.update(cx_a, |p, cx| p.share(cx)).await.unwrap(); project_a.update(cx_a, |p, cx| p.share(cx)).await.unwrap();
// Join the worktree as client B. // Join the worktree as client B.