fix upx and udp conn counter (#131)

* fix upx in workflow
* fix udp conn counter
This commit is contained in:
Sijie.Sun
2024-06-04 18:50:30 +08:00
committed by GitHub
parent df17a7bb68
commit f9e6264f31
10 changed files with 158 additions and 136 deletions
+1 -1
View File
@@ -121,7 +121,7 @@ jobs:
TAG=$GITHUB_SHA TAG=$GITHUB_SHA
fi fi
if [[ $OS =~ ^windows.*$ ]]; then if [[ ! $OS =~ ^macos.*$ ]]; then
upx --lzma --best ./target/$TARGET/release/easytier-core"$SUFFIX" upx --lzma --best ./target/$TARGET/release/easytier-core"$SUFFIX"
upx --lzma --best ./target/$TARGET/release/easytier-cli"$SUFFIX" upx --lzma --best ./target/$TARGET/release/easytier-cli"$SUFFIX"
fi fi
+1 -1
View File
@@ -1008,8 +1008,8 @@ pub mod tests {
use tokio::net::UdpSocket; use tokio::net::UdpSocket;
use crate::connector::udp_hole_punch::UdpHolePunchListener; use crate::connector::udp_hole_punch::UdpHolePunchListener;
use crate::peers::tests::wait_for_condition;
use crate::rpc::{NatType, StunInfo}; use crate::rpc::{NatType, StunInfo};
use crate::tunnel::common::tests::wait_for_condition;
use crate::{ use crate::{
common::{error::Error, stun::StunInfoCollectorTrait}, common::{error::Error, stun::StunInfoCollectorTrait},
+2 -3
View File
@@ -352,9 +352,8 @@ impl PeerCenterInstance {
mod tests { mod tests {
use crate::{ use crate::{
peer_center::server::get_global_data, peer_center::server::get_global_data,
peers::tests::{ peers::tests::{connect_peer_manager, create_mock_peer_manager, wait_route_appear},
connect_peer_manager, create_mock_peer_manager, wait_for_condition, wait_route_appear, tunnel::common::tests::wait_for_condition,
},
}; };
use super::*; use super::*;
+2 -1
View File
@@ -758,9 +758,10 @@ mod tests {
peers::{ peers::{
peer_manager::RouteAlgoType, peer_manager::RouteAlgoType,
peer_rpc::tests::{MockService, TestRpcService, TestRpcServiceClient}, peer_rpc::tests::{MockService, TestRpcService, TestRpcServiceClient},
tests::{connect_peer_manager, wait_for_condition, wait_route_appear}, tests::{connect_peer_manager, wait_route_appear},
}, },
rpc::NatType, rpc::NatType,
tunnel::common::tests::wait_for_condition,
tunnel::{TunnelConnector, TunnelListener}, tunnel::{TunnelConnector, TunnelListener},
}; };
+2 -1
View File
@@ -1467,9 +1467,10 @@ mod tests {
peers::{ peers::{
peer_manager::{PeerManager, RouteAlgoType}, peer_manager::{PeerManager, RouteAlgoType},
route_trait::{NextHopPolicy, Route, RouteCostCalculatorInterface}, route_trait::{NextHopPolicy, Route, RouteCostCalculatorInterface},
tests::{connect_peer_manager, wait_for_condition}, tests::connect_peer_manager,
}, },
rpc::NatType, rpc::NatType,
tunnel::common::tests::wait_for_condition,
}; };
use super::PeerRoute; use super::PeerRoute;
+3 -6
View File
@@ -557,14 +557,11 @@ pub mod tests {
common::{error::Error, new_peer_id, PeerId}, common::{error::Error, new_peer_id, PeerId},
peers::{ peers::{
peer_rpc::PeerRpcManager, peer_rpc::PeerRpcManager,
tests::{ tests::{connect_peer_manager, create_mock_peer_manager, wait_route_appear},
connect_peer_manager, create_mock_peer_manager, wait_for_condition,
wait_route_appear,
},
}, },
tunnel::{ tunnel::{
packet_def::ZCPacket, ring::create_ring_tunnel_pair, Tunnel, ZCPacketSink, common::tests::wait_for_condition, packet_def::ZCPacket, ring::create_ring_tunnel_pair,
ZCPacketStream, Tunnel, ZCPacketSink, ZCPacketStream,
}, },
}; };
-17
View File
@@ -1,7 +1,5 @@
use std::sync::Arc; use std::sync::Arc;
use futures::Future;
use crate::{ use crate::{
common::{error::Error, global_ctx::tests::get_mock_global_ctx, PeerId}, common::{error::Error, global_ctx::tests::get_mock_global_ctx, PeerId},
tunnel::ring::create_ring_tunnel_pair, tunnel::ring::create_ring_tunnel_pair,
@@ -58,18 +56,3 @@ pub async fn wait_route_appear(
wait_route_appear_with_cost(peer_mgr.clone(), target_peer.my_peer_id(), None).await?; wait_route_appear_with_cost(peer_mgr.clone(), target_peer.my_peer_id(), None).await?;
wait_route_appear_with_cost(target_peer, peer_mgr.my_peer_id(), None).await wait_route_appear_with_cost(target_peer, peer_mgr.my_peer_id(), None).await
} }
pub async fn wait_for_condition<F, FRet>(mut condition: F, timeout: std::time::Duration) -> ()
where
F: FnMut() -> FRet + Send,
FRet: Future<Output = bool>,
{
let now = std::time::Instant::now();
while now.elapsed() < timeout {
if condition().await {
return;
}
tokio::time::sleep(std::time::Duration::from_millis(50)).await;
}
assert!(condition().await, "Timeout")
}
+1 -1
View File
@@ -13,7 +13,7 @@ use crate::{
netns::{NetNS, ROOT_NETNS_NAME}, netns::{NetNS, ROOT_NETNS_NAME},
}, },
instance::instance::Instance, instance::instance::Instance,
peers::tests::wait_for_condition, tunnel::common::tests::wait_for_condition,
tunnel::{ring::RingTunnelConnector, tcp::TcpTunnelConnector, udp::UdpTunnelConnector}, tunnel::{ring::RingTunnelConnector, tcp::TcpTunnelConnector, udp::UdpTunnelConnector},
}; };
+16 -1
View File
@@ -419,7 +419,7 @@ pub fn reserve_buf(buf: &mut BytesMut, min_size: usize, max_size: usize) {
pub mod tests { pub mod tests {
use std::time::Instant; use std::time::Instant;
use futures::{SinkExt, StreamExt, TryStreamExt}; use futures::{Future, SinkExt, StreamExt, TryStreamExt};
use tokio_util::bytes::{BufMut, Bytes, BytesMut}; use tokio_util::bytes::{BufMut, Bytes, BytesMut};
use crate::{ use crate::{
@@ -595,4 +595,19 @@ pub mod tests {
.with_env_filter(filter) .with_env_filter(filter)
.init(); .init();
} }
pub async fn wait_for_condition<F, FRet>(mut condition: F, timeout: std::time::Duration) -> ()
where
F: FnMut() -> FRet + Send,
FRet: Future<Output = bool>,
{
let now = std::time::Instant::now();
while now.elapsed() < timeout {
if condition().await {
return;
}
tokio::time::sleep(std::time::Duration::from_millis(50)).await;
}
assert!(condition().await, "Timeout")
}
} }
+130 -104
View File
@@ -35,8 +35,8 @@ use super::{
pub const UDP_DATA_MTU: usize = 2000; pub const UDP_DATA_MTU: usize = 2000;
type UdpCloseEventSender = UnboundedSender<Option<TunnelError>>; type UdpCloseEventSender = UnboundedSender<(SocketAddr, Option<TunnelError>)>;
type UdpCloseEventReceiver = UnboundedReceiver<Option<TunnelError>>; type UdpCloseEventReceiver = UnboundedReceiver<(SocketAddr, Option<TunnelError>)>;
fn new_udp_packet<F>(f: F, udp_body: Option<&mut [u8]>) -> ZCPacket fn new_udp_packet<F>(f: F, udp_body: Option<&mut [u8]>) -> ZCPacket
where where
@@ -151,6 +151,33 @@ async fn forward_from_ring_to_udp(
} }
} }
async fn udp_recv_from_socket_forward_task<F>(socket: Arc<UdpSocket>, f: F)
where
F: Fn(ZCPacket, SocketAddr) -> (),
{
let mut buf = BytesMut::new();
loop {
reserve_buf(&mut buf, UDP_DATA_MTU, UDP_DATA_MTU * 16);
let (dg_size, addr) = socket.recv_buf_from(&mut buf).await.unwrap();
tracing::trace!(
"udp recv packet: {:?}, buf: {:?}, size: {}",
addr,
buf,
dg_size
);
let zc_packet = match get_zcpacket_from_buf(buf.split()) {
Ok(v) => v,
Err(e) => {
tracing::warn!(?e, "udp get zc packet from buf error");
continue;
}
};
f(zc_packet, addr);
}
}
struct UdpConnection { struct UdpConnection {
socket: Arc<UdpSocket>, socket: Arc<UdpSocket>,
conn_id: u32, conn_id: u32,
@@ -173,7 +200,7 @@ impl UdpConnection {
let forward_task = tokio::spawn(async move { let forward_task = tokio::spawn(async move {
let close_event_sender = close_event_sender; let close_event_sender = close_event_sender;
let err = forward_from_ring_to_udp(ring_recv, &s, &dst_addr, conn_id).await; let err = forward_from_ring_to_udp(ring_recv, &s, &dst_addr, conn_id).await;
if let Err(e) = close_event_sender.send(err) { if let Err(e) = close_event_sender.send((dst_addr, err)) {
tracing::error!(?e, "udp send close event error"); tracing::error!(?e, "udp send close event error");
} }
}); });
@@ -186,6 +213,27 @@ impl UdpConnection {
forward_task, forward_task,
} }
} }
pub fn handle_packet_from_remote(&self, zc_packet: ZCPacket) -> Result<(), TunnelError> {
let header = zc_packet.udp_tunnel_header().unwrap();
let conn_id = header.conn_id.get();
if header.msg_type != UdpPacketType::Data as u8 {
return Err(TunnelError::InvalidPacket("not data packet".to_owned()));
}
if self.conn_id != conn_id {
return Err(TunnelError::ConnIdNotMatch(self.conn_id, conn_id));
}
if !self.ring_sender.has_empty_slot() {
return Err(TunnelError::BufferFull);
}
self.ring_sender.push_no_check(zc_packet)?;
Ok(())
}
} }
impl Drop for UdpConnection { impl Drop for UdpConnection {
@@ -275,40 +323,16 @@ impl UdpTunnelListenerData {
} }
} }
async fn try_forward_packet( fn do_forward_one_packet_to_conn(&self, zc_packet: ZCPacket, addr: SocketAddr) {
self: &Self,
remote_addr: &SocketAddr,
conn_id: u32,
p: ZCPacket,
) -> Result<(), TunnelError> {
let Some(conn) = self.sock_map.get(remote_addr) else {
return Err(TunnelError::InternalError(
"udp connection not found".to_owned(),
));
};
if conn.conn_id != conn_id {
return Err(TunnelError::ConnIdNotMatch(conn.conn_id, conn_id));
}
if !conn.ring_sender.has_empty_slot() {
return Err(TunnelError::BufferFull);
}
conn.ring_sender.push_no_check(p)?;
Ok(())
}
async fn process_forward_packet(&self, zc_packet: ZCPacket, addr: &SocketAddr) {
let header = zc_packet.udp_tunnel_header().unwrap(); let header = zc_packet.udp_tunnel_header().unwrap();
if header.msg_type == UdpPacketType::Syn as u8 { if header.msg_type == UdpPacketType::Syn as u8 {
tokio::spawn(Self::handle_new_connect(self.clone(), *addr, zc_packet)); tokio::spawn(Self::handle_new_connect(self.clone(), addr, zc_packet));
} else if header.msg_type != UdpPacketType::HolePunch as u8 { } else if header.msg_type != UdpPacketType::HolePunch as u8 {
if let Err(e) = self let Some(conn) = self.sock_map.get(&addr) else {
.try_forward_packet(addr, header.conn_id.get(), zc_packet) tracing::trace!(?header, "udp forward packet error, connection not found");
.await return;
{ };
if let Err(e) = conn.handle_packet_from_remote(zc_packet) {
tracing::trace!(?e, "udp forward packet error"); tracing::trace!(?e, "udp forward packet error");
} }
} }
@@ -316,26 +340,10 @@ impl UdpTunnelListenerData {
async fn do_forward_task(self: Self) { async fn do_forward_task(self: Self) {
let socket = self.socket.as_ref().unwrap().clone(); let socket = self.socket.as_ref().unwrap().clone();
let mut buf = BytesMut::new(); udp_recv_from_socket_forward_task(socket, |zc_packet, addr| {
loop { self.do_forward_one_packet_to_conn(zc_packet, addr);
reserve_buf(&mut buf, UDP_DATA_MTU, UDP_DATA_MTU * 16); })
let (dg_size, addr) = socket.recv_buf_from(&mut buf).await.unwrap(); .await;
tracing::trace!(
"udp recv packet: {:?}, buf: {:?}, size: {}",
addr,
buf,
dg_size
);
let zc_packet = match get_zcpacket_from_buf(buf.split()) {
Ok(v) => v,
Err(e) => {
tracing::warn!(?e, "udp get zc packet from buf error");
continue;
}
};
self.process_forward_packet(zc_packet, &addr).await;
}
} }
} }
@@ -346,7 +354,7 @@ pub struct UdpTunnelListener {
conn_recv: Receiver<Box<dyn Tunnel>>, conn_recv: Receiver<Box<dyn Tunnel>>,
data: UdpTunnelListenerData, data: UdpTunnelListenerData,
forward_tasks: Arc<std::sync::Mutex<JoinSet<()>>>, forward_tasks: Arc<std::sync::Mutex<JoinSet<()>>>,
close_event_recv: UdpCloseEventReceiver, close_event_recv: Option<UdpCloseEventReceiver>,
} }
impl UdpTunnelListener { impl UdpTunnelListener {
@@ -359,7 +367,7 @@ impl UdpTunnelListener {
conn_recv, conn_recv,
data: UdpTunnelListenerData::new(addr, conn_send, close_event_send), data: UdpTunnelListenerData::new(addr, conn_send, close_event_send),
forward_tasks: Arc::new(std::sync::Mutex::new(JoinSet::new())), forward_tasks: Arc::new(std::sync::Mutex::new(JoinSet::new())),
close_event_recv, close_event_recv: Some(close_event_recv),
} }
} }
@@ -398,6 +406,17 @@ impl TunnelListener for UdpTunnelListener {
.unwrap() .unwrap()
.spawn(self.data.clone().do_forward_task()); .spawn(self.data.clone().do_forward_task());
let sock_map = Arc::downgrade(&self.data.sock_map.clone());
let mut close_recv = self.close_event_recv.take().unwrap();
self.forward_tasks.lock().unwrap().spawn(async move {
while let Some((dst_addr, err)) = close_recv.recv().await {
if let Some(err) = err {
tracing::error!(?err, "udp close event error");
}
sock_map.upgrade().map(|v| v.remove(&dst_addr));
}
});
join_joinset_background(self.forward_tasks.clone(), "UdpTunnelListener".to_owned()); join_joinset_background(self.forward_tasks.clone(), "UdpTunnelListener".to_owned());
Ok(()) Ok(())
@@ -538,62 +557,44 @@ impl UdpTunnelConnector {
"udp build tunnel for connector" "udp build tunnel for connector"
); );
let (close_event_send, mut close_event_recv) = tokio::sync::mpsc::unbounded_channel(); let (close_event_sender, mut close_event_recv) = tokio::sync::mpsc::unbounded_channel();
// forward from ring to udp
let socket_sender = socket.clone();
let ring_recv = RingStream::new(ring_for_send_udp.clone()); let ring_recv = RingStream::new(ring_for_send_udp.clone());
tokio::spawn(async move {
let err = forward_from_ring_to_udp(ring_recv, &socket_sender, &dst_addr, conn_id).await;
tracing::debug!(?err, "udp forward from ring to udp done");
close_event_send.send(err).unwrap();
});
let socket_recv = socket.clone();
let ring_sender = RingSink::new(ring_for_recv_udp.clone()); let ring_sender = RingSink::new(ring_for_recv_udp.clone());
tokio::spawn(async move { let udp_conn = UdpConnection::new(
let mut buf = BytesMut::new(); socket.clone(),
loop { conn_id,
reserve_buf(&mut buf, UDP_DATA_MTU, UDP_DATA_MTU * 16); dst_addr,
let ret; ring_sender,
ring_recv,
close_event_sender,
);
let socket_clone = socket.clone();
tokio::spawn(
async move {
tokio::select! { tokio::select! {
_ = close_event_recv.recv() => { _ = close_event_recv.recv() => {
tracing::debug!("connector udp close event"); tracing::debug!("connector udp close event");
break; return;
} }
recv_res = socket_recv.recv_buf_from(&mut buf) => ret = Some(recv_res.unwrap()), _ = udp_recv_from_socket_forward_task(socket_clone, |zc_packet, addr| {
} tracing::debug!(?addr, "connector udp forward task done");
let (dg_size, addr) = ret.unwrap(); if let Err(e) = udp_conn.handle_packet_from_remote(zc_packet) {
tracing::trace!( tracing::trace!(?e, ?addr, "udp forward packet error");
"connector udp recv packet: {:?}, buf: {:?}, size: {}", }
addr, }) => {
buf, tracing::debug!("connector udp forward task done");
dg_size return;
);
let zc_packet = match get_zcpacket_from_buf(buf.split()) {
Ok(v) => v,
Err(e) => {
tracing::warn!(?e, "connector udp get zc packet from buf error");
continue;
}
};
let header = zc_packet.udp_tunnel_header().unwrap();
if header.conn_id.get() != conn_id {
tracing::trace!(
"connector udp conn id not match: {:?}, {:?}",
header.conn_id.get(),
conn_id
);
}
if header.msg_type == UdpPacketType::Data as u8 {
if let Err(e) = ring_sender.push_no_check(zc_packet) {
tracing::trace!(?e, "udp forward packet error");
} }
} }
} }
}.instrument(tracing::info_span!("udp connector forward from udp to ring", ?ring_for_recv_udp))); .instrument(tracing::info_span!(
"udp forward from udp to ring",
?conn_id,
?dst_addr,
)),
);
Ok(Box::new(TunnelWrapper::new( Ok(Box::new(TunnelWrapper::new(
Box::new(RingStream::new(ring_for_recv_udp)), Box::new(RingStream::new(ring_for_recv_udp)),
@@ -713,7 +714,7 @@ mod tests {
check_scheme_and_get_socket_addr, check_scheme_and_get_socket_addr,
common::{ common::{
get_interface_name_by_ip, get_interface_name_by_ip,
tests::{_tunnel_bench, _tunnel_echo_server, _tunnel_pingpong}, tests::{_tunnel_bench, _tunnel_echo_server, _tunnel_pingpong, wait_for_condition},
}, },
TunnelConnector, TunnelConnector,
}, },
@@ -723,7 +724,7 @@ mod tests {
async fn udp_pingpong() { async fn udp_pingpong() {
let listener = UdpTunnelListener::new("udp://0.0.0.0:5556".parse().unwrap()); let listener = UdpTunnelListener::new("udp://0.0.0.0:5556".parse().unwrap());
let connector = UdpTunnelConnector::new("udp://127.0.0.1:5556".parse().unwrap()); let connector = UdpTunnelConnector::new("udp://127.0.0.1:5556".parse().unwrap());
_tunnel_pingpong(listener, connector).await _tunnel_pingpong(listener, connector).await;
} }
#[tokio::test] #[tokio::test]
@@ -911,4 +912,29 @@ mod tests {
let port = listener.local_url().port().unwrap(); let port = listener.local_url().port().unwrap();
assert!(port > 0); assert!(port > 0);
} }
#[tokio::test]
async fn test_conn_counter() {
let mut listener = UdpTunnelListener::new("udp://0.0.0.0:5556".parse().unwrap());
let mut connector = UdpTunnelConnector::new("udp://127.0.0.1:5556".parse().unwrap());
tokio::spawn(async move {
tokio::time::sleep(tokio::time::Duration::from_secs(1)).await;
let _c1 = connector.connect().await.unwrap();
let _c2 = connector.connect().await.unwrap();
});
let conn_counter = listener.get_conn_counter();
listener.listen().await.unwrap();
let c1 = listener.accept().await.unwrap();
assert_eq!(conn_counter.get(), 1);
let c2 = listener.accept().await.unwrap();
assert_eq!(conn_counter.get(), 2);
drop(c2);
wait_for_condition(|| async { conn_counter.get() == 1 }, Duration::from_secs(1)).await;
drop(c1);
wait_for_condition(|| async { conn_counter.get() == 0 }, Duration::from_secs(1)).await;
}
} }