Initial commit: MikroTik btest server & client in Rust

Full reimplementation of the MikroTik Bandwidth Test protocol:
- Server mode: accepts connections from MikroTik devices on port 2000
- Client mode: connects to MikroTik btest servers
- TCP and UDP protocols with bidirectional support
- MD5 challenge-response authentication
- Dynamic speed adjustment (1.5x algorithm)
- Status exchange matching original C pselect() behavior
- Docker support with multi-stage build

Tested against MikroTik RouterOS achieving:
- 1.05 Gbps server RX (single connection)
- 530 Mbps client TCP download
- 840 Mbps client TCP upload
- 433 Mbps client UDP download

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
Siavash Sameni
2026-03-31 11:56:34 +04:00
commit d9007dc169
18 changed files with 2801 additions and 0 deletions

456
src/server.rs Normal file
View File

@@ -0,0 +1,456 @@
use std::net::SocketAddr;
use std::sync::atomic::Ordering;
use std::sync::Arc;
use std::time::{Duration, Instant};
use tokio::io::{AsyncReadExt, AsyncWriteExt};
use tokio::net::{TcpListener, TcpStream, UdpSocket};
use crate::auth;
use crate::bandwidth::{self, BandwidthState};
use crate::protocol::*;
pub async fn run_server(
port: u16,
auth_user: Option<String>,
auth_pass: Option<String>,
) -> Result<()> {
let addr = format!("0.0.0.0:{}", port);
let listener = TcpListener::bind(&addr).await?;
tracing::info!("btest server listening on {}", addr);
let udp_port_offset = Arc::new(std::sync::atomic::AtomicU16::new(0));
loop {
let (stream, peer) = listener.accept().await?;
tracing::info!("New connection from {}", peer);
let auth_user = auth_user.clone();
let auth_pass = auth_pass.clone();
let udp_offset = udp_port_offset.clone();
tokio::spawn(async move {
if let Err(e) = handle_client(stream, peer, auth_user, auth_pass, udp_offset).await {
tracing::error!("Client {} error: {}", peer, e);
}
});
}
}
async fn handle_client(
mut stream: TcpStream,
peer: SocketAddr,
auth_user: Option<String>,
auth_pass: Option<String>,
udp_port_offset: Arc<std::sync::atomic::AtomicU16>,
) -> Result<()> {
stream.set_nodelay(true)?;
send_hello(&mut stream).await?;
let cmd = recv_command(&mut stream).await?;
tracing::info!(
"Client {} command: proto={} dir={} tx_size={} remote_speed={} local_speed={}",
peer,
if cmd.is_udp() { "UDP" } else { "TCP" },
match cmd.direction {
CMD_DIR_RX => "RX",
CMD_DIR_TX => "TX",
CMD_DIR_BOTH => "BOTH",
_ => "?",
},
cmd.tx_size,
cmd.remote_tx_speed,
cmd.local_tx_speed,
);
auth::server_authenticate(
&mut stream,
auth_user.as_deref(),
auth_pass.as_deref(),
)
.await?;
if cmd.is_udp() {
run_udp_test_server(&mut stream, peer, &cmd, udp_port_offset).await
} else {
run_tcp_test_server(stream, cmd).await
}
}
// --- TCP Test Server ---
async fn run_tcp_test_server(stream: TcpStream, cmd: Command) -> Result<()> {
let state = BandwidthState::new();
let tx_size = cmd.tx_size as usize;
let server_should_tx = cmd.server_tx();
let server_should_rx = cmd.server_rx();
let tx_speed = cmd.remote_tx_speed;
let (reader, writer) = stream.into_split();
// IMPORTANT: Do NOT drop unused halves - dropping sends TCP FIN
let mut _writer_keepalive = None;
let mut _reader_keepalive = None;
let state_tx = state.clone();
let tx_handle = if server_should_tx {
Some(tokio::spawn(async move {
tcp_tx_loop(writer, tx_size, tx_speed, state_tx).await
}))
} else {
_writer_keepalive = Some(writer);
None
};
let state_rx = state.clone();
let rx_handle = if server_should_rx {
Some(tokio::spawn(async move {
tcp_rx_loop(reader, state_rx).await
}))
} else {
_reader_keepalive = Some(reader);
None
};
status_report_loop(&cmd, &state).await;
state.running.store(false, Ordering::SeqCst);
if let Some(h) = tx_handle { let _ = h.await; }
if let Some(h) = rx_handle { let _ = h.await; }
Ok(())
}
async fn tcp_tx_loop(
mut writer: tokio::net::tcp::OwnedWriteHalf,
tx_size: usize,
tx_speed: u32,
state: Arc<BandwidthState>,
) {
tokio::time::sleep(Duration::from_millis(100)).await;
let mut packet = vec![0u8; tx_size];
packet[0] = STATUS_MSG_TYPE;
let mut interval = bandwidth::calc_send_interval(tx_speed, tx_size as u16);
let mut next_send = Instant::now();
while state.running.load(Ordering::Relaxed) {
if writer.write_all(&packet).await.is_err() {
break;
}
state.tx_bytes.fetch_add(tx_size as u64, Ordering::Relaxed);
if state.tx_speed_changed.load(Ordering::Relaxed) {
state.tx_speed_changed.store(false, Ordering::Relaxed);
let new_speed = state.tx_speed.load(Ordering::Relaxed);
interval = bandwidth::calc_send_interval(new_speed, tx_size as u16);
next_send = Instant::now();
}
match interval {
Some(iv) => {
next_send += iv;
let now = Instant::now();
if next_send > now {
tokio::time::sleep(next_send - now).await;
}
}
None => {
tokio::task::yield_now().await;
}
}
}
}
async fn tcp_rx_loop(mut reader: tokio::net::tcp::OwnedReadHalf, state: Arc<BandwidthState>) {
let mut buf = vec![0u8; 65536];
while state.running.load(Ordering::Relaxed) {
match reader.read(&mut buf).await {
Ok(0) | Err(_) => break,
Ok(n) => {
state.rx_bytes.fetch_add(n as u64, Ordering::Relaxed);
}
}
}
}
// --- UDP Test Server ---
async fn run_udp_test_server(
stream: &mut TcpStream,
peer: SocketAddr,
cmd: &Command,
udp_port_offset: Arc<std::sync::atomic::AtomicU16>,
) -> Result<()> {
let offset = udp_port_offset.fetch_add(1, Ordering::SeqCst);
let server_udp_port = BTEST_UDP_PORT_START + offset;
let client_udp_port = server_udp_port + BTEST_PORT_CLIENT_OFFSET;
stream.write_all(&server_udp_port.to_be_bytes()).await?;
stream.flush().await?;
tracing::info!(
"UDP test: server_port={}, client_port={}, peer={}",
server_udp_port, client_udp_port, peer,
);
let udp = UdpSocket::bind(format!("0.0.0.0:{}", server_udp_port)).await?;
let client_udp_addr: SocketAddr =
format!("{}:{}", peer.ip(), client_udp_port).parse().unwrap();
udp.connect(client_udp_addr).await?;
let state = BandwidthState::new();
let tx_size = cmd.tx_size as usize;
let server_should_tx = cmd.server_tx();
let server_should_rx = cmd.server_rx();
let tx_speed = cmd.remote_tx_speed;
let udp = Arc::new(udp);
let state_tx = state.clone();
let udp_tx = udp.clone();
let tx_handle = if server_should_tx {
Some(tokio::spawn(async move {
udp_tx_loop(&udp_tx, tx_size, tx_speed, state_tx).await
}))
} else {
None
};
let state_rx = state.clone();
let udp_rx = udp.clone();
let rx_handle = if server_should_rx {
Some(tokio::spawn(async move {
udp_rx_loop(&udp_rx, state_rx).await
}))
} else {
None
};
// Status exchange using select! to match C pselect() behavior
udp_status_loop(stream, cmd, &state).await;
state.running.store(false, Ordering::SeqCst);
if let Some(h) = tx_handle { let _ = h.await; }
if let Some(h) = rx_handle { let _ = h.await; }
Ok(())
}
async fn udp_tx_loop(
socket: &UdpSocket,
tx_size: usize,
initial_tx_speed: u32,
state: Arc<BandwidthState>,
) {
let mut seq: u32 = 0;
let mut packet = vec![0u8; tx_size];
let mut interval = bandwidth::calc_send_interval(initial_tx_speed, tx_size as u16);
let mut next_send = Instant::now();
let mut consecutive_errors: u32 = 0;
while state.running.load(Ordering::Relaxed) {
packet[0..4].copy_from_slice(&seq.to_be_bytes());
match socket.send(&packet).await {
Ok(n) => {
seq = seq.wrapping_add(1);
state.tx_bytes.fetch_add(n as u64, Ordering::Relaxed);
consecutive_errors = 0;
}
Err(_) => {
consecutive_errors += 1;
if consecutive_errors > 1000 {
tracing::warn!("UDP TX: too many consecutive send errors, stopping");
break;
}
// Back off on ENOBUFS/EAGAIN
tokio::time::sleep(Duration::from_micros(200)).await;
continue;
}
}
// Pick up dynamic speed changes from status loop
if state.tx_speed_changed.load(Ordering::Relaxed) {
state.tx_speed_changed.store(false, Ordering::Relaxed);
let new_speed = state.tx_speed.load(Ordering::Relaxed);
interval = bandwidth::calc_send_interval(new_speed, tx_size as u16);
next_send = Instant::now();
tracing::debug!("TX speed adjusted to {} bps ({:.2} Mbps)",
new_speed, new_speed as f64 / 1_000_000.0);
}
match interval {
Some(iv) => {
next_send += iv;
let now = Instant::now();
if next_send > now {
tokio::time::sleep(next_send - now).await;
}
}
None => {
// Unlimited: yield every 64 packets to keep system responsive
if seq % 64 == 0 {
tokio::task::yield_now().await;
}
}
}
}
}
async fn udp_rx_loop(socket: &UdpSocket, state: Arc<BandwidthState>) {
let mut buf = vec![0u8; 65536];
let mut last_seq: Option<u32> = None;
while state.running.load(Ordering::Relaxed) {
match tokio::time::timeout(Duration::from_secs(5), socket.recv(&mut buf)).await {
Ok(Ok(n)) if n >= 4 => {
state.rx_bytes.fetch_add(n as u64, Ordering::Relaxed);
state.rx_packets.fetch_add(1, Ordering::Relaxed);
let seq = u32::from_be_bytes([buf[0], buf[1], buf[2], buf[3]]);
if let Some(last) = last_seq {
let expected = last.wrapping_add(1);
if seq > expected {
let lost = seq - expected;
state.rx_lost_packets.fetch_add(lost as u64, Ordering::Relaxed);
}
}
last_seq = Some(seq);
state.last_udp_seq.store(seq, Ordering::Relaxed);
}
Ok(Ok(_)) => {}
Ok(Err(e)) => {
tracing::debug!("UDP recv error: {}", e);
tokio::time::sleep(Duration::from_millis(10)).await;
}
Err(_) => {
tracing::debug!("UDP RX timeout");
}
}
}
}
// --- Status Reporting ---
async fn status_report_loop(cmd: &Command, state: &BandwidthState) {
let mut seq: u32 = 0;
let mut interval = tokio::time::interval(Duration::from_secs(1));
loop {
interval.tick().await;
if !state.running.load(Ordering::Relaxed) {
break;
}
seq += 1;
if cmd.server_tx() {
let tx = state.tx_bytes.swap(0, Ordering::Relaxed);
bandwidth::print_status(seq, "TX", tx, Duration::from_secs(1), None);
}
if cmd.server_rx() {
let rx = state.rx_bytes.swap(0, Ordering::Relaxed);
let lost = state.rx_lost_packets.swap(0, Ordering::Relaxed);
let lost_opt = if cmd.is_udp() { Some(lost) } else { None };
bandwidth::print_status(seq, "RX", rx, Duration::from_secs(1), lost_opt);
}
}
}
/// UDP status exchange loop - matches C pselect() behavior exactly:
/// 1. Wait up to 1 second for client status (like pselect with 1s timeout)
/// 2. If client sent status, read and process it
/// 3. ALWAYS send our status (unconditional, matching C line 1048)
/// 4. Reset counters and print stats
/// This sequential approach prevents the ticker from being starved.
async fn udp_status_loop(
stream: &mut TcpStream,
cmd: &Command,
state: &BandwidthState,
) {
let mut seq: u32 = 0;
let (mut reader, mut writer) = tokio::io::split(stream);
let mut status_buf = [0u8; STATUS_MSG_SIZE];
let mut next_status = Instant::now() + Duration::from_secs(1);
loop {
if !state.running.load(Ordering::Relaxed) {
break;
}
// Step 1: Wait for client status OR timeout (like C pselect)
let now = Instant::now();
let wait_time = if next_status > now {
next_status - now
} else {
Duration::ZERO
};
// Try to read client's status within the remaining time window
match tokio::time::timeout(wait_time, reader.read_exact(&mut status_buf)).await {
Ok(Ok(_)) => {
let client_status = StatusMessage::deserialize(&status_buf);
tracing::debug!(
"RECV status: raw={:02x?} seq={} bytes_received={}",
&status_buf, client_status.seq, client_status.bytes_received,
);
if client_status.bytes_received > 0 && cmd.server_tx() {
let new_speed =
((client_status.bytes_received as u64 * 8 * 3) / 2) as u32;
state.tx_speed.store(new_speed, Ordering::Relaxed);
state.tx_speed_changed.store(true, Ordering::Relaxed);
tracing::debug!(
"Speed adjust: client got {} bytes → our TX {:.2} Mbps",
client_status.bytes_received,
new_speed as f64 / 1_000_000.0,
);
}
if Instant::now() < next_status {
continue;
}
}
Ok(Err(e)) => {
tracing::debug!("Client TCP read error: {}", e);
state.running.store(false, Ordering::SeqCst);
break;
}
Err(_) => {
// Timeout - 1 second elapsed
}
}
// Step 2: ALWAYS send our status every 1 second
seq += 1;
next_status = Instant::now() + Duration::from_secs(1);
let rx_bytes = state.rx_bytes.swap(0, Ordering::Relaxed);
let tx_bytes = state.tx_bytes.swap(0, Ordering::Relaxed);
let lost = state.rx_lost_packets.swap(0, Ordering::Relaxed);
let status = StatusMessage {
seq,
bytes_received: rx_bytes as u32,
};
let serialized = status.serialize();
tracing::debug!(
"SEND status: raw={:02x?} seq={} bytes_received={} ({:.2} Mbps)",
&serialized, seq, rx_bytes, rx_bytes as f64 * 8.0 / 1_000_000.0,
);
if writer.write_all(&serialized).await.is_err() {
state.running.store(false, Ordering::SeqCst);
break;
}
let _ = writer.flush().await;
// Print local stats
if cmd.server_tx() {
bandwidth::print_status(seq, "TX", tx_bytes, Duration::from_secs(1), None);
}
if cmd.server_rx() {
bandwidth::print_status(seq, "RX", rx_bytes, Duration::from_secs(1), Some(lost));
}
}
}