From: Stefano Brivio <sbrivio@redhat.com>
To: Jon Maloy <jmaloy@redhat.com>
Cc: david@gibson.dropbear.id.au, passt-dev@passt.top,
Yumei Huang <yuhuang@redhat.com>
Subject: Re: [PATCH v3] tcp: Use SO_MEMINFO for accurate send buffer overhead accounting
Date: Thu, 07 May 2026 11:48:44 +0200 (CEST) [thread overview]
Message-ID: <20260507114842.4f4c85b6@elisabeth> (raw)
In-Reply-To: <20260425195818.572409-1-jmaloy@redhat.com>
On Sat, 25 Apr 2026 15:58:18 -0400
Jon Maloy <jmaloy@redhat.com> wrote:
> The TCP window advertised to the guest/container must balance two
> competing needs: large enough to trigger kernel socket buffer
> auto-tuning, but not so large that sendmsg() partially fails, causing
> retransmissions.
>
> The current approach uses the difference (SNDBUF_GET() - SIOCOUTQ), but
> SNDBUF_GET() returns a scaled value that only roughly accounts for
> per-skb overhead. The clamped_scale approximation doesn't accurately
> track the actual per-segment overhead, which can lead to both excessive
> retransmissions and reduced throughput.
>
> We now introduce the use of SO_MEMINFO to obtain SK_MEMINFO_SNDBUF and
> SK_MEMINFO_WMEM_QUEUED from the kernel. The latter is presented in the
> kernel's own accounting units, i.e. including the sk_buff overhead,
> and matches exactly what the kernel's own sk_stream_memory_free()
> function is using.
>
> When data is queued and the overhead ratio is observable, we calculate
> the per-segment overhead as (wmem_queued - sendq) / num_segments, then
> determine how many additional segments should fit in the remaining
> buffer space, considering the calculated per-mss overhead. This approach
> treats segments as discrete quantities, and produces a more accurate
> estimate of available buffer space than a linear scaling factor does.
>
> When the ratio cannot be observed, e.g. because the queue is empty or
> we are in a transient state, we fall back to the existing clamped_scale
> calculation (scaling between 100% and 75% of buffer capacity).
>
> When SO_MEMINFO succeeds, we also use SK_MEMINFO_SNDBUF directly to
> set SNDBUF, avoiding a separate SO_SNDBUF getsockopt() call.
>
> If SO_MEMINFO is unavailable, we fall back to the pre-existing
> SNDBUF_GET() - SIOCOUTQ calculation.
>
> Link: https://bugs.passt.top/show_bug.cgi?id=138
> Link: https://github.com/containers/podman/issues/28219
> Signed-off-by: Jon Maloy <jmaloy@redhat.com>
I finally tested this in a low (but not negligible) RTT setup (~200 to
~500 µs) and it looks extremely reliable there as well. I asked the
reporter of https://github.com/containers/podman/issues/28219 to also
test this but I think we can start merging this meanwhile.
Applied, with an additional tag:
Analysed-by: Yumei Huang <yuhuang@redhat.com>
as the analysis / tests behind this approach partially came from Yumei.
> ---
>
> v2: Updated according to feedback from Stefano. Segment-based discrete
> overhead calculation instead of linear ratio.
>
> v3: Addressed Stefano's v2 feedback:
> - Extracted window calculation into tcp_wnd_from_sndbuf()
> - Use wmem_queued instead of SIOCOUTQ for fallback and SWS check
> ---
> tcp.c | 137 ++++++++++++++++++++++++++++++++++-------------------
> tcp_conn.h | 2 +-
> 2 files changed, 89 insertions(+), 50 deletions(-)
>
> diff --git a/tcp.c b/tcp.c
> index 43b8fdb..61160cf 100644
> --- a/tcp.c
> +++ b/tcp.c
> @@ -295,6 +295,7 @@
> #include <arpa/inet.h>
>
> #include <linux/sockios.h>
> +#include <linux/sock_diag.h>
>
> #include "checksum.h"
> #include "util.h"
> @@ -1017,6 +1018,90 @@ size_t tcp_fill_headers(const struct ctx *c, struct tcp_tap_conn *conn,
> return MAX(l3len + sizeof(struct ethhdr), ETH_ZLEN);
> }
>
> +/**
> + * tcp_wnd_from_sndbuf() - Calculate window from available send buffer space
> + * @s: Socket file descriptor
> + * @conn: Connection pointer
> + * @tinfo: tcp_info from kernel
> + *
> + * Return: window value to advertise, not scaled
> + */
> +static uint32_t tcp_wnd_from_sndbuf(int s, struct tcp_tap_conn *conn,
> + const struct tcp_info_linux *tinfo)
> +{
> + uint32_t rtt_ms_ceiling = DIV_ROUND_UP(tinfo->tcpi_rtt, 1000);
> + uint32_t mem[SK_MEMINFO_VARS];
> + socklen_t mem_sl = sizeof(mem);
> + int mss = MSS_GET(conn);
> + uint32_t limit, sendq;
> +
> + if (ioctl(s, SIOCOUTQ, &sendq)) {
> + debug_perror("SIOCOUTQ on socket %i, assuming 0", s);
> + sendq = 0;
> + }
> +
> + if (getsockopt(s, SOL_SOCKET, SO_MEMINFO, &mem, &mem_sl)) {
> + tcp_get_sndbuf(conn);
> +
> + if (sendq > SNDBUF_GET(conn)) /* Due to memory pressure? */
> + limit = 0;
> + else
> + limit = SNDBUF_GET(conn) - sendq;
> + } else {
> + uint32_t sndbuf = mem[SK_MEMINFO_SNDBUF];
> + uint32_t wmemq = mem[SK_MEMINFO_WMEM_QUEUED];
> + uint32_t scaled = clamped_scale(sndbuf, sndbuf, SNDBUF_SMALL,
> + SNDBUF_BIG, 75);
> +
> + SNDBUF_SET(conn, MIN(INT_MAX, scaled));
> +
> + if (wmemq > sndbuf) {
> + limit = 0;
> + } else if (!sendq || !mss || wmemq <= sendq) {
> + limit = SNDBUF_GET(conn) - wmemq;
> + } else {
> + uint32_t used_segs = MAX(sendq / mss, 1);
> + uint32_t overhead = (wmemq - sendq) / used_segs;
> + uint32_t remaining = sndbuf - wmemq;
> + uint32_t avail_segs = remaining / (mss + overhead);
> +
> + limit = avail_segs * mss;
> + }
> + }
> +
> + /* If the sender uses mechanisms to prevent Silly Window
> + * Syndrome (SWS, described in RFC 813 Section 3) it's critical
> + * that, should the window ever become less than the MSS, we
> + * advertise a new value once it increases again to be above it.
> + *
> + * The mechanism to avoid SWS in the kernel is, implicitly,
> + * implemented by Nagle's algorithm (which was proposed after
> + * RFC 813).
> + *
> + * To this end, for simplicity, approximate a window value below
> + * the MSS to zero, as we already have mechanisms in place to
> + * force updates after the window becomes zero. This matches the
> + * suggestion from RFC 813, Section 4.
> + *
> + * But don't do this if, either:
> + *
> + * - there's nothing in the outbound queue: the size of the
> + * sending buffer is limiting us, and it won't increase if we
> + * don't send data, so there's no point in waiting, or
> + *
> + * - we haven't sent data in a while (somewhat arbitrarily, ten
> + * times the RTT), as that might indicate that the receiver
> + * will only process data in batches that are large enough,
> + * but we won't send enough to fill one because we're stuck
> + * with pending data in the outbound queue
> + */
> + if (limit < (uint32_t)MSS_GET(conn) && sendq &&
> + tinfo->tcpi_last_data_sent < rtt_ms_ceiling * 10)
> + limit = 0;
> +
> + return MIN(tinfo->tcpi_snd_wnd, limit);
> +}
> +
> /**
> * tcp_update_seqack_wnd() - Update ACK sequence and window to guest/tap
> * @c: Execution context
> @@ -1124,56 +1209,10 @@ int tcp_update_seqack_wnd(const struct ctx *c, struct tcp_tap_conn *conn,
> }
> }
>
> - if ((conn->flags & LOCAL) || tcp_rtt_dst_low(conn)) {
> + if ((conn->flags & LOCAL) || tcp_rtt_dst_low(conn))
> new_wnd_to_tap = tinfo->tcpi_snd_wnd;
> - } else {
> - unsigned rtt_ms_ceiling = DIV_ROUND_UP(tinfo->tcpi_rtt, 1000);
> - uint32_t sendq;
> - int limit;
> -
> - if (ioctl(s, SIOCOUTQ, &sendq)) {
> - debug_perror("SIOCOUTQ on socket %i, assuming 0", s);
> - sendq = 0;
> - }
> - tcp_get_sndbuf(conn);
> -
> - if ((int)sendq > SNDBUF_GET(conn)) /* Due to memory pressure? */
> - limit = 0;
> - else
> - limit = SNDBUF_GET(conn) - (int)sendq;
> -
> - /* If the sender uses mechanisms to prevent Silly Window
> - * Syndrome (SWS, described in RFC 813 Section 3) it's critical
> - * that, should the window ever become less than the MSS, we
> - * advertise a new value once it increases again to be above it.
> - *
> - * The mechanism to avoid SWS in the kernel is, implicitly,
> - * implemented by Nagle's algorithm (which was proposed after
> - * RFC 813).
> - *
> - * To this end, for simplicity, approximate a window value below
> - * the MSS to zero, as we already have mechanisms in place to
> - * force updates after the window becomes zero. This matches the
> - * suggestion from RFC 813, Section 4.
> - *
> - * But don't do this if, either:
> - *
> - * - there's nothing in the outbound queue: the size of the
> - * sending buffer is limiting us, and it won't increase if we
> - * don't send data, so there's no point in waiting, or
> - *
> - * - we haven't sent data in a while (somewhat arbitrarily, ten
> - * times the RTT), as that might indicate that the receiver
> - * will only process data in batches that are large enough,
> - * but we won't send enough to fill one because we're stuck
> - * with pending data in the outbound queue
> - */
> - if (limit < MSS_GET(conn) && sendq &&
> - tinfo->tcpi_last_data_sent < rtt_ms_ceiling * 10)
> - limit = 0;
> -
> - new_wnd_to_tap = MIN((int)tinfo->tcpi_snd_wnd, limit);
> - }
> + else
> + new_wnd_to_tap = tcp_wnd_from_sndbuf(s, conn, tinfo);
>
> new_wnd_to_tap = MIN(new_wnd_to_tap, MAX_WINDOW);
> if (!(conn->events & ESTABLISHED))
> diff --git a/tcp_conn.h b/tcp_conn.h
> index 6985426..9f5bee0 100644
> --- a/tcp_conn.h
> +++ b/tcp_conn.h
> @@ -98,7 +98,7 @@ struct tcp_tap_conn {
> #define SNDBUF_BITS 24
> unsigned int sndbuf :SNDBUF_BITS;
> #define SNDBUF_SET(conn, bytes) (conn->sndbuf = ((bytes) >> (32 - SNDBUF_BITS)))
> -#define SNDBUF_GET(conn) (conn->sndbuf << (32 - SNDBUF_BITS))
> +#define SNDBUF_GET(conn) ((uint32_t)(conn->sndbuf << (32 - SNDBUF_BITS)))
>
> uint8_t seq_dup_ack_approx;
>
--
Stefano
prev parent reply other threads:[~2026-05-07 9:48 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-25 17:58 Jon Maloy
2026-05-07 9:48 ` Stefano Brivio [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260507114842.4f4c85b6@elisabeth \
--to=sbrivio@redhat.com \
--cc=david@gibson.dropbear.id.au \
--cc=jmaloy@redhat.com \
--cc=passt-dev@passt.top \
--cc=yuhuang@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
Code repositories for project(s) associated with this public inbox
https://passt.top/passt
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for IMAP folder(s).