From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mail-qk1-x72d.google.com (mail-qk1-x72d.google.com [IPv6:2607:f8b0:4864:20::72d]) by mail.toke.dk (Postfix) with ESMTPS id 8C26EA140C7 for ; Fri, 30 Jun 2023 08:23:05 +0200 (CEST) Authentication-Results: mail.toke.dk; dkim=pass (2048-bit key; unprotected) header.d=gmail.com header.i=@gmail.com header.a=rsa-sha256 header.s=20221208 header.b=UrbQQozd Received: by mail-qk1-x72d.google.com with SMTP id af79cd13be357-765ab83a724so31329385a.0 for ; Thu, 29 Jun 2023 23:23:05 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20221208; t=1688106183; x=1690698183; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date :message-id:reply-to; bh=zn+IvaySyr+8hFvUNeN9VI08I1VCqfQ46AdL4PsISXU=; b=UrbQQozdD4eke3QB7kAQ1U09uC5C1K7t1KWj4dClLEbkFbySTkn25po2x7hX/+2noa 2eYNarAqMglEbUzyPVXiSYtKOrjOWwGmZszkV4i8UXRVkUCi0iqExVVPzV1G0BcW6K2V X59RLxONdtSEHrB7MHpJ6UVkiGAD18SmgMFY+lf7xYZsEB4KR+VKWmjyZAkfyc6O+NZb VthpvwtzFPYp5gC2XUT+yusqibmINTZv9t8/2SnhlpLSH7t/jgqKXFmQOt2pm11XMksa lohYMDbswXvtmtAbqP3QPh3483yuiAFsHUvmx5KSbvTbXCtRJ7bi4sNcXTehhhJEjoI4 YpZQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20221208; t=1688106183; x=1690698183; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=zn+IvaySyr+8hFvUNeN9VI08I1VCqfQ46AdL4PsISXU=; b=FlzEK+vj7pKOH6vRHxHysupjgadJlzslgTIm9Zzp3hdJ57BoKmB0uhMJBVAAz56K8n AqDs6i3PZJ7y84Ku0O9e8F7lCEWP9iFkoNCEQDOFR+yl9f8TZrQOBq4Praw8z8kJtXbu t9lhybe2lX5xGCjN0Qqwsksgfsl5LFRW6AqNWbGBE3vdubVXMd2vHjjdrSl1gf/Ip7Lx 7J7iqMlNwo2Or/bOYMGFYR6Q8maF72sMKdt6lzutV4iFtTTsk0y/eGH5RsPJutIj1BAS BKLlbXyW9Jn8pYvtBDXWg2kzXZ4zokdW5e+bpJBeOqRV0MKCYKf+aeRUNqhA7DXWmIB2 0JOw== X-Gm-Message-State: ABy/qLaRsAY+zdaDVegjivNpQRrxwX0LY19w+UFP3djmoC6fAB66FsIb kKba8Zvh5+MYn6p2tqbFW+kaLBBiMmAgP47Re7k= X-Google-Smtp-Source: APBJJlE1bURf+fbpyVzyq1EsCUd1lFm37Y+cPQ2XI3+qEUL7UjlapvLLAb5L3cCQrIwEq+YhQyf0SNm2MZn0O+WmAy4= X-Received: by 2002:a0c:f7d2:0:b0:635:fa38:5216 with SMTP id f18-20020a0cf7d2000000b00635fa385216mr2161758qvo.0.1688106183065; Thu, 29 Jun 2023 23:23:03 -0700 (PDT) MIME-Version: 1.0 References: <20230621170244.1283336-1-sdf@google.com> <20230621170244.1283336-4-sdf@google.com> <57b9fc14-c02e-f0e5-148d-a549ebab6cf6@brouer.com> <435d1630-c3f4-97fb-b6fe-9795d1f0bf33@redhat.com> <871qhuh5ec.fsf@toke.dk> <87r0pufpf2.fsf@toke.dk> In-Reply-To: <87r0pufpf2.fsf@toke.dk> From: Magnus Karlsson Date: Fri, 30 Jun 2023 08:22:51 +0200 Message-ID: To: =?UTF-8?B?VG9rZSBIw7hpbGFuZC1Kw7hyZ2Vuc2Vu?= Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable Message-ID-Hash: JKWAJ7YXRWIEWP65LH5CCKT3EDTIBXYQ X-Message-ID-Hash: JKWAJ7YXRWIEWP65LH5CCKT3EDTIBXYQ X-MailFrom: magnus.karlsson@gmail.com X-Mailman-Rule-Misses: dmarc-mitigation; no-senders; approved; emergency; loop; banned-address; member-moderation; nonmember-moderation; administrivia; implicit-dest; max-recipients; max-size; news-moderation; no-subject; digests; suspicious-header CC: Stanislav Fomichev , Jesper Dangaard Brouer , brouer@redhat.com, bpf@vger.kernel.org, ast@kernel.org, daniel@iogearbox.net, andrii@kernel.org, martin.lau@linux.dev, song@kernel.org, yhs@fb.com, john.fastabend@gmail.com, kpsingh@kernel.org, haoluo@google.com, jolsa@kernel.org, =?UTF-8?B?QmrDtnJuIFTDtnBlbA==?= , "Karlsson, Magnus" , "xdp-hints@xdp-project.net" X-Mailman-Version: 3.3.8 Precedence: list Subject: [xdp-hints] Re: [RFC bpf-next v2 03/11] xsk: Support XDP_TX_METADATA_LEN List-Id: XDP hardware hints design discussion Archived-At: List-Archive: List-Help: List-Owner: List-Post: List-Subscribe: List-Unsubscribe: On Thu, 29 Jun 2023 at 14:01, Toke H=C3=B8iland-J=C3=B8rgensen wrote: > > Magnus Karlsson writes: > > > On Thu, 29 Jun 2023 at 13:30, Toke H=C3=B8iland-J=C3=B8rgensen wrote: > >> > >> Stanislav Fomichev writes: > >> > >> > On Wed, Jun 28, 2023 at 1:09=E2=80=AFAM Magnus Karlsson > >> > wrote: > >> >> > >> >> On Mon, 26 Jun 2023 at 19:06, Stanislav Fomichev w= rote: > >> >> > > >> >> > On Sat, Jun 24, 2023 at 2:02=E2=80=AFAM Jesper Dangaard Brouer > >> >> > wrote: > >> >> > > > >> >> > > > >> >> > > > >> >> > > On 23/06/2023 19.41, Stanislav Fomichev wrote: > >> >> > > > On Fri, Jun 23, 2023 at 3:24=E2=80=AFAM Jesper Dangaard Broue= r > >> >> > > > wrote: > >> >> > > >> > >> >> > > >> > >> >> > > >> > >> >> > > >> On 22/06/2023 19.55, Stanislav Fomichev wrote: > >> >> > > >>> On Thu, Jun 22, 2023 at 2:11=E2=80=AFAM Jesper D. Brouer wrote: > >> >> > > >>>> > >> >> > > >>>> > >> >> > > >>>> This needs to be reviewed by AF_XDP maintainers Magnus and= Bj=C3=B8rn (Cc) > >> >> > > >>>> > >> >> > > >>>> On 21/06/2023 19.02, Stanislav Fomichev wrote: > >> >> > > >>>>> For zerocopy mode, tx_desc->addr can point to the arbitra= ry offset > >> >> > > >>>>> and carry some TX metadata in the headroom. For copy mode= , there > >> >> > > >>>>> is no way currently to populate skb metadata. > >> >> > > >>>>> > >> >> > > >>>>> Introduce new XDP_TX_METADATA_LEN that indicates how many= bytes > >> >> > > >>>>> to treat as metadata. Metadata bytes come prior to tx_des= c address > >> >> > > >>>>> (same as in RX case). > >> >> > > >>>> > >> >> > > >>>> From looking at the code, this introduces a socket opti= on for this TX > >> >> > > >>>> metadata length (tx_metadata_len). > >> >> > > >>>> This implies the same fixed TX metadata size is used for a= ll packets. > >> >> > > >>>> Maybe describe this in patch desc. > >> >> > > >>> > >> >> > > >>> I was planning to do a proper documentation page once we se= ttle on all > >> >> > > >>> the details (similar to the one we have for rx). > >> >> > > >>> > >> >> > > >>>> What is the plan for dealing with cases that doesn't popul= ate same/full > >> >> > > >>>> TX metadata size ? > >> >> > > >>> > >> >> > > >>> Do we need to support that? I was assuming that the TX layo= ut would be > >> >> > > >>> fixed between the userspace and BPF. > >> >> > > >> > >> >> > > >> I hope you don't mean fixed layout, as the whole point is ad= ding > >> >> > > >> flexibility and extensibility. > >> >> > > > > >> >> > > > I do mean a fixed layout between the userspace (af_xdp) and d= evtx program. > >> >> > > > At least fixed max size of the metadata. The userspace and th= e bpf > >> >> > > > prog can then use this fixed space to implement some flexibil= ity > >> >> > > > (btf_ids, versioned structs, bitmasks, tlv, etc). > >> >> > > > If we were to make the metalen vary per packet, we'd have to = signal > >> >> > > > its size per packet. Probably not worth it? > >> >> > > > >> >> > > Existing XDP metadata implementation also expand in a fixed/lim= ited > >> >> > > sized memory area, but communicate size per packet in this area= (also > >> >> > > for validation purposes). BUT for AF_XDP we don't have room fo= r another > >> >> > > pointer or size in the AF_XDP descriptor (see struct xdp_desc). > >> >> > > > >> >> > > > >> >> > > > > >> >> > > >>> If every packet would have a different metadata length, it = seems like > >> >> > > >>> a nightmare to parse? > >> >> > > >>> > >> >> > > >> > >> >> > > >> No parsing is really needed. We can simply use BTF IDs and = type cast in > >> >> > > >> BPF-prog. Both BPF-prog and userspace have access to the loc= al BTF ids, > >> >> > > >> see [1] and [2]. > >> >> > > >> > >> >> > > >> It seems we are talking slightly past each-other(?). Let me= rephrase > >> >> > > >> and reframe the question, what is your *plan* for dealing wi= th different > >> >> > > >> *types* of TX metadata. The different struct *types* will o= f-cause have > >> >> > > >> different sizes, but that is okay as long as they fit into t= he maximum > >> >> > > >> size set by this new socket option XDP_TX_METADATA_LEN. > >> >> > > >> Thus, in principle I'm fine with XSK having configured a fix= ed headroom > >> >> > > >> for metadata, but we need a plan for handling more than one = type and > >> >> > > >> perhaps a xsk desc indicator/flag for knowing TX metadata is= n't random > >> >> > > >> data ("leftover" since last time this mem was used). > >> >> > > > > >> >> > > > Yeah, I think the above correctly catches my expectation here= . Some > >> >> > > > headroom is reserved via XDP_TX_METADATA_LEN and the flexibil= ity is > >> >> > > > offloaded to the bpf program via btf_id/tlv/etc. > >> >> > > > > >> >> > > > Regarding leftover metadata: can we assume the userspace will= take > >> >> > > > care of setting it up? > >> >> > > > > >> >> > > >> With this kfunc approach, then things in-principle, becomes = a contract > >> >> > > >> between the "local" TX-hook BPF-prog and AF_XDP userspace. = These two > >> >> > > >> components can as illustrated here [1]+[2] can coordinate ba= sed on local > >> >> > > >> BPF-prog BTF IDs. This approach works as-is today, but patc= hset > >> >> > > >> selftests examples don't use this and instead have a very st= atic > >> >> > > >> approach (that people will copy-paste). > >> >> > > >> > >> >> > > >> An unsolved problem with TX-hook is that it can also get pac= kets from > >> >> > > >> XDP_REDIRECT and even normal SKBs gets processed (right?). = How does the > >> >> > > >> BPF-prog know if metadata is valid and intended to be used f= or e.g. > >> >> > > >> requesting the timestamp? (imagine metadata size happen to m= atch) > >> >> > > > > >> >> > > > My assumption was the bpf program can do ifindex/netns filter= ing. Plus > >> >> > > > maybe check that the meta_len is the one that's expected. > >> >> > > > Will that be enough to handle XDP_REDIRECT? > >> >> > > > >> >> > > I don't think so, using the meta_len (+ ifindex/netns) to commu= nicate > >> >> > > activation of TX hardware hints is too weak and not enough. Th= is is an > >> >> > > implicit API for BPF-programmers to understand and can lead to = implicit > >> >> > > activation. > >> >> > > > >> >> > > Think about what will happen for your AF_XDP send use-case. Fo= r > >> >> > > performance reasons AF_XDP don't zero out frame memory. Thus, = meta_len > >> >> > > is fixed even if not used (and can contain garbage), it can by = accident > >> >> > > create hard-to-debug situations. As discussed with Magnus+Mary= am > >> >> > > before, we found it was practical (and faster than mem zero) to= extend > >> >> > > AF_XDP descriptor (see struct xdp_desc) with some flags to > >> >> > > indicate/communicate this frame comes with TX metadata hints. > >> >> > > >> >> > What is that "if not used" situation? Can the metadata itself hav= e > >> >> > is_used bit? The userspace has to initialize at least that bit. > >> >> > We can definitely add that extra "has_metadata" bit to the descri= ptor, > >> >> > but I'm trying to understand whether we can do without it. > >> >> > >> >> To me, this "has_metadata" bit in the descriptor is just an > >> >> optimization. If it is 0, then there is no need to go and check the > >> >> metadata field and you save some performance. Regardless of this bi= t, > >> >> you need some way to say "is_used" for each metadata entry (at leas= t > >> >> when the number of metadata entries is >1). Three options come to m= ind > >> >> each with their pros and cons. > >> >> > >> >> #1: Let each metadata entry have an invalid state. Not possible for > >> >> every metadata and requires the user/kernel to go scan through ever= y > >> >> entry for every packet. > >> >> > >> >> #2: Have a field of bits at the start of the metadata section (clos= est > >> >> to packet data) that signifies if a metadata entry is valid or not.= If > >> >> there are N metadata entries in the metadata area, then N bits in t= his > >> >> field would be used to signify if the corresponding metadata is use= d > >> >> or not. Only requires the user/kernel to scan the valid entries plu= s > >> >> one access for the "is_used" bits. > >> >> > >> >> #3: Have N bits in the AF_XDP descriptor options field instead of t= he > >> >> N bits in the metadata area of #2. Faster but would consume many > >> >> precious bits in the fixed descriptor and cap the number of metadat= a > >> >> entries possible at around 8. E.g., 8 for Rx, 8 for Tx, 1 for the > >> >> multi-buffer work, and 15 for some future use. Depends on how darin= g > >> >> we are. > >> >> > >> >> The "has_metadata" bit suggestion can be combined with 1 or 2. > >> >> Approach 3 is just a fine grained extension of the idea itself. > >> >> > >> >> IMO, the best approach unfortunately depends on the metadata itself= . > >> >> If it is rarely valid, you want something like the "has_metadata" b= it. > >> >> If it is nearly always valid and used, approach #1 (if possible for > >> >> the metadata) should be the fastest. The decision also depends on t= he > >> >> number of metadata entries you have per packet. Sorry that I do not > >> >> have a good answer. My feeling is that we need something like #1 or > >> >> #2, or maybe both, then if needed we can add the "has_metadata" bit= or > >> >> bits (#3) optimization. Can we do this encoding and choice (#1, #2,= or > >> >> a combo) in the eBPF program itself? Would provide us with the > >> >> flexibility, if possible. > >> > > >> > Here is my take on it, lmk if I'm missing something: > >> > > >> > af_xdp users call this new setsockopt(XDP_TX_METADATA_LEN) when they > >> > plan to use metadata on tx. > >> > This essentially requires allocating a fixed headroom to carry the m= etadata. > >> > af_xdp machinery exports this fixed len into the bpf programs someho= w > >> > (devtx_frame.meta_len in this series). > >> > Then it's up to the userspace and bpf program to agree on the layout= . > >> > If not every packet is expected to carry the metadata, there might b= e > >> > some bitmask in the metadata area to indicate that. > >> > > >> > Iow, the metadata isn't interpreted by the kernel. It's up to the pr= og > >> > to interpret it and call appropriate kfunc to enable some offload. > >> > >> The reason for the flag on RX is mostly performance: there's a > >> substantial performance hit from reading the metadata area because it'= s > >> not cache-hot; we want to avoid that when no metadata is in use. Putti= ng > >> the flag inside the metadata area itself doesn't work for this, becaus= e > >> then you incur the cache miss just to read the flag. > > > > Not necessarily. Let us say that the flag is 4 bytes. Increase the > > start address of the packet buffer with 4 and the flags field will be > > on the same cache line as the first 60 bytes of the packet data > > (assuming a 64 byte cache line size and the flags field is closest to > > the start of the packet data). As long as you write something in those > > first 60 bytes of packet data that cache line will be brought in and > > will likely be in the cache when you access the bits in the metadata > > field. The trick works similarly for Rx by setting the umem headroom > > accordingly. > > Yeah, a trick like that was what I was alluding to with the "could" in > this bit: > > >> but I see no reason it could not also occur on TX (it'll mostly > >> depend on data alignment I guess?). > > right below the text you quoted ;) Ouch! Sorry Toke. Was a bit too trigger-happy there. > > But you are correct in that dedicating a bit in the descriptor will > > make sure it is always hot, while the trick above is dependent on the > > app wanting to read or write the first cache line worth of packet > > data. > > Exactly; which is why I think it's worth the flag bit :) > > -Toke >