[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]
Subject: Re: [virtio-comment] Re: [PATCH v2 05/11] transport-fabrics: introduce Keyed Transmission
On Thu, Jun 01, 2023 at 05:02:45PM +0800, zhenwei pi wrote: > > > On 6/1/23 00:20, Stefan Hajnoczi wrote: > > On Thu, May 04, 2023 at 04:19:04PM +0800, zhenwei pi wrote: > > > Keyed transmission is used for message oriented communication(Ex RDMA), > > > also add virtio-blk read/write 8K example. > > > > > > Signed-off-by: zhenwei pi <pizhenwei@bytedance.com> > > > --- > > > transport-fabrics.tex | 178 ++++++++++++++++++++++++++++++++++++++++++ > > > 1 file changed, 178 insertions(+) > > > > > > diff --git a/transport-fabrics.tex b/transport-fabrics.tex > > > index c02cf26..7711321 100644 > > > --- a/transport-fabrics.tex > > > +++ b/transport-fabrics.tex > > > @@ -317,3 +317,181 @@ \subsubsection{Buffer Mapping Definition}\label{sec:Virtio Transport Options / V > > > |......| > > > +------+ -> 8193 > > > \end{lstlisting} > > > + > > > +\paragraph{Keyed Transmission}\label{sec:Virtio Transport Options / Virtio Over Fabrics / Transmission Protocol / Commands Definition / Keyed Transmission} > > > +Command and Segment Descriptors are transmitted in a message within a > > > +connection, and buffer is transmitted by remote memory access. The layout in message: > > > > With RDMA it is theoretically possible to implement virtqueues without > > messages in the data path (i.e. by using something similar to vring with > > RDMA). Why did you decide to use a mixed messages + RDMA approach > > instead of a 100% RDMA approach? > > > > Hi, > > To reduce networking RTT. From my experience, a single RDMA message(event > based) uses at least 6us. > This approach has a chance to send a command(include data segments) by 1 > networking RTT, and receive a completion(include data segments) in 1 > networking RTT. I tried to design a 100% RDMA approach(mapping a vring to > the remote side, the remote side accesses this vring by RDMA READ/WRITE), > but I failed to find an idea to achieve. The goal is to minimize the number of RDMA transfers. Each area of memory should be located on the system that is polling constantly (busy waiting) and the other side occassionally sends an RDMA WRITE request. This idea requires bi-directional RDMA where both initiator and target make memory accessible to the other side. Is this possible? The target owns the Available Ring, a descriptor table similar to those used by the Split and Packed Virtqueue layouts that is used by the driver to submit virtqueue buffers to the device. The target sends a key to the Available Ring to the initiator during virtqueue setup. The initiator sends RDMA WRITEs that fill in virtqueue descriptors. Indirect descriptors are supported, but the target will need to use RDMA READs to load the indirect descriptor table, so there is overhead. Even regular non-indirect descriptors have overhead because an RDMA READ is required to read the payload. The best approach for small virtqueue elements is to inline the payload in the Available Ring descriptor so no additional RDMA transfers are needed (this achieves similar effect to your approach of using messages + RDMA, but with pure RDMA). The target polls the Available Ring to detect available buffers. The initiator sends a key to the Used Ring to the target during virtqueue setup. The target sends RDMA WRITEs that fill in used elements. The initiator polls the Used Ring to detect used buffers. I'm not sure if the Used Ring makes sense as RDMA memory. Maybe it's better to send a message over the reliable connection instead so that Used Buffer Notifications can support interrupts and not just polling. This is a new virtqueue layout. It's only worthwhile implementing it if the Available Ring RDMA performance is significantly better than the current approach. Stefan
Attachment:
signature.asc
Description: PGP signature
[Date Prev] | [Thread Prev] | [Thread Next] | [Date Next] -- [Date Index] | [Thread Index] | [List Home]