-----Original Message-----
> >
https://lore.kernel.org/qemu-devel/CAMGffEn-DKpMZ4tA71MJYdyemg0Zda
> > > > > 15
> > > > > > > wVAqk81vXtKzx-LfJQ(a)mail.gmail.com/
> > > > > > >
> > > > > > > Appreciate a lot for everyone helping on the
testings.
> > > > > > >
> > > > > > > > InfiniBand controller: Mellanox Technologies
MT27800
> > > > > > > > Family [ConnectX-5]
> > > > > > > >
> > > > > > > > which doesn't meet our purpose. I can choose
RDMA or TCP
> > > > > > > > for VM migration. RDMA traffic is through
InfiniBand and
> > > > > > > > TCP through Ethernet on these two hosts. One is
standby
> > > > > > > > while the other
> > is active.
> > > > > > > >
> > > > > > > > Now I'll try on a server with more recent
Ethernet and
> > > > > > > > InfiniBand network adapters. One of them has:
> > > > > > > > BCM57414 NetXtreme-E 10Gb/25Gb RDMA Ethernet
Controller
> > > > > > > > (rev
> > > > > > > > 01)
> > > > > > > >
> > > > > > > > The comparison between RDMA and TCP on the same
NIC
> > > > > > > > could make more
> > > > > > > sense.
> > > > > > >
> > > > > > > It looks to me NICs are powerful now, but again as I
> > > > > > > mentioned I don't think it's a reason we need
to deprecate
> > > > > > > rdma, especially if QEMU's rdma migration has the
chance
> > > > > > > to be refactored
> > using rsocket.
> > > > > > >
> > > > > > > Is there anyone who started looking into that
direction?
> > > > > > > Would it make sense we start some PoC now?
> > > > > > >
> > > > > >
> > > > > > My team has finished the PoC refactoring which works
well.
> > > > > >
> > > > > > Progress:
> > > > > > 1. Implement io/channel-rdma.c, 2. Add unit test
> > > > > > tests/unit/test-io-channel-rdma.c and verifying it is
> > > > > > successful, 3. Remove the original code from
migration/rdma.c, 4.
> > > > > > Rewrite the rdma_start_outgoing_migration and
> > > > > > rdma_start_incoming_migration logic, 5. Remove all
rdma_xxx
> > > > > > functions from migration/ram.c. (to prevent RDMA live
> > > > > > migration from polluting the
> > > > > core logic of live migration), 6. The soft-RoCE implemented
> > > > > by software is used to test the RDMA live migration. It's
successful.
> > > > > >
> > > > > > We will be submit the patchset later.
> > > > >
> > > > > That's great news, thank you!
> > > > >
> > > > > --
> > > > > Peter Xu
> > > >
> > > > For rdma programming, the current mainstream implementation is
> > > > to use
> > rdma_cm to establish a connection, and then use verbs to transmit data.
> > > >
> > > > rdma_cm and ibverbs create two FDs respectively. The two FDs
> > > > have different responsibilities. rdma_cm fd is used to notify
> > > > connection establishment events, and verbs fd is used to notify
> > > > new CQEs. When
> > poll/epoll monitoring is directly performed on the rdma_cm fd, only
> > a pollin event can be monitored, which means that an rdma_cm event
> > occurs. When the verbs fd is directly polled/epolled, only the
> > pollin event can be listened, which indicates that a new CQE is generated.
> > > >
> > > > Rsocket is a sub-module attached to the rdma_cm library and
> > > > provides rdma calls that are completely similar to socket
interfaces.
> > > > However, this library returns only the rdma_cm fd for listening
> > > > to link
> > setup-related events and does not expose the verbs fd (readable and
> > writable events for listening to data). Only the rpoll interface
> > provided by the RSocket can be used to listen to related events.
> > However, QEMU uses the ppoll interface to listen to the rdma_cm fd
(gotten by raccept API).
> > > > And cannot listen to the verbs fd event.
I'm confused, the rs_poll_arm
:https://github.com/linux-rdma/rdma-core/blob/master/librdmacm/rsocket.c#
L3290
For STREAM, rpoll setup fd for both cq fd and cm fd.
Right. But the question is QEMU do not use rpoll but gilb's ppoll. :(
Regards,
-Gonglei