Trace Agent for virtio-trace ============================ Trace agent is a user tool for sending trace data of a guest to a Host in low overhead. Trace agent has the following functions: - splice a page of ring-buffer to read_pipe without memory copying - splice the page from write_pipe to virtio-console without memory copying - write trace data to stdout by using -o option - controlled by start/stop orders from a Host The trace agent operates as follows: 1) Initialize all structures. 2) Create a read/write thread per CPU. Each thread is bound to a CPU. The read/write threads hold it. 3) A controller thread does poll() for a start order of a host. 4) After the controller of the trace agent receives a start order from a host, the controller wake read/write threads. 5) The read/write threads start to read trace data from ring-buffers and write the data to virtio-serial. 6) If the controller receives a stop order from a host, the read/write threads stop to read trace data. Files ===== README: this file Makefile: Makefile of trace agent for virtio-trace trace-agent.c: includes main function, sets up for operating trace agent trace-agent.h: includes all structures and some macros trace-agent-ctl.c: includes controller function for read/write threads trace-agent-rw.c: includes read/write threads function Setup ===== To use this trace agent for virtio-trace, we need to prepare some virtio-serial I/Fs. 1) Make FIFO in a host virtio-trace uses virtio-serial pipe as trace data paths as to the number of CPUs and a control path, so FIFO (named pipe) should be created as follows: # mkdir /tmp/virtio-trace/ # mkfifo /tmp/virtio-trace/trace-path-cpu{0,1,2,...,X}.{in,out} # mkfifo /tmp/virtio-trace/agent-ctl-path.{in,out} For example, if a guest use three CPUs, the names are trace-path-cpu{0,1,2}.{in.out} and agent-ctl-path.{in,out}. 2) Set up of virtio-serial pipe in a host Add qemu option to use virtio-serial pipe. ##virtio-serial device## -device virtio-serial-pci,id=virtio-serial0\ ##control path## -chardev pipe,id=charchannel0,path=/tmp/virtio-trace/agent-ctl-path\ -device virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,\ id=channel0,name=agent-ctl-path\ ##data path## -chardev pipe,id=charchannel1,path=/tmp/virtio-trace/trace-path-cpu0\ -device virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel0,\ id=channel1,name=trace-path-cpu0\ ... If you manage guests with libvirt, add the following tags to domain XML files. Then, libvirt passes the same command option to qemu.
... Here, chardev names are restricted to trace-path-cpuX and agent-ctl-path. For example, if a guest use three CPUs, chardev names should be trace-path-cpu0, trace-path-cpu1, trace-path-cpu2, and agent-ctl-path. 3) Boot the guest You can find some chardev in /dev/virtio-ports/ in the guest. Run === 0) Build trace agent in a guest $ make 1) Enable ftrace in the guest # echo 1 > /sys/kernel/debug/tracing/events/sched/enable 2) Run trace agent in the guest This agent must be operated as root. # ./trace-agent read/write threads in the agent wait for start order from host. If you add -o option, trace data are output via stdout in the guest. 3) Open FIFO in a host # cat /tmp/virtio-trace/trace-path-cpu0.out If a host does not open these, trace data get stuck in buffers of virtio. Then, the guest will stop by specification of chardev in QEMU. This blocking mode may be solved in the future. 4) Start to read trace data by ordering from a host A host injects read start order to the guest via virtio-serial. # echo 1 > /tmp/virtio-trace/agent-ctl-path.in 5) Stop to read trace data by ordering from a host A host injects read stop order to the guest via virtio-serial. # echo 0 > /tmp/virtio-trace/agent-ctl-path.in >space:mode:
authorSriharsha Basavapatna <sriharsha.basavapatna@broadcom.com>2017-01-09 16:00:44 +0530
committerJ. Bruce Fields <bfields@redhat.com>2017-01-12 16:14:47 -0500
commitce1ca7d2d140a1f4aaffd297ac487f246963dd2f (patch)
tree5aeab5e0cb6ef404c894a18251588278b9432a2c /tools/virtio/asm
parent546125d1614264d26080817d0c8cddb9b25081fa (diff)
svcrdma: avoid duplicate dma unmapping during error recovery
In rdma_read_chunk_frmr() when ib_post_send() fails, the error code path invokes ib_dma_unmap_sg() to unmap the sg list. It then invokes svc_rdma_put_frmr() which in turn tries to unmap the same sg list through ib_dma_unmap_sg() again. This second unmap is invalid and could lead to problems when the iova being unmapped is subsequently reused. Remove the call to unmap in rdma_read_chunk_frmr() and let svc_rdma_put_frmr() handle it. Fixes: 412a15c0fe53 ("svcrdma: Port to new memory registration API") Cc: stable@vger.kernel.org Signed-off-by: Sriharsha Basavapatna <sriharsha.basavapatna@broadcom.com> Reviewed-by: Chuck Lever <chuck.lever@oracle.com> Reviewed-by: Yuval Shaia <yuval.shaia@oracle.com> Signed-off-by: J. Bruce Fields <bfields@redhat.com>
Diffstat (limited to 'tools/virtio/asm')