mirror of https://github.com/xemu-project/xemu.git
hw/rdma: Replace QList by GQueue
RdmaProtectedQList provides a thread-safe queue of int64_t on top of a QList. rdma_protected_qlist_destroy() calls qlist_destroy_obj() directly. qlist_destroy_obj() is actually for use by qobject_destroy() only. The next commit will make that obvious. The minimal fix would be calling qobject_unref() instead. But QList is actually a bad fit here. It's designed for representing JSON arrays. We're better off with a GQueue here. Replace. Cc: Yuval Shaia <yuval.shaia.ml@gmail.com> Cc: Marcel Apfelbaum <marcel.apfelbaum@gmail.com> Signed-off-by: Markus Armbruster <armbru@redhat.com> Message-Id: <20201211171152.146877-8-armbru@redhat.com>
This commit is contained in:
parent
88e25b1e6d
commit
bce800869b
|
@ -78,7 +78,7 @@ static void clean_recv_mads(RdmaBackendDev *backend_dev)
|
||||||
unsigned long cqe_ctx_id;
|
unsigned long cqe_ctx_id;
|
||||||
|
|
||||||
do {
|
do {
|
||||||
cqe_ctx_id = rdma_protected_qlist_pop_int64(&backend_dev->
|
cqe_ctx_id = rdma_protected_gqueue_pop_int64(&backend_dev->
|
||||||
recv_mads_list);
|
recv_mads_list);
|
||||||
if (cqe_ctx_id != -ENOENT) {
|
if (cqe_ctx_id != -ENOENT) {
|
||||||
qatomic_inc(&backend_dev->rdma_dev_res->stats.missing_cqe);
|
qatomic_inc(&backend_dev->rdma_dev_res->stats.missing_cqe);
|
||||||
|
@ -597,7 +597,7 @@ static unsigned int save_mad_recv_buffer(RdmaBackendDev *backend_dev,
|
||||||
bctx->up_ctx = ctx;
|
bctx->up_ctx = ctx;
|
||||||
bctx->sge = *sge;
|
bctx->sge = *sge;
|
||||||
|
|
||||||
rdma_protected_qlist_append_int64(&backend_dev->recv_mads_list, bctx_id);
|
rdma_protected_gqueue_append_int64(&backend_dev->recv_mads_list, bctx_id);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@ -1111,7 +1111,7 @@ static void process_incoming_mad_req(RdmaBackendDev *backend_dev,
|
||||||
|
|
||||||
trace_mad_message("recv", msg->umad.mad, msg->umad_len);
|
trace_mad_message("recv", msg->umad.mad, msg->umad_len);
|
||||||
|
|
||||||
cqe_ctx_id = rdma_protected_qlist_pop_int64(&backend_dev->recv_mads_list);
|
cqe_ctx_id = rdma_protected_gqueue_pop_int64(&backend_dev->recv_mads_list);
|
||||||
if (cqe_ctx_id == -ENOENT) {
|
if (cqe_ctx_id == -ENOENT) {
|
||||||
rdma_warn_report("No more free MADs buffers, waiting for a while");
|
rdma_warn_report("No more free MADs buffers, waiting for a while");
|
||||||
sleep(THR_POLL_TO);
|
sleep(THR_POLL_TO);
|
||||||
|
@ -1185,7 +1185,7 @@ static int mad_init(RdmaBackendDev *backend_dev, CharBackend *mad_chr_be)
|
||||||
return -EIO;
|
return -EIO;
|
||||||
}
|
}
|
||||||
|
|
||||||
rdma_protected_qlist_init(&backend_dev->recv_mads_list);
|
rdma_protected_gqueue_init(&backend_dev->recv_mads_list);
|
||||||
|
|
||||||
enable_rdmacm_mux_async(backend_dev);
|
enable_rdmacm_mux_async(backend_dev);
|
||||||
|
|
||||||
|
@ -1205,7 +1205,7 @@ static void mad_fini(RdmaBackendDev *backend_dev)
|
||||||
{
|
{
|
||||||
disable_rdmacm_mux_async(backend_dev);
|
disable_rdmacm_mux_async(backend_dev);
|
||||||
qemu_chr_fe_disconnect(backend_dev->rdmacm_mux.chr_be);
|
qemu_chr_fe_disconnect(backend_dev->rdmacm_mux.chr_be);
|
||||||
rdma_protected_qlist_destroy(&backend_dev->recv_mads_list);
|
rdma_protected_gqueue_destroy(&backend_dev->recv_mads_list);
|
||||||
}
|
}
|
||||||
|
|
||||||
int rdma_backend_get_gid_index(RdmaBackendDev *backend_dev,
|
int rdma_backend_get_gid_index(RdmaBackendDev *backend_dev,
|
||||||
|
|
|
@ -43,7 +43,7 @@ typedef struct RdmaBackendDev {
|
||||||
struct ibv_context *context;
|
struct ibv_context *context;
|
||||||
struct ibv_comp_channel *channel;
|
struct ibv_comp_channel *channel;
|
||||||
uint8_t port_num;
|
uint8_t port_num;
|
||||||
RdmaProtectedQList recv_mads_list;
|
RdmaProtectedGQueue recv_mads_list;
|
||||||
RdmaCmMux rdmacm_mux;
|
RdmaCmMux rdmacm_mux;
|
||||||
} RdmaBackendDev;
|
} RdmaBackendDev;
|
||||||
|
|
||||||
|
|
|
@ -14,8 +14,6 @@
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include "qemu/osdep.h"
|
#include "qemu/osdep.h"
|
||||||
#include "qapi/qmp/qlist.h"
|
|
||||||
#include "qapi/qmp/qnum.h"
|
|
||||||
#include "trace.h"
|
#include "trace.h"
|
||||||
#include "rdma_utils.h"
|
#include "rdma_utils.h"
|
||||||
|
|
||||||
|
@ -54,41 +52,46 @@ void rdma_pci_dma_unmap(PCIDevice *dev, void *buffer, dma_addr_t len)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void rdma_protected_qlist_init(RdmaProtectedQList *list)
|
void rdma_protected_gqueue_init(RdmaProtectedGQueue *list)
|
||||||
{
|
{
|
||||||
qemu_mutex_init(&list->lock);
|
qemu_mutex_init(&list->lock);
|
||||||
list->list = qlist_new();
|
list->list = g_queue_new();
|
||||||
}
|
}
|
||||||
|
|
||||||
void rdma_protected_qlist_destroy(RdmaProtectedQList *list)
|
void rdma_protected_gqueue_destroy(RdmaProtectedGQueue *list)
|
||||||
{
|
{
|
||||||
if (list->list) {
|
if (list->list) {
|
||||||
qlist_destroy_obj(QOBJECT(list->list));
|
g_queue_free_full(list->list, g_free);
|
||||||
qemu_mutex_destroy(&list->lock);
|
qemu_mutex_destroy(&list->lock);
|
||||||
list->list = NULL;
|
list->list = NULL;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void rdma_protected_qlist_append_int64(RdmaProtectedQList *list, int64_t value)
|
void rdma_protected_gqueue_append_int64(RdmaProtectedGQueue *list,
|
||||||
|
int64_t value)
|
||||||
{
|
{
|
||||||
qemu_mutex_lock(&list->lock);
|
qemu_mutex_lock(&list->lock);
|
||||||
qlist_append_int(list->list, value);
|
g_queue_push_tail(list->list, g_memdup(&value, sizeof(value)));
|
||||||
qemu_mutex_unlock(&list->lock);
|
qemu_mutex_unlock(&list->lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
int64_t rdma_protected_qlist_pop_int64(RdmaProtectedQList *list)
|
int64_t rdma_protected_gqueue_pop_int64(RdmaProtectedGQueue *list)
|
||||||
{
|
{
|
||||||
QObject *obj;
|
int64_t *valp;
|
||||||
|
int64_t val;
|
||||||
|
|
||||||
qemu_mutex_lock(&list->lock);
|
qemu_mutex_lock(&list->lock);
|
||||||
obj = qlist_pop(list->list);
|
|
||||||
|
valp = g_queue_pop_head(list->list);
|
||||||
qemu_mutex_unlock(&list->lock);
|
qemu_mutex_unlock(&list->lock);
|
||||||
|
|
||||||
if (!obj) {
|
if (!valp) {
|
||||||
return -ENOENT;
|
return -ENOENT;
|
||||||
}
|
}
|
||||||
|
|
||||||
return qnum_get_uint(qobject_to(QNum, obj));
|
val = *valp;
|
||||||
|
g_free(valp);
|
||||||
|
return val;
|
||||||
}
|
}
|
||||||
|
|
||||||
void rdma_protected_gslist_init(RdmaProtectedGSList *list)
|
void rdma_protected_gslist_init(RdmaProtectedGSList *list)
|
||||||
|
|
|
@ -28,10 +28,10 @@
|
||||||
#define rdma_info_report(fmt, ...) \
|
#define rdma_info_report(fmt, ...) \
|
||||||
info_report("%s: " fmt, "rdma", ## __VA_ARGS__)
|
info_report("%s: " fmt, "rdma", ## __VA_ARGS__)
|
||||||
|
|
||||||
typedef struct RdmaProtectedQList {
|
typedef struct RdmaProtectedGQueue {
|
||||||
QemuMutex lock;
|
QemuMutex lock;
|
||||||
QList *list;
|
GQueue *list;
|
||||||
} RdmaProtectedQList;
|
} RdmaProtectedGQueue;
|
||||||
|
|
||||||
typedef struct RdmaProtectedGSList {
|
typedef struct RdmaProtectedGSList {
|
||||||
QemuMutex lock;
|
QemuMutex lock;
|
||||||
|
@ -40,10 +40,11 @@ typedef struct RdmaProtectedGSList {
|
||||||
|
|
||||||
void *rdma_pci_dma_map(PCIDevice *dev, dma_addr_t addr, dma_addr_t plen);
|
void *rdma_pci_dma_map(PCIDevice *dev, dma_addr_t addr, dma_addr_t plen);
|
||||||
void rdma_pci_dma_unmap(PCIDevice *dev, void *buffer, dma_addr_t len);
|
void rdma_pci_dma_unmap(PCIDevice *dev, void *buffer, dma_addr_t len);
|
||||||
void rdma_protected_qlist_init(RdmaProtectedQList *list);
|
void rdma_protected_gqueue_init(RdmaProtectedGQueue *list);
|
||||||
void rdma_protected_qlist_destroy(RdmaProtectedQList *list);
|
void rdma_protected_gqueue_destroy(RdmaProtectedGQueue *list);
|
||||||
void rdma_protected_qlist_append_int64(RdmaProtectedQList *list, int64_t value);
|
void rdma_protected_gqueue_append_int64(RdmaProtectedGQueue *list,
|
||||||
int64_t rdma_protected_qlist_pop_int64(RdmaProtectedQList *list);
|
int64_t value);
|
||||||
|
int64_t rdma_protected_gqueue_pop_int64(RdmaProtectedGQueue *list);
|
||||||
void rdma_protected_gslist_init(RdmaProtectedGSList *list);
|
void rdma_protected_gslist_init(RdmaProtectedGSList *list);
|
||||||
void rdma_protected_gslist_destroy(RdmaProtectedGSList *list);
|
void rdma_protected_gslist_destroy(RdmaProtectedGSList *list);
|
||||||
void rdma_protected_gslist_append_int32(RdmaProtectedGSList *list,
|
void rdma_protected_gslist_append_int32(RdmaProtectedGSList *list,
|
||||||
|
|
Loading…
Reference in New Issue