flush_waiters(0),
completed_op_seq(0), queue_op_seq(0) {
- qpair = spdk_nvme_ctrlr_alloc_io_qpair(ctrlr, SPDK_NVME_QPRIO_URGENT);
+ qpair = spdk_nvme_ctrlr_alloc_io_qpair(ctrlr, NULL, SPDK_NVME_QPRIO_URGENT);
PerfCountersBuilder b(g_ceph_context, string("NVMEDevice-AIOThread-"+stringify(this)),
l_bluestore_nvmedevice_first, l_bluestore_nvmedevice_last);
b.add_time_avg(l_bluestore_nvmedevice_aio_write_lat, "aio_write_lat", "Average write completing latency");
if (data_buf_mempool.empty()) {
for (uint16_t i = 0; i < data_buffer_default_num; i++) {
- void *b = spdk_zmalloc(data_buffer_size, CEPH_PAGE_SIZE, NULL);
+ void *b = spdk_dma_zmalloc(data_buffer_size, CEPH_PAGE_SIZE, NULL);
if (!b) {
derr << __func__ << " failed to create memory pool for nvme data buffer" << dendl;
assert(b);
spdk_env_opts_init(&opts);
opts.name = "ceph-osd";
opts.core_mask = coremask_arg;
- opts.dpdk_master_core = m_core_arg;
- opts.dpdk_mem_size = mem_size_arg;
+ opts.master_core = m_core_arg;
+ opts.mem_size = mem_size_arg;
spdk_env_init(&opts);
spdk_nvme_retry_count = g_ceph_context->_conf->bdev_nvme_retry_count;