Merge remote-tracking branch 'kwolf/for-anthony' into staging

This commit is contained in:
Anthony Liguori 2011-06-15 09:02:18 -05:00
commit 1f8bedba79
16 changed files with 423 additions and 73 deletions

View File

@ -137,11 +137,12 @@ QEMUBH *qemu_bh_new(QEMUBHFunc *cb, void *opaque)
int qemu_bh_poll(void) int qemu_bh_poll(void)
{ {
QEMUBH *bh, **bhp; QEMUBH *bh, **bhp, *next;
int ret; int ret;
ret = 0; ret = 0;
for (bh = async_context->first_bh; bh; bh = bh->next) { for (bh = async_context->first_bh; bh; bh = next) {
next = bh->next;
if (!bh->deleted && bh->scheduled) { if (!bh->deleted && bh->scheduled) {
bh->scheduled = 0; bh->scheduled = 0;
if (!bh->idle) if (!bh->idle)

View File

@ -110,7 +110,7 @@ int bdrv_check(BlockDriverState *bs, BdrvCheckResult *res);
typedef struct BlockDriverAIOCB BlockDriverAIOCB; typedef struct BlockDriverAIOCB BlockDriverAIOCB;
typedef void BlockDriverCompletionFunc(void *opaque, int ret); typedef void BlockDriverCompletionFunc(void *opaque, int ret);
typedef void BlockDriverDirtyHandler(BlockDriverState *bs, int64_t sector, typedef void BlockDriverDirtyHandler(BlockDriverState *bs, int64_t sector,
int sector_num); int sector_num);
BlockDriverAIOCB *bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num, BlockDriverAIOCB *bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num,
QEMUIOVector *iov, int nb_sectors, QEMUIOVector *iov, int nb_sectors,
BlockDriverCompletionFunc *cb, void *opaque); BlockDriverCompletionFunc *cb, void *opaque);
@ -118,7 +118,7 @@ BlockDriverAIOCB *bdrv_aio_writev(BlockDriverState *bs, int64_t sector_num,
QEMUIOVector *iov, int nb_sectors, QEMUIOVector *iov, int nb_sectors,
BlockDriverCompletionFunc *cb, void *opaque); BlockDriverCompletionFunc *cb, void *opaque);
BlockDriverAIOCB *bdrv_aio_flush(BlockDriverState *bs, BlockDriverAIOCB *bdrv_aio_flush(BlockDriverState *bs,
BlockDriverCompletionFunc *cb, void *opaque); BlockDriverCompletionFunc *cb, void *opaque);
void bdrv_aio_cancel(BlockDriverAIOCB *acb); void bdrv_aio_cancel(BlockDriverAIOCB *acb);
typedef struct BlockRequest { typedef struct BlockRequest {
@ -150,7 +150,7 @@ void bdrv_close_all(void);
int bdrv_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); int bdrv_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors);
int bdrv_has_zero_init(BlockDriverState *bs); int bdrv_has_zero_init(BlockDriverState *bs);
int bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors, int bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors,
int *pnum); int *pnum);
#define BIOS_ATA_TRANSLATION_AUTO 0 #define BIOS_ATA_TRANSLATION_AUTO 0
#define BIOS_ATA_TRANSLATION_NONE 1 #define BIOS_ATA_TRANSLATION_NONE 1

View File

@ -496,6 +496,8 @@ typedef struct QCowAIOCB {
uint64_t cluster_offset; uint64_t cluster_offset;
uint8_t *cluster_data; uint8_t *cluster_data;
struct iovec hd_iov; struct iovec hd_iov;
bool is_write;
QEMUBH *bh;
QEMUIOVector hd_qiov; QEMUIOVector hd_qiov;
BlockDriverAIOCB *hd_aiocb; BlockDriverAIOCB *hd_aiocb;
} QCowAIOCB; } QCowAIOCB;
@ -525,6 +527,8 @@ static QCowAIOCB *qcow_aio_setup(BlockDriverState *bs,
acb->hd_aiocb = NULL; acb->hd_aiocb = NULL;
acb->sector_num = sector_num; acb->sector_num = sector_num;
acb->qiov = qiov; acb->qiov = qiov;
acb->is_write = is_write;
if (qiov->niov > 1) { if (qiov->niov > 1) {
acb->buf = acb->orig_buf = qemu_blockalign(bs, qiov->size); acb->buf = acb->orig_buf = qemu_blockalign(bs, qiov->size);
if (is_write) if (is_write)
@ -538,6 +542,38 @@ static QCowAIOCB *qcow_aio_setup(BlockDriverState *bs,
return acb; return acb;
} }
static void qcow_aio_read_cb(void *opaque, int ret);
static void qcow_aio_write_cb(void *opaque, int ret);
static void qcow_aio_rw_bh(void *opaque)
{
QCowAIOCB *acb = opaque;
qemu_bh_delete(acb->bh);
acb->bh = NULL;
if (acb->is_write) {
qcow_aio_write_cb(opaque, 0);
} else {
qcow_aio_read_cb(opaque, 0);
}
}
static int qcow_schedule_bh(QEMUBHFunc *cb, QCowAIOCB *acb)
{
if (acb->bh) {
return -EIO;
}
acb->bh = qemu_bh_new(cb, acb);
if (!acb->bh) {
return -EIO;
}
qemu_bh_schedule(acb->bh);
return 0;
}
static void qcow_aio_read_cb(void *opaque, int ret) static void qcow_aio_read_cb(void *opaque, int ret)
{ {
QCowAIOCB *acb = opaque; QCowAIOCB *acb = opaque;
@ -640,12 +676,21 @@ static BlockDriverAIOCB *qcow_aio_readv(BlockDriverState *bs,
BlockDriverCompletionFunc *cb, void *opaque) BlockDriverCompletionFunc *cb, void *opaque)
{ {
QCowAIOCB *acb; QCowAIOCB *acb;
int ret;
acb = qcow_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 0); acb = qcow_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 0);
if (!acb) if (!acb)
return NULL; return NULL;
qcow_aio_read_cb(acb, 0); ret = qcow_schedule_bh(qcow_aio_rw_bh, acb);
if (ret < 0) {
if (acb->qiov->niov > 1) {
qemu_vfree(acb->orig_buf);
}
qemu_aio_release(acb);
return NULL;
}
return &acb->common; return &acb->common;
} }
@ -725,6 +770,7 @@ static BlockDriverAIOCB *qcow_aio_writev(BlockDriverState *bs,
{ {
BDRVQcowState *s = bs->opaque; BDRVQcowState *s = bs->opaque;
QCowAIOCB *acb; QCowAIOCB *acb;
int ret;
s->cluster_cache_offset = -1; /* disable compressed cache */ s->cluster_cache_offset = -1; /* disable compressed cache */
@ -733,7 +779,15 @@ static BlockDriverAIOCB *qcow_aio_writev(BlockDriverState *bs,
return NULL; return NULL;
qcow_aio_write_cb(acb, 0); ret = qcow_schedule_bh(qcow_aio_rw_bh, acb);
if (ret < 0) {
if (acb->qiov->niov > 1) {
qemu_vfree(acb->orig_buf);
}
qemu_aio_release(acb);
return NULL;
}
return &acb->common; return &acb->common;
} }

View File

@ -796,8 +796,8 @@ int qcow2_alloc_cluster_offset(BlockDriverState *bs, uint64_t offset,
m->depends_on = old_alloc; m->depends_on = old_alloc;
m->nb_clusters = 0; m->nb_clusters = 0;
*num = 0; *num = 0;
ret = 0;
goto fail; goto out_wait_dependency;
} }
} }
} }
@ -812,7 +812,6 @@ int qcow2_alloc_cluster_offset(BlockDriverState *bs, uint64_t offset,
cluster_offset = qcow2_alloc_clusters(bs, nb_clusters * s->cluster_size); cluster_offset = qcow2_alloc_clusters(bs, nb_clusters * s->cluster_size);
if (cluster_offset < 0) { if (cluster_offset < 0) {
QLIST_REMOVE(m, next_in_flight);
ret = cluster_offset; ret = cluster_offset;
goto fail; goto fail;
} }
@ -825,7 +824,7 @@ int qcow2_alloc_cluster_offset(BlockDriverState *bs, uint64_t offset,
out: out:
ret = qcow2_cache_put(bs, s->l2_table_cache, (void**) &l2_table); ret = qcow2_cache_put(bs, s->l2_table_cache, (void**) &l2_table);
if (ret < 0) { if (ret < 0) {
return ret; goto fail_put;
} }
m->nb_available = MIN(nb_clusters << (s->cluster_bits - 9), n_end); m->nb_available = MIN(nb_clusters << (s->cluster_bits - 9), n_end);
@ -835,8 +834,13 @@ out:
return 0; return 0;
out_wait_dependency:
return qcow2_cache_put(bs, s->l2_table_cache, (void**) &l2_table);
fail: fail:
qcow2_cache_put(bs, s->l2_table_cache, (void**) &l2_table); qcow2_cache_put(bs, s->l2_table_cache, (void**) &l2_table);
fail_put:
QLIST_REMOVE(m, next_in_flight);
return ret; return ret;
} }

View File

@ -378,6 +378,7 @@ typedef struct QCowAIOCB {
uint64_t bytes_done; uint64_t bytes_done;
uint64_t cluster_offset; uint64_t cluster_offset;
uint8_t *cluster_data; uint8_t *cluster_data;
bool is_write;
BlockDriverAIOCB *hd_aiocb; BlockDriverAIOCB *hd_aiocb;
QEMUIOVector hd_qiov; QEMUIOVector hd_qiov;
QEMUBH *bh; QEMUBH *bh;
@ -399,12 +400,19 @@ static AIOPool qcow2_aio_pool = {
}; };
static void qcow2_aio_read_cb(void *opaque, int ret); static void qcow2_aio_read_cb(void *opaque, int ret);
static void qcow2_aio_read_bh(void *opaque) static void qcow2_aio_write_cb(void *opaque, int ret);
static void qcow2_aio_rw_bh(void *opaque)
{ {
QCowAIOCB *acb = opaque; QCowAIOCB *acb = opaque;
qemu_bh_delete(acb->bh); qemu_bh_delete(acb->bh);
acb->bh = NULL; acb->bh = NULL;
qcow2_aio_read_cb(opaque, 0);
if (acb->is_write) {
qcow2_aio_write_cb(opaque, 0);
} else {
qcow2_aio_read_cb(opaque, 0);
}
} }
static int qcow2_schedule_bh(QEMUBHFunc *cb, QCowAIOCB *acb) static int qcow2_schedule_bh(QEMUBHFunc *cb, QCowAIOCB *acb)
@ -493,14 +501,14 @@ static void qcow2_aio_read_cb(void *opaque, int ret)
goto done; goto done;
} }
} else { } else {
ret = qcow2_schedule_bh(qcow2_aio_read_bh, acb); ret = qcow2_schedule_bh(qcow2_aio_rw_bh, acb);
if (ret < 0) if (ret < 0)
goto done; goto done;
} }
} else { } else {
/* Note: in this case, no need to wait */ /* Note: in this case, no need to wait */
qemu_iovec_memset(&acb->hd_qiov, 0, 512 * acb->cur_nr_sectors); qemu_iovec_memset(&acb->hd_qiov, 0, 512 * acb->cur_nr_sectors);
ret = qcow2_schedule_bh(qcow2_aio_read_bh, acb); ret = qcow2_schedule_bh(qcow2_aio_rw_bh, acb);
if (ret < 0) if (ret < 0)
goto done; goto done;
} }
@ -515,7 +523,7 @@ static void qcow2_aio_read_cb(void *opaque, int ret)
s->cluster_cache + index_in_cluster * 512, s->cluster_cache + index_in_cluster * 512,
512 * acb->cur_nr_sectors); 512 * acb->cur_nr_sectors);
ret = qcow2_schedule_bh(qcow2_aio_read_bh, acb); ret = qcow2_schedule_bh(qcow2_aio_rw_bh, acb);
if (ret < 0) if (ret < 0)
goto done; goto done;
} else { } else {
@ -572,6 +580,7 @@ static QCowAIOCB *qcow2_aio_setup(BlockDriverState *bs, int64_t sector_num,
acb->hd_aiocb = NULL; acb->hd_aiocb = NULL;
acb->sector_num = sector_num; acb->sector_num = sector_num;
acb->qiov = qiov; acb->qiov = qiov;
acb->is_write = is_write;
qemu_iovec_init(&acb->hd_qiov, qiov->niov); qemu_iovec_init(&acb->hd_qiov, qiov->niov);
@ -591,17 +600,22 @@ static BlockDriverAIOCB *qcow2_aio_readv(BlockDriverState *bs,
void *opaque) void *opaque)
{ {
QCowAIOCB *acb; QCowAIOCB *acb;
int ret;
acb = qcow2_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 0); acb = qcow2_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 0);
if (!acb) if (!acb)
return NULL; return NULL;
qcow2_aio_read_cb(acb, 0); ret = qcow2_schedule_bh(qcow2_aio_rw_bh, acb);
if (ret < 0) {
qemu_iovec_destroy(&acb->hd_qiov);
qemu_aio_release(acb);
return NULL;
}
return &acb->common; return &acb->common;
} }
static void qcow2_aio_write_cb(void *opaque, int ret);
static void run_dependent_requests(QCowL2Meta *m) static void run_dependent_requests(QCowL2Meta *m)
{ {
QCowAIOCB *req; QCowAIOCB *req;
@ -724,6 +738,7 @@ static BlockDriverAIOCB *qcow2_aio_writev(BlockDriverState *bs,
{ {
BDRVQcowState *s = bs->opaque; BDRVQcowState *s = bs->opaque;
QCowAIOCB *acb; QCowAIOCB *acb;
int ret;
s->cluster_cache_offset = -1; /* disable compressed cache */ s->cluster_cache_offset = -1; /* disable compressed cache */
@ -731,7 +746,13 @@ static BlockDriverAIOCB *qcow2_aio_writev(BlockDriverState *bs,
if (!acb) if (!acb)
return NULL; return NULL;
qcow2_aio_write_cb(acb, 0); ret = qcow2_schedule_bh(qcow2_aio_rw_bh, acb);
if (ret < 0) {
qemu_iovec_destroy(&acb->hd_qiov);
qemu_aio_release(acb);
return NULL;
}
return &acb->common; return &acb->common;
} }

View File

@ -227,7 +227,6 @@ static int qemu_rbd_create(const char *filename, QEMUOptionParameter *options)
char name[RBD_MAX_IMAGE_NAME_SIZE]; char name[RBD_MAX_IMAGE_NAME_SIZE];
char snap_buf[RBD_MAX_SNAP_NAME_SIZE]; char snap_buf[RBD_MAX_SNAP_NAME_SIZE];
char conf[RBD_MAX_CONF_SIZE]; char conf[RBD_MAX_CONF_SIZE];
char *snap = NULL;
rados_t cluster; rados_t cluster;
rados_ioctx_t io_ctx; rados_ioctx_t io_ctx;
int ret; int ret;
@ -238,9 +237,6 @@ static int qemu_rbd_create(const char *filename, QEMUOptionParameter *options)
conf, sizeof(conf)) < 0) { conf, sizeof(conf)) < 0) {
return -EINVAL; return -EINVAL;
} }
if (snap_buf[0] != '\0') {
snap = snap_buf;
}
/* Read out options */ /* Read out options */
while (options && options->name) { while (options && options->name) {

View File

@ -152,6 +152,7 @@ typedef struct {
/* Buffer for new allocated block. */ /* Buffer for new allocated block. */
void *block_buffer; void *block_buffer;
void *orig_buf; void *orig_buf;
bool is_write;
int header_modified; int header_modified;
BlockDriverAIOCB *hd_aiocb; BlockDriverAIOCB *hd_aiocb;
struct iovec hd_iov; struct iovec hd_iov;
@ -504,6 +505,8 @@ static VdiAIOCB *vdi_aio_setup(BlockDriverState *bs, int64_t sector_num,
acb->hd_aiocb = NULL; acb->hd_aiocb = NULL;
acb->sector_num = sector_num; acb->sector_num = sector_num;
acb->qiov = qiov; acb->qiov = qiov;
acb->is_write = is_write;
if (qiov->niov > 1) { if (qiov->niov > 1) {
acb->buf = qemu_blockalign(bs, qiov->size); acb->buf = qemu_blockalign(bs, qiov->size);
acb->orig_buf = acb->buf; acb->orig_buf = acb->buf;
@ -542,14 +545,20 @@ static int vdi_schedule_bh(QEMUBHFunc *cb, VdiAIOCB *acb)
} }
static void vdi_aio_read_cb(void *opaque, int ret); static void vdi_aio_read_cb(void *opaque, int ret);
static void vdi_aio_write_cb(void *opaque, int ret);
static void vdi_aio_read_bh(void *opaque) static void vdi_aio_rw_bh(void *opaque)
{ {
VdiAIOCB *acb = opaque; VdiAIOCB *acb = opaque;
logout("\n"); logout("\n");
qemu_bh_delete(acb->bh); qemu_bh_delete(acb->bh);
acb->bh = NULL; acb->bh = NULL;
vdi_aio_read_cb(opaque, 0);
if (acb->is_write) {
vdi_aio_write_cb(opaque, 0);
} else {
vdi_aio_read_cb(opaque, 0);
}
} }
static void vdi_aio_read_cb(void *opaque, int ret) static void vdi_aio_read_cb(void *opaque, int ret)
@ -597,7 +606,7 @@ static void vdi_aio_read_cb(void *opaque, int ret)
if (bmap_entry == VDI_UNALLOCATED) { if (bmap_entry == VDI_UNALLOCATED) {
/* Block not allocated, return zeros, no need to wait. */ /* Block not allocated, return zeros, no need to wait. */
memset(acb->buf, 0, n_sectors * SECTOR_SIZE); memset(acb->buf, 0, n_sectors * SECTOR_SIZE);
ret = vdi_schedule_bh(vdi_aio_read_bh, acb); ret = vdi_schedule_bh(vdi_aio_rw_bh, acb);
if (ret < 0) { if (ret < 0) {
goto done; goto done;
} }
@ -630,12 +639,23 @@ static BlockDriverAIOCB *vdi_aio_readv(BlockDriverState *bs,
BlockDriverCompletionFunc *cb, void *opaque) BlockDriverCompletionFunc *cb, void *opaque)
{ {
VdiAIOCB *acb; VdiAIOCB *acb;
int ret;
logout("\n"); logout("\n");
acb = vdi_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 0); acb = vdi_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 0);
if (!acb) { if (!acb) {
return NULL; return NULL;
} }
vdi_aio_read_cb(acb, 0);
ret = vdi_schedule_bh(vdi_aio_rw_bh, acb);
if (ret < 0) {
if (acb->qiov->niov > 1) {
qemu_vfree(acb->orig_buf);
}
qemu_aio_release(acb);
return NULL;
}
return &acb->common; return &acb->common;
} }
@ -789,12 +809,23 @@ static BlockDriverAIOCB *vdi_aio_writev(BlockDriverState *bs,
BlockDriverCompletionFunc *cb, void *opaque) BlockDriverCompletionFunc *cb, void *opaque)
{ {
VdiAIOCB *acb; VdiAIOCB *acb;
int ret;
logout("\n"); logout("\n");
acb = vdi_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 1); acb = vdi_aio_setup(bs, sector_num, qiov, nb_sectors, cb, opaque, 1);
if (!acb) { if (!acb) {
return NULL; return NULL;
} }
vdi_aio_write_cb(acb, 0);
ret = vdi_schedule_bh(vdi_aio_rw_bh, acb);
if (ret < 0) {
if (acb->qiov->niov > 1) {
qemu_vfree(acb->orig_buf);
}
qemu_aio_release(acb);
return NULL;
}
return &acb->common; return &acb->common;
} }

View File

@ -203,8 +203,8 @@ struct BlockDriverState {
void *private; void *private;
}; };
#define CHANGE_MEDIA 0x01 #define CHANGE_MEDIA 0x01
#define CHANGE_SIZE 0x02 #define CHANGE_SIZE 0x02
struct BlockDriverAIOCB { struct BlockDriverAIOCB {
AIOPool *pool; AIOPool *pool;

View File

@ -47,6 +47,7 @@ typedef struct {
target_phys_addr_t sg_cur_byte; target_phys_addr_t sg_cur_byte;
QEMUIOVector iov; QEMUIOVector iov;
QEMUBH *bh; QEMUBH *bh;
DMAIOFunc *io_func;
} DMAAIOCB; } DMAAIOCB;
static void dma_bdrv_cb(void *opaque, int ret); static void dma_bdrv_cb(void *opaque, int ret);
@ -116,13 +117,8 @@ static void dma_bdrv_cb(void *opaque, int ret)
return; return;
} }
if (dbs->is_write) { dbs->acb = dbs->io_func(dbs->bs, dbs->sector_num, &dbs->iov,
dbs->acb = bdrv_aio_writev(dbs->bs, dbs->sector_num, &dbs->iov, dbs->iov.size / 512, dma_bdrv_cb, dbs);
dbs->iov.size / 512, dma_bdrv_cb, dbs);
} else {
dbs->acb = bdrv_aio_readv(dbs->bs, dbs->sector_num, &dbs->iov,
dbs->iov.size / 512, dma_bdrv_cb, dbs);
}
if (!dbs->acb) { if (!dbs->acb) {
dma_bdrv_unmap(dbs); dma_bdrv_unmap(dbs);
qemu_iovec_destroy(&dbs->iov); qemu_iovec_destroy(&dbs->iov);
@ -144,12 +140,12 @@ static AIOPool dma_aio_pool = {
.cancel = dma_aio_cancel, .cancel = dma_aio_cancel,
}; };
static BlockDriverAIOCB *dma_bdrv_io( BlockDriverAIOCB *dma_bdrv_io(
BlockDriverState *bs, QEMUSGList *sg, uint64_t sector_num, BlockDriverState *bs, QEMUSGList *sg, uint64_t sector_num,
BlockDriverCompletionFunc *cb, void *opaque, DMAIOFunc *io_func, BlockDriverCompletionFunc *cb,
int is_write) void *opaque, int is_write)
{ {
DMAAIOCB *dbs = qemu_aio_get(&dma_aio_pool, bs, cb, opaque); DMAAIOCB *dbs = qemu_aio_get(&dma_aio_pool, bs, cb, opaque);
dbs->acb = NULL; dbs->acb = NULL;
dbs->bs = bs; dbs->bs = bs;
@ -158,6 +154,7 @@ static BlockDriverAIOCB *dma_bdrv_io(
dbs->sg_cur_index = 0; dbs->sg_cur_index = 0;
dbs->sg_cur_byte = 0; dbs->sg_cur_byte = 0;
dbs->is_write = is_write; dbs->is_write = is_write;
dbs->io_func = io_func;
dbs->bh = NULL; dbs->bh = NULL;
qemu_iovec_init(&dbs->iov, sg->nsg); qemu_iovec_init(&dbs->iov, sg->nsg);
dma_bdrv_cb(dbs, 0); dma_bdrv_cb(dbs, 0);
@ -173,12 +170,12 @@ BlockDriverAIOCB *dma_bdrv_read(BlockDriverState *bs,
QEMUSGList *sg, uint64_t sector, QEMUSGList *sg, uint64_t sector,
void (*cb)(void *opaque, int ret), void *opaque) void (*cb)(void *opaque, int ret), void *opaque)
{ {
return dma_bdrv_io(bs, sg, sector, cb, opaque, 0); return dma_bdrv_io(bs, sg, sector, bdrv_aio_readv, cb, opaque, 0);
} }
BlockDriverAIOCB *dma_bdrv_write(BlockDriverState *bs, BlockDriverAIOCB *dma_bdrv_write(BlockDriverState *bs,
QEMUSGList *sg, uint64_t sector, QEMUSGList *sg, uint64_t sector,
void (*cb)(void *opaque, int ret), void *opaque) void (*cb)(void *opaque, int ret), void *opaque)
{ {
return dma_bdrv_io(bs, sg, sector, cb, opaque, 1); return dma_bdrv_io(bs, sg, sector, bdrv_aio_writev, cb, opaque, 1);
} }

8
dma.h
View File

@ -32,6 +32,14 @@ void qemu_sglist_add(QEMUSGList *qsg, target_phys_addr_t base,
target_phys_addr_t len); target_phys_addr_t len);
void qemu_sglist_destroy(QEMUSGList *qsg); void qemu_sglist_destroy(QEMUSGList *qsg);
typedef BlockDriverAIOCB *DMAIOFunc(BlockDriverState *bs, int64_t sector_num,
QEMUIOVector *iov, int nb_sectors,
BlockDriverCompletionFunc *cb, void *opaque);
BlockDriverAIOCB *dma_bdrv_io(BlockDriverState *bs,
QEMUSGList *sg, uint64_t sector_num,
DMAIOFunc *io_func, BlockDriverCompletionFunc *cb,
void *opaque, int is_write);
BlockDriverAIOCB *dma_bdrv_read(BlockDriverState *bs, BlockDriverAIOCB *dma_bdrv_read(BlockDriverState *bs,
QEMUSGList *sg, uint64_t sector, QEMUSGList *sg, uint64_t sector,
BlockDriverCompletionFunc *cb, void *opaque); BlockDriverCompletionFunc *cb, void *opaque);

View File

@ -78,7 +78,7 @@ static void ide_identify(IDEState *s)
{ {
uint16_t *p; uint16_t *p;
unsigned int oldsize; unsigned int oldsize;
IDEDevice *dev; IDEDevice *dev = s->unit ? s->bus->slave : s->bus->master;
if (s->identify_set) { if (s->identify_set) {
memcpy(s->io_buffer, s->identify_data, sizeof(s->identify_data)); memcpy(s->io_buffer, s->identify_data, sizeof(s->identify_data));
@ -124,6 +124,9 @@ static void ide_identify(IDEState *s)
put_le16(p + 66, 120); put_le16(p + 66, 120);
put_le16(p + 67, 120); put_le16(p + 67, 120);
put_le16(p + 68, 120); put_le16(p + 68, 120);
if (dev && dev->conf.discard_granularity) {
put_le16(p + 69, (1 << 14)); /* determinate TRIM behavior */
}
if (s->ncq_queues) { if (s->ncq_queues) {
put_le16(p + 75, s->ncq_queues - 1); put_le16(p + 75, s->ncq_queues - 1);
@ -154,9 +157,12 @@ static void ide_identify(IDEState *s)
put_le16(p + 101, s->nb_sectors >> 16); put_le16(p + 101, s->nb_sectors >> 16);
put_le16(p + 102, s->nb_sectors >> 32); put_le16(p + 102, s->nb_sectors >> 32);
put_le16(p + 103, s->nb_sectors >> 48); put_le16(p + 103, s->nb_sectors >> 48);
dev = s->unit ? s->bus->slave : s->bus->master;
if (dev && dev->conf.physical_block_size) if (dev && dev->conf.physical_block_size)
put_le16(p + 106, 0x6000 | get_physical_block_exp(&dev->conf)); put_le16(p + 106, 0x6000 | get_physical_block_exp(&dev->conf));
if (dev && dev->conf.discard_granularity) {
put_le16(p + 169, 1); /* TRIM support */
}
memcpy(s->identify_data, p, sizeof(s->identify_data)); memcpy(s->identify_data, p, sizeof(s->identify_data));
s->identify_set = 1; s->identify_set = 1;
@ -299,6 +305,74 @@ static void ide_set_signature(IDEState *s)
} }
} }
typedef struct TrimAIOCB {
BlockDriverAIOCB common;
QEMUBH *bh;
int ret;
} TrimAIOCB;
static void trim_aio_cancel(BlockDriverAIOCB *acb)
{
TrimAIOCB *iocb = container_of(acb, TrimAIOCB, common);
qemu_bh_delete(iocb->bh);
iocb->bh = NULL;
qemu_aio_release(iocb);
}
static AIOPool trim_aio_pool = {
.aiocb_size = sizeof(TrimAIOCB),
.cancel = trim_aio_cancel,
};
static void ide_trim_bh_cb(void *opaque)
{
TrimAIOCB *iocb = opaque;
iocb->common.cb(iocb->common.opaque, iocb->ret);
qemu_bh_delete(iocb->bh);
iocb->bh = NULL;
qemu_aio_release(iocb);
}
BlockDriverAIOCB *ide_issue_trim(BlockDriverState *bs,
int64_t sector_num, QEMUIOVector *qiov, int nb_sectors,
BlockDriverCompletionFunc *cb, void *opaque)
{
TrimAIOCB *iocb;
int i, j, ret;
iocb = qemu_aio_get(&trim_aio_pool, bs, cb, opaque);
iocb->bh = qemu_bh_new(ide_trim_bh_cb, iocb);
iocb->ret = 0;
for (j = 0; j < qiov->niov; j++) {
uint64_t *buffer = qiov->iov[j].iov_base;
for (i = 0; i < qiov->iov[j].iov_len / 8; i++) {
/* 6-byte LBA + 2-byte range per entry */
uint64_t entry = le64_to_cpu(buffer[i]);
uint64_t sector = entry & 0x0000ffffffffffffULL;
uint16_t count = entry >> 48;
if (count == 0) {
break;
}
ret = bdrv_discard(bs, sector, count);
if (!iocb->ret) {
iocb->ret = ret;
}
}
}
qemu_bh_schedule(iocb->bh);
return &iocb->common;
}
static inline void ide_abort_command(IDEState *s) static inline void ide_abort_command(IDEState *s)
{ {
s->status = READY_STAT | ERR_STAT; s->status = READY_STAT | ERR_STAT;
@ -446,7 +520,7 @@ static int ide_handle_rw_error(IDEState *s, int error, int op)
if ((error == ENOSPC && action == BLOCK_ERR_STOP_ENOSPC) if ((error == ENOSPC && action == BLOCK_ERR_STOP_ENOSPC)
|| action == BLOCK_ERR_STOP_ANY) { || action == BLOCK_ERR_STOP_ANY) {
s->bus->dma->ops->set_unit(s->bus->dma, s->unit); s->bus->dma->ops->set_unit(s->bus->dma, s->unit);
s->bus->dma->ops->add_status(s->bus->dma, op); s->bus->error_status = op;
bdrv_mon_event(s->bs, BDRV_ACTION_STOP, is_read); bdrv_mon_event(s->bs, BDRV_ACTION_STOP, is_read);
vm_stop(VMSTOP_DISKFULL); vm_stop(VMSTOP_DISKFULL);
} else { } else {
@ -472,8 +546,11 @@ handle_rw_error:
if (ret < 0) { if (ret < 0) {
int op = BM_STATUS_DMA_RETRY; int op = BM_STATUS_DMA_RETRY;
if (s->is_read) if (s->dma_cmd == IDE_DMA_READ)
op |= BM_STATUS_RETRY_READ; op |= BM_STATUS_RETRY_READ;
else if (s->dma_cmd == IDE_DMA_TRIM)
op |= BM_STATUS_RETRY_TRIM;
if (ide_handle_rw_error(s, -ret, op)) { if (ide_handle_rw_error(s, -ret, op)) {
return; return;
} }
@ -482,7 +559,7 @@ handle_rw_error:
n = s->io_buffer_size >> 9; n = s->io_buffer_size >> 9;
sector_num = ide_get_sector(s); sector_num = ide_get_sector(s);
if (n > 0) { if (n > 0) {
dma_buf_commit(s, s->is_read); dma_buf_commit(s, ide_cmd_is_read(s));
sector_num += n; sector_num += n;
ide_set_sector(s, sector_num); ide_set_sector(s, sector_num);
s->nsector -= n; s->nsector -= n;
@ -499,23 +576,30 @@ handle_rw_error:
n = s->nsector; n = s->nsector;
s->io_buffer_index = 0; s->io_buffer_index = 0;
s->io_buffer_size = n * 512; s->io_buffer_size = n * 512;
if (s->bus->dma->ops->prepare_buf(s->bus->dma, s->is_read) == 0) { if (s->bus->dma->ops->prepare_buf(s->bus->dma, ide_cmd_is_read(s)) == 0) {
/* The PRDs were too short. Reset the Active bit, but don't raise an /* The PRDs were too short. Reset the Active bit, but don't raise an
* interrupt. */ * interrupt. */
goto eot; goto eot;
} }
#ifdef DEBUG_AIO #ifdef DEBUG_AIO
printf("ide_dma_cb: sector_num=%" PRId64 " n=%d, is_read=%d\n", printf("ide_dma_cb: sector_num=%" PRId64 " n=%d, cmd_cmd=%d\n",
sector_num, n, s->is_read); sector_num, n, s->dma_cmd);
#endif #endif
if (s->is_read) { switch (s->dma_cmd) {
case IDE_DMA_READ:
s->bus->dma->aiocb = dma_bdrv_read(s->bs, &s->sg, sector_num, s->bus->dma->aiocb = dma_bdrv_read(s->bs, &s->sg, sector_num,
ide_dma_cb, s); ide_dma_cb, s);
} else { break;
case IDE_DMA_WRITE:
s->bus->dma->aiocb = dma_bdrv_write(s->bs, &s->sg, sector_num, s->bus->dma->aiocb = dma_bdrv_write(s->bs, &s->sg, sector_num,
ide_dma_cb, s); ide_dma_cb, s);
break;
case IDE_DMA_TRIM:
s->bus->dma->aiocb = dma_bdrv_io(s->bs, &s->sg, sector_num,
ide_issue_trim, ide_dma_cb, s, 1);
break;
} }
if (!s->bus->dma->aiocb) { if (!s->bus->dma->aiocb) {
@ -528,12 +612,12 @@ eot:
ide_set_inactive(s); ide_set_inactive(s);
} }
static void ide_sector_start_dma(IDEState *s, int is_read) static void ide_sector_start_dma(IDEState *s, enum ide_dma_cmd dma_cmd)
{ {
s->status = READY_STAT | SEEK_STAT | DRQ_STAT | BUSY_STAT; s->status = READY_STAT | SEEK_STAT | DRQ_STAT | BUSY_STAT;
s->io_buffer_index = 0; s->io_buffer_index = 0;
s->io_buffer_size = 0; s->io_buffer_size = 0;
s->is_read = is_read; s->dma_cmd = dma_cmd;
s->bus->dma->ops->start_dma(s->bus->dma, s, ide_dma_cb); s->bus->dma->ops->start_dma(s->bus->dma, s, ide_dma_cb);
} }
@ -815,6 +899,18 @@ void ide_exec_cmd(IDEBus *bus, uint32_t val)
return; return;
switch(val) { switch(val) {
case WIN_DSM:
switch (s->feature) {
case DSM_TRIM:
if (!s->bs) {
goto abort_cmd;
}
ide_sector_start_dma(s, IDE_DMA_TRIM);
break;
default:
goto abort_cmd;
}
break;
case WIN_IDENTIFY: case WIN_IDENTIFY:
if (s->bs && s->drive_kind != IDE_CD) { if (s->bs && s->drive_kind != IDE_CD) {
if (s->drive_kind != IDE_CFATA) if (s->drive_kind != IDE_CFATA)
@ -916,7 +1012,7 @@ void ide_exec_cmd(IDEBus *bus, uint32_t val)
if (!s->bs) if (!s->bs)
goto abort_cmd; goto abort_cmd;
ide_cmd_lba48_transform(s, lba48); ide_cmd_lba48_transform(s, lba48);
ide_sector_start_dma(s, 1); ide_sector_start_dma(s, IDE_DMA_READ);
break; break;
case WIN_WRITEDMA_EXT: case WIN_WRITEDMA_EXT:
lba48 = 1; lba48 = 1;
@ -925,7 +1021,7 @@ void ide_exec_cmd(IDEBus *bus, uint32_t val)
if (!s->bs) if (!s->bs)
goto abort_cmd; goto abort_cmd;
ide_cmd_lba48_transform(s, lba48); ide_cmd_lba48_transform(s, lba48);
ide_sector_start_dma(s, 0); ide_sector_start_dma(s, IDE_DMA_WRITE);
s->media_changed = 1; s->media_changed = 1;
break; break;
case WIN_READ_NATIVE_MAX_EXT: case WIN_READ_NATIVE_MAX_EXT:
@ -1837,7 +1933,8 @@ static bool ide_drive_pio_state_needed(void *opaque)
{ {
IDEState *s = opaque; IDEState *s = opaque;
return (s->status & DRQ_STAT) != 0; return ((s->status & DRQ_STAT) != 0)
|| (s->bus->error_status & BM_STATUS_PIO_RETRY);
} }
static bool ide_atapi_gesn_needed(void *opaque) static bool ide_atapi_gesn_needed(void *opaque)
@ -1847,6 +1944,13 @@ static bool ide_atapi_gesn_needed(void *opaque)
return s->events.new_media || s->events.eject_request; return s->events.new_media || s->events.eject_request;
} }
static bool ide_error_needed(void *opaque)
{
IDEBus *bus = opaque;
return (bus->error_status != 0);
}
/* Fields for GET_EVENT_STATUS_NOTIFICATION ATAPI command */ /* Fields for GET_EVENT_STATUS_NOTIFICATION ATAPI command */
const VMStateDescription vmstate_ide_atapi_gesn_state = { const VMStateDescription vmstate_ide_atapi_gesn_state = {
.name ="ide_drive/atapi/gesn_state", .name ="ide_drive/atapi/gesn_state",
@ -1856,6 +1960,7 @@ const VMStateDescription vmstate_ide_atapi_gesn_state = {
.fields = (VMStateField []) { .fields = (VMStateField []) {
VMSTATE_BOOL(events.new_media, IDEState), VMSTATE_BOOL(events.new_media, IDEState),
VMSTATE_BOOL(events.eject_request, IDEState), VMSTATE_BOOL(events.eject_request, IDEState),
VMSTATE_END_OF_LIST()
} }
}; };
@ -1921,6 +2026,17 @@ const VMStateDescription vmstate_ide_drive = {
} }
}; };
const VMStateDescription vmstate_ide_error_status = {
.name ="ide_bus/error",
.version_id = 1,
.minimum_version_id = 1,
.minimum_version_id_old = 1,
.fields = (VMStateField []) {
VMSTATE_INT32(error_status, IDEBus),
VMSTATE_END_OF_LIST()
}
};
const VMStateDescription vmstate_ide_bus = { const VMStateDescription vmstate_ide_bus = {
.name = "ide_bus", .name = "ide_bus",
.version_id = 1, .version_id = 1,
@ -1930,6 +2046,14 @@ const VMStateDescription vmstate_ide_bus = {
VMSTATE_UINT8(cmd, IDEBus), VMSTATE_UINT8(cmd, IDEBus),
VMSTATE_UINT8(unit, IDEBus), VMSTATE_UINT8(unit, IDEBus),
VMSTATE_END_OF_LIST() VMSTATE_END_OF_LIST()
},
.subsections = (VMStateSubsection []) {
{
.vmsd = &vmstate_ide_error_status,
.needed = ide_error_needed,
}, {
/* empty */
}
} }
}; };

View File

@ -62,7 +62,11 @@ typedef struct IDEDMAOps IDEDMAOps;
*/ */
#define CFA_REQ_EXT_ERROR_CODE 0x03 /* CFA Request Extended Error Code */ #define CFA_REQ_EXT_ERROR_CODE 0x03 /* CFA Request Extended Error Code */
/* /*
* 0x04->0x07 Reserved * 0x04->0x05 Reserved
*/
#define WIN_DSM 0x06
/*
* 0x07 Reserved
*/ */
#define WIN_SRST 0x08 /* ATAPI soft reset command */ #define WIN_SRST 0x08 /* ATAPI soft reset command */
#define WIN_DEVICE_RESET 0x08 #define WIN_DEVICE_RESET 0x08
@ -190,6 +194,9 @@ typedef struct IDEDMAOps IDEDMAOps;
#define IDE_DMA_BUF_SECTORS 256 #define IDE_DMA_BUF_SECTORS 256
/* feature values for Data Set Management */
#define DSM_TRIM 0x01
#if (IDE_DMA_BUF_SECTORS < MAX_MULT_SECTORS) #if (IDE_DMA_BUF_SECTORS < MAX_MULT_SECTORS)
#error "IDE_DMA_BUF_SECTORS must be bigger or equal to MAX_MULT_SECTORS" #error "IDE_DMA_BUF_SECTORS must be bigger or equal to MAX_MULT_SECTORS"
#endif #endif
@ -379,6 +386,15 @@ struct unreported_events {
bool new_media; bool new_media;
}; };
enum ide_dma_cmd {
IDE_DMA_READ,
IDE_DMA_WRITE,
IDE_DMA_TRIM,
};
#define ide_cmd_is_read(s) \
((s)->dma_cmd == IDE_DMA_READ)
/* NOTE: IDEState represents in fact one drive */ /* NOTE: IDEState represents in fact one drive */
struct IDEState { struct IDEState {
IDEBus *bus; IDEBus *bus;
@ -446,7 +462,7 @@ struct IDEState {
uint32_t mdata_size; uint32_t mdata_size;
uint8_t *mdata_storage; uint8_t *mdata_storage;
int media_changed; int media_changed;
int is_read; enum ide_dma_cmd dma_cmd;
/* SMART */ /* SMART */
uint8_t smart_enabled; uint8_t smart_enabled;
uint8_t smart_autosave; uint8_t smart_autosave;
@ -486,6 +502,8 @@ struct IDEBus {
uint8_t unit; uint8_t unit;
uint8_t cmd; uint8_t cmd;
qemu_irq irq; qemu_irq irq;
int error_status;
}; };
struct IDEDevice { struct IDEDevice {
@ -505,10 +523,17 @@ struct IDEDeviceInfo {
#define BM_STATUS_DMAING 0x01 #define BM_STATUS_DMAING 0x01
#define BM_STATUS_ERROR 0x02 #define BM_STATUS_ERROR 0x02
#define BM_STATUS_INT 0x04 #define BM_STATUS_INT 0x04
/* FIXME These are not status register bits */
#define BM_STATUS_DMA_RETRY 0x08 #define BM_STATUS_DMA_RETRY 0x08
#define BM_STATUS_PIO_RETRY 0x10 #define BM_STATUS_PIO_RETRY 0x10
#define BM_STATUS_RETRY_READ 0x20 #define BM_STATUS_RETRY_READ 0x20
#define BM_STATUS_RETRY_FLUSH 0x40 #define BM_STATUS_RETRY_FLUSH 0x40
#define BM_STATUS_RETRY_TRIM 0x80
#define BM_MIGRATION_COMPAT_STATUS_BITS \
(BM_STATUS_DMA_RETRY | BM_STATUS_PIO_RETRY | \
BM_STATUS_RETRY_READ | BM_STATUS_RETRY_FLUSH)
#define BM_CMD_START 0x01 #define BM_CMD_START 0x01
#define BM_CMD_READ 0x08 #define BM_CMD_READ 0x08
@ -575,6 +600,9 @@ void ide_transfer_start(IDEState *s, uint8_t *buf, int size,
EndTransferFunc *end_transfer_func); EndTransferFunc *end_transfer_func);
void ide_transfer_stop(IDEState *s); void ide_transfer_stop(IDEState *s);
void ide_set_inactive(IDEState *s); void ide_set_inactive(IDEState *s);
BlockDriverAIOCB *ide_issue_trim(BlockDriverState *bs,
int64_t sector_num, QEMUIOVector *qiov, int nb_sectors,
BlockDriverCompletionFunc *cb, void *opaque);
/* hw/ide/atapi.c */ /* hw/ide/atapi.c */
void ide_atapi_cmd(IDEState *s); void ide_atapi_cmd(IDEState *s);

View File

@ -145,12 +145,21 @@ static void pmac_ide_transfer_cb(void *opaque, int ret)
io->addr += io->len; io->addr += io->len;
io->len = 0; io->len = 0;
if (s->is_read) switch (s->dma_cmd) {
case IDE_DMA_READ:
m->aiocb = dma_bdrv_read(s->bs, &s->sg, sector_num, m->aiocb = dma_bdrv_read(s->bs, &s->sg, sector_num,
pmac_ide_transfer_cb, io); pmac_ide_transfer_cb, io);
else break;
case IDE_DMA_WRITE:
m->aiocb = dma_bdrv_write(s->bs, &s->sg, sector_num, m->aiocb = dma_bdrv_write(s->bs, &s->sg, sector_num,
pmac_ide_transfer_cb, io); pmac_ide_transfer_cb, io);
break;
case IDE_DMA_TRIM:
m->aiocb = dma_bdrv_io(s->bs, &s->sg, sector_num,
ide_issue_trim, pmac_ide_transfer_cb, s, 1);
break;
}
if (!m->aiocb) if (!m->aiocb)
pmac_ide_transfer_cb(io, -1); pmac_ide_transfer_cb(io, -1);
} }

View File

@ -169,7 +169,7 @@ static int bmdma_set_inactive(IDEDMA *dma)
return 0; return 0;
} }
static void bmdma_restart_dma(BMDMAState *bm, int is_read) static void bmdma_restart_dma(BMDMAState *bm, enum ide_dma_cmd dma_cmd)
{ {
IDEState *s = bmdma_active_if(bm); IDEState *s = bmdma_active_if(bm);
@ -177,33 +177,48 @@ static void bmdma_restart_dma(BMDMAState *bm, int is_read)
s->io_buffer_index = 0; s->io_buffer_index = 0;
s->io_buffer_size = 0; s->io_buffer_size = 0;
s->nsector = bm->nsector; s->nsector = bm->nsector;
s->is_read = is_read; s->dma_cmd = dma_cmd;
bm->cur_addr = bm->addr; bm->cur_addr = bm->addr;
bm->dma_cb = ide_dma_cb; bm->dma_cb = ide_dma_cb;
bmdma_start_dma(&bm->dma, s, bm->dma_cb); bmdma_start_dma(&bm->dma, s, bm->dma_cb);
} }
/* TODO This should be common IDE code */
static void bmdma_restart_bh(void *opaque) static void bmdma_restart_bh(void *opaque)
{ {
BMDMAState *bm = opaque; BMDMAState *bm = opaque;
IDEBus *bus = bm->bus;
int is_read; int is_read;
int error_status;
qemu_bh_delete(bm->bh); qemu_bh_delete(bm->bh);
bm->bh = NULL; bm->bh = NULL;
is_read = !!(bm->status & BM_STATUS_RETRY_READ); if (bm->unit == (uint8_t) -1) {
return;
}
if (bm->status & BM_STATUS_DMA_RETRY) { is_read = !!(bus->error_status & BM_STATUS_RETRY_READ);
bm->status &= ~(BM_STATUS_DMA_RETRY | BM_STATUS_RETRY_READ);
bmdma_restart_dma(bm, is_read); /* The error status must be cleared before resubmitting the request: The
} else if (bm->status & BM_STATUS_PIO_RETRY) { * request may fail again, and this case can only be distinguished if the
bm->status &= ~(BM_STATUS_PIO_RETRY | BM_STATUS_RETRY_READ); * called function can set a new error status. */
error_status = bus->error_status;
bus->error_status = 0;
if (error_status & BM_STATUS_DMA_RETRY) {
if (error_status & BM_STATUS_RETRY_TRIM) {
bmdma_restart_dma(bm, IDE_DMA_TRIM);
} else {
bmdma_restart_dma(bm, is_read ? IDE_DMA_READ : IDE_DMA_WRITE);
}
} else if (error_status & BM_STATUS_PIO_RETRY) {
if (is_read) { if (is_read) {
ide_sector_read(bmdma_active_if(bm)); ide_sector_read(bmdma_active_if(bm));
} else { } else {
ide_sector_write(bmdma_active_if(bm)); ide_sector_write(bmdma_active_if(bm));
} }
} else if (bm->status & BM_STATUS_RETRY_FLUSH) { } else if (error_status & BM_STATUS_RETRY_FLUSH) {
ide_flush_cache(bmdma_active_if(bm)); ide_flush_cache(bmdma_active_if(bm));
} }
} }
@ -351,6 +366,43 @@ static bool ide_bmdma_current_needed(void *opaque)
return (bm->cur_prd_len != 0); return (bm->cur_prd_len != 0);
} }
static bool ide_bmdma_status_needed(void *opaque)
{
BMDMAState *bm = opaque;
/* Older versions abused some bits in the status register for internal
* error state. If any of these bits are set, we must add a subsection to
* transfer the real status register */
uint8_t abused_bits = BM_MIGRATION_COMPAT_STATUS_BITS;
return ((bm->status & abused_bits) != 0);
}
static void ide_bmdma_pre_save(void *opaque)
{
BMDMAState *bm = opaque;
uint8_t abused_bits = BM_MIGRATION_COMPAT_STATUS_BITS;
bm->migration_compat_status =
(bm->status & ~abused_bits) | (bm->bus->error_status & abused_bits);
}
/* This function accesses bm->bus->error_status which is loaded only after
* BMDMA itself. This is why the function is called from ide_pci_post_load
* instead of being registered with VMState where it would run too early. */
static int ide_bmdma_post_load(void *opaque, int version_id)
{
BMDMAState *bm = opaque;
uint8_t abused_bits = BM_MIGRATION_COMPAT_STATUS_BITS;
if (bm->status == 0) {
bm->status = bm->migration_compat_status & ~abused_bits;
bm->bus->error_status |= bm->migration_compat_status & abused_bits;
}
return 0;
}
static const VMStateDescription vmstate_bmdma_current = { static const VMStateDescription vmstate_bmdma_current = {
.name = "ide bmdma_current", .name = "ide bmdma_current",
.version_id = 1, .version_id = 1,
@ -365,15 +417,26 @@ static const VMStateDescription vmstate_bmdma_current = {
} }
}; };
const VMStateDescription vmstate_bmdma_status = {
.name ="ide bmdma/status",
.version_id = 1,
.minimum_version_id = 1,
.minimum_version_id_old = 1,
.fields = (VMStateField []) {
VMSTATE_UINT8(status, BMDMAState),
VMSTATE_END_OF_LIST()
}
};
static const VMStateDescription vmstate_bmdma = { static const VMStateDescription vmstate_bmdma = {
.name = "ide bmdma", .name = "ide bmdma",
.version_id = 3, .version_id = 3,
.minimum_version_id = 0, .minimum_version_id = 0,
.minimum_version_id_old = 0, .minimum_version_id_old = 0,
.pre_save = ide_bmdma_pre_save,
.fields = (VMStateField []) { .fields = (VMStateField []) {
VMSTATE_UINT8(cmd, BMDMAState), VMSTATE_UINT8(cmd, BMDMAState),
VMSTATE_UINT8(status, BMDMAState), VMSTATE_UINT8(migration_compat_status, BMDMAState),
VMSTATE_UINT32(addr, BMDMAState), VMSTATE_UINT32(addr, BMDMAState),
VMSTATE_INT64(sector_num, BMDMAState), VMSTATE_INT64(sector_num, BMDMAState),
VMSTATE_UINT32(nsector, BMDMAState), VMSTATE_UINT32(nsector, BMDMAState),
@ -384,6 +447,9 @@ static const VMStateDescription vmstate_bmdma = {
{ {
.vmsd = &vmstate_bmdma_current, .vmsd = &vmstate_bmdma_current,
.needed = ide_bmdma_current_needed, .needed = ide_bmdma_current_needed,
}, {
.vmsd = &vmstate_bmdma_status,
.needed = ide_bmdma_status_needed,
}, { }, {
/* empty */ /* empty */
} }
@ -399,7 +465,9 @@ static int ide_pci_post_load(void *opaque, int version_id)
/* current versions always store 0/1, but older version /* current versions always store 0/1, but older version
stored bigger values. We only need last bit */ stored bigger values. We only need last bit */
d->bmdma[i].unit &= 1; d->bmdma[i].unit &= 1;
ide_bmdma_post_load(&d->bmdma[i], -1);
} }
return 0; return 0;
} }

View File

@ -22,6 +22,10 @@ typedef struct BMDMAState {
IORange addr_ioport; IORange addr_ioport;
QEMUBH *bh; QEMUBH *bh;
qemu_irq irq; qemu_irq irq;
/* Bit 0-2 and 7: BM status register
* Bit 3-6: bus->error_status */
uint8_t migration_compat_status;
} BMDMAState; } BMDMAState;
typedef struct PCIIDEState { typedef struct PCIIDEState {

View File

@ -125,6 +125,11 @@ static int ide_dev_initfn(IDEDevice *dev, IDEDriveKind kind)
const char *serial; const char *serial;
DriveInfo *dinfo; DriveInfo *dinfo;
if (dev->conf.discard_granularity && dev->conf.discard_granularity != 512) {
error_report("discard_granularity must be 512 for ide");
return -1;
}
serial = dev->serial; serial = dev->serial;
if (!serial) { if (!serial) {
/* try to fall back to value set with legacy -drive serial=... */ /* try to fall back to value set with legacy -drive serial=... */