|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * virtio_pmem.c: Virtio pmem Driver | 
|  | * | 
|  | * Discovers persistent memory range information | 
|  | * from host and provides a virtio based flushing | 
|  | * interface. | 
|  | */ | 
|  | #include "virtio_pmem.h" | 
|  | #include "nd.h" | 
|  |  | 
|  | /* The interrupt handler */ | 
|  | void virtio_pmem_host_ack(struct virtqueue *vq) | 
|  | { | 
|  | struct virtio_pmem *vpmem = vq->vdev->priv; | 
|  | struct virtio_pmem_request *req_data, *req_buf; | 
|  | unsigned long flags; | 
|  | unsigned int len; | 
|  |  | 
|  | spin_lock_irqsave(&vpmem->pmem_lock, flags); | 
|  | while ((req_data = virtqueue_get_buf(vq, &len)) != NULL) { | 
|  | req_data->done = true; | 
|  | wake_up(&req_data->host_acked); | 
|  |  | 
|  | if (!list_empty(&vpmem->req_list)) { | 
|  | req_buf = list_first_entry(&vpmem->req_list, | 
|  | struct virtio_pmem_request, list); | 
|  | req_buf->wq_buf_avail = true; | 
|  | wake_up(&req_buf->wq_buf); | 
|  | list_del(&req_buf->list); | 
|  | } | 
|  | } | 
|  | spin_unlock_irqrestore(&vpmem->pmem_lock, flags); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(virtio_pmem_host_ack); | 
|  |  | 
|  | /* The request submission function */ | 
|  | static int virtio_pmem_flush(struct nd_region *nd_region) | 
|  | { | 
|  | struct virtio_device *vdev = nd_region->provider_data; | 
|  | struct virtio_pmem *vpmem  = vdev->priv; | 
|  | struct virtio_pmem_request *req_data; | 
|  | struct scatterlist *sgs[2], sg, ret; | 
|  | unsigned long flags; | 
|  | int err, err1; | 
|  |  | 
|  | might_sleep(); | 
|  | req_data = kmalloc(sizeof(*req_data), GFP_KERNEL); | 
|  | if (!req_data) | 
|  | return -ENOMEM; | 
|  |  | 
|  | req_data->done = false; | 
|  | init_waitqueue_head(&req_data->host_acked); | 
|  | init_waitqueue_head(&req_data->wq_buf); | 
|  | INIT_LIST_HEAD(&req_data->list); | 
|  | req_data->req.type = cpu_to_le32(VIRTIO_PMEM_REQ_TYPE_FLUSH); | 
|  | sg_init_one(&sg, &req_data->req, sizeof(req_data->req)); | 
|  | sgs[0] = &sg; | 
|  | sg_init_one(&ret, &req_data->resp.ret, sizeof(req_data->resp)); | 
|  | sgs[1] = &ret; | 
|  |  | 
|  | spin_lock_irqsave(&vpmem->pmem_lock, flags); | 
|  | /* | 
|  | * If virtqueue_add_sgs returns -ENOSPC then req_vq virtual | 
|  | * queue does not have free descriptor. We add the request | 
|  | * to req_list and wait for host_ack to wake us up when free | 
|  | * slots are available. | 
|  | */ | 
|  | while ((err = virtqueue_add_sgs(vpmem->req_vq, sgs, 1, 1, req_data, | 
|  | GFP_ATOMIC)) == -ENOSPC) { | 
|  |  | 
|  | dev_info(&vdev->dev, "failed to send command to virtio pmem device, no free slots in the virtqueue\n"); | 
|  | req_data->wq_buf_avail = false; | 
|  | list_add_tail(&req_data->list, &vpmem->req_list); | 
|  | spin_unlock_irqrestore(&vpmem->pmem_lock, flags); | 
|  |  | 
|  | /* A host response results in "host_ack" getting called */ | 
|  | wait_event(req_data->wq_buf, req_data->wq_buf_avail); | 
|  | spin_lock_irqsave(&vpmem->pmem_lock, flags); | 
|  | } | 
|  | err1 = virtqueue_kick(vpmem->req_vq); | 
|  | spin_unlock_irqrestore(&vpmem->pmem_lock, flags); | 
|  | /* | 
|  | * virtqueue_add_sgs failed with error different than -ENOSPC, we can't | 
|  | * do anything about that. | 
|  | */ | 
|  | if (err || !err1) { | 
|  | dev_info(&vdev->dev, "failed to send command to virtio pmem device\n"); | 
|  | err = -EIO; | 
|  | } else { | 
|  | /* A host repsonse results in "host_ack" getting called */ | 
|  | wait_event(req_data->host_acked, req_data->done); | 
|  | err = le32_to_cpu(req_data->resp.ret); | 
|  | } | 
|  |  | 
|  | kfree(req_data); | 
|  | return err; | 
|  | }; | 
|  |  | 
|  | /* The asynchronous flush callback function */ | 
|  | int async_pmem_flush(struct nd_region *nd_region, struct bio *bio) | 
|  | { | 
|  | /* | 
|  | * Create child bio for asynchronous flush and chain with | 
|  | * parent bio. Otherwise directly call nd_region flush. | 
|  | */ | 
|  | if (bio && bio->bi_iter.bi_sector != -1) { | 
|  | struct bio *child = bio_alloc(bio->bi_bdev, 0, | 
|  | REQ_OP_WRITE | REQ_PREFLUSH, | 
|  | GFP_ATOMIC); | 
|  |  | 
|  | if (!child) | 
|  | return -ENOMEM; | 
|  | bio_clone_blkg_association(child, bio); | 
|  | child->bi_iter.bi_sector = -1; | 
|  | bio_chain(child, bio); | 
|  | submit_bio(child); | 
|  | return 0; | 
|  | } | 
|  | if (virtio_pmem_flush(nd_region)) | 
|  | return -EIO; | 
|  |  | 
|  | return 0; | 
|  | }; | 
|  | EXPORT_SYMBOL_GPL(async_pmem_flush); | 
|  | MODULE_LICENSE("GPL"); |