| /* SPDX-License-Identifier: (GPL-2.0 OR CDDL-1.0) */ |
| /* |
| * vboxguest core guest-device handling code, VBoxGuest.cpp in upstream svn. |
| * |
| * Copyright (C) 2007-2016 Oracle Corporation |
| */ |
| |
| #include <linux/device.h> |
| #include <linux/io.h> |
| #include <linux/mm.h> |
| #include <linux/sched.h> |
| #include <linux/sizes.h> |
| #include <linux/slab.h> |
| #include <linux/vbox_err.h> |
| #include <linux/vbox_utils.h> |
| #include <linux/vmalloc.h> |
| #include "vboxguest_core.h" |
| #include "vboxguest_version.h" |
| |
| /* Get the pointer to the first HGCM parameter. */ |
| #define VBG_IOCTL_HGCM_CALL_PARMS(a) \ |
| ((struct vmmdev_hgcm_function_parameter *)( \ |
| (u8 *)(a) + sizeof(struct vbg_ioctl_hgcm_call))) |
| /* Get the pointer to the first HGCM parameter in a 32-bit request. */ |
| #define VBG_IOCTL_HGCM_CALL_PARMS32(a) \ |
| ((struct vmmdev_hgcm_function_parameter32 *)( \ |
| (u8 *)(a) + sizeof(struct vbg_ioctl_hgcm_call))) |
| |
| #define GUEST_MAPPINGS_TRIES 5 |
| |
| #define VBG_KERNEL_REQUEST \ |
| (VMMDEV_REQUESTOR_KERNEL | VMMDEV_REQUESTOR_USR_DRV | \ |
| VMMDEV_REQUESTOR_CON_DONT_KNOW | VMMDEV_REQUESTOR_TRUST_NOT_GIVEN) |
| |
| /** |
| * Reserves memory in which the VMM can relocate any guest mappings |
| * that are floating around. |
| * |
| * This operation is a little bit tricky since the VMM might not accept |
| * just any address because of address clashes between the three contexts |
| * it operates in, so we try several times. |
| * |
| * Failure to reserve the guest mappings is ignored. |
| * |
| * @gdev: The Guest extension device. |
| */ |
| static void vbg_guest_mappings_init(struct vbg_dev *gdev) |
| { |
| struct vmmdev_hypervisorinfo *req; |
| void *guest_mappings[GUEST_MAPPINGS_TRIES]; |
| struct page **pages = NULL; |
| u32 size, hypervisor_size; |
| int i, rc; |
| |
| /* Query the required space. */ |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_GET_HYPERVISOR_INFO, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return; |
| |
| req->hypervisor_start = 0; |
| req->hypervisor_size = 0; |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) |
| goto out; |
| |
| /* |
| * The VMM will report back if there is nothing it wants to map, like |
| * for instance in VT-x and AMD-V mode. |
| */ |
| if (req->hypervisor_size == 0) |
| goto out; |
| |
| hypervisor_size = req->hypervisor_size; |
| /* Add 4M so that we can align the vmap to 4MiB as the host requires. */ |
| size = PAGE_ALIGN(req->hypervisor_size) + SZ_4M; |
| |
| pages = kmalloc_array(size >> PAGE_SHIFT, sizeof(*pages), GFP_KERNEL); |
| if (!pages) |
| goto out; |
| |
| gdev->guest_mappings_dummy_page = alloc_page(GFP_HIGHUSER); |
| if (!gdev->guest_mappings_dummy_page) |
| goto out; |
| |
| for (i = 0; i < (size >> PAGE_SHIFT); i++) |
| pages[i] = gdev->guest_mappings_dummy_page; |
| |
| /* |
| * Try several times, the VMM might not accept some addresses because |
| * of address clashes between the three contexts. |
| */ |
| for (i = 0; i < GUEST_MAPPINGS_TRIES; i++) { |
| guest_mappings[i] = vmap(pages, (size >> PAGE_SHIFT), |
| VM_MAP, PAGE_KERNEL_RO); |
| if (!guest_mappings[i]) |
| break; |
| |
| req->header.request_type = VMMDEVREQ_SET_HYPERVISOR_INFO; |
| req->header.rc = VERR_INTERNAL_ERROR; |
| req->hypervisor_size = hypervisor_size; |
| req->hypervisor_start = |
| (unsigned long)PTR_ALIGN(guest_mappings[i], SZ_4M); |
| |
| rc = vbg_req_perform(gdev, req); |
| if (rc >= 0) { |
| gdev->guest_mappings = guest_mappings[i]; |
| break; |
| } |
| } |
| |
| /* Free vmap's from failed attempts. */ |
| while (--i >= 0) |
| vunmap(guest_mappings[i]); |
| |
| /* On failure free the dummy-page backing the vmap */ |
| if (!gdev->guest_mappings) { |
| __free_page(gdev->guest_mappings_dummy_page); |
| gdev->guest_mappings_dummy_page = NULL; |
| } |
| |
| out: |
| vbg_req_free(req, sizeof(*req)); |
| kfree(pages); |
| } |
| |
| /** |
| * Undo what vbg_guest_mappings_init did. |
| * |
| * @gdev: The Guest extension device. |
| */ |
| static void vbg_guest_mappings_exit(struct vbg_dev *gdev) |
| { |
| struct vmmdev_hypervisorinfo *req; |
| int rc; |
| |
| if (!gdev->guest_mappings) |
| return; |
| |
| /* |
| * Tell the host that we're going to free the memory we reserved for |
| * it, the free it up. (Leak the memory if anything goes wrong here.) |
| */ |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_HYPERVISOR_INFO, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return; |
| |
| req->hypervisor_start = 0; |
| req->hypervisor_size = 0; |
| |
| rc = vbg_req_perform(gdev, req); |
| |
| vbg_req_free(req, sizeof(*req)); |
| |
| if (rc < 0) { |
| vbg_err("%s error: %d\n", __func__, rc); |
| return; |
| } |
| |
| vunmap(gdev->guest_mappings); |
| gdev->guest_mappings = NULL; |
| |
| __free_page(gdev->guest_mappings_dummy_page); |
| gdev->guest_mappings_dummy_page = NULL; |
| } |
| |
| /** |
| * Report the guest information to the host. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| */ |
| static int vbg_report_guest_info(struct vbg_dev *gdev) |
| { |
| /* |
| * Allocate and fill in the two guest info reports. |
| */ |
| struct vmmdev_guest_info *req1 = NULL; |
| struct vmmdev_guest_info2 *req2 = NULL; |
| int rc, ret = -ENOMEM; |
| |
| req1 = vbg_req_alloc(sizeof(*req1), VMMDEVREQ_REPORT_GUEST_INFO, |
| VBG_KERNEL_REQUEST); |
| req2 = vbg_req_alloc(sizeof(*req2), VMMDEVREQ_REPORT_GUEST_INFO2, |
| VBG_KERNEL_REQUEST); |
| if (!req1 || !req2) |
| goto out_free; |
| |
| req1->interface_version = VMMDEV_VERSION; |
| req1->os_type = VMMDEV_OSTYPE_LINUX26; |
| #if __BITS_PER_LONG == 64 |
| req1->os_type |= VMMDEV_OSTYPE_X64; |
| #endif |
| |
| req2->additions_major = VBG_VERSION_MAJOR; |
| req2->additions_minor = VBG_VERSION_MINOR; |
| req2->additions_build = VBG_VERSION_BUILD; |
| req2->additions_revision = VBG_SVN_REV; |
| req2->additions_features = |
| VMMDEV_GUEST_INFO2_ADDITIONS_FEATURES_REQUESTOR_INFO; |
| strscpy(req2->name, VBG_VERSION_STRING, |
| sizeof(req2->name)); |
| |
| /* |
| * There are two protocols here: |
| * 1. INFO2 + INFO1. Supported by >=3.2.51. |
| * 2. INFO1 and optionally INFO2. The old protocol. |
| * |
| * We try protocol 2 first. It will fail with VERR_NOT_SUPPORTED |
| * if not supported by the VMMDev (message ordering requirement). |
| */ |
| rc = vbg_req_perform(gdev, req2); |
| if (rc >= 0) { |
| rc = vbg_req_perform(gdev, req1); |
| } else if (rc == VERR_NOT_SUPPORTED || rc == VERR_NOT_IMPLEMENTED) { |
| rc = vbg_req_perform(gdev, req1); |
| if (rc >= 0) { |
| rc = vbg_req_perform(gdev, req2); |
| if (rc == VERR_NOT_IMPLEMENTED) |
| rc = VINF_SUCCESS; |
| } |
| } |
| ret = vbg_status_code_to_errno(rc); |
| |
| out_free: |
| vbg_req_free(req2, sizeof(*req2)); |
| vbg_req_free(req1, sizeof(*req1)); |
| return ret; |
| } |
| |
| /** |
| * Report the guest driver status to the host. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @active: Flag whether the driver is now active or not. |
| */ |
| static int vbg_report_driver_status(struct vbg_dev *gdev, bool active) |
| { |
| struct vmmdev_guest_status *req; |
| int rc; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_REPORT_GUEST_STATUS, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return -ENOMEM; |
| |
| req->facility = VBOXGUEST_FACILITY_TYPE_VBOXGUEST_DRIVER; |
| if (active) |
| req->status = VBOXGUEST_FACILITY_STATUS_ACTIVE; |
| else |
| req->status = VBOXGUEST_FACILITY_STATUS_INACTIVE; |
| req->flags = 0; |
| |
| rc = vbg_req_perform(gdev, req); |
| if (rc == VERR_NOT_IMPLEMENTED) /* Compatibility with older hosts. */ |
| rc = VINF_SUCCESS; |
| |
| vbg_req_free(req, sizeof(*req)); |
| |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| /** |
| * Inflate the balloon by one chunk. The caller owns the balloon mutex. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @chunk_idx: Index of the chunk. |
| */ |
| static int vbg_balloon_inflate(struct vbg_dev *gdev, u32 chunk_idx) |
| { |
| struct vmmdev_memballoon_change *req = gdev->mem_balloon.change_req; |
| struct page **pages; |
| int i, rc, ret; |
| |
| pages = kmalloc_array(VMMDEV_MEMORY_BALLOON_CHUNK_PAGES, |
| sizeof(*pages), |
| GFP_KERNEL | __GFP_NOWARN); |
| if (!pages) |
| return -ENOMEM; |
| |
| req->header.size = sizeof(*req); |
| req->inflate = true; |
| req->pages = VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; |
| |
| for (i = 0; i < VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; i++) { |
| pages[i] = alloc_page(GFP_KERNEL | __GFP_NOWARN); |
| if (!pages[i]) { |
| ret = -ENOMEM; |
| goto out_error; |
| } |
| |
| req->phys_page[i] = page_to_phys(pages[i]); |
| } |
| |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) { |
| vbg_err("%s error, rc: %d\n", __func__, rc); |
| ret = vbg_status_code_to_errno(rc); |
| goto out_error; |
| } |
| |
| gdev->mem_balloon.pages[chunk_idx] = pages; |
| |
| return 0; |
| |
| out_error: |
| while (--i >= 0) |
| __free_page(pages[i]); |
| kfree(pages); |
| |
| return ret; |
| } |
| |
| /** |
| * Deflate the balloon by one chunk. The caller owns the balloon mutex. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @chunk_idx: Index of the chunk. |
| */ |
| static int vbg_balloon_deflate(struct vbg_dev *gdev, u32 chunk_idx) |
| { |
| struct vmmdev_memballoon_change *req = gdev->mem_balloon.change_req; |
| struct page **pages = gdev->mem_balloon.pages[chunk_idx]; |
| int i, rc; |
| |
| req->header.size = sizeof(*req); |
| req->inflate = false; |
| req->pages = VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; |
| |
| for (i = 0; i < VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; i++) |
| req->phys_page[i] = page_to_phys(pages[i]); |
| |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) { |
| vbg_err("%s error, rc: %d\n", __func__, rc); |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| for (i = 0; i < VMMDEV_MEMORY_BALLOON_CHUNK_PAGES; i++) |
| __free_page(pages[i]); |
| kfree(pages); |
| gdev->mem_balloon.pages[chunk_idx] = NULL; |
| |
| return 0; |
| } |
| |
| /** |
| * Respond to VMMDEV_EVENT_BALLOON_CHANGE_REQUEST events, query the size |
| * the host wants the balloon to be and adjust accordingly. |
| */ |
| static void vbg_balloon_work(struct work_struct *work) |
| { |
| struct vbg_dev *gdev = |
| container_of(work, struct vbg_dev, mem_balloon.work); |
| struct vmmdev_memballoon_info *req = gdev->mem_balloon.get_req; |
| u32 i, chunks; |
| int rc, ret; |
| |
| /* |
| * Setting this bit means that we request the value from the host and |
| * change the guest memory balloon according to the returned value. |
| */ |
| req->event_ack = VMMDEV_EVENT_BALLOON_CHANGE_REQUEST; |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) { |
| vbg_err("%s error, rc: %d)\n", __func__, rc); |
| return; |
| } |
| |
| /* |
| * The host always returns the same maximum amount of chunks, so |
| * we do this once. |
| */ |
| if (!gdev->mem_balloon.max_chunks) { |
| gdev->mem_balloon.pages = |
| devm_kcalloc(gdev->dev, req->phys_mem_chunks, |
| sizeof(struct page **), GFP_KERNEL); |
| if (!gdev->mem_balloon.pages) |
| return; |
| |
| gdev->mem_balloon.max_chunks = req->phys_mem_chunks; |
| } |
| |
| chunks = req->balloon_chunks; |
| if (chunks > gdev->mem_balloon.max_chunks) { |
| vbg_err("%s: illegal balloon size %u (max=%u)\n", |
| __func__, chunks, gdev->mem_balloon.max_chunks); |
| return; |
| } |
| |
| if (chunks > gdev->mem_balloon.chunks) { |
| /* inflate */ |
| for (i = gdev->mem_balloon.chunks; i < chunks; i++) { |
| ret = vbg_balloon_inflate(gdev, i); |
| if (ret < 0) |
| return; |
| |
| gdev->mem_balloon.chunks++; |
| } |
| } else { |
| /* deflate */ |
| for (i = gdev->mem_balloon.chunks; i-- > chunks;) { |
| ret = vbg_balloon_deflate(gdev, i); |
| if (ret < 0) |
| return; |
| |
| gdev->mem_balloon.chunks--; |
| } |
| } |
| } |
| |
| /** |
| * Callback for heartbeat timer. |
| */ |
| static void vbg_heartbeat_timer(struct timer_list *t) |
| { |
| struct vbg_dev *gdev = from_timer(gdev, t, heartbeat_timer); |
| |
| vbg_req_perform(gdev, gdev->guest_heartbeat_req); |
| mod_timer(&gdev->heartbeat_timer, |
| msecs_to_jiffies(gdev->heartbeat_interval_ms)); |
| } |
| |
| /** |
| * Configure the host to check guest's heartbeat |
| * and get heartbeat interval from the host. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @enabled: Set true to enable guest heartbeat checks on host. |
| */ |
| static int vbg_heartbeat_host_config(struct vbg_dev *gdev, bool enabled) |
| { |
| struct vmmdev_heartbeat *req; |
| int rc; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_HEARTBEAT_CONFIGURE, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return -ENOMEM; |
| |
| req->enabled = enabled; |
| req->interval_ns = 0; |
| rc = vbg_req_perform(gdev, req); |
| do_div(req->interval_ns, 1000000); /* ns -> ms */ |
| gdev->heartbeat_interval_ms = req->interval_ns; |
| vbg_req_free(req, sizeof(*req)); |
| |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| /** |
| * Initializes the heartbeat timer. This feature may be disabled by the host. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| */ |
| static int vbg_heartbeat_init(struct vbg_dev *gdev) |
| { |
| int ret; |
| |
| /* Make sure that heartbeat checking is disabled if we fail. */ |
| ret = vbg_heartbeat_host_config(gdev, false); |
| if (ret < 0) |
| return ret; |
| |
| ret = vbg_heartbeat_host_config(gdev, true); |
| if (ret < 0) |
| return ret; |
| |
| gdev->guest_heartbeat_req = vbg_req_alloc( |
| sizeof(*gdev->guest_heartbeat_req), |
| VMMDEVREQ_GUEST_HEARTBEAT, |
| VBG_KERNEL_REQUEST); |
| if (!gdev->guest_heartbeat_req) |
| return -ENOMEM; |
| |
| vbg_info("%s: Setting up heartbeat to trigger every %d milliseconds\n", |
| __func__, gdev->heartbeat_interval_ms); |
| mod_timer(&gdev->heartbeat_timer, 0); |
| |
| return 0; |
| } |
| |
| /** |
| * Cleanup hearbeat code, stop HB timer and disable host heartbeat checking. |
| * @gdev: The Guest extension device. |
| */ |
| static void vbg_heartbeat_exit(struct vbg_dev *gdev) |
| { |
| del_timer_sync(&gdev->heartbeat_timer); |
| vbg_heartbeat_host_config(gdev, false); |
| vbg_req_free(gdev->guest_heartbeat_req, |
| sizeof(*gdev->guest_heartbeat_req)); |
| } |
| |
| /** |
| * Applies a change to the bit usage tracker. |
| * Return: true if the mask changed, false if not. |
| * @tracker: The bit usage tracker. |
| * @changed: The bits to change. |
| * @previous: The previous value of the bits. |
| */ |
| static bool vbg_track_bit_usage(struct vbg_bit_usage_tracker *tracker, |
| u32 changed, u32 previous) |
| { |
| bool global_change = false; |
| |
| while (changed) { |
| u32 bit = ffs(changed) - 1; |
| u32 bitmask = BIT(bit); |
| |
| if (bitmask & previous) { |
| tracker->per_bit_usage[bit] -= 1; |
| if (tracker->per_bit_usage[bit] == 0) { |
| global_change = true; |
| tracker->mask &= ~bitmask; |
| } |
| } else { |
| tracker->per_bit_usage[bit] += 1; |
| if (tracker->per_bit_usage[bit] == 1) { |
| global_change = true; |
| tracker->mask |= bitmask; |
| } |
| } |
| |
| changed &= ~bitmask; |
| } |
| |
| return global_change; |
| } |
| |
| /** |
| * Init and termination worker for resetting the (host) event filter on the host |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @fixed_events: Fixed events (init time). |
| */ |
| static int vbg_reset_host_event_filter(struct vbg_dev *gdev, |
| u32 fixed_events) |
| { |
| struct vmmdev_mask *req; |
| int rc; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_CTL_GUEST_FILTER_MASK, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return -ENOMEM; |
| |
| req->not_mask = U32_MAX & ~fixed_events; |
| req->or_mask = fixed_events; |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) |
| vbg_err("%s error, rc: %d\n", __func__, rc); |
| |
| vbg_req_free(req, sizeof(*req)); |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| /** |
| * Changes the event filter mask for the given session. |
| * |
| * This is called in response to VBG_IOCTL_CHANGE_FILTER_MASK as well as to |
| * do session cleanup. Takes the session mutex. |
| * |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @session: The session. |
| * @or_mask: The events to add. |
| * @not_mask: The events to remove. |
| * @session_termination: Set if we're called by the session cleanup code. |
| * This tweaks the error handling so we perform |
| * proper session cleanup even if the host |
| * misbehaves. |
| */ |
| static int vbg_set_session_event_filter(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| u32 or_mask, u32 not_mask, |
| bool session_termination) |
| { |
| struct vmmdev_mask *req; |
| u32 changed, previous; |
| int rc, ret = 0; |
| |
| /* |
| * Allocate a request buffer before taking the spinlock, when |
| * the session is being terminated the requestor is the kernel, |
| * as we're cleaning up. |
| */ |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_CTL_GUEST_FILTER_MASK, |
| session_termination ? VBG_KERNEL_REQUEST : |
| session->requestor); |
| if (!req) { |
| if (!session_termination) |
| return -ENOMEM; |
| /* Ignore allocation failure, we must do session cleanup. */ |
| } |
| |
| mutex_lock(&gdev->session_mutex); |
| |
| /* Apply the changes to the session mask. */ |
| previous = session->event_filter; |
| session->event_filter |= or_mask; |
| session->event_filter &= ~not_mask; |
| |
| /* If anything actually changed, update the global usage counters. */ |
| changed = previous ^ session->event_filter; |
| if (!changed) |
| goto out; |
| |
| vbg_track_bit_usage(&gdev->event_filter_tracker, changed, previous); |
| or_mask = gdev->fixed_events | gdev->event_filter_tracker.mask; |
| |
| if (gdev->event_filter_host == or_mask || !req) |
| goto out; |
| |
| gdev->event_filter_host = or_mask; |
| req->or_mask = or_mask; |
| req->not_mask = ~or_mask; |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) { |
| ret = vbg_status_code_to_errno(rc); |
| |
| /* Failed, roll back (unless it's session termination time). */ |
| gdev->event_filter_host = U32_MAX; |
| if (session_termination) |
| goto out; |
| |
| vbg_track_bit_usage(&gdev->event_filter_tracker, changed, |
| session->event_filter); |
| session->event_filter = previous; |
| } |
| |
| out: |
| mutex_unlock(&gdev->session_mutex); |
| vbg_req_free(req, sizeof(*req)); |
| |
| return ret; |
| } |
| |
| /** |
| * Init and termination worker for set guest capabilities to zero on the host. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| */ |
| static int vbg_reset_host_capabilities(struct vbg_dev *gdev) |
| { |
| struct vmmdev_mask *req; |
| int rc; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_GUEST_CAPABILITIES, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return -ENOMEM; |
| |
| req->not_mask = U32_MAX; |
| req->or_mask = 0; |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) |
| vbg_err("%s error, rc: %d\n", __func__, rc); |
| |
| vbg_req_free(req, sizeof(*req)); |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| /** |
| * Set guest capabilities on the host. |
| * Must be called with gdev->session_mutex hold. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @session: The session. |
| * @session_termination: Set if we're called by the session cleanup code. |
| */ |
| static int vbg_set_host_capabilities(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| bool session_termination) |
| { |
| struct vmmdev_mask *req; |
| u32 caps; |
| int rc; |
| |
| WARN_ON(!mutex_is_locked(&gdev->session_mutex)); |
| |
| caps = gdev->acquired_guest_caps | gdev->set_guest_caps_tracker.mask; |
| |
| if (gdev->guest_caps_host == caps) |
| return 0; |
| |
| /* On termination the requestor is the kernel, as we're cleaning up. */ |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_GUEST_CAPABILITIES, |
| session_termination ? VBG_KERNEL_REQUEST : |
| session->requestor); |
| if (!req) { |
| gdev->guest_caps_host = U32_MAX; |
| return -ENOMEM; |
| } |
| |
| req->or_mask = caps; |
| req->not_mask = ~caps; |
| rc = vbg_req_perform(gdev, req); |
| vbg_req_free(req, sizeof(*req)); |
| |
| gdev->guest_caps_host = (rc >= 0) ? caps : U32_MAX; |
| |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| /** |
| * Acquire (get exclusive access) guest capabilities for a session. |
| * Takes the session mutex. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @session: The session. |
| * @flags: Flags (VBGL_IOC_AGC_FLAGS_XXX). |
| * @or_mask: The capabilities to add. |
| * @not_mask: The capabilities to remove. |
| * @session_termination: Set if we're called by the session cleanup code. |
| * This tweaks the error handling so we perform |
| * proper session cleanup even if the host |
| * misbehaves. |
| */ |
| static int vbg_acquire_session_capabilities(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| u32 or_mask, u32 not_mask, |
| u32 flags, bool session_termination) |
| { |
| unsigned long irqflags; |
| bool wakeup = false; |
| int ret = 0; |
| |
| mutex_lock(&gdev->session_mutex); |
| |
| if (gdev->set_guest_caps_tracker.mask & or_mask) { |
| vbg_err("%s error: cannot acquire caps which are currently set\n", |
| __func__); |
| ret = -EINVAL; |
| goto out; |
| } |
| |
| /* |
| * Mark any caps in the or_mask as now being in acquire-mode. Note |
| * once caps are in acquire_mode they always stay in this mode. |
| * This impacts event handling, so we take the event-lock. |
| */ |
| spin_lock_irqsave(&gdev->event_spinlock, irqflags); |
| gdev->acquire_mode_guest_caps |= or_mask; |
| spin_unlock_irqrestore(&gdev->event_spinlock, irqflags); |
| |
| /* If we only have to switch the caps to acquire mode, we're done. */ |
| if (flags & VBGL_IOC_AGC_FLAGS_CONFIG_ACQUIRE_MODE) |
| goto out; |
| |
| not_mask &= ~or_mask; /* or_mask takes priority over not_mask */ |
| not_mask &= session->acquired_guest_caps; |
| or_mask &= ~session->acquired_guest_caps; |
| |
| if (or_mask == 0 && not_mask == 0) |
| goto out; |
| |
| if (gdev->acquired_guest_caps & or_mask) { |
| ret = -EBUSY; |
| goto out; |
| } |
| |
| gdev->acquired_guest_caps |= or_mask; |
| gdev->acquired_guest_caps &= ~not_mask; |
| /* session->acquired_guest_caps impacts event handling, take the lock */ |
| spin_lock_irqsave(&gdev->event_spinlock, irqflags); |
| session->acquired_guest_caps |= or_mask; |
| session->acquired_guest_caps &= ~not_mask; |
| spin_unlock_irqrestore(&gdev->event_spinlock, irqflags); |
| |
| ret = vbg_set_host_capabilities(gdev, session, session_termination); |
| /* Roll back on failure, unless it's session termination time. */ |
| if (ret < 0 && !session_termination) { |
| gdev->acquired_guest_caps &= ~or_mask; |
| gdev->acquired_guest_caps |= not_mask; |
| spin_lock_irqsave(&gdev->event_spinlock, irqflags); |
| session->acquired_guest_caps &= ~or_mask; |
| session->acquired_guest_caps |= not_mask; |
| spin_unlock_irqrestore(&gdev->event_spinlock, irqflags); |
| } |
| |
| /* |
| * If we added a capability, check if that means some other thread in |
| * our session should be unblocked because there are events pending |
| * (the result of vbg_get_allowed_event_mask_for_session() may change). |
| * |
| * HACK ALERT! When the seamless support capability is added we generate |
| * a seamless change event so that the ring-3 client can sync with |
| * the seamless state. |
| */ |
| if (ret == 0 && or_mask != 0) { |
| spin_lock_irqsave(&gdev->event_spinlock, irqflags); |
| |
| if (or_mask & VMMDEV_GUEST_SUPPORTS_SEAMLESS) |
| gdev->pending_events |= |
| VMMDEV_EVENT_SEAMLESS_MODE_CHANGE_REQUEST; |
| |
| if (gdev->pending_events) |
| wakeup = true; |
| |
| spin_unlock_irqrestore(&gdev->event_spinlock, irqflags); |
| |
| if (wakeup) |
| wake_up(&gdev->event_wq); |
| } |
| |
| out: |
| mutex_unlock(&gdev->session_mutex); |
| |
| return ret; |
| } |
| |
| /** |
| * Sets the guest capabilities for a session. Takes the session mutex. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @session: The session. |
| * @or_mask: The capabilities to add. |
| * @not_mask: The capabilities to remove. |
| * @session_termination: Set if we're called by the session cleanup code. |
| * This tweaks the error handling so we perform |
| * proper session cleanup even if the host |
| * misbehaves. |
| */ |
| static int vbg_set_session_capabilities(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| u32 or_mask, u32 not_mask, |
| bool session_termination) |
| { |
| u32 changed, previous; |
| int ret = 0; |
| |
| mutex_lock(&gdev->session_mutex); |
| |
| if (gdev->acquire_mode_guest_caps & or_mask) { |
| vbg_err("%s error: cannot set caps which are in acquire_mode\n", |
| __func__); |
| ret = -EBUSY; |
| goto out; |
| } |
| |
| /* Apply the changes to the session mask. */ |
| previous = session->set_guest_caps; |
| session->set_guest_caps |= or_mask; |
| session->set_guest_caps &= ~not_mask; |
| |
| /* If anything actually changed, update the global usage counters. */ |
| changed = previous ^ session->set_guest_caps; |
| if (!changed) |
| goto out; |
| |
| vbg_track_bit_usage(&gdev->set_guest_caps_tracker, changed, previous); |
| |
| ret = vbg_set_host_capabilities(gdev, session, session_termination); |
| /* Roll back on failure, unless it's session termination time. */ |
| if (ret < 0 && !session_termination) { |
| vbg_track_bit_usage(&gdev->set_guest_caps_tracker, changed, |
| session->set_guest_caps); |
| session->set_guest_caps = previous; |
| } |
| |
| out: |
| mutex_unlock(&gdev->session_mutex); |
| |
| return ret; |
| } |
| |
| /** |
| * vbg_query_host_version get the host feature mask and version information. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| */ |
| static int vbg_query_host_version(struct vbg_dev *gdev) |
| { |
| struct vmmdev_host_version *req; |
| int rc, ret; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_GET_HOST_VERSION, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return -ENOMEM; |
| |
| rc = vbg_req_perform(gdev, req); |
| ret = vbg_status_code_to_errno(rc); |
| if (ret) { |
| vbg_err("%s error: %d\n", __func__, rc); |
| goto out; |
| } |
| |
| snprintf(gdev->host_version, sizeof(gdev->host_version), "%u.%u.%ur%u", |
| req->major, req->minor, req->build, req->revision); |
| gdev->host_features = req->features; |
| |
| vbg_info("vboxguest: host-version: %s %#x\n", gdev->host_version, |
| gdev->host_features); |
| |
| if (!(req->features & VMMDEV_HVF_HGCM_PHYS_PAGE_LIST)) { |
| vbg_err("vboxguest: Error host too old (does not support page-lists)\n"); |
| ret = -ENODEV; |
| } |
| |
| out: |
| vbg_req_free(req, sizeof(*req)); |
| return ret; |
| } |
| |
| /** |
| * Initializes the VBoxGuest device extension when the |
| * device driver is loaded. |
| * |
| * The native code locates the VMMDev on the PCI bus and retrieve |
| * the MMIO and I/O port ranges, this function will take care of |
| * mapping the MMIO memory (if present). Upon successful return |
| * the native code should set up the interrupt handler. |
| * |
| * Return: 0 or negative errno value. |
| * |
| * @gdev: The Guest extension device. |
| * @fixed_events: Events that will be enabled upon init and no client |
| * will ever be allowed to mask. |
| */ |
| int vbg_core_init(struct vbg_dev *gdev, u32 fixed_events) |
| { |
| int ret = -ENOMEM; |
| |
| gdev->fixed_events = fixed_events | VMMDEV_EVENT_HGCM; |
| gdev->event_filter_host = U32_MAX; /* forces a report */ |
| gdev->guest_caps_host = U32_MAX; /* forces a report */ |
| |
| init_waitqueue_head(&gdev->event_wq); |
| init_waitqueue_head(&gdev->hgcm_wq); |
| spin_lock_init(&gdev->event_spinlock); |
| mutex_init(&gdev->session_mutex); |
| mutex_init(&gdev->cancel_req_mutex); |
| timer_setup(&gdev->heartbeat_timer, vbg_heartbeat_timer, 0); |
| INIT_WORK(&gdev->mem_balloon.work, vbg_balloon_work); |
| |
| gdev->mem_balloon.get_req = |
| vbg_req_alloc(sizeof(*gdev->mem_balloon.get_req), |
| VMMDEVREQ_GET_MEMBALLOON_CHANGE_REQ, |
| VBG_KERNEL_REQUEST); |
| gdev->mem_balloon.change_req = |
| vbg_req_alloc(sizeof(*gdev->mem_balloon.change_req), |
| VMMDEVREQ_CHANGE_MEMBALLOON, |
| VBG_KERNEL_REQUEST); |
| gdev->cancel_req = |
| vbg_req_alloc(sizeof(*(gdev->cancel_req)), |
| VMMDEVREQ_HGCM_CANCEL2, |
| VBG_KERNEL_REQUEST); |
| gdev->ack_events_req = |
| vbg_req_alloc(sizeof(*gdev->ack_events_req), |
| VMMDEVREQ_ACKNOWLEDGE_EVENTS, |
| VBG_KERNEL_REQUEST); |
| gdev->mouse_status_req = |
| vbg_req_alloc(sizeof(*gdev->mouse_status_req), |
| VMMDEVREQ_GET_MOUSE_STATUS, |
| VBG_KERNEL_REQUEST); |
| |
| if (!gdev->mem_balloon.get_req || !gdev->mem_balloon.change_req || |
| !gdev->cancel_req || !gdev->ack_events_req || |
| !gdev->mouse_status_req) |
| goto err_free_reqs; |
| |
| ret = vbg_query_host_version(gdev); |
| if (ret) |
| goto err_free_reqs; |
| |
| ret = vbg_report_guest_info(gdev); |
| if (ret) { |
| vbg_err("vboxguest: vbg_report_guest_info error: %d\n", ret); |
| goto err_free_reqs; |
| } |
| |
| ret = vbg_reset_host_event_filter(gdev, gdev->fixed_events); |
| if (ret) { |
| vbg_err("vboxguest: Error setting fixed event filter: %d\n", |
| ret); |
| goto err_free_reqs; |
| } |
| |
| ret = vbg_reset_host_capabilities(gdev); |
| if (ret) { |
| vbg_err("vboxguest: Error clearing guest capabilities: %d\n", |
| ret); |
| goto err_free_reqs; |
| } |
| |
| ret = vbg_core_set_mouse_status(gdev, 0); |
| if (ret) { |
| vbg_err("vboxguest: Error clearing mouse status: %d\n", ret); |
| goto err_free_reqs; |
| } |
| |
| /* These may fail without requiring the driver init to fail. */ |
| vbg_guest_mappings_init(gdev); |
| vbg_heartbeat_init(gdev); |
| |
| /* All Done! */ |
| ret = vbg_report_driver_status(gdev, true); |
| if (ret < 0) |
| vbg_err("vboxguest: Error reporting driver status: %d\n", ret); |
| |
| return 0; |
| |
| err_free_reqs: |
| vbg_req_free(gdev->mouse_status_req, |
| sizeof(*gdev->mouse_status_req)); |
| vbg_req_free(gdev->ack_events_req, |
| sizeof(*gdev->ack_events_req)); |
| vbg_req_free(gdev->cancel_req, |
| sizeof(*gdev->cancel_req)); |
| vbg_req_free(gdev->mem_balloon.change_req, |
| sizeof(*gdev->mem_balloon.change_req)); |
| vbg_req_free(gdev->mem_balloon.get_req, |
| sizeof(*gdev->mem_balloon.get_req)); |
| return ret; |
| } |
| |
| /** |
| * Call this on exit to clean-up vboxguest-core managed resources. |
| * |
| * The native code should call this before the driver is loaded, |
| * but don't call this on shutdown. |
| * @gdev: The Guest extension device. |
| */ |
| void vbg_core_exit(struct vbg_dev *gdev) |
| { |
| vbg_heartbeat_exit(gdev); |
| vbg_guest_mappings_exit(gdev); |
| |
| /* Clear the host flags (mouse status etc). */ |
| vbg_reset_host_event_filter(gdev, 0); |
| vbg_reset_host_capabilities(gdev); |
| vbg_core_set_mouse_status(gdev, 0); |
| |
| vbg_req_free(gdev->mouse_status_req, |
| sizeof(*gdev->mouse_status_req)); |
| vbg_req_free(gdev->ack_events_req, |
| sizeof(*gdev->ack_events_req)); |
| vbg_req_free(gdev->cancel_req, |
| sizeof(*gdev->cancel_req)); |
| vbg_req_free(gdev->mem_balloon.change_req, |
| sizeof(*gdev->mem_balloon.change_req)); |
| vbg_req_free(gdev->mem_balloon.get_req, |
| sizeof(*gdev->mem_balloon.get_req)); |
| } |
| |
| /** |
| * Creates a VBoxGuest user session. |
| * |
| * vboxguest_linux.c calls this when userspace opens the char-device. |
| * Return: A pointer to the new session or an ERR_PTR on error. |
| * @gdev: The Guest extension device. |
| * @requestor: VMMDEV_REQUESTOR_* flags |
| */ |
| struct vbg_session *vbg_core_open_session(struct vbg_dev *gdev, u32 requestor) |
| { |
| struct vbg_session *session; |
| |
| session = kzalloc(sizeof(*session), GFP_KERNEL); |
| if (!session) |
| return ERR_PTR(-ENOMEM); |
| |
| session->gdev = gdev; |
| session->requestor = requestor; |
| |
| return session; |
| } |
| |
| /** |
| * Closes a VBoxGuest session. |
| * @session: The session to close (and free). |
| */ |
| void vbg_core_close_session(struct vbg_session *session) |
| { |
| struct vbg_dev *gdev = session->gdev; |
| int i, rc; |
| |
| vbg_acquire_session_capabilities(gdev, session, 0, U32_MAX, 0, true); |
| vbg_set_session_capabilities(gdev, session, 0, U32_MAX, true); |
| vbg_set_session_event_filter(gdev, session, 0, U32_MAX, true); |
| |
| for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) { |
| if (!session->hgcm_client_ids[i]) |
| continue; |
| |
| /* requestor is kernel here, as we're cleaning up. */ |
| vbg_hgcm_disconnect(gdev, VBG_KERNEL_REQUEST, |
| session->hgcm_client_ids[i], &rc); |
| } |
| |
| kfree(session); |
| } |
| |
| static int vbg_ioctl_chk(struct vbg_ioctl_hdr *hdr, size_t in_size, |
| size_t out_size) |
| { |
| if (hdr->size_in != (sizeof(*hdr) + in_size) || |
| hdr->size_out != (sizeof(*hdr) + out_size)) |
| return -EINVAL; |
| |
| return 0; |
| } |
| |
| static int vbg_ioctl_driver_version_info( |
| struct vbg_ioctl_driver_version_info *info) |
| { |
| const u16 vbg_maj_version = VBG_IOC_VERSION >> 16; |
| u16 min_maj_version, req_maj_version; |
| |
| if (vbg_ioctl_chk(&info->hdr, sizeof(info->u.in), sizeof(info->u.out))) |
| return -EINVAL; |
| |
| req_maj_version = info->u.in.req_version >> 16; |
| min_maj_version = info->u.in.min_version >> 16; |
| |
| if (info->u.in.min_version > info->u.in.req_version || |
| min_maj_version != req_maj_version) |
| return -EINVAL; |
| |
| if (info->u.in.min_version <= VBG_IOC_VERSION && |
| min_maj_version == vbg_maj_version) { |
| info->u.out.session_version = VBG_IOC_VERSION; |
| } else { |
| info->u.out.session_version = U32_MAX; |
| info->hdr.rc = VERR_VERSION_MISMATCH; |
| } |
| |
| info->u.out.driver_version = VBG_IOC_VERSION; |
| info->u.out.driver_revision = 0; |
| info->u.out.reserved1 = 0; |
| info->u.out.reserved2 = 0; |
| |
| return 0; |
| } |
| |
| /* Must be called with the event_lock held */ |
| static u32 vbg_get_allowed_event_mask_for_session(struct vbg_dev *gdev, |
| struct vbg_session *session) |
| { |
| u32 acquire_mode_caps = gdev->acquire_mode_guest_caps; |
| u32 session_acquired_caps = session->acquired_guest_caps; |
| u32 allowed_events = VMMDEV_EVENT_VALID_EVENT_MASK; |
| |
| if ((acquire_mode_caps & VMMDEV_GUEST_SUPPORTS_GRAPHICS) && |
| !(session_acquired_caps & VMMDEV_GUEST_SUPPORTS_GRAPHICS)) |
| allowed_events &= ~VMMDEV_EVENT_DISPLAY_CHANGE_REQUEST; |
| |
| if ((acquire_mode_caps & VMMDEV_GUEST_SUPPORTS_SEAMLESS) && |
| !(session_acquired_caps & VMMDEV_GUEST_SUPPORTS_SEAMLESS)) |
| allowed_events &= ~VMMDEV_EVENT_SEAMLESS_MODE_CHANGE_REQUEST; |
| |
| return allowed_events; |
| } |
| |
| static bool vbg_wait_event_cond(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| u32 event_mask) |
| { |
| unsigned long flags; |
| bool wakeup; |
| u32 events; |
| |
| spin_lock_irqsave(&gdev->event_spinlock, flags); |
| |
| events = gdev->pending_events & event_mask; |
| events &= vbg_get_allowed_event_mask_for_session(gdev, session); |
| wakeup = events || session->cancel_waiters; |
| |
| spin_unlock_irqrestore(&gdev->event_spinlock, flags); |
| |
| return wakeup; |
| } |
| |
| /* Must be called with the event_lock held */ |
| static u32 vbg_consume_events_locked(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| u32 event_mask) |
| { |
| u32 events = gdev->pending_events & event_mask; |
| |
| events &= vbg_get_allowed_event_mask_for_session(gdev, session); |
| gdev->pending_events &= ~events; |
| return events; |
| } |
| |
| static int vbg_ioctl_wait_for_events(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_wait_for_events *wait) |
| { |
| u32 timeout_ms = wait->u.in.timeout_ms; |
| u32 event_mask = wait->u.in.events; |
| unsigned long flags; |
| long timeout; |
| int ret = 0; |
| |
| if (vbg_ioctl_chk(&wait->hdr, sizeof(wait->u.in), sizeof(wait->u.out))) |
| return -EINVAL; |
| |
| if (timeout_ms == U32_MAX) |
| timeout = MAX_SCHEDULE_TIMEOUT; |
| else |
| timeout = msecs_to_jiffies(timeout_ms); |
| |
| wait->u.out.events = 0; |
| do { |
| timeout = wait_event_interruptible_timeout( |
| gdev->event_wq, |
| vbg_wait_event_cond(gdev, session, event_mask), |
| timeout); |
| |
| spin_lock_irqsave(&gdev->event_spinlock, flags); |
| |
| if (timeout < 0 || session->cancel_waiters) { |
| ret = -EINTR; |
| } else if (timeout == 0) { |
| ret = -ETIMEDOUT; |
| } else { |
| wait->u.out.events = |
| vbg_consume_events_locked(gdev, session, event_mask); |
| } |
| |
| spin_unlock_irqrestore(&gdev->event_spinlock, flags); |
| |
| /* |
| * Someone else may have consumed the event(s) first, in |
| * which case we go back to waiting. |
| */ |
| } while (ret == 0 && wait->u.out.events == 0); |
| |
| return ret; |
| } |
| |
| static int vbg_ioctl_interrupt_all_wait_events(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_hdr *hdr) |
| { |
| unsigned long flags; |
| |
| if (hdr->size_in != sizeof(*hdr) || hdr->size_out != sizeof(*hdr)) |
| return -EINVAL; |
| |
| spin_lock_irqsave(&gdev->event_spinlock, flags); |
| session->cancel_waiters = true; |
| spin_unlock_irqrestore(&gdev->event_spinlock, flags); |
| |
| wake_up(&gdev->event_wq); |
| |
| return 0; |
| } |
| |
| /** |
| * Checks if the VMM request is allowed in the context of the given session. |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @session: The calling session. |
| * @req: The request. |
| */ |
| static int vbg_req_allowed(struct vbg_dev *gdev, struct vbg_session *session, |
| const struct vmmdev_request_header *req) |
| { |
| const struct vmmdev_guest_status *guest_status; |
| bool trusted_apps_only; |
| |
| switch (req->request_type) { |
| /* Trusted users apps only. */ |
| case VMMDEVREQ_QUERY_CREDENTIALS: |
| case VMMDEVREQ_REPORT_CREDENTIALS_JUDGEMENT: |
| case VMMDEVREQ_REGISTER_SHARED_MODULE: |
| case VMMDEVREQ_UNREGISTER_SHARED_MODULE: |
| case VMMDEVREQ_WRITE_COREDUMP: |
| case VMMDEVREQ_GET_CPU_HOTPLUG_REQ: |
| case VMMDEVREQ_SET_CPU_HOTPLUG_STATUS: |
| case VMMDEVREQ_CHECK_SHARED_MODULES: |
| case VMMDEVREQ_GET_PAGE_SHARING_STATUS: |
| case VMMDEVREQ_DEBUG_IS_PAGE_SHARED: |
| case VMMDEVREQ_REPORT_GUEST_STATS: |
| case VMMDEVREQ_REPORT_GUEST_USER_STATE: |
| case VMMDEVREQ_GET_STATISTICS_CHANGE_REQ: |
| trusted_apps_only = true; |
| break; |
| |
| /* Anyone. */ |
| case VMMDEVREQ_GET_MOUSE_STATUS: |
| case VMMDEVREQ_SET_MOUSE_STATUS: |
| case VMMDEVREQ_SET_POINTER_SHAPE: |
| case VMMDEVREQ_GET_HOST_VERSION: |
| case VMMDEVREQ_IDLE: |
| case VMMDEVREQ_GET_HOST_TIME: |
| case VMMDEVREQ_SET_POWER_STATUS: |
| case VMMDEVREQ_ACKNOWLEDGE_EVENTS: |
| case VMMDEVREQ_CTL_GUEST_FILTER_MASK: |
| case VMMDEVREQ_REPORT_GUEST_STATUS: |
| case VMMDEVREQ_GET_DISPLAY_CHANGE_REQ: |
| case VMMDEVREQ_VIDEMODE_SUPPORTED: |
| case VMMDEVREQ_GET_HEIGHT_REDUCTION: |
| case VMMDEVREQ_GET_DISPLAY_CHANGE_REQ2: |
| case VMMDEVREQ_VIDEMODE_SUPPORTED2: |
| case VMMDEVREQ_VIDEO_ACCEL_ENABLE: |
| case VMMDEVREQ_VIDEO_ACCEL_FLUSH: |
| case VMMDEVREQ_VIDEO_SET_VISIBLE_REGION: |
| case VMMDEVREQ_VIDEO_UPDATE_MONITOR_POSITIONS: |
| case VMMDEVREQ_GET_DISPLAY_CHANGE_REQEX: |
| case VMMDEVREQ_GET_DISPLAY_CHANGE_REQ_MULTI: |
| case VMMDEVREQ_GET_SEAMLESS_CHANGE_REQ: |
| case VMMDEVREQ_GET_VRDPCHANGE_REQ: |
| case VMMDEVREQ_LOG_STRING: |
| case VMMDEVREQ_GET_SESSION_ID: |
| trusted_apps_only = false; |
| break; |
| |
| /* Depends on the request parameters... */ |
| case VMMDEVREQ_REPORT_GUEST_CAPABILITIES: |
| guest_status = (const struct vmmdev_guest_status *)req; |
| switch (guest_status->facility) { |
| case VBOXGUEST_FACILITY_TYPE_ALL: |
| case VBOXGUEST_FACILITY_TYPE_VBOXGUEST_DRIVER: |
| vbg_err("Denying userspace vmm report guest cap. call facility %#08x\n", |
| guest_status->facility); |
| return -EPERM; |
| case VBOXGUEST_FACILITY_TYPE_VBOX_SERVICE: |
| trusted_apps_only = true; |
| break; |
| case VBOXGUEST_FACILITY_TYPE_VBOX_TRAY_CLIENT: |
| case VBOXGUEST_FACILITY_TYPE_SEAMLESS: |
| case VBOXGUEST_FACILITY_TYPE_GRAPHICS: |
| default: |
| trusted_apps_only = false; |
| break; |
| } |
| break; |
| |
| /* Anything else is not allowed. */ |
| default: |
| vbg_err("Denying userspace vmm call type %#08x\n", |
| req->request_type); |
| return -EPERM; |
| } |
| |
| if (trusted_apps_only && |
| (session->requestor & VMMDEV_REQUESTOR_USER_DEVICE)) { |
| vbg_err("Denying userspace vmm call type %#08x through vboxuser device node\n", |
| req->request_type); |
| return -EPERM; |
| } |
| |
| return 0; |
| } |
| |
| static int vbg_ioctl_vmmrequest(struct vbg_dev *gdev, |
| struct vbg_session *session, void *data) |
| { |
| struct vbg_ioctl_hdr *hdr = data; |
| int ret; |
| |
| if (hdr->size_in != hdr->size_out) |
| return -EINVAL; |
| |
| if (hdr->size_in > VMMDEV_MAX_VMMDEVREQ_SIZE) |
| return -E2BIG; |
| |
| if (hdr->type == VBG_IOCTL_HDR_TYPE_DEFAULT) |
| return -EINVAL; |
| |
| ret = vbg_req_allowed(gdev, session, data); |
| if (ret < 0) |
| return ret; |
| |
| vbg_req_perform(gdev, data); |
| WARN_ON(hdr->rc == VINF_HGCM_ASYNC_EXECUTE); |
| |
| return 0; |
| } |
| |
| static int vbg_ioctl_hgcm_connect(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_hgcm_connect *conn) |
| { |
| u32 client_id; |
| int i, ret; |
| |
| if (vbg_ioctl_chk(&conn->hdr, sizeof(conn->u.in), sizeof(conn->u.out))) |
| return -EINVAL; |
| |
| /* Find a free place in the sessions clients array and claim it */ |
| mutex_lock(&gdev->session_mutex); |
| for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) { |
| if (!session->hgcm_client_ids[i]) { |
| session->hgcm_client_ids[i] = U32_MAX; |
| break; |
| } |
| } |
| mutex_unlock(&gdev->session_mutex); |
| |
| if (i >= ARRAY_SIZE(session->hgcm_client_ids)) |
| return -EMFILE; |
| |
| ret = vbg_hgcm_connect(gdev, session->requestor, &conn->u.in.loc, |
| &client_id, &conn->hdr.rc); |
| |
| mutex_lock(&gdev->session_mutex); |
| if (ret == 0 && conn->hdr.rc >= 0) { |
| conn->u.out.client_id = client_id; |
| session->hgcm_client_ids[i] = client_id; |
| } else { |
| conn->u.out.client_id = 0; |
| session->hgcm_client_ids[i] = 0; |
| } |
| mutex_unlock(&gdev->session_mutex); |
| |
| return ret; |
| } |
| |
| static int vbg_ioctl_hgcm_disconnect(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_hgcm_disconnect *disconn) |
| { |
| u32 client_id; |
| int i, ret; |
| |
| if (vbg_ioctl_chk(&disconn->hdr, sizeof(disconn->u.in), 0)) |
| return -EINVAL; |
| |
| client_id = disconn->u.in.client_id; |
| if (client_id == 0 || client_id == U32_MAX) |
| return -EINVAL; |
| |
| mutex_lock(&gdev->session_mutex); |
| for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) { |
| if (session->hgcm_client_ids[i] == client_id) { |
| session->hgcm_client_ids[i] = U32_MAX; |
| break; |
| } |
| } |
| mutex_unlock(&gdev->session_mutex); |
| |
| if (i >= ARRAY_SIZE(session->hgcm_client_ids)) |
| return -EINVAL; |
| |
| ret = vbg_hgcm_disconnect(gdev, session->requestor, client_id, |
| &disconn->hdr.rc); |
| |
| mutex_lock(&gdev->session_mutex); |
| if (ret == 0 && disconn->hdr.rc >= 0) |
| session->hgcm_client_ids[i] = 0; |
| else |
| session->hgcm_client_ids[i] = client_id; |
| mutex_unlock(&gdev->session_mutex); |
| |
| return ret; |
| } |
| |
| static bool vbg_param_valid(enum vmmdev_hgcm_function_parameter_type type) |
| { |
| switch (type) { |
| case VMMDEV_HGCM_PARM_TYPE_32BIT: |
| case VMMDEV_HGCM_PARM_TYPE_64BIT: |
| case VMMDEV_HGCM_PARM_TYPE_LINADDR: |
| case VMMDEV_HGCM_PARM_TYPE_LINADDR_IN: |
| case VMMDEV_HGCM_PARM_TYPE_LINADDR_OUT: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| static int vbg_ioctl_hgcm_call(struct vbg_dev *gdev, |
| struct vbg_session *session, bool f32bit, |
| struct vbg_ioctl_hgcm_call *call) |
| { |
| size_t actual_size; |
| u32 client_id; |
| int i, ret; |
| |
| if (call->hdr.size_in < sizeof(*call)) |
| return -EINVAL; |
| |
| if (call->hdr.size_in != call->hdr.size_out) |
| return -EINVAL; |
| |
| if (call->parm_count > VMMDEV_HGCM_MAX_PARMS) |
| return -E2BIG; |
| |
| client_id = call->client_id; |
| if (client_id == 0 || client_id == U32_MAX) |
| return -EINVAL; |
| |
| actual_size = sizeof(*call); |
| if (f32bit) |
| actual_size += call->parm_count * |
| sizeof(struct vmmdev_hgcm_function_parameter32); |
| else |
| actual_size += call->parm_count * |
| sizeof(struct vmmdev_hgcm_function_parameter); |
| if (call->hdr.size_in < actual_size) { |
| vbg_debug("VBG_IOCTL_HGCM_CALL: hdr.size_in %d required size is %zd\n", |
| call->hdr.size_in, actual_size); |
| return -EINVAL; |
| } |
| call->hdr.size_out = actual_size; |
| |
| /* Validate parameter types */ |
| if (f32bit) { |
| struct vmmdev_hgcm_function_parameter32 *parm = |
| VBG_IOCTL_HGCM_CALL_PARMS32(call); |
| |
| for (i = 0; i < call->parm_count; i++) |
| if (!vbg_param_valid(parm[i].type)) |
| return -EINVAL; |
| } else { |
| struct vmmdev_hgcm_function_parameter *parm = |
| VBG_IOCTL_HGCM_CALL_PARMS(call); |
| |
| for (i = 0; i < call->parm_count; i++) |
| if (!vbg_param_valid(parm[i].type)) |
| return -EINVAL; |
| } |
| |
| /* |
| * Validate the client id. |
| */ |
| mutex_lock(&gdev->session_mutex); |
| for (i = 0; i < ARRAY_SIZE(session->hgcm_client_ids); i++) |
| if (session->hgcm_client_ids[i] == client_id) |
| break; |
| mutex_unlock(&gdev->session_mutex); |
| if (i >= ARRAY_SIZE(session->hgcm_client_ids)) { |
| vbg_debug("VBG_IOCTL_HGCM_CALL: INVALID handle. u32Client=%#08x\n", |
| client_id); |
| return -EINVAL; |
| } |
| |
| if (IS_ENABLED(CONFIG_COMPAT) && f32bit) |
| ret = vbg_hgcm_call32(gdev, session->requestor, client_id, |
| call->function, call->timeout_ms, |
| VBG_IOCTL_HGCM_CALL_PARMS32(call), |
| call->parm_count, &call->hdr.rc); |
| else |
| ret = vbg_hgcm_call(gdev, session->requestor, client_id, |
| call->function, call->timeout_ms, |
| VBG_IOCTL_HGCM_CALL_PARMS(call), |
| call->parm_count, &call->hdr.rc); |
| |
| if (ret == -E2BIG) { |
| /* E2BIG needs to be reported through the hdr.rc field. */ |
| call->hdr.rc = VERR_OUT_OF_RANGE; |
| ret = 0; |
| } |
| |
| if (ret && ret != -EINTR && ret != -ETIMEDOUT) |
| vbg_err("VBG_IOCTL_HGCM_CALL error: %d\n", ret); |
| |
| return ret; |
| } |
| |
| static int vbg_ioctl_log(struct vbg_ioctl_log *log) |
| { |
| if (log->hdr.size_out != sizeof(log->hdr)) |
| return -EINVAL; |
| |
| vbg_info("%.*s", (int)(log->hdr.size_in - sizeof(log->hdr)), |
| log->u.in.msg); |
| |
| return 0; |
| } |
| |
| static int vbg_ioctl_change_filter_mask(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_change_filter *filter) |
| { |
| u32 or_mask, not_mask; |
| |
| if (vbg_ioctl_chk(&filter->hdr, sizeof(filter->u.in), 0)) |
| return -EINVAL; |
| |
| or_mask = filter->u.in.or_mask; |
| not_mask = filter->u.in.not_mask; |
| |
| if ((or_mask | not_mask) & ~VMMDEV_EVENT_VALID_EVENT_MASK) |
| return -EINVAL; |
| |
| return vbg_set_session_event_filter(gdev, session, or_mask, not_mask, |
| false); |
| } |
| |
| static int vbg_ioctl_acquire_guest_capabilities(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_acquire_guest_caps *caps) |
| { |
| u32 flags, or_mask, not_mask; |
| |
| if (vbg_ioctl_chk(&caps->hdr, sizeof(caps->u.in), 0)) |
| return -EINVAL; |
| |
| flags = caps->u.in.flags; |
| or_mask = caps->u.in.or_mask; |
| not_mask = caps->u.in.not_mask; |
| |
| if (flags & ~VBGL_IOC_AGC_FLAGS_VALID_MASK) |
| return -EINVAL; |
| |
| if ((or_mask | not_mask) & ~VMMDEV_GUEST_CAPABILITIES_MASK) |
| return -EINVAL; |
| |
| return vbg_acquire_session_capabilities(gdev, session, or_mask, |
| not_mask, flags, false); |
| } |
| |
| static int vbg_ioctl_change_guest_capabilities(struct vbg_dev *gdev, |
| struct vbg_session *session, struct vbg_ioctl_set_guest_caps *caps) |
| { |
| u32 or_mask, not_mask; |
| int ret; |
| |
| if (vbg_ioctl_chk(&caps->hdr, sizeof(caps->u.in), sizeof(caps->u.out))) |
| return -EINVAL; |
| |
| or_mask = caps->u.in.or_mask; |
| not_mask = caps->u.in.not_mask; |
| |
| if ((or_mask | not_mask) & ~VMMDEV_GUEST_CAPABILITIES_MASK) |
| return -EINVAL; |
| |
| ret = vbg_set_session_capabilities(gdev, session, or_mask, not_mask, |
| false); |
| if (ret) |
| return ret; |
| |
| caps->u.out.session_caps = session->set_guest_caps; |
| caps->u.out.global_caps = gdev->guest_caps_host; |
| |
| return 0; |
| } |
| |
| static int vbg_ioctl_check_balloon(struct vbg_dev *gdev, |
| struct vbg_ioctl_check_balloon *balloon_info) |
| { |
| if (vbg_ioctl_chk(&balloon_info->hdr, 0, sizeof(balloon_info->u.out))) |
| return -EINVAL; |
| |
| balloon_info->u.out.balloon_chunks = gdev->mem_balloon.chunks; |
| /* |
| * Under Linux we handle VMMDEV_EVENT_BALLOON_CHANGE_REQUEST |
| * events entirely in the kernel, see vbg_core_isr(). |
| */ |
| balloon_info->u.out.handle_in_r3 = false; |
| |
| return 0; |
| } |
| |
| static int vbg_ioctl_write_core_dump(struct vbg_dev *gdev, |
| struct vbg_session *session, |
| struct vbg_ioctl_write_coredump *dump) |
| { |
| struct vmmdev_write_core_dump *req; |
| |
| if (vbg_ioctl_chk(&dump->hdr, sizeof(dump->u.in), 0)) |
| return -EINVAL; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_WRITE_COREDUMP, |
| session->requestor); |
| if (!req) |
| return -ENOMEM; |
| |
| req->flags = dump->u.in.flags; |
| dump->hdr.rc = vbg_req_perform(gdev, req); |
| |
| vbg_req_free(req, sizeof(*req)); |
| return 0; |
| } |
| |
| /** |
| * Common IOCtl for user to kernel communication. |
| * Return: 0 or negative errno value. |
| * @session: The client session. |
| * @req: The requested function. |
| * @data: The i/o data buffer, minimum size sizeof(struct vbg_ioctl_hdr). |
| */ |
| int vbg_core_ioctl(struct vbg_session *session, unsigned int req, void *data) |
| { |
| unsigned int req_no_size = req & ~IOCSIZE_MASK; |
| struct vbg_dev *gdev = session->gdev; |
| struct vbg_ioctl_hdr *hdr = data; |
| bool f32bit = false; |
| |
| hdr->rc = VINF_SUCCESS; |
| if (!hdr->size_out) |
| hdr->size_out = hdr->size_in; |
| |
| /* |
| * hdr->version and hdr->size_in / hdr->size_out minimum size are |
| * already checked by vbg_misc_device_ioctl(). |
| */ |
| |
| /* For VMMDEV_REQUEST hdr->type != VBG_IOCTL_HDR_TYPE_DEFAULT */ |
| if (req_no_size == VBG_IOCTL_VMMDEV_REQUEST(0) || |
| req == VBG_IOCTL_VMMDEV_REQUEST_BIG || |
| req == VBG_IOCTL_VMMDEV_REQUEST_BIG_ALT) |
| return vbg_ioctl_vmmrequest(gdev, session, data); |
| |
| if (hdr->type != VBG_IOCTL_HDR_TYPE_DEFAULT) |
| return -EINVAL; |
| |
| /* Fixed size requests. */ |
| switch (req) { |
| case VBG_IOCTL_DRIVER_VERSION_INFO: |
| return vbg_ioctl_driver_version_info(data); |
| case VBG_IOCTL_HGCM_CONNECT: |
| return vbg_ioctl_hgcm_connect(gdev, session, data); |
| case VBG_IOCTL_HGCM_DISCONNECT: |
| return vbg_ioctl_hgcm_disconnect(gdev, session, data); |
| case VBG_IOCTL_WAIT_FOR_EVENTS: |
| return vbg_ioctl_wait_for_events(gdev, session, data); |
| case VBG_IOCTL_INTERRUPT_ALL_WAIT_FOR_EVENTS: |
| return vbg_ioctl_interrupt_all_wait_events(gdev, session, data); |
| case VBG_IOCTL_CHANGE_FILTER_MASK: |
| return vbg_ioctl_change_filter_mask(gdev, session, data); |
| case VBG_IOCTL_ACQUIRE_GUEST_CAPABILITIES: |
| return vbg_ioctl_acquire_guest_capabilities(gdev, session, data); |
| case VBG_IOCTL_CHANGE_GUEST_CAPABILITIES: |
| return vbg_ioctl_change_guest_capabilities(gdev, session, data); |
| case VBG_IOCTL_CHECK_BALLOON: |
| return vbg_ioctl_check_balloon(gdev, data); |
| case VBG_IOCTL_WRITE_CORE_DUMP: |
| return vbg_ioctl_write_core_dump(gdev, session, data); |
| } |
| |
| /* Variable sized requests. */ |
| switch (req_no_size) { |
| #ifdef CONFIG_COMPAT |
| case VBG_IOCTL_HGCM_CALL_32(0): |
| f32bit = true; |
| fallthrough; |
| #endif |
| case VBG_IOCTL_HGCM_CALL(0): |
| return vbg_ioctl_hgcm_call(gdev, session, f32bit, data); |
| case VBG_IOCTL_LOG(0): |
| case VBG_IOCTL_LOG_ALT(0): |
| return vbg_ioctl_log(data); |
| } |
| |
| vbg_err_ratelimited("Userspace made an unknown ioctl req %#08x\n", req); |
| return -ENOTTY; |
| } |
| |
| /** |
| * Report guest supported mouse-features to the host. |
| * |
| * Return: 0 or negative errno value. |
| * @gdev: The Guest extension device. |
| * @features: The set of features to report to the host. |
| */ |
| int vbg_core_set_mouse_status(struct vbg_dev *gdev, u32 features) |
| { |
| struct vmmdev_mouse_status *req; |
| int rc; |
| |
| req = vbg_req_alloc(sizeof(*req), VMMDEVREQ_SET_MOUSE_STATUS, |
| VBG_KERNEL_REQUEST); |
| if (!req) |
| return -ENOMEM; |
| |
| req->mouse_features = features; |
| req->pointer_pos_x = 0; |
| req->pointer_pos_y = 0; |
| |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) |
| vbg_err("%s error, rc: %d\n", __func__, rc); |
| |
| vbg_req_free(req, sizeof(*req)); |
| return vbg_status_code_to_errno(rc); |
| } |
| |
| /** Core interrupt service routine. */ |
| irqreturn_t vbg_core_isr(int irq, void *dev_id) |
| { |
| struct vbg_dev *gdev = dev_id; |
| struct vmmdev_events *req = gdev->ack_events_req; |
| bool mouse_position_changed = false; |
| unsigned long flags; |
| u32 events = 0; |
| int rc; |
| |
| if (!gdev->mmio->V.V1_04.have_events) |
| return IRQ_NONE; |
| |
| /* Get and acknowlegde events. */ |
| req->header.rc = VERR_INTERNAL_ERROR; |
| req->events = 0; |
| rc = vbg_req_perform(gdev, req); |
| if (rc < 0) { |
| vbg_err("Error performing events req, rc: %d\n", rc); |
| return IRQ_NONE; |
| } |
| |
| events = req->events; |
| |
| if (events & VMMDEV_EVENT_MOUSE_POSITION_CHANGED) { |
| mouse_position_changed = true; |
| events &= ~VMMDEV_EVENT_MOUSE_POSITION_CHANGED; |
| } |
| |
| if (events & VMMDEV_EVENT_HGCM) { |
| wake_up(&gdev->hgcm_wq); |
| events &= ~VMMDEV_EVENT_HGCM; |
| } |
| |
| if (events & VMMDEV_EVENT_BALLOON_CHANGE_REQUEST) { |
| schedule_work(&gdev->mem_balloon.work); |
| events &= ~VMMDEV_EVENT_BALLOON_CHANGE_REQUEST; |
| } |
| |
| if (events) { |
| spin_lock_irqsave(&gdev->event_spinlock, flags); |
| gdev->pending_events |= events; |
| spin_unlock_irqrestore(&gdev->event_spinlock, flags); |
| |
| wake_up(&gdev->event_wq); |
| } |
| |
| if (mouse_position_changed) |
| vbg_linux_mouse_event(gdev); |
| |
| return IRQ_HANDLED; |
| } |