| // SPDX-License-Identifier: MIT |
| /* |
| * Copyright (C) 2019 Google, Inc. |
| * |
| * Authors: |
| * Sean Paul <seanpaul@chromium.org> |
| */ |
| #include <linux/average.h> |
| #include <linux/bitops.h> |
| #include <linux/slab.h> |
| #include <linux/workqueue.h> |
| |
| #include <drm/drm_atomic.h> |
| #include <drm/drm_atomic_helper.h> |
| #include <drm/drm_connector.h> |
| #include <drm/drm_crtc.h> |
| #include <drm/drm_device.h> |
| #include <drm/drm_mode_config.h> |
| #include <drm/drm_modeset_lock.h> |
| #include <drm/drm_print.h> |
| #include <drm/drm_self_refresh_helper.h> |
| |
| /** |
| * DOC: overview |
| * |
| * This helper library provides an easy way for drivers to leverage the atomic |
| * framework to implement panel self refresh (SR) support. Drivers are |
| * responsible for initializing and cleaning up the SR helpers on load/unload |
| * (see &drm_self_refresh_helper_init/&drm_self_refresh_helper_cleanup). |
| * The connector is responsible for setting |
| * &drm_connector_state.self_refresh_aware to true at runtime if it is SR-aware |
| * (meaning it knows how to initiate self refresh on the panel). |
| * |
| * Once a crtc has enabled SR using &drm_self_refresh_helper_init, the |
| * helpers will monitor activity and call back into the driver to enable/disable |
| * SR as appropriate. The best way to think about this is that it's a DPMS |
| * on/off request with &drm_crtc_state.self_refresh_active set in crtc state |
| * that tells you to disable/enable SR on the panel instead of power-cycling it. |
| * |
| * During SR, drivers may choose to fully disable their crtc/encoder/bridge |
| * hardware (in which case no driver changes are necessary), or they can inspect |
| * &drm_crtc_state.self_refresh_active if they want to enter low power mode |
| * without full disable (in case full disable/enable is too slow). |
| * |
| * SR will be deactivated if there are any atomic updates affecting the |
| * pipe that is in SR mode. If a crtc is driving multiple connectors, all |
| * connectors must be SR aware and all will enter/exit SR mode at the same time. |
| * |
| * If the crtc and connector are SR aware, but the panel connected does not |
| * support it (or is otherwise unable to enter SR), the driver should fail |
| * atomic_check when &drm_crtc_state.self_refresh_active is true. |
| */ |
| |
| #define SELF_REFRESH_AVG_SEED_MS 200 |
| |
| DECLARE_EWMA(psr_time, 4, 4) |
| |
| struct drm_self_refresh_data { |
| struct drm_crtc *crtc; |
| struct delayed_work entry_work; |
| |
| struct mutex avg_mutex; |
| struct ewma_psr_time entry_avg_ms; |
| struct ewma_psr_time exit_avg_ms; |
| }; |
| |
| static void drm_self_refresh_helper_entry_work(struct work_struct *work) |
| { |
| struct drm_self_refresh_data *sr_data = container_of( |
| to_delayed_work(work), |
| struct drm_self_refresh_data, entry_work); |
| struct drm_crtc *crtc = sr_data->crtc; |
| struct drm_device *dev = crtc->dev; |
| struct drm_modeset_acquire_ctx ctx; |
| struct drm_atomic_state *state; |
| struct drm_connector *conn; |
| struct drm_connector_state *conn_state; |
| struct drm_crtc_state *crtc_state; |
| int i, ret = 0; |
| |
| drm_modeset_acquire_init(&ctx, 0); |
| |
| state = drm_atomic_state_alloc(dev); |
| if (!state) { |
| ret = -ENOMEM; |
| goto out_drop_locks; |
| } |
| |
| retry: |
| state->acquire_ctx = &ctx; |
| |
| crtc_state = drm_atomic_get_crtc_state(state, crtc); |
| if (IS_ERR(crtc_state)) { |
| ret = PTR_ERR(crtc_state); |
| goto out; |
| } |
| |
| if (!crtc_state->enable) |
| goto out; |
| |
| ret = drm_atomic_add_affected_connectors(state, crtc); |
| if (ret) |
| goto out; |
| |
| for_each_new_connector_in_state(state, conn, conn_state, i) { |
| if (!conn_state->self_refresh_aware) |
| goto out; |
| } |
| |
| crtc_state->active = false; |
| crtc_state->self_refresh_active = true; |
| |
| ret = drm_atomic_commit(state); |
| if (ret) |
| goto out; |
| |
| out: |
| if (ret == -EDEADLK) { |
| drm_atomic_state_clear(state); |
| ret = drm_modeset_backoff(&ctx); |
| if (!ret) |
| goto retry; |
| } |
| |
| drm_atomic_state_put(state); |
| |
| out_drop_locks: |
| drm_modeset_drop_locks(&ctx); |
| drm_modeset_acquire_fini(&ctx); |
| } |
| |
| /** |
| * drm_self_refresh_helper_update_avg_times - Updates a crtc's SR time averages |
| * @state: the state which has just been applied to hardware |
| * @commit_time_ms: the amount of time in ms that this commit took to complete |
| * @new_self_refresh_mask: bitmask of crtc's that have self_refresh_active in |
| * new state |
| * |
| * Called after &drm_mode_config_funcs.atomic_commit_tail, this function will |
| * update the average entry/exit self refresh times on self refresh transitions. |
| * These averages will be used when calculating how long to delay before |
| * entering self refresh mode after activity. |
| */ |
| void |
| drm_self_refresh_helper_update_avg_times(struct drm_atomic_state *state, |
| unsigned int commit_time_ms, |
| unsigned int new_self_refresh_mask) |
| { |
| struct drm_crtc *crtc; |
| struct drm_crtc_state *old_crtc_state; |
| int i; |
| |
| for_each_old_crtc_in_state(state, crtc, old_crtc_state, i) { |
| bool new_self_refresh_active = new_self_refresh_mask & BIT(i); |
| struct drm_self_refresh_data *sr_data = crtc->self_refresh_data; |
| struct ewma_psr_time *time; |
| |
| if (old_crtc_state->self_refresh_active == |
| new_self_refresh_active) |
| continue; |
| |
| if (new_self_refresh_active) |
| time = &sr_data->entry_avg_ms; |
| else |
| time = &sr_data->exit_avg_ms; |
| |
| mutex_lock(&sr_data->avg_mutex); |
| ewma_psr_time_add(time, commit_time_ms); |
| mutex_unlock(&sr_data->avg_mutex); |
| } |
| } |
| EXPORT_SYMBOL(drm_self_refresh_helper_update_avg_times); |
| |
| /** |
| * drm_self_refresh_helper_alter_state - Alters the atomic state for SR exit |
| * @state: the state currently being checked |
| * |
| * Called at the end of atomic check. This function checks the state for flags |
| * incompatible with self refresh exit and changes them. This is a bit |
| * disingenuous since userspace is expecting one thing and we're giving it |
| * another. However in order to keep self refresh entirely hidden from |
| * userspace, this is required. |
| * |
| * At the end, we queue up the self refresh entry work so we can enter PSR after |
| * the desired delay. |
| */ |
| void drm_self_refresh_helper_alter_state(struct drm_atomic_state *state) |
| { |
| struct drm_crtc *crtc; |
| struct drm_crtc_state *crtc_state; |
| int i; |
| |
| if (state->async_update || !state->allow_modeset) { |
| for_each_old_crtc_in_state(state, crtc, crtc_state, i) { |
| if (crtc_state->self_refresh_active) { |
| state->async_update = false; |
| state->allow_modeset = true; |
| break; |
| } |
| } |
| } |
| |
| for_each_new_crtc_in_state(state, crtc, crtc_state, i) { |
| struct drm_self_refresh_data *sr_data; |
| unsigned int delay; |
| |
| /* Don't trigger the entry timer when we're already in SR */ |
| if (crtc_state->self_refresh_active) |
| continue; |
| |
| sr_data = crtc->self_refresh_data; |
| if (!sr_data) |
| continue; |
| |
| mutex_lock(&sr_data->avg_mutex); |
| delay = (ewma_psr_time_read(&sr_data->entry_avg_ms) + |
| ewma_psr_time_read(&sr_data->exit_avg_ms)) * 2; |
| mutex_unlock(&sr_data->avg_mutex); |
| |
| mod_delayed_work(system_wq, &sr_data->entry_work, |
| msecs_to_jiffies(delay)); |
| } |
| } |
| EXPORT_SYMBOL(drm_self_refresh_helper_alter_state); |
| |
| /** |
| * drm_self_refresh_helper_init - Initializes self refresh helpers for a crtc |
| * @crtc: the crtc which supports self refresh supported displays |
| * |
| * Returns zero if successful or -errno on failure |
| */ |
| int drm_self_refresh_helper_init(struct drm_crtc *crtc) |
| { |
| struct drm_self_refresh_data *sr_data = crtc->self_refresh_data; |
| |
| /* Helper is already initialized */ |
| if (WARN_ON(sr_data)) |
| return -EINVAL; |
| |
| sr_data = kzalloc(sizeof(*sr_data), GFP_KERNEL); |
| if (!sr_data) |
| return -ENOMEM; |
| |
| INIT_DELAYED_WORK(&sr_data->entry_work, |
| drm_self_refresh_helper_entry_work); |
| sr_data->crtc = crtc; |
| mutex_init(&sr_data->avg_mutex); |
| ewma_psr_time_init(&sr_data->entry_avg_ms); |
| ewma_psr_time_init(&sr_data->exit_avg_ms); |
| |
| /* |
| * Seed the averages so they're non-zero (and sufficiently large |
| * for even poorly performing panels). As time goes on, this will be |
| * averaged out and the values will trend to their true value. |
| */ |
| ewma_psr_time_add(&sr_data->entry_avg_ms, SELF_REFRESH_AVG_SEED_MS); |
| ewma_psr_time_add(&sr_data->exit_avg_ms, SELF_REFRESH_AVG_SEED_MS); |
| |
| crtc->self_refresh_data = sr_data; |
| return 0; |
| } |
| EXPORT_SYMBOL(drm_self_refresh_helper_init); |
| |
| /** |
| * drm_self_refresh_helper_cleanup - Cleans up self refresh helpers for a crtc |
| * @crtc: the crtc to cleanup |
| */ |
| void drm_self_refresh_helper_cleanup(struct drm_crtc *crtc) |
| { |
| struct drm_self_refresh_data *sr_data = crtc->self_refresh_data; |
| |
| /* Helper is already uninitialized */ |
| if (!sr_data) |
| return; |
| |
| crtc->self_refresh_data = NULL; |
| |
| cancel_delayed_work_sync(&sr_data->entry_work); |
| kfree(sr_data); |
| } |
| EXPORT_SYMBOL(drm_self_refresh_helper_cleanup); |