| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright (C) 2017 NVIDIA CORPORATION. All rights reserved. |
| */ |
| |
| #include <linux/clk.h> |
| #include <linux/delay.h> |
| #include <linux/dma-mapping.h> |
| #include <linux/host1x.h> |
| #include <linux/module.h> |
| #include <linux/of.h> |
| #include <linux/of_device.h> |
| #include <linux/of_graph.h> |
| #include <linux/platform_device.h> |
| #include <linux/pm_runtime.h> |
| #include <linux/reset.h> |
| |
| #include <drm/drm_atomic.h> |
| #include <drm/drm_atomic_helper.h> |
| #include <drm/drm_blend.h> |
| #include <drm/drm_fourcc.h> |
| #include <drm/drm_framebuffer.h> |
| #include <drm/drm_probe_helper.h> |
| |
| #include "drm.h" |
| #include "dc.h" |
| #include "plane.h" |
| |
| #define NFB 24 |
| |
| static const u32 tegra_shared_plane_formats[] = { |
| DRM_FORMAT_ARGB1555, |
| DRM_FORMAT_RGB565, |
| DRM_FORMAT_RGBA5551, |
| DRM_FORMAT_ARGB8888, |
| DRM_FORMAT_ABGR8888, |
| /* new on Tegra114 */ |
| DRM_FORMAT_ABGR4444, |
| DRM_FORMAT_ABGR1555, |
| DRM_FORMAT_BGRA5551, |
| DRM_FORMAT_XRGB1555, |
| DRM_FORMAT_RGBX5551, |
| DRM_FORMAT_XBGR1555, |
| DRM_FORMAT_BGRX5551, |
| DRM_FORMAT_BGR565, |
| DRM_FORMAT_XRGB8888, |
| DRM_FORMAT_XBGR8888, |
| /* planar formats */ |
| DRM_FORMAT_UYVY, |
| DRM_FORMAT_YUYV, |
| DRM_FORMAT_YUV420, |
| DRM_FORMAT_YUV422, |
| }; |
| |
| static const u64 tegra_shared_plane_modifiers[] = { |
| DRM_FORMAT_MOD_LINEAR, |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0), |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1), |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2), |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3), |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4), |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5), |
| /* |
| * The GPU sector layout is only supported on Tegra194, but these will |
| * be filtered out later on by ->format_mod_supported() on SoCs where |
| * it isn't supported. |
| */ |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(0) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT, |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(1) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT, |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(2) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT, |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(3) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT, |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(4) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT, |
| DRM_FORMAT_MOD_NVIDIA_16BX2_BLOCK(5) | DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT, |
| /* sentinel */ |
| DRM_FORMAT_MOD_INVALID |
| }; |
| |
| static inline unsigned int tegra_plane_offset(struct tegra_plane *plane, |
| unsigned int offset) |
| { |
| if (offset >= 0x500 && offset <= 0x581) { |
| offset = 0x000 + (offset - 0x500); |
| return plane->offset + offset; |
| } |
| |
| if (offset >= 0x700 && offset <= 0x73c) { |
| offset = 0x180 + (offset - 0x700); |
| return plane->offset + offset; |
| } |
| |
| if (offset >= 0x800 && offset <= 0x83e) { |
| offset = 0x1c0 + (offset - 0x800); |
| return plane->offset + offset; |
| } |
| |
| dev_WARN(plane->dc->dev, "invalid offset: %x\n", offset); |
| |
| return plane->offset + offset; |
| } |
| |
| static inline u32 tegra_plane_readl(struct tegra_plane *plane, |
| unsigned int offset) |
| { |
| return tegra_dc_readl(plane->dc, tegra_plane_offset(plane, offset)); |
| } |
| |
| static inline void tegra_plane_writel(struct tegra_plane *plane, u32 value, |
| unsigned int offset) |
| { |
| tegra_dc_writel(plane->dc, value, tegra_plane_offset(plane, offset)); |
| } |
| |
| static int tegra_windowgroup_enable(struct tegra_windowgroup *wgrp) |
| { |
| int err = 0; |
| |
| mutex_lock(&wgrp->lock); |
| |
| if (wgrp->usecount == 0) { |
| err = host1x_client_resume(wgrp->parent); |
| if (err < 0) { |
| dev_err(wgrp->parent->dev, "failed to resume: %d\n", err); |
| goto unlock; |
| } |
| |
| reset_control_deassert(wgrp->rst); |
| } |
| |
| wgrp->usecount++; |
| |
| unlock: |
| mutex_unlock(&wgrp->lock); |
| return err; |
| } |
| |
| static void tegra_windowgroup_disable(struct tegra_windowgroup *wgrp) |
| { |
| int err; |
| |
| mutex_lock(&wgrp->lock); |
| |
| if (wgrp->usecount == 1) { |
| err = reset_control_assert(wgrp->rst); |
| if (err < 0) { |
| pr_err("failed to assert reset for window group %u\n", |
| wgrp->index); |
| } |
| |
| host1x_client_suspend(wgrp->parent); |
| } |
| |
| wgrp->usecount--; |
| mutex_unlock(&wgrp->lock); |
| } |
| |
| int tegra_display_hub_prepare(struct tegra_display_hub *hub) |
| { |
| unsigned int i; |
| |
| /* |
| * XXX Enabling/disabling windowgroups needs to happen when the owner |
| * display controller is disabled. There's currently no good point at |
| * which this could be executed, so unconditionally enable all window |
| * groups for now. |
| */ |
| for (i = 0; i < hub->soc->num_wgrps; i++) { |
| struct tegra_windowgroup *wgrp = &hub->wgrps[i]; |
| |
| /* Skip orphaned window group whose parent DC is disabled */ |
| if (wgrp->parent) |
| tegra_windowgroup_enable(wgrp); |
| } |
| |
| return 0; |
| } |
| |
| void tegra_display_hub_cleanup(struct tegra_display_hub *hub) |
| { |
| unsigned int i; |
| |
| /* |
| * XXX Remove this once window groups can be more fine-grainedly |
| * enabled and disabled. |
| */ |
| for (i = 0; i < hub->soc->num_wgrps; i++) { |
| struct tegra_windowgroup *wgrp = &hub->wgrps[i]; |
| |
| /* Skip orphaned window group whose parent DC is disabled */ |
| if (wgrp->parent) |
| tegra_windowgroup_disable(wgrp); |
| } |
| } |
| |
| static void tegra_shared_plane_update(struct tegra_plane *plane) |
| { |
| struct tegra_dc *dc = plane->dc; |
| unsigned long timeout; |
| u32 mask, value; |
| |
| mask = COMMON_UPDATE | WIN_A_UPDATE << plane->base.index; |
| tegra_dc_writel(dc, mask, DC_CMD_STATE_CONTROL); |
| |
| timeout = jiffies + msecs_to_jiffies(1000); |
| |
| while (time_before(jiffies, timeout)) { |
| value = tegra_dc_readl(dc, DC_CMD_STATE_CONTROL); |
| if ((value & mask) == 0) |
| break; |
| |
| usleep_range(100, 400); |
| } |
| } |
| |
| static void tegra_shared_plane_activate(struct tegra_plane *plane) |
| { |
| struct tegra_dc *dc = plane->dc; |
| unsigned long timeout; |
| u32 mask, value; |
| |
| mask = COMMON_ACTREQ | WIN_A_ACT_REQ << plane->base.index; |
| tegra_dc_writel(dc, mask, DC_CMD_STATE_CONTROL); |
| |
| timeout = jiffies + msecs_to_jiffies(1000); |
| |
| while (time_before(jiffies, timeout)) { |
| value = tegra_dc_readl(dc, DC_CMD_STATE_CONTROL); |
| if ((value & mask) == 0) |
| break; |
| |
| usleep_range(100, 400); |
| } |
| } |
| |
| static unsigned int |
| tegra_shared_plane_get_owner(struct tegra_plane *plane, struct tegra_dc *dc) |
| { |
| unsigned int offset = |
| tegra_plane_offset(plane, DC_WIN_CORE_WINDOWGROUP_SET_CONTROL); |
| |
| return tegra_dc_readl(dc, offset) & OWNER_MASK; |
| } |
| |
| static bool tegra_dc_owns_shared_plane(struct tegra_dc *dc, |
| struct tegra_plane *plane) |
| { |
| struct device *dev = dc->dev; |
| |
| if (tegra_shared_plane_get_owner(plane, dc) == dc->pipe) { |
| if (plane->dc == dc) |
| return true; |
| |
| dev_WARN(dev, "head %u owns window %u but is not attached\n", |
| dc->pipe, plane->index); |
| } |
| |
| return false; |
| } |
| |
| static int tegra_shared_plane_set_owner(struct tegra_plane *plane, |
| struct tegra_dc *new) |
| { |
| unsigned int offset = |
| tegra_plane_offset(plane, DC_WIN_CORE_WINDOWGROUP_SET_CONTROL); |
| struct tegra_dc *old = plane->dc, *dc = new ? new : old; |
| struct device *dev = new ? new->dev : old->dev; |
| unsigned int owner, index = plane->index; |
| u32 value; |
| |
| value = tegra_dc_readl(dc, offset); |
| owner = value & OWNER_MASK; |
| |
| if (new && (owner != OWNER_MASK && owner != new->pipe)) { |
| dev_WARN(dev, "window %u owned by head %u\n", index, owner); |
| return -EBUSY; |
| } |
| |
| /* |
| * This seems to happen whenever the head has been disabled with one |
| * or more windows being active. This is harmless because we'll just |
| * reassign the window to the new head anyway. |
| */ |
| if (old && owner == OWNER_MASK) |
| dev_dbg(dev, "window %u not owned by head %u but %u\n", index, |
| old->pipe, owner); |
| |
| value &= ~OWNER_MASK; |
| |
| if (new) |
| value |= OWNER(new->pipe); |
| else |
| value |= OWNER_MASK; |
| |
| tegra_dc_writel(dc, value, offset); |
| |
| plane->dc = new; |
| |
| return 0; |
| } |
| |
| static void tegra_shared_plane_setup_scaler(struct tegra_plane *plane) |
| { |
| static const unsigned int coeffs[192] = { |
| 0x00000000, 0x3c70e400, 0x3bb037e4, 0x0c51cc9c, |
| 0x00100001, 0x3bf0dbfa, 0x3d00f406, 0x3fe003ff, |
| 0x00300002, 0x3b80cbf5, 0x3da1040d, 0x3fb003fe, |
| 0x00400002, 0x3b20bff1, 0x3e511015, 0x3f9003fc, |
| 0x00500002, 0x3ad0b3ed, 0x3f21201d, 0x3f5003fb, |
| 0x00500003, 0x3aa0a3e9, 0x3ff13026, 0x3f2007f9, |
| 0x00500403, 0x3a7097e6, 0x00e1402f, 0x3ee007f7, |
| 0x00500403, 0x3a608be4, 0x01d14c38, 0x3ea00bf6, |
| 0x00500403, 0x3a507fe2, 0x02e15c42, 0x3e500ff4, |
| 0x00500402, 0x3a6073e1, 0x03f16c4d, 0x3e000ff2, |
| 0x00400402, 0x3a706be0, 0x05117858, 0x3db013f0, |
| 0x00300402, 0x3a905fe0, 0x06318863, 0x3d6017ee, |
| 0x00300402, 0x3ab057e0, 0x0771986e, 0x3d001beb, |
| 0x00200001, 0x3af04fe1, 0x08a1a47a, 0x3cb023e9, |
| 0x00100001, 0x3b2047e2, 0x09e1b485, 0x3c6027e7, |
| 0x00100000, 0x3b703fe2, 0x0b11c091, 0x3c002fe6, |
| 0x3f203800, 0x0391103f, 0x3ff0a014, 0x0811606c, |
| 0x3f2037ff, 0x0351083c, 0x03e11842, 0x3f203c00, |
| 0x3f302fff, 0x03010439, 0x04311c45, 0x3f104401, |
| 0x3f302fff, 0x02c0fc35, 0x04812448, 0x3f104802, |
| 0x3f4027ff, 0x0270f832, 0x04c1284b, 0x3f205003, |
| 0x3f4023ff, 0x0230f030, 0x0511304e, 0x3f205403, |
| 0x3f601fff, 0x01f0e82d, 0x05613451, 0x3f205c04, |
| 0x3f701bfe, 0x01b0e02a, 0x05a13c54, 0x3f306006, |
| 0x3f7017fe, 0x0170d827, 0x05f14057, 0x3f406807, |
| 0x3f8017ff, 0x0140d424, 0x0641445a, 0x3f406c08, |
| 0x3fa013ff, 0x0100cc22, 0x0681485d, 0x3f507409, |
| 0x3fa00fff, 0x00d0c41f, 0x06d14c60, 0x3f607c0b, |
| 0x3fc00fff, 0x0090bc1c, 0x07115063, 0x3f80840c, |
| 0x3fd00bff, 0x0070b41a, 0x07515465, 0x3f908c0e, |
| 0x3fe007ff, 0x0040b018, 0x07915868, 0x3fb0900f, |
| 0x3ff00400, 0x0010a816, 0x07d15c6a, 0x3fd09811, |
| 0x00a04c0e, 0x0460f442, 0x0240a827, 0x05c15859, |
| 0x0090440d, 0x0440f040, 0x0480fc43, 0x00b05010, |
| 0x0080400c, 0x0410ec3e, 0x04910044, 0x00d05411, |
| 0x0070380b, 0x03f0e83d, 0x04b10846, 0x00e05812, |
| 0x0060340a, 0x03d0e43b, 0x04d10c48, 0x00f06013, |
| 0x00503009, 0x03b0e039, 0x04e11449, 0x01106415, |
| 0x00402c08, 0x0390d838, 0x05011c4b, 0x01206c16, |
| 0x00302807, 0x0370d436, 0x0511204c, 0x01407018, |
| 0x00302406, 0x0340d034, 0x0531244e, 0x01507419, |
| 0x00202005, 0x0320cc32, 0x05412c50, 0x01707c1b, |
| 0x00101c04, 0x0300c431, 0x05613451, 0x0180801d, |
| 0x00101803, 0x02e0c02f, 0x05713853, 0x01a0881e, |
| 0x00101002, 0x02b0bc2d, 0x05814054, 0x01c08c20, |
| 0x00000c02, 0x02a0b82c, 0x05914455, 0x01e09421, |
| 0x00000801, 0x0280b02a, 0x05a14c57, 0x02009c23, |
| 0x00000400, 0x0260ac28, 0x05b15458, 0x0220a025, |
| }; |
| unsigned int ratio, row, column; |
| |
| for (ratio = 0; ratio <= 2; ratio++) { |
| for (row = 0; row <= 15; row++) { |
| for (column = 0; column <= 3; column++) { |
| unsigned int index = (ratio << 6) + (row << 2) + column; |
| u32 value; |
| |
| value = COEFF_INDEX(index) | COEFF_DATA(coeffs[index]); |
| tegra_plane_writel(plane, value, |
| DC_WIN_WINDOWGROUP_SET_INPUT_SCALER_COEFF); |
| } |
| } |
| } |
| } |
| |
| static void tegra_dc_assign_shared_plane(struct tegra_dc *dc, |
| struct tegra_plane *plane) |
| { |
| u32 value; |
| int err; |
| |
| if (!tegra_dc_owns_shared_plane(dc, plane)) { |
| err = tegra_shared_plane_set_owner(plane, dc); |
| if (err < 0) |
| return; |
| } |
| |
| value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_LINEBUF_CONFIG); |
| value |= MODE_FOUR_LINES; |
| tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_LINEBUF_CONFIG); |
| |
| value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_FETCH_METER); |
| value = SLOTS(1); |
| tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_FETCH_METER); |
| |
| /* disable watermark */ |
| value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLA); |
| value &= ~LATENCY_CTL_MODE_ENABLE; |
| tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLA); |
| |
| value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLB); |
| value |= WATERMARK_MASK; |
| tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_LATENCY_CTLB); |
| |
| /* pipe meter */ |
| value = tegra_plane_readl(plane, DC_WIN_CORE_PRECOMP_WGRP_PIPE_METER); |
| value = PIPE_METER_INT(0) | PIPE_METER_FRAC(0); |
| tegra_plane_writel(plane, value, DC_WIN_CORE_PRECOMP_WGRP_PIPE_METER); |
| |
| /* mempool entries */ |
| value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_WGRP_POOL_CONFIG); |
| value = MEMPOOL_ENTRIES(0x331); |
| tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_WGRP_POOL_CONFIG); |
| |
| value = tegra_plane_readl(plane, DC_WIN_CORE_IHUB_THREAD_GROUP); |
| value &= ~THREAD_NUM_MASK; |
| value |= THREAD_NUM(plane->base.index); |
| value |= THREAD_GROUP_ENABLE; |
| tegra_plane_writel(plane, value, DC_WIN_CORE_IHUB_THREAD_GROUP); |
| |
| tegra_shared_plane_setup_scaler(plane); |
| |
| tegra_shared_plane_update(plane); |
| tegra_shared_plane_activate(plane); |
| } |
| |
| static void tegra_dc_remove_shared_plane(struct tegra_dc *dc, |
| struct tegra_plane *plane) |
| { |
| tegra_shared_plane_set_owner(plane, NULL); |
| } |
| |
| static int tegra_shared_plane_atomic_check(struct drm_plane *plane, |
| struct drm_atomic_state *state) |
| { |
| struct drm_plane_state *new_plane_state = drm_atomic_get_new_plane_state(state, |
| plane); |
| struct tegra_plane_state *plane_state = to_tegra_plane_state(new_plane_state); |
| struct tegra_shared_plane *tegra = to_tegra_shared_plane(plane); |
| struct tegra_bo_tiling *tiling = &plane_state->tiling; |
| struct tegra_dc *dc = to_tegra_dc(new_plane_state->crtc); |
| int err; |
| |
| /* no need for further checks if the plane is being disabled */ |
| if (!new_plane_state->crtc || !new_plane_state->fb) |
| return 0; |
| |
| err = tegra_plane_format(new_plane_state->fb->format->format, |
| &plane_state->format, |
| &plane_state->swap); |
| if (err < 0) |
| return err; |
| |
| err = tegra_fb_get_tiling(new_plane_state->fb, tiling); |
| if (err < 0) |
| return err; |
| |
| if (tiling->mode == TEGRA_BO_TILING_MODE_BLOCK && |
| !dc->soc->supports_block_linear) { |
| DRM_ERROR("hardware doesn't support block linear mode\n"); |
| return -EINVAL; |
| } |
| |
| if (tiling->sector_layout == TEGRA_BO_SECTOR_LAYOUT_GPU && |
| !dc->soc->supports_sector_layout) { |
| DRM_ERROR("hardware doesn't support GPU sector layout\n"); |
| return -EINVAL; |
| } |
| |
| /* |
| * Tegra doesn't support different strides for U and V planes so we |
| * error out if the user tries to display a framebuffer with such a |
| * configuration. |
| */ |
| if (new_plane_state->fb->format->num_planes > 2) { |
| if (new_plane_state->fb->pitches[2] != new_plane_state->fb->pitches[1]) { |
| DRM_ERROR("unsupported UV-plane configuration\n"); |
| return -EINVAL; |
| } |
| } |
| |
| /* XXX scaling is not yet supported, add a check here */ |
| |
| err = tegra_plane_state_add(&tegra->base, new_plane_state); |
| if (err < 0) |
| return err; |
| |
| return 0; |
| } |
| |
| static void tegra_shared_plane_atomic_disable(struct drm_plane *plane, |
| struct drm_atomic_state *state) |
| { |
| struct drm_plane_state *old_state = drm_atomic_get_old_plane_state(state, |
| plane); |
| struct tegra_plane *p = to_tegra_plane(plane); |
| struct tegra_dc *dc; |
| u32 value; |
| int err; |
| |
| /* rien ne va plus */ |
| if (!old_state || !old_state->crtc) |
| return; |
| |
| dc = to_tegra_dc(old_state->crtc); |
| |
| err = host1x_client_resume(&dc->client); |
| if (err < 0) { |
| dev_err(dc->dev, "failed to resume: %d\n", err); |
| return; |
| } |
| |
| /* |
| * XXX Legacy helpers seem to sometimes call ->atomic_disable() even |
| * on planes that are already disabled. Make sure we fallback to the |
| * head for this particular state instead of crashing. |
| */ |
| if (WARN_ON(p->dc == NULL)) |
| p->dc = dc; |
| |
| value = tegra_plane_readl(p, DC_WIN_WIN_OPTIONS); |
| value &= ~WIN_ENABLE; |
| tegra_plane_writel(p, value, DC_WIN_WIN_OPTIONS); |
| |
| tegra_dc_remove_shared_plane(dc, p); |
| |
| host1x_client_suspend(&dc->client); |
| } |
| |
| static inline u32 compute_phase_incr(fixed20_12 in, unsigned int out) |
| { |
| u64 tmp, tmp1, tmp2; |
| |
| tmp = (u64)dfixed_trunc(in); |
| tmp2 = (u64)out; |
| tmp1 = (tmp << NFB) + (tmp2 >> 1); |
| do_div(tmp1, tmp2); |
| |
| return lower_32_bits(tmp1); |
| } |
| |
| static void tegra_shared_plane_atomic_update(struct drm_plane *plane, |
| struct drm_atomic_state *state) |
| { |
| struct drm_plane_state *new_state = drm_atomic_get_new_plane_state(state, |
| plane); |
| struct tegra_plane_state *tegra_plane_state = to_tegra_plane_state(new_state); |
| struct tegra_dc *dc = to_tegra_dc(new_state->crtc); |
| unsigned int zpos = new_state->normalized_zpos; |
| struct drm_framebuffer *fb = new_state->fb; |
| struct tegra_plane *p = to_tegra_plane(plane); |
| u32 value, min_width, bypass = 0; |
| dma_addr_t base, addr_flag = 0; |
| unsigned int bpc, planes; |
| bool yuv; |
| int err; |
| |
| /* rien ne va plus */ |
| if (!new_state->crtc || !new_state->fb) |
| return; |
| |
| if (!new_state->visible) { |
| tegra_shared_plane_atomic_disable(plane, state); |
| return; |
| } |
| |
| err = host1x_client_resume(&dc->client); |
| if (err < 0) { |
| dev_err(dc->dev, "failed to resume: %d\n", err); |
| return; |
| } |
| |
| yuv = tegra_plane_format_is_yuv(tegra_plane_state->format, &planes, &bpc); |
| |
| tegra_dc_assign_shared_plane(dc, p); |
| |
| tegra_plane_writel(p, VCOUNTER, DC_WIN_CORE_ACT_CONTROL); |
| |
| /* blending */ |
| value = BLEND_FACTOR_DST_ALPHA_ZERO | BLEND_FACTOR_SRC_ALPHA_K2 | |
| BLEND_FACTOR_DST_COLOR_NEG_K1_TIMES_SRC | |
| BLEND_FACTOR_SRC_COLOR_K1_TIMES_SRC; |
| tegra_plane_writel(p, value, DC_WIN_BLEND_MATCH_SELECT); |
| |
| value = BLEND_FACTOR_DST_ALPHA_ZERO | BLEND_FACTOR_SRC_ALPHA_K2 | |
| BLEND_FACTOR_DST_COLOR_NEG_K1_TIMES_SRC | |
| BLEND_FACTOR_SRC_COLOR_K1_TIMES_SRC; |
| tegra_plane_writel(p, value, DC_WIN_BLEND_NOMATCH_SELECT); |
| |
| value = K2(255) | K1(255) | WINDOW_LAYER_DEPTH(255 - zpos); |
| tegra_plane_writel(p, value, DC_WIN_BLEND_LAYER_CONTROL); |
| |
| /* scaling */ |
| min_width = min(new_state->src_w >> 16, new_state->crtc_w); |
| |
| value = tegra_plane_readl(p, DC_WINC_PRECOMP_WGRP_PIPE_CAPC); |
| |
| if (min_width < MAX_PIXELS_5TAP444(value)) { |
| value = HORIZONTAL_TAPS_5 | VERTICAL_TAPS_5; |
| } else { |
| value = tegra_plane_readl(p, DC_WINC_PRECOMP_WGRP_PIPE_CAPE); |
| |
| if (min_width < MAX_PIXELS_2TAP444(value)) |
| value = HORIZONTAL_TAPS_2 | VERTICAL_TAPS_2; |
| else |
| dev_err(dc->dev, "invalid minimum width: %u\n", min_width); |
| } |
| |
| value = HORIZONTAL_TAPS_5 | VERTICAL_TAPS_5; |
| tegra_plane_writel(p, value, DC_WIN_WINDOWGROUP_SET_CONTROL_INPUT_SCALER); |
| |
| if (new_state->src_w != new_state->crtc_w << 16) { |
| fixed20_12 width = dfixed_init(new_state->src_w >> 16); |
| u32 incr = compute_phase_incr(width, new_state->crtc_w) & ~0x1; |
| u32 init = (1 << (NFB - 1)) + (incr >> 1); |
| |
| tegra_plane_writel(p, incr, DC_WIN_SET_INPUT_SCALER_HPHASE_INCR); |
| tegra_plane_writel(p, init, DC_WIN_SET_INPUT_SCALER_H_START_PHASE); |
| } else { |
| bypass |= INPUT_SCALER_HBYPASS; |
| } |
| |
| if (new_state->src_h != new_state->crtc_h << 16) { |
| fixed20_12 height = dfixed_init(new_state->src_h >> 16); |
| u32 incr = compute_phase_incr(height, new_state->crtc_h) & ~0x1; |
| u32 init = (1 << (NFB - 1)) + (incr >> 1); |
| |
| tegra_plane_writel(p, incr, DC_WIN_SET_INPUT_SCALER_VPHASE_INCR); |
| tegra_plane_writel(p, init, DC_WIN_SET_INPUT_SCALER_V_START_PHASE); |
| } else { |
| bypass |= INPUT_SCALER_VBYPASS; |
| } |
| |
| tegra_plane_writel(p, bypass, DC_WIN_WINDOWGROUP_SET_INPUT_SCALER_USAGE); |
| |
| /* disable compression */ |
| tegra_plane_writel(p, 0, DC_WINBUF_CDE_CONTROL); |
| |
| #ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT |
| /* |
| * Physical address bit 39 in Tegra194 is used as a switch for special |
| * logic that swizzles the memory using either the legacy Tegra or the |
| * dGPU sector layout. |
| */ |
| if (tegra_plane_state->tiling.sector_layout == TEGRA_BO_SECTOR_LAYOUT_GPU) |
| addr_flag = BIT_ULL(39); |
| #endif |
| |
| base = tegra_plane_state->iova[0] + fb->offsets[0]; |
| base |= addr_flag; |
| |
| tegra_plane_writel(p, tegra_plane_state->format, DC_WIN_COLOR_DEPTH); |
| tegra_plane_writel(p, 0, DC_WIN_PRECOMP_WGRP_PARAMS); |
| |
| value = V_POSITION(new_state->crtc_y) | |
| H_POSITION(new_state->crtc_x); |
| tegra_plane_writel(p, value, DC_WIN_POSITION); |
| |
| value = V_SIZE(new_state->crtc_h) | H_SIZE(new_state->crtc_w); |
| tegra_plane_writel(p, value, DC_WIN_SIZE); |
| |
| value = WIN_ENABLE | COLOR_EXPAND; |
| tegra_plane_writel(p, value, DC_WIN_WIN_OPTIONS); |
| |
| value = V_SIZE(new_state->src_h >> 16) | H_SIZE(new_state->src_w >> 16); |
| tegra_plane_writel(p, value, DC_WIN_CROPPED_SIZE); |
| |
| tegra_plane_writel(p, upper_32_bits(base), DC_WINBUF_START_ADDR_HI); |
| tegra_plane_writel(p, lower_32_bits(base), DC_WINBUF_START_ADDR); |
| |
| value = PITCH(fb->pitches[0]); |
| tegra_plane_writel(p, value, DC_WIN_PLANAR_STORAGE); |
| |
| if (yuv && planes > 1) { |
| base = tegra_plane_state->iova[1] + fb->offsets[1]; |
| base |= addr_flag; |
| |
| tegra_plane_writel(p, upper_32_bits(base), DC_WINBUF_START_ADDR_HI_U); |
| tegra_plane_writel(p, lower_32_bits(base), DC_WINBUF_START_ADDR_U); |
| |
| if (planes > 2) { |
| base = tegra_plane_state->iova[2] + fb->offsets[2]; |
| base |= addr_flag; |
| |
| tegra_plane_writel(p, upper_32_bits(base), DC_WINBUF_START_ADDR_HI_V); |
| tegra_plane_writel(p, lower_32_bits(base), DC_WINBUF_START_ADDR_V); |
| } |
| |
| value = PITCH_U(fb->pitches[1]); |
| |
| if (planes > 2) |
| value |= PITCH_V(fb->pitches[2]); |
| |
| tegra_plane_writel(p, value, DC_WIN_PLANAR_STORAGE_UV); |
| } else { |
| tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_U); |
| tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_HI_U); |
| tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_V); |
| tegra_plane_writel(p, 0, DC_WINBUF_START_ADDR_HI_V); |
| tegra_plane_writel(p, 0, DC_WIN_PLANAR_STORAGE_UV); |
| } |
| |
| value = CLAMP_BEFORE_BLEND | INPUT_RANGE_FULL; |
| |
| if (yuv) { |
| if (bpc < 12) |
| value |= DEGAMMA_YUV8_10; |
| else |
| value |= DEGAMMA_YUV12; |
| |
| /* XXX parameterize */ |
| value |= COLOR_SPACE_YUV_2020; |
| } else { |
| if (!tegra_plane_format_is_indexed(tegra_plane_state->format)) |
| value |= DEGAMMA_SRGB; |
| } |
| |
| tegra_plane_writel(p, value, DC_WIN_SET_PARAMS); |
| |
| value = OFFSET_X(new_state->src_y >> 16) | |
| OFFSET_Y(new_state->src_x >> 16); |
| tegra_plane_writel(p, value, DC_WINBUF_CROPPED_POINT); |
| |
| if (dc->soc->supports_block_linear) { |
| unsigned long height = tegra_plane_state->tiling.value; |
| |
| /* XXX */ |
| switch (tegra_plane_state->tiling.mode) { |
| case TEGRA_BO_TILING_MODE_PITCH: |
| value = DC_WINBUF_SURFACE_KIND_BLOCK_HEIGHT(0) | |
| DC_WINBUF_SURFACE_KIND_PITCH; |
| break; |
| |
| /* XXX not supported on Tegra186 and later */ |
| case TEGRA_BO_TILING_MODE_TILED: |
| value = DC_WINBUF_SURFACE_KIND_TILED; |
| break; |
| |
| case TEGRA_BO_TILING_MODE_BLOCK: |
| value = DC_WINBUF_SURFACE_KIND_BLOCK_HEIGHT(height) | |
| DC_WINBUF_SURFACE_KIND_BLOCK; |
| break; |
| } |
| |
| tegra_plane_writel(p, value, DC_WINBUF_SURFACE_KIND); |
| } |
| |
| /* disable gamut CSC */ |
| value = tegra_plane_readl(p, DC_WIN_WINDOW_SET_CONTROL); |
| value &= ~CONTROL_CSC_ENABLE; |
| tegra_plane_writel(p, value, DC_WIN_WINDOW_SET_CONTROL); |
| |
| host1x_client_suspend(&dc->client); |
| } |
| |
| static const struct drm_plane_helper_funcs tegra_shared_plane_helper_funcs = { |
| .prepare_fb = tegra_plane_prepare_fb, |
| .cleanup_fb = tegra_plane_cleanup_fb, |
| .atomic_check = tegra_shared_plane_atomic_check, |
| .atomic_update = tegra_shared_plane_atomic_update, |
| .atomic_disable = tegra_shared_plane_atomic_disable, |
| }; |
| |
| struct drm_plane *tegra_shared_plane_create(struct drm_device *drm, |
| struct tegra_dc *dc, |
| unsigned int wgrp, |
| unsigned int index) |
| { |
| enum drm_plane_type type = DRM_PLANE_TYPE_OVERLAY; |
| struct tegra_drm *tegra = drm->dev_private; |
| struct tegra_display_hub *hub = tegra->hub; |
| struct tegra_shared_plane *plane; |
| unsigned int possible_crtcs; |
| unsigned int num_formats; |
| const u64 *modifiers; |
| struct drm_plane *p; |
| const u32 *formats; |
| int err; |
| |
| plane = kzalloc(sizeof(*plane), GFP_KERNEL); |
| if (!plane) |
| return ERR_PTR(-ENOMEM); |
| |
| plane->base.offset = 0x0a00 + 0x0300 * index; |
| plane->base.index = index; |
| |
| plane->wgrp = &hub->wgrps[wgrp]; |
| plane->wgrp->parent = &dc->client; |
| |
| p = &plane->base.base; |
| |
| /* planes can be assigned to arbitrary CRTCs */ |
| possible_crtcs = BIT(tegra->num_crtcs) - 1; |
| |
| num_formats = ARRAY_SIZE(tegra_shared_plane_formats); |
| formats = tegra_shared_plane_formats; |
| modifiers = tegra_shared_plane_modifiers; |
| |
| err = drm_universal_plane_init(drm, p, possible_crtcs, |
| &tegra_plane_funcs, formats, |
| num_formats, modifiers, type, NULL); |
| if (err < 0) { |
| kfree(plane); |
| return ERR_PTR(err); |
| } |
| |
| drm_plane_helper_add(p, &tegra_shared_plane_helper_funcs); |
| drm_plane_create_zpos_property(p, 0, 0, 255); |
| |
| return p; |
| } |
| |
| static struct drm_private_state * |
| tegra_display_hub_duplicate_state(struct drm_private_obj *obj) |
| { |
| struct tegra_display_hub_state *state; |
| |
| state = kmemdup(obj->state, sizeof(*state), GFP_KERNEL); |
| if (!state) |
| return NULL; |
| |
| __drm_atomic_helper_private_obj_duplicate_state(obj, &state->base); |
| |
| return &state->base; |
| } |
| |
| static void tegra_display_hub_destroy_state(struct drm_private_obj *obj, |
| struct drm_private_state *state) |
| { |
| struct tegra_display_hub_state *hub_state = |
| to_tegra_display_hub_state(state); |
| |
| kfree(hub_state); |
| } |
| |
| static const struct drm_private_state_funcs tegra_display_hub_state_funcs = { |
| .atomic_duplicate_state = tegra_display_hub_duplicate_state, |
| .atomic_destroy_state = tegra_display_hub_destroy_state, |
| }; |
| |
| static struct tegra_display_hub_state * |
| tegra_display_hub_get_state(struct tegra_display_hub *hub, |
| struct drm_atomic_state *state) |
| { |
| struct drm_private_state *priv; |
| |
| priv = drm_atomic_get_private_obj_state(state, &hub->base); |
| if (IS_ERR(priv)) |
| return ERR_CAST(priv); |
| |
| return to_tegra_display_hub_state(priv); |
| } |
| |
| int tegra_display_hub_atomic_check(struct drm_device *drm, |
| struct drm_atomic_state *state) |
| { |
| struct tegra_drm *tegra = drm->dev_private; |
| struct tegra_display_hub_state *hub_state; |
| struct drm_crtc_state *old, *new; |
| struct drm_crtc *crtc; |
| unsigned int i; |
| |
| if (!tegra->hub) |
| return 0; |
| |
| hub_state = tegra_display_hub_get_state(tegra->hub, state); |
| if (IS_ERR(hub_state)) |
| return PTR_ERR(hub_state); |
| |
| /* |
| * The display hub display clock needs to be fed by the display clock |
| * with the highest frequency to ensure proper functioning of all the |
| * displays. |
| * |
| * Note that this isn't used before Tegra186, but it doesn't hurt and |
| * conditionalizing it would make the code less clean. |
| */ |
| for_each_oldnew_crtc_in_state(state, crtc, old, new, i) { |
| struct tegra_dc_state *dc = to_dc_state(new); |
| |
| if (new->active) { |
| if (!hub_state->clk || dc->pclk > hub_state->rate) { |
| hub_state->dc = to_tegra_dc(dc->base.crtc); |
| hub_state->clk = hub_state->dc->clk; |
| hub_state->rate = dc->pclk; |
| } |
| } |
| } |
| |
| return 0; |
| } |
| |
| static void tegra_display_hub_update(struct tegra_dc *dc) |
| { |
| u32 value; |
| int err; |
| |
| err = host1x_client_resume(&dc->client); |
| if (err < 0) { |
| dev_err(dc->dev, "failed to resume: %d\n", err); |
| return; |
| } |
| |
| value = tegra_dc_readl(dc, DC_CMD_IHUB_COMMON_MISC_CTL); |
| value &= ~LATENCY_EVENT; |
| tegra_dc_writel(dc, value, DC_CMD_IHUB_COMMON_MISC_CTL); |
| |
| value = tegra_dc_readl(dc, DC_DISP_IHUB_COMMON_DISPLAY_FETCH_METER); |
| value = CURS_SLOTS(1) | WGRP_SLOTS(1); |
| tegra_dc_writel(dc, value, DC_DISP_IHUB_COMMON_DISPLAY_FETCH_METER); |
| |
| tegra_dc_writel(dc, COMMON_UPDATE, DC_CMD_STATE_CONTROL); |
| tegra_dc_readl(dc, DC_CMD_STATE_CONTROL); |
| tegra_dc_writel(dc, COMMON_ACTREQ, DC_CMD_STATE_CONTROL); |
| tegra_dc_readl(dc, DC_CMD_STATE_CONTROL); |
| |
| host1x_client_suspend(&dc->client); |
| } |
| |
| void tegra_display_hub_atomic_commit(struct drm_device *drm, |
| struct drm_atomic_state *state) |
| { |
| struct tegra_drm *tegra = drm->dev_private; |
| struct tegra_display_hub *hub = tegra->hub; |
| struct tegra_display_hub_state *hub_state; |
| struct device *dev = hub->client.dev; |
| int err; |
| |
| hub_state = to_tegra_display_hub_state(hub->base.state); |
| |
| if (hub_state->clk) { |
| err = clk_set_rate(hub_state->clk, hub_state->rate); |
| if (err < 0) |
| dev_err(dev, "failed to set rate of %pC to %lu Hz\n", |
| hub_state->clk, hub_state->rate); |
| |
| err = clk_set_parent(hub->clk_disp, hub_state->clk); |
| if (err < 0) |
| dev_err(dev, "failed to set parent of %pC to %pC: %d\n", |
| hub->clk_disp, hub_state->clk, err); |
| } |
| |
| if (hub_state->dc) |
| tegra_display_hub_update(hub_state->dc); |
| } |
| |
| static int tegra_display_hub_init(struct host1x_client *client) |
| { |
| struct tegra_display_hub *hub = to_tegra_display_hub(client); |
| struct drm_device *drm = dev_get_drvdata(client->host); |
| struct tegra_drm *tegra = drm->dev_private; |
| struct tegra_display_hub_state *state; |
| |
| state = kzalloc(sizeof(*state), GFP_KERNEL); |
| if (!state) |
| return -ENOMEM; |
| |
| drm_atomic_private_obj_init(drm, &hub->base, &state->base, |
| &tegra_display_hub_state_funcs); |
| |
| tegra->hub = hub; |
| |
| return 0; |
| } |
| |
| static int tegra_display_hub_exit(struct host1x_client *client) |
| { |
| struct drm_device *drm = dev_get_drvdata(client->host); |
| struct tegra_drm *tegra = drm->dev_private; |
| |
| drm_atomic_private_obj_fini(&tegra->hub->base); |
| tegra->hub = NULL; |
| |
| return 0; |
| } |
| |
| static int tegra_display_hub_runtime_suspend(struct host1x_client *client) |
| { |
| struct tegra_display_hub *hub = to_tegra_display_hub(client); |
| struct device *dev = client->dev; |
| unsigned int i = hub->num_heads; |
| int err; |
| |
| err = reset_control_assert(hub->rst); |
| if (err < 0) |
| return err; |
| |
| while (i--) |
| clk_disable_unprepare(hub->clk_heads[i]); |
| |
| clk_disable_unprepare(hub->clk_hub); |
| clk_disable_unprepare(hub->clk_dsc); |
| clk_disable_unprepare(hub->clk_disp); |
| |
| pm_runtime_put_sync(dev); |
| |
| return 0; |
| } |
| |
| static int tegra_display_hub_runtime_resume(struct host1x_client *client) |
| { |
| struct tegra_display_hub *hub = to_tegra_display_hub(client); |
| struct device *dev = client->dev; |
| unsigned int i; |
| int err; |
| |
| err = pm_runtime_resume_and_get(dev); |
| if (err < 0) { |
| dev_err(dev, "failed to get runtime PM: %d\n", err); |
| return err; |
| } |
| |
| err = clk_prepare_enable(hub->clk_disp); |
| if (err < 0) |
| goto put_rpm; |
| |
| err = clk_prepare_enable(hub->clk_dsc); |
| if (err < 0) |
| goto disable_disp; |
| |
| err = clk_prepare_enable(hub->clk_hub); |
| if (err < 0) |
| goto disable_dsc; |
| |
| for (i = 0; i < hub->num_heads; i++) { |
| err = clk_prepare_enable(hub->clk_heads[i]); |
| if (err < 0) |
| goto disable_heads; |
| } |
| |
| err = reset_control_deassert(hub->rst); |
| if (err < 0) |
| goto disable_heads; |
| |
| return 0; |
| |
| disable_heads: |
| while (i--) |
| clk_disable_unprepare(hub->clk_heads[i]); |
| |
| clk_disable_unprepare(hub->clk_hub); |
| disable_dsc: |
| clk_disable_unprepare(hub->clk_dsc); |
| disable_disp: |
| clk_disable_unprepare(hub->clk_disp); |
| put_rpm: |
| pm_runtime_put_sync(dev); |
| return err; |
| } |
| |
| static const struct host1x_client_ops tegra_display_hub_ops = { |
| .init = tegra_display_hub_init, |
| .exit = tegra_display_hub_exit, |
| .suspend = tegra_display_hub_runtime_suspend, |
| .resume = tegra_display_hub_runtime_resume, |
| }; |
| |
| static int tegra_display_hub_probe(struct platform_device *pdev) |
| { |
| u64 dma_mask = dma_get_mask(pdev->dev.parent); |
| struct device_node *child = NULL; |
| struct tegra_display_hub *hub; |
| struct clk *clk; |
| unsigned int i; |
| int err; |
| |
| err = dma_coerce_mask_and_coherent(&pdev->dev, dma_mask); |
| if (err < 0) { |
| dev_err(&pdev->dev, "failed to set DMA mask: %d\n", err); |
| return err; |
| } |
| |
| hub = devm_kzalloc(&pdev->dev, sizeof(*hub), GFP_KERNEL); |
| if (!hub) |
| return -ENOMEM; |
| |
| hub->soc = of_device_get_match_data(&pdev->dev); |
| |
| hub->clk_disp = devm_clk_get(&pdev->dev, "disp"); |
| if (IS_ERR(hub->clk_disp)) { |
| err = PTR_ERR(hub->clk_disp); |
| return err; |
| } |
| |
| if (hub->soc->supports_dsc) { |
| hub->clk_dsc = devm_clk_get(&pdev->dev, "dsc"); |
| if (IS_ERR(hub->clk_dsc)) { |
| err = PTR_ERR(hub->clk_dsc); |
| return err; |
| } |
| } |
| |
| hub->clk_hub = devm_clk_get(&pdev->dev, "hub"); |
| if (IS_ERR(hub->clk_hub)) { |
| err = PTR_ERR(hub->clk_hub); |
| return err; |
| } |
| |
| hub->rst = devm_reset_control_get(&pdev->dev, "misc"); |
| if (IS_ERR(hub->rst)) { |
| err = PTR_ERR(hub->rst); |
| return err; |
| } |
| |
| hub->wgrps = devm_kcalloc(&pdev->dev, hub->soc->num_wgrps, |
| sizeof(*hub->wgrps), GFP_KERNEL); |
| if (!hub->wgrps) |
| return -ENOMEM; |
| |
| for (i = 0; i < hub->soc->num_wgrps; i++) { |
| struct tegra_windowgroup *wgrp = &hub->wgrps[i]; |
| char id[8]; |
| |
| snprintf(id, sizeof(id), "wgrp%u", i); |
| mutex_init(&wgrp->lock); |
| wgrp->usecount = 0; |
| wgrp->index = i; |
| |
| wgrp->rst = devm_reset_control_get(&pdev->dev, id); |
| if (IS_ERR(wgrp->rst)) |
| return PTR_ERR(wgrp->rst); |
| |
| err = reset_control_assert(wgrp->rst); |
| if (err < 0) |
| return err; |
| } |
| |
| hub->num_heads = of_get_child_count(pdev->dev.of_node); |
| |
| hub->clk_heads = devm_kcalloc(&pdev->dev, hub->num_heads, sizeof(clk), |
| GFP_KERNEL); |
| if (!hub->clk_heads) |
| return -ENOMEM; |
| |
| for (i = 0; i < hub->num_heads; i++) { |
| child = of_get_next_child(pdev->dev.of_node, child); |
| if (!child) { |
| dev_err(&pdev->dev, "failed to find node for head %u\n", |
| i); |
| return -ENODEV; |
| } |
| |
| clk = devm_get_clk_from_child(&pdev->dev, child, "dc"); |
| if (IS_ERR(clk)) { |
| dev_err(&pdev->dev, "failed to get clock for head %u\n", |
| i); |
| of_node_put(child); |
| return PTR_ERR(clk); |
| } |
| |
| hub->clk_heads[i] = clk; |
| } |
| |
| of_node_put(child); |
| |
| /* XXX: enable clock across reset? */ |
| err = reset_control_assert(hub->rst); |
| if (err < 0) |
| return err; |
| |
| platform_set_drvdata(pdev, hub); |
| pm_runtime_enable(&pdev->dev); |
| |
| INIT_LIST_HEAD(&hub->client.list); |
| hub->client.ops = &tegra_display_hub_ops; |
| hub->client.dev = &pdev->dev; |
| |
| err = host1x_client_register(&hub->client); |
| if (err < 0) |
| dev_err(&pdev->dev, "failed to register host1x client: %d\n", |
| err); |
| |
| err = devm_of_platform_populate(&pdev->dev); |
| if (err < 0) |
| goto unregister; |
| |
| return err; |
| |
| unregister: |
| host1x_client_unregister(&hub->client); |
| pm_runtime_disable(&pdev->dev); |
| return err; |
| } |
| |
| static void tegra_display_hub_remove(struct platform_device *pdev) |
| { |
| struct tegra_display_hub *hub = platform_get_drvdata(pdev); |
| unsigned int i; |
| |
| host1x_client_unregister(&hub->client); |
| |
| for (i = 0; i < hub->soc->num_wgrps; i++) { |
| struct tegra_windowgroup *wgrp = &hub->wgrps[i]; |
| |
| mutex_destroy(&wgrp->lock); |
| } |
| |
| pm_runtime_disable(&pdev->dev); |
| } |
| |
| static const struct tegra_display_hub_soc tegra186_display_hub = { |
| .num_wgrps = 6, |
| .supports_dsc = true, |
| }; |
| |
| static const struct tegra_display_hub_soc tegra194_display_hub = { |
| .num_wgrps = 6, |
| .supports_dsc = false, |
| }; |
| |
| static const struct of_device_id tegra_display_hub_of_match[] = { |
| { |
| .compatible = "nvidia,tegra194-display", |
| .data = &tegra194_display_hub |
| }, { |
| .compatible = "nvidia,tegra186-display", |
| .data = &tegra186_display_hub |
| }, { |
| /* sentinel */ |
| } |
| }; |
| MODULE_DEVICE_TABLE(of, tegra_display_hub_of_match); |
| |
| struct platform_driver tegra_display_hub_driver = { |
| .driver = { |
| .name = "tegra-display-hub", |
| .of_match_table = tegra_display_hub_of_match, |
| }, |
| .probe = tegra_display_hub_probe, |
| .remove_new = tegra_display_hub_remove, |
| }; |