mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-01-18 02:46:06 +00:00
fad3dad832
- Fix documentation of DP tunnel functions (Imre) - DP MST read sideband messaging cap (Jani) - Preparation patches for Adaptive Sync SDP Support for DP (Mitul) Driver Changes: i915 core (non-display): - Documentation improvements (Nirmoy) - Add includes for BUG_ON/BUILD_BUG_ON in i915_memcpy.c (Joonas) - Do not print 'pxp init failed with 0' when it succeed (Jose) - Clean-up, including removal of dead code for unsupported platforms (Lucas) - Adding new DG2 PCI ID (Ravi) {i915,xe} display: - Spelling fix (Colin Ian) - Document CDCLK components (Gustavo) - Lunar Lake display enabling, including cdclk and other refactors (Gustavo, Bala) - BIOS/VBT/opregion related refactor (Jani, Ville, RK) - Save a few bytes of memory using {kstrdup,kfree}_const variant (Christophe) - Digital port related refactor/clean-up (Ville) - Fix 2s boot time regression on DP panel replay init (Animesh) - Remove redundant drm_rect_visible() overlay use (Arthur) - DSC HW state readout fixes (Imre) - Remove duplication on audio enable/disable on SDVO and g4x+ DP (Ville) - Disable AuxCCS framebuffers if built for Xe (Juha-Pekka) - Fix DSI init order (Ville) - DRRS related refactor and fixes (Bhanuprakash) - Fix DSB vblank waits with VRR (Ville) - General improvements on register name and use of REG_BIT (Ville) - Some display power well related improvements (Ville) - FBC changes for better w/a handling (Ville) - Make crtc disable more atomic (Ville) - Fix hwmon locking inversion in sysfs getter (Janusz) - Increase DP idle pattern wait timeout to 2ms (Shekhar) - PSR related fixes and improvents (Jouni) - Start using container_of_const() for some extra const safety (Ville) - Use drm_printer more on display code (Ville) - Fix Jasper Lake boot freeze (Jonathon) - Update Pipe src size check in skl_update_scaler (Ankit) - Enable MST mode for 128b/132b single-stream sideband (Jani) - Pass encoder around more for port/phy checks (Jani) - Some initial work to make display code more independent from i915 (Jani) - Pre-populate the cursor physical dma address (Ville) - Do not bump min backlight brightness to max on enable (Gareth) - Fix MTL supported DP rates - removal of UHBR13.5 (Arun) - Fix the computation for compressed_bpp for DISPLAY < 1 (Ankit) - Bigjoiner modeset sequence redesign and MST support (Ville) - Enable Adaptive Sync SDP Support for DP (Mitul) - Implemnt vblank sycnhronized mbus joining changes (Ville, Stanislav) - HDCP related fixes (Suraj) - Fix i915_display_info debugfs when connectors are not active (Ville) - Clean up on Xe compat layer (Jani) - Add jitter WAs for MST/FEC/DSC links (Imre) - DMC wakelock implementation (Luca) -----BEGIN PGP SIGNATURE----- iQEzBAABCAAdFiEEbSBwaO7dZQkcLOKj+mJfZA7rE8oFAmYfzQEACgkQ+mJfZA7r E8qYvAf/T8KrEewHOWz7NOaKcFRCNYaF4QTdVOfgHUYBX5NPDF/xzwFdHCL8QWQu bwKwE2b94VEyruG3DYwTMd8GNcDxrsOrmU0IZe3PVkm+BvHLTmrOqL6BlCd85zXF 02IuE+LCaWREmmpLMcsDMxsaaq8yp+cw9/F0jJDrH6LiyfxFriefxyZYpGYjRCuv 8GP1fHXLFV2yys4rveR/+y9xIhgy82mVcg3/Kfk0+er7gALkY6Vc0N38wedET9MZ ZPfVidBeaTkIKcCDFKnFzGjG+9rNQ7NFrXyS7Hl97VolGt2l03qGGPNW1PouDiUx 7Y8CJOc+1k9wyBMKl0a/NQBRAqSZBQ== =JvZN -----END PGP SIGNATURE----- Merge tag 'drm-intel-next-2024-04-17-1' of https://anongit.freedesktop.org/git/drm/drm-intel into drm-next Core Changes (DRM): - Fix documentation of DP tunnel functions (Imre) - DP MST read sideband messaging cap (Jani) - Preparation patches for Adaptive Sync SDP Support for DP (Mitul) Driver Changes: i915 core (non-display): - Documentation improvements (Nirmoy) - Add includes for BUG_ON/BUILD_BUG_ON in i915_memcpy.c (Joonas) - Do not print 'pxp init failed with 0' when it succeed (Jose) - Clean-up, including removal of dead code for unsupported platforms (Lucas) - Adding new DG2 PCI ID (Ravi) {i915,xe} display: - Spelling fix (Colin Ian) - Document CDCLK components (Gustavo) - Lunar Lake display enabling, including cdclk and other refactors (Gustavo, Bala) - BIOS/VBT/opregion related refactor (Jani, Ville, RK) - Save a few bytes of memory using {kstrdup,kfree}_const variant (Christophe) - Digital port related refactor/clean-up (Ville) - Fix 2s boot time regression on DP panel replay init (Animesh) - Remove redundant drm_rect_visible() overlay use (Arthur) - DSC HW state readout fixes (Imre) - Remove duplication on audio enable/disable on SDVO and g4x+ DP (Ville) - Disable AuxCCS framebuffers if built for Xe (Juha-Pekka) - Fix DSI init order (Ville) - DRRS related refactor and fixes (Bhanuprakash) - Fix DSB vblank waits with VRR (Ville) - General improvements on register name and use of REG_BIT (Ville) - Some display power well related improvements (Ville) - FBC changes for better w/a handling (Ville) - Make crtc disable more atomic (Ville) - Fix hwmon locking inversion in sysfs getter (Janusz) - Increase DP idle pattern wait timeout to 2ms (Shekhar) - PSR related fixes and improvents (Jouni) - Start using container_of_const() for some extra const safety (Ville) - Use drm_printer more on display code (Ville) - Fix Jasper Lake boot freeze (Jonathon) - Update Pipe src size check in skl_update_scaler (Ankit) - Enable MST mode for 128b/132b single-stream sideband (Jani) - Pass encoder around more for port/phy checks (Jani) - Some initial work to make display code more independent from i915 (Jani) - Pre-populate the cursor physical dma address (Ville) - Do not bump min backlight brightness to max on enable (Gareth) - Fix MTL supported DP rates - removal of UHBR13.5 (Arun) - Fix the computation for compressed_bpp for DISPLAY < 1 (Ankit) - Bigjoiner modeset sequence redesign and MST support (Ville) - Enable Adaptive Sync SDP Support for DP (Mitul) - Implemnt vblank sycnhronized mbus joining changes (Ville, Stanislav) - HDCP related fixes (Suraj) - Fix i915_display_info debugfs when connectors are not active (Ville) - Clean up on Xe compat layer (Jani) - Add jitter WAs for MST/FEC/DSC links (Imre) - DMC wakelock implementation (Luca) Signed-off-by: Dave Airlie <airlied@redhat.com> # -----BEGIN PGP SIGNATURE----- # # iQEzBAABCAAdFiEEbSBwaO7dZQkcLOKj+mJfZA7rE8oFAmYfzQEACgkQ+mJfZA7r # E8qYvAf/T8KrEewHOWz7NOaKcFRCNYaF4QTdVOfgHUYBX5NPDF/xzwFdHCL8QWQu # bwKwE2b94VEyruG3DYwTMd8GNcDxrsOrmU0IZe3PVkm+BvHLTmrOqL6BlCd85zXF # 02IuE+LCaWREmmpLMcsDMxsaaq8yp+cw9/F0jJDrH6LiyfxFriefxyZYpGYjRCuv # 8GP1fHXLFV2yys4rveR/+y9xIhgy82mVcg3/Kfk0+er7gALkY6Vc0N38wedET9MZ # ZPfVidBeaTkIKcCDFKnFzGjG+9rNQ7NFrXyS7Hl97VolGt2l03qGGPNW1PouDiUx # 7Y8CJOc+1k9wyBMKl0a/NQBRAqSZBQ== # =JvZN # -----END PGP SIGNATURE----- # gpg: Signature made Wed 17 Apr 2024 23:22:09 AEST # gpg: using RSA key 6D207068EEDD65091C2CE2A3FA625F640EEB13CA # gpg: Good signature from "Rodrigo Vivi <rodrigo.vivi@intel.com>" [unknown] # gpg: aka "Rodrigo Vivi <rodrigo.vivi@gmail.com>" [unknown] # gpg: WARNING: This key is not certified with a trusted signature! # gpg: There is no indication that the signature belongs to the owner. # Primary key fingerprint: 6D20 7068 EEDD 6509 1C2C E2A3 FA62 5F64 0EEB 13CA From: Rodrigo Vivi <rodrigo.vivi@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/Zh_Q72gYKMMbge9A@intel.com
1951 lines
50 KiB
C
1951 lines
50 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2023 Intel Corporation
|
|
*/
|
|
|
|
#include <linux/ref_tracker.h>
|
|
#include <linux/types.h>
|
|
|
|
#include <drm/drm_atomic_state_helper.h>
|
|
|
|
#include <drm/drm_atomic.h>
|
|
#include <drm/drm_print.h>
|
|
#include <drm/display/drm_dp.h>
|
|
#include <drm/display/drm_dp_helper.h>
|
|
#include <drm/display/drm_dp_tunnel.h>
|
|
|
|
#define to_group(__private_obj) \
|
|
container_of(__private_obj, struct drm_dp_tunnel_group, base)
|
|
|
|
#define to_group_state(__private_state) \
|
|
container_of(__private_state, struct drm_dp_tunnel_group_state, base)
|
|
|
|
#define is_dp_tunnel_private_obj(__obj) \
|
|
((__obj)->funcs == &tunnel_group_funcs)
|
|
|
|
#define for_each_new_group_in_state(__state, __new_group_state, __i) \
|
|
for ((__i) = 0; \
|
|
(__i) < (__state)->num_private_objs; \
|
|
(__i)++) \
|
|
for_each_if ((__state)->private_objs[__i].ptr && \
|
|
is_dp_tunnel_private_obj((__state)->private_objs[__i].ptr) && \
|
|
((__new_group_state) = \
|
|
to_group_state((__state)->private_objs[__i].new_state), 1))
|
|
|
|
#define for_each_old_group_in_state(__state, __old_group_state, __i) \
|
|
for ((__i) = 0; \
|
|
(__i) < (__state)->num_private_objs; \
|
|
(__i)++) \
|
|
for_each_if ((__state)->private_objs[__i].ptr && \
|
|
is_dp_tunnel_private_obj((__state)->private_objs[__i].ptr) && \
|
|
((__old_group_state) = \
|
|
to_group_state((__state)->private_objs[__i].old_state), 1))
|
|
|
|
#define for_each_tunnel_in_group(__group, __tunnel) \
|
|
list_for_each_entry(__tunnel, &(__group)->tunnels, node)
|
|
|
|
#define for_each_tunnel_state(__group_state, __tunnel_state) \
|
|
list_for_each_entry(__tunnel_state, &(__group_state)->tunnel_states, node)
|
|
|
|
#define for_each_tunnel_state_safe(__group_state, __tunnel_state, __tunnel_state_tmp) \
|
|
list_for_each_entry_safe(__tunnel_state, __tunnel_state_tmp, \
|
|
&(__group_state)->tunnel_states, node)
|
|
|
|
#define kbytes_to_mbits(__kbytes) \
|
|
DIV_ROUND_UP((__kbytes) * 8, 1000)
|
|
|
|
#define DPTUN_BW_ARG(__bw) ((__bw) < 0 ? (__bw) : kbytes_to_mbits(__bw))
|
|
|
|
#define __tun_prn(__tunnel, __level, __type, __fmt, ...) \
|
|
drm_##__level##__type((__tunnel)->group->mgr->dev, \
|
|
"[DPTUN %s][%s] " __fmt, \
|
|
drm_dp_tunnel_name(__tunnel), \
|
|
(__tunnel)->aux->name, ## \
|
|
__VA_ARGS__)
|
|
|
|
#define tun_dbg(__tunnel, __fmt, ...) \
|
|
__tun_prn(__tunnel, dbg, _kms, __fmt, ## __VA_ARGS__)
|
|
|
|
#define tun_dbg_stat(__tunnel, __err, __fmt, ...) do { \
|
|
if (__err) \
|
|
__tun_prn(__tunnel, dbg, _kms, __fmt " (Failed, err: %pe)\n", \
|
|
## __VA_ARGS__, ERR_PTR(__err)); \
|
|
else \
|
|
__tun_prn(__tunnel, dbg, _kms, __fmt " (Ok)\n", \
|
|
## __VA_ARGS__); \
|
|
} while (0)
|
|
|
|
#define tun_dbg_atomic(__tunnel, __fmt, ...) \
|
|
__tun_prn(__tunnel, dbg, _atomic, __fmt, ## __VA_ARGS__)
|
|
|
|
#define tun_grp_dbg(__group, __fmt, ...) \
|
|
drm_dbg_kms((__group)->mgr->dev, \
|
|
"[DPTUN %s] " __fmt, \
|
|
drm_dp_tunnel_group_name(__group), ## \
|
|
__VA_ARGS__)
|
|
|
|
#define DP_TUNNELING_BASE DP_TUNNELING_OUI
|
|
|
|
#define __DPTUN_REG_RANGE(__start, __size) \
|
|
GENMASK_ULL((__start) + (__size) - 1, (__start))
|
|
|
|
#define DPTUN_REG_RANGE(__addr, __size) \
|
|
__DPTUN_REG_RANGE((__addr) - DP_TUNNELING_BASE, (__size))
|
|
|
|
#define DPTUN_REG(__addr) DPTUN_REG_RANGE(__addr, 1)
|
|
|
|
#define DPTUN_INFO_REG_MASK ( \
|
|
DPTUN_REG_RANGE(DP_TUNNELING_OUI, DP_TUNNELING_OUI_BYTES) | \
|
|
DPTUN_REG_RANGE(DP_TUNNELING_DEV_ID, DP_TUNNELING_DEV_ID_BYTES) | \
|
|
DPTUN_REG(DP_TUNNELING_HW_REV) | \
|
|
DPTUN_REG(DP_TUNNELING_SW_REV_MAJOR) | \
|
|
DPTUN_REG(DP_TUNNELING_SW_REV_MINOR) | \
|
|
DPTUN_REG(DP_TUNNELING_CAPABILITIES) | \
|
|
DPTUN_REG(DP_IN_ADAPTER_INFO) | \
|
|
DPTUN_REG(DP_USB4_DRIVER_ID) | \
|
|
DPTUN_REG(DP_USB4_DRIVER_BW_CAPABILITY) | \
|
|
DPTUN_REG(DP_IN_ADAPTER_TUNNEL_INFORMATION) | \
|
|
DPTUN_REG(DP_BW_GRANULARITY) | \
|
|
DPTUN_REG(DP_ESTIMATED_BW) | \
|
|
DPTUN_REG(DP_ALLOCATED_BW) | \
|
|
DPTUN_REG(DP_TUNNELING_MAX_LINK_RATE) | \
|
|
DPTUN_REG(DP_TUNNELING_MAX_LANE_COUNT) | \
|
|
DPTUN_REG(DP_DPTX_BW_ALLOCATION_MODE_CONTROL))
|
|
|
|
static const DECLARE_BITMAP(dptun_info_regs, 64) = {
|
|
DPTUN_INFO_REG_MASK & -1UL,
|
|
#if BITS_PER_LONG == 32
|
|
DPTUN_INFO_REG_MASK >> 32,
|
|
#endif
|
|
};
|
|
|
|
struct drm_dp_tunnel_regs {
|
|
u8 buf[HWEIGHT64(DPTUN_INFO_REG_MASK)];
|
|
};
|
|
|
|
struct drm_dp_tunnel_group;
|
|
|
|
struct drm_dp_tunnel {
|
|
struct drm_dp_tunnel_group *group;
|
|
|
|
struct list_head node;
|
|
|
|
struct kref kref;
|
|
struct ref_tracker *tracker;
|
|
struct drm_dp_aux *aux;
|
|
char name[8];
|
|
|
|
int bw_granularity;
|
|
int estimated_bw;
|
|
int allocated_bw;
|
|
|
|
int max_dprx_rate;
|
|
u8 max_dprx_lane_count;
|
|
|
|
u8 adapter_id;
|
|
|
|
bool bw_alloc_supported:1;
|
|
bool bw_alloc_enabled:1;
|
|
bool has_io_error:1;
|
|
bool destroyed:1;
|
|
};
|
|
|
|
struct drm_dp_tunnel_group_state;
|
|
|
|
struct drm_dp_tunnel_state {
|
|
struct drm_dp_tunnel_group_state *group_state;
|
|
|
|
struct drm_dp_tunnel_ref tunnel_ref;
|
|
|
|
struct list_head node;
|
|
|
|
u32 stream_mask;
|
|
int *stream_bw;
|
|
};
|
|
|
|
struct drm_dp_tunnel_group_state {
|
|
struct drm_private_state base;
|
|
|
|
struct list_head tunnel_states;
|
|
};
|
|
|
|
struct drm_dp_tunnel_group {
|
|
struct drm_private_obj base;
|
|
struct drm_dp_tunnel_mgr *mgr;
|
|
|
|
struct list_head tunnels;
|
|
|
|
/* available BW including the allocated_bw of all tunnels in the group */
|
|
int available_bw;
|
|
|
|
u8 drv_group_id;
|
|
char name[8];
|
|
|
|
bool active:1;
|
|
};
|
|
|
|
struct drm_dp_tunnel_mgr {
|
|
struct drm_device *dev;
|
|
|
|
int group_count;
|
|
struct drm_dp_tunnel_group *groups;
|
|
wait_queue_head_t bw_req_queue;
|
|
|
|
#ifdef CONFIG_DRM_DISPLAY_DP_TUNNEL_STATE_DEBUG
|
|
struct ref_tracker_dir ref_tracker;
|
|
#endif
|
|
};
|
|
|
|
/*
|
|
* The following helpers provide a way to read out the tunneling DPCD
|
|
* registers with a minimal amount of AUX transfers (1 transfer per contiguous
|
|
* range, as permitted by the 16 byte per transfer AUX limit), not accessing
|
|
* other registers to avoid any read side-effects.
|
|
*/
|
|
static int next_reg_area(int *offset)
|
|
{
|
|
*offset = find_next_bit(dptun_info_regs, 64, *offset);
|
|
|
|
return find_next_zero_bit(dptun_info_regs, 64, *offset + 1) - *offset;
|
|
}
|
|
|
|
#define tunnel_reg_ptr(__regs, __address) ({ \
|
|
WARN_ON(!test_bit((__address) - DP_TUNNELING_BASE, dptun_info_regs)); \
|
|
&(__regs)->buf[bitmap_weight(dptun_info_regs, (__address) - DP_TUNNELING_BASE)]; \
|
|
})
|
|
|
|
static int read_tunnel_regs(struct drm_dp_aux *aux, struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
int offset = 0;
|
|
int len;
|
|
|
|
while ((len = next_reg_area(&offset))) {
|
|
int address = DP_TUNNELING_BASE + offset;
|
|
|
|
if (drm_dp_dpcd_read(aux, address, tunnel_reg_ptr(regs, address), len) < 0)
|
|
return -EIO;
|
|
|
|
offset += len;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static u8 tunnel_reg(const struct drm_dp_tunnel_regs *regs, int address)
|
|
{
|
|
return *tunnel_reg_ptr(regs, address);
|
|
}
|
|
|
|
static u8 tunnel_reg_drv_group_id(const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
u8 drv_id = tunnel_reg(regs, DP_USB4_DRIVER_ID) & DP_USB4_DRIVER_ID_MASK;
|
|
u8 group_id = tunnel_reg(regs, DP_IN_ADAPTER_TUNNEL_INFORMATION) & DP_GROUP_ID_MASK;
|
|
|
|
if (!group_id)
|
|
return 0;
|
|
|
|
return (drv_id << DP_GROUP_ID_BITS) | group_id;
|
|
}
|
|
|
|
/* Return granularity in kB/s units */
|
|
static int tunnel_reg_bw_granularity(const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
int gr = tunnel_reg(regs, DP_BW_GRANULARITY) & DP_BW_GRANULARITY_MASK;
|
|
|
|
if (gr > 2)
|
|
return -1;
|
|
|
|
return (250000 << gr) / 8;
|
|
}
|
|
|
|
static int tunnel_reg_max_dprx_rate(const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
u8 bw_code = tunnel_reg(regs, DP_TUNNELING_MAX_LINK_RATE);
|
|
|
|
return drm_dp_bw_code_to_link_rate(bw_code);
|
|
}
|
|
|
|
static int tunnel_reg_max_dprx_lane_count(const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
return tunnel_reg(regs, DP_TUNNELING_MAX_LANE_COUNT) &
|
|
DP_TUNNELING_MAX_LANE_COUNT_MASK;
|
|
}
|
|
|
|
static bool tunnel_reg_bw_alloc_supported(const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
u8 cap_mask = DP_TUNNELING_SUPPORT | DP_IN_BW_ALLOCATION_MODE_SUPPORT;
|
|
|
|
if ((tunnel_reg(regs, DP_TUNNELING_CAPABILITIES) & cap_mask) != cap_mask)
|
|
return false;
|
|
|
|
return tunnel_reg(regs, DP_USB4_DRIVER_BW_CAPABILITY) &
|
|
DP_USB4_DRIVER_BW_ALLOCATION_MODE_SUPPORT;
|
|
}
|
|
|
|
static bool tunnel_reg_bw_alloc_enabled(const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
return tunnel_reg(regs, DP_DPTX_BW_ALLOCATION_MODE_CONTROL) &
|
|
DP_DISPLAY_DRIVER_BW_ALLOCATION_MODE_ENABLE;
|
|
}
|
|
|
|
static u8 tunnel_group_drv_id(u8 drv_group_id)
|
|
{
|
|
return drv_group_id >> DP_GROUP_ID_BITS;
|
|
}
|
|
|
|
static u8 tunnel_group_id(u8 drv_group_id)
|
|
{
|
|
return drv_group_id & DP_GROUP_ID_MASK;
|
|
}
|
|
|
|
const char *drm_dp_tunnel_name(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel->name;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_name);
|
|
|
|
static const char *drm_dp_tunnel_group_name(const struct drm_dp_tunnel_group *group)
|
|
{
|
|
return group->name;
|
|
}
|
|
|
|
static struct drm_dp_tunnel_group *
|
|
lookup_or_alloc_group(struct drm_dp_tunnel_mgr *mgr, u8 drv_group_id)
|
|
{
|
|
struct drm_dp_tunnel_group *group = NULL;
|
|
int i;
|
|
|
|
for (i = 0; i < mgr->group_count; i++) {
|
|
/*
|
|
* A tunnel group with 0 group ID shouldn't have more than one
|
|
* tunnels.
|
|
*/
|
|
if (tunnel_group_id(drv_group_id) &&
|
|
mgr->groups[i].drv_group_id == drv_group_id)
|
|
return &mgr->groups[i];
|
|
|
|
if (!group && !mgr->groups[i].active)
|
|
group = &mgr->groups[i];
|
|
}
|
|
|
|
if (!group) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: Can't allocate more tunnel groups\n");
|
|
return NULL;
|
|
}
|
|
|
|
group->drv_group_id = drv_group_id;
|
|
group->active = true;
|
|
|
|
/*
|
|
* The group name format here and elsewhere: Driver-ID:Group-ID:*
|
|
* (* standing for all DP-Adapters/tunnels in the group).
|
|
*/
|
|
snprintf(group->name, sizeof(group->name), "%d:%d:*",
|
|
tunnel_group_drv_id(drv_group_id) & ((1 << DP_GROUP_ID_BITS) - 1),
|
|
tunnel_group_id(drv_group_id) & ((1 << DP_USB4_DRIVER_ID_BITS) - 1));
|
|
|
|
return group;
|
|
}
|
|
|
|
static void free_group(struct drm_dp_tunnel_group *group)
|
|
{
|
|
struct drm_dp_tunnel_mgr *mgr = group->mgr;
|
|
|
|
if (drm_WARN_ON(mgr->dev, !list_empty(&group->tunnels)))
|
|
return;
|
|
|
|
group->drv_group_id = 0;
|
|
group->available_bw = -1;
|
|
group->active = false;
|
|
}
|
|
|
|
static struct drm_dp_tunnel *
|
|
tunnel_get(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
kref_get(&tunnel->kref);
|
|
|
|
return tunnel;
|
|
}
|
|
|
|
static void free_tunnel(struct kref *kref)
|
|
{
|
|
struct drm_dp_tunnel *tunnel = container_of(kref, typeof(*tunnel), kref);
|
|
struct drm_dp_tunnel_group *group = tunnel->group;
|
|
|
|
list_del(&tunnel->node);
|
|
if (list_empty(&group->tunnels))
|
|
free_group(group);
|
|
|
|
kfree(tunnel);
|
|
}
|
|
|
|
static void tunnel_put(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
kref_put(&tunnel->kref, free_tunnel);
|
|
}
|
|
|
|
#ifdef CONFIG_DRM_DISPLAY_DP_TUNNEL_STATE_DEBUG
|
|
static void track_tunnel_ref(struct drm_dp_tunnel *tunnel,
|
|
struct ref_tracker **tracker)
|
|
{
|
|
ref_tracker_alloc(&tunnel->group->mgr->ref_tracker,
|
|
tracker, GFP_KERNEL);
|
|
}
|
|
|
|
static void untrack_tunnel_ref(struct drm_dp_tunnel *tunnel,
|
|
struct ref_tracker **tracker)
|
|
{
|
|
ref_tracker_free(&tunnel->group->mgr->ref_tracker,
|
|
tracker);
|
|
}
|
|
#else
|
|
static void track_tunnel_ref(struct drm_dp_tunnel *tunnel,
|
|
struct ref_tracker **tracker)
|
|
{
|
|
}
|
|
|
|
static void untrack_tunnel_ref(struct drm_dp_tunnel *tunnel,
|
|
struct ref_tracker **tracker)
|
|
{
|
|
}
|
|
#endif
|
|
|
|
/**
|
|
* drm_dp_tunnel_get - Get a reference for a DP tunnel
|
|
* @tunnel: Tunnel object
|
|
* @tracker: Debug tracker for the reference
|
|
*
|
|
* Get a reference for @tunnel, along with a debug tracker to help locating
|
|
* the source of a reference leak/double reference put etc. issue.
|
|
*
|
|
* The reference must be dropped after use calling drm_dp_tunnel_put()
|
|
* passing @tunnel and *@tracker returned from here.
|
|
*
|
|
* Returns @tunnel - as a convenience - along with *@tracker.
|
|
*/
|
|
struct drm_dp_tunnel *
|
|
drm_dp_tunnel_get(struct drm_dp_tunnel *tunnel,
|
|
struct ref_tracker **tracker)
|
|
{
|
|
track_tunnel_ref(tunnel, tracker);
|
|
|
|
return tunnel_get(tunnel);
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_get);
|
|
|
|
/**
|
|
* drm_dp_tunnel_put - Put a reference for a DP tunnel
|
|
* @tunnel: Tunnel object
|
|
* @tracker: Debug tracker for the reference
|
|
*
|
|
* Put a reference for @tunnel along with its debug *@tracker, which
|
|
* was obtained with drm_dp_tunnel_get().
|
|
*/
|
|
void drm_dp_tunnel_put(struct drm_dp_tunnel *tunnel,
|
|
struct ref_tracker **tracker)
|
|
{
|
|
untrack_tunnel_ref(tunnel, tracker);
|
|
|
|
tunnel_put(tunnel);
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_put);
|
|
|
|
static bool add_tunnel_to_group(struct drm_dp_tunnel_mgr *mgr,
|
|
u8 drv_group_id,
|
|
struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_group *group;
|
|
|
|
group = lookup_or_alloc_group(mgr, drv_group_id);
|
|
if (!group)
|
|
return false;
|
|
|
|
tunnel->group = group;
|
|
list_add(&tunnel->node, &group->tunnels);
|
|
|
|
return true;
|
|
}
|
|
|
|
static struct drm_dp_tunnel *
|
|
create_tunnel(struct drm_dp_tunnel_mgr *mgr,
|
|
struct drm_dp_aux *aux,
|
|
const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
u8 drv_group_id = tunnel_reg_drv_group_id(regs);
|
|
struct drm_dp_tunnel *tunnel;
|
|
|
|
tunnel = kzalloc(sizeof(*tunnel), GFP_KERNEL);
|
|
if (!tunnel)
|
|
return NULL;
|
|
|
|
INIT_LIST_HEAD(&tunnel->node);
|
|
|
|
kref_init(&tunnel->kref);
|
|
|
|
tunnel->aux = aux;
|
|
|
|
tunnel->adapter_id = tunnel_reg(regs, DP_IN_ADAPTER_INFO) & DP_IN_ADAPTER_NUMBER_MASK;
|
|
|
|
snprintf(tunnel->name, sizeof(tunnel->name), "%d:%d:%d",
|
|
tunnel_group_drv_id(drv_group_id) & ((1 << DP_GROUP_ID_BITS) - 1),
|
|
tunnel_group_id(drv_group_id) & ((1 << DP_USB4_DRIVER_ID_BITS) - 1),
|
|
tunnel->adapter_id & ((1 << DP_IN_ADAPTER_NUMBER_BITS) - 1));
|
|
|
|
tunnel->bw_granularity = tunnel_reg_bw_granularity(regs);
|
|
tunnel->allocated_bw = tunnel_reg(regs, DP_ALLOCATED_BW) *
|
|
tunnel->bw_granularity;
|
|
/*
|
|
* An initial allocated BW of 0 indicates an undefined state: the
|
|
* actual allocation is determined by the TBT CM, usually following a
|
|
* legacy allocation policy (based on the max DPRX caps). From the
|
|
* driver's POV the state becomes defined only after the first
|
|
* allocation request.
|
|
*/
|
|
if (!tunnel->allocated_bw)
|
|
tunnel->allocated_bw = -1;
|
|
|
|
tunnel->bw_alloc_supported = tunnel_reg_bw_alloc_supported(regs);
|
|
tunnel->bw_alloc_enabled = tunnel_reg_bw_alloc_enabled(regs);
|
|
|
|
if (!add_tunnel_to_group(mgr, drv_group_id, tunnel)) {
|
|
kfree(tunnel);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
track_tunnel_ref(tunnel, &tunnel->tracker);
|
|
|
|
return tunnel;
|
|
}
|
|
|
|
static void destroy_tunnel(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
untrack_tunnel_ref(tunnel, &tunnel->tracker);
|
|
tunnel_put(tunnel);
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_set_io_error - Set the IO error flag for a DP tunnel
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Set the IO error flag for @tunnel. Drivers can call this function upon
|
|
* detecting a failure that affects the tunnel functionality, for instance
|
|
* after a DP AUX transfer failure on the port @tunnel is connected to.
|
|
*
|
|
* This disables further management of @tunnel, including any related
|
|
* AUX accesses for tunneling DPCD registers, returning error to the
|
|
* initiators of these. The driver is supposed to drop this tunnel and -
|
|
* optionally - recreate it.
|
|
*/
|
|
void drm_dp_tunnel_set_io_error(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
tunnel->has_io_error = true;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_set_io_error);
|
|
|
|
#define SKIP_DPRX_CAPS_CHECK BIT(0)
|
|
#define ALLOW_ALLOCATED_BW_CHANGE BIT(1)
|
|
static bool tunnel_regs_are_valid(struct drm_dp_tunnel_mgr *mgr,
|
|
const struct drm_dp_tunnel_regs *regs,
|
|
unsigned int flags)
|
|
{
|
|
u8 drv_group_id = tunnel_reg_drv_group_id(regs);
|
|
bool check_dprx = !(flags & SKIP_DPRX_CAPS_CHECK);
|
|
bool ret = true;
|
|
|
|
if (!tunnel_reg_bw_alloc_supported(regs)) {
|
|
if (tunnel_group_id(drv_group_id)) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: A non-zero group ID is only allowed with BWA support\n");
|
|
ret = false;
|
|
}
|
|
|
|
if (tunnel_reg(regs, DP_ALLOCATED_BW)) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: BW is allocated without BWA support\n");
|
|
ret = false;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
if (!tunnel_group_id(drv_group_id)) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: BWA support requires a non-zero group ID\n");
|
|
ret = false;
|
|
}
|
|
|
|
if (check_dprx && hweight8(tunnel_reg_max_dprx_lane_count(regs)) != 1) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: Invalid DPRX lane count: %d\n",
|
|
tunnel_reg_max_dprx_lane_count(regs));
|
|
|
|
ret = false;
|
|
}
|
|
|
|
if (check_dprx && !tunnel_reg_max_dprx_rate(regs)) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: DPRX rate is 0\n");
|
|
|
|
ret = false;
|
|
}
|
|
|
|
if (tunnel_reg_bw_granularity(regs) < 0) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: Invalid BW granularity\n");
|
|
|
|
ret = false;
|
|
}
|
|
|
|
if (tunnel_reg(regs, DP_ALLOCATED_BW) > tunnel_reg(regs, DP_ESTIMATED_BW)) {
|
|
drm_dbg_kms(mgr->dev,
|
|
"DPTUN: Allocated BW %d > estimated BW %d Mb/s\n",
|
|
DPTUN_BW_ARG(tunnel_reg(regs, DP_ALLOCATED_BW) *
|
|
tunnel_reg_bw_granularity(regs)),
|
|
DPTUN_BW_ARG(tunnel_reg(regs, DP_ESTIMATED_BW) *
|
|
tunnel_reg_bw_granularity(regs)));
|
|
|
|
ret = false;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int tunnel_allocated_bw(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return max(tunnel->allocated_bw, 0);
|
|
}
|
|
|
|
static bool tunnel_info_changes_are_valid(struct drm_dp_tunnel *tunnel,
|
|
const struct drm_dp_tunnel_regs *regs,
|
|
unsigned int flags)
|
|
{
|
|
u8 new_drv_group_id = tunnel_reg_drv_group_id(regs);
|
|
bool ret = true;
|
|
|
|
if (tunnel->bw_alloc_supported != tunnel_reg_bw_alloc_supported(regs)) {
|
|
tun_dbg(tunnel,
|
|
"BW alloc support has changed %s -> %s\n",
|
|
str_yes_no(tunnel->bw_alloc_supported),
|
|
str_yes_no(tunnel_reg_bw_alloc_supported(regs)));
|
|
|
|
ret = false;
|
|
}
|
|
|
|
if (tunnel->group->drv_group_id != new_drv_group_id) {
|
|
tun_dbg(tunnel,
|
|
"Driver/group ID has changed %d:%d:* -> %d:%d:*\n",
|
|
tunnel_group_drv_id(tunnel->group->drv_group_id),
|
|
tunnel_group_id(tunnel->group->drv_group_id),
|
|
tunnel_group_drv_id(new_drv_group_id),
|
|
tunnel_group_id(new_drv_group_id));
|
|
|
|
ret = false;
|
|
}
|
|
|
|
if (!tunnel->bw_alloc_supported)
|
|
return ret;
|
|
|
|
if (tunnel->bw_granularity != tunnel_reg_bw_granularity(regs)) {
|
|
tun_dbg(tunnel,
|
|
"BW granularity has changed: %d -> %d Mb/s\n",
|
|
DPTUN_BW_ARG(tunnel->bw_granularity),
|
|
DPTUN_BW_ARG(tunnel_reg_bw_granularity(regs)));
|
|
|
|
ret = false;
|
|
}
|
|
|
|
/*
|
|
* On some devices at least the BW alloc mode enabled status is always
|
|
* reported as 0, so skip checking that here.
|
|
*/
|
|
|
|
if (!(flags & ALLOW_ALLOCATED_BW_CHANGE) &&
|
|
tunnel_allocated_bw(tunnel) !=
|
|
tunnel_reg(regs, DP_ALLOCATED_BW) * tunnel->bw_granularity) {
|
|
tun_dbg(tunnel,
|
|
"Allocated BW has changed: %d -> %d Mb/s\n",
|
|
DPTUN_BW_ARG(tunnel->allocated_bw),
|
|
DPTUN_BW_ARG(tunnel_reg(regs, DP_ALLOCATED_BW) * tunnel->bw_granularity));
|
|
|
|
ret = false;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
read_and_verify_tunnel_regs(struct drm_dp_tunnel *tunnel,
|
|
struct drm_dp_tunnel_regs *regs,
|
|
unsigned int flags)
|
|
{
|
|
int err;
|
|
|
|
err = read_tunnel_regs(tunnel->aux, regs);
|
|
if (err < 0) {
|
|
drm_dp_tunnel_set_io_error(tunnel);
|
|
|
|
return err;
|
|
}
|
|
|
|
if (!tunnel_regs_are_valid(tunnel->group->mgr, regs, flags))
|
|
return -EINVAL;
|
|
|
|
if (!tunnel_info_changes_are_valid(tunnel, regs, flags))
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool update_dprx_caps(struct drm_dp_tunnel *tunnel, const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
bool changed = false;
|
|
|
|
if (tunnel_reg_max_dprx_rate(regs) != tunnel->max_dprx_rate) {
|
|
tunnel->max_dprx_rate = tunnel_reg_max_dprx_rate(regs);
|
|
changed = true;
|
|
}
|
|
|
|
if (tunnel_reg_max_dprx_lane_count(regs) != tunnel->max_dprx_lane_count) {
|
|
tunnel->max_dprx_lane_count = tunnel_reg_max_dprx_lane_count(regs);
|
|
changed = true;
|
|
}
|
|
|
|
return changed;
|
|
}
|
|
|
|
static int dev_id_len(const u8 *dev_id, int max_len)
|
|
{
|
|
while (max_len && dev_id[max_len - 1] == '\0')
|
|
max_len--;
|
|
|
|
return max_len;
|
|
}
|
|
|
|
static int get_max_dprx_bw(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
int max_dprx_bw = drm_dp_max_dprx_data_rate(tunnel->max_dprx_rate,
|
|
tunnel->max_dprx_lane_count);
|
|
|
|
/*
|
|
* A BW request of roundup(max_dprx_bw, tunnel->bw_granularity) results in
|
|
* an allocation of max_dprx_bw. A BW request above this rounded-up
|
|
* value will fail.
|
|
*/
|
|
return min(roundup(max_dprx_bw, tunnel->bw_granularity),
|
|
MAX_DP_REQUEST_BW * tunnel->bw_granularity);
|
|
}
|
|
|
|
static int get_max_tunnel_bw(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return min(get_max_dprx_bw(tunnel), tunnel->group->available_bw);
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_detect - Detect DP tunnel on the link
|
|
* @mgr: Tunnel manager
|
|
* @aux: DP AUX on which the tunnel will be detected
|
|
*
|
|
* Detect if there is any DP tunnel on the link and add it to the tunnel
|
|
* group's tunnel list.
|
|
*
|
|
* Returns a pointer to a tunnel on success, or an ERR_PTR() error on
|
|
* failure.
|
|
*/
|
|
struct drm_dp_tunnel *
|
|
drm_dp_tunnel_detect(struct drm_dp_tunnel_mgr *mgr,
|
|
struct drm_dp_aux *aux)
|
|
{
|
|
struct drm_dp_tunnel_regs regs;
|
|
struct drm_dp_tunnel *tunnel;
|
|
int err;
|
|
|
|
err = read_tunnel_regs(aux, ®s);
|
|
if (err)
|
|
return ERR_PTR(err);
|
|
|
|
if (!(tunnel_reg(®s, DP_TUNNELING_CAPABILITIES) &
|
|
DP_TUNNELING_SUPPORT))
|
|
return ERR_PTR(-ENODEV);
|
|
|
|
/* The DPRX caps are valid only after enabling BW alloc mode. */
|
|
if (!tunnel_regs_are_valid(mgr, ®s, SKIP_DPRX_CAPS_CHECK))
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
tunnel = create_tunnel(mgr, aux, ®s);
|
|
if (!tunnel)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
tun_dbg(tunnel,
|
|
"OUI:%*phD DevID:%*pE Rev-HW:%d.%d SW:%d.%d PR-Sup:%s BWA-Sup:%s BWA-En:%s\n",
|
|
DP_TUNNELING_OUI_BYTES,
|
|
tunnel_reg_ptr(®s, DP_TUNNELING_OUI),
|
|
dev_id_len(tunnel_reg_ptr(®s, DP_TUNNELING_DEV_ID), DP_TUNNELING_DEV_ID_BYTES),
|
|
tunnel_reg_ptr(®s, DP_TUNNELING_DEV_ID),
|
|
(tunnel_reg(®s, DP_TUNNELING_HW_REV) & DP_TUNNELING_HW_REV_MAJOR_MASK) >>
|
|
DP_TUNNELING_HW_REV_MAJOR_SHIFT,
|
|
(tunnel_reg(®s, DP_TUNNELING_HW_REV) & DP_TUNNELING_HW_REV_MINOR_MASK) >>
|
|
DP_TUNNELING_HW_REV_MINOR_SHIFT,
|
|
tunnel_reg(®s, DP_TUNNELING_SW_REV_MAJOR),
|
|
tunnel_reg(®s, DP_TUNNELING_SW_REV_MINOR),
|
|
str_yes_no(tunnel_reg(®s, DP_TUNNELING_CAPABILITIES) &
|
|
DP_PANEL_REPLAY_OPTIMIZATION_SUPPORT),
|
|
str_yes_no(tunnel->bw_alloc_supported),
|
|
str_yes_no(tunnel->bw_alloc_enabled));
|
|
|
|
return tunnel;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_detect);
|
|
|
|
/**
|
|
* drm_dp_tunnel_destroy - Destroy tunnel object
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Remove the tunnel from the tunnel topology and destroy it.
|
|
*
|
|
* Returns 0 on success, -ENODEV if the tunnel has been destroyed already.
|
|
*/
|
|
int drm_dp_tunnel_destroy(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
if (!tunnel)
|
|
return 0;
|
|
|
|
if (drm_WARN_ON(tunnel->group->mgr->dev, tunnel->destroyed))
|
|
return -ENODEV;
|
|
|
|
tun_dbg(tunnel, "destroying\n");
|
|
|
|
tunnel->destroyed = true;
|
|
destroy_tunnel(tunnel);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_destroy);
|
|
|
|
static int check_tunnel(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
if (tunnel->destroyed)
|
|
return -ENODEV;
|
|
|
|
if (tunnel->has_io_error)
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int group_allocated_bw(struct drm_dp_tunnel_group *group)
|
|
{
|
|
struct drm_dp_tunnel *tunnel;
|
|
int group_allocated_bw = 0;
|
|
|
|
for_each_tunnel_in_group(group, tunnel) {
|
|
if (check_tunnel(tunnel) == 0 &&
|
|
tunnel->bw_alloc_enabled)
|
|
group_allocated_bw += tunnel_allocated_bw(tunnel);
|
|
}
|
|
|
|
return group_allocated_bw;
|
|
}
|
|
|
|
/*
|
|
* The estimated BW reported by the TBT Connection Manager for each tunnel in
|
|
* a group includes the BW already allocated for the given tunnel and the
|
|
* unallocated BW which is free to be used by any tunnel in the group.
|
|
*/
|
|
static int group_free_bw(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel->estimated_bw - tunnel_allocated_bw(tunnel);
|
|
}
|
|
|
|
static int calc_group_available_bw(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return group_allocated_bw(tunnel->group) +
|
|
group_free_bw(tunnel);
|
|
}
|
|
|
|
static int update_group_available_bw(struct drm_dp_tunnel *tunnel,
|
|
const struct drm_dp_tunnel_regs *regs)
|
|
{
|
|
struct drm_dp_tunnel *tunnel_iter;
|
|
int group_available_bw;
|
|
bool changed;
|
|
|
|
tunnel->estimated_bw = tunnel_reg(regs, DP_ESTIMATED_BW) * tunnel->bw_granularity;
|
|
|
|
if (calc_group_available_bw(tunnel) == tunnel->group->available_bw)
|
|
return 0;
|
|
|
|
for_each_tunnel_in_group(tunnel->group, tunnel_iter) {
|
|
int err;
|
|
|
|
if (tunnel_iter == tunnel)
|
|
continue;
|
|
|
|
if (check_tunnel(tunnel_iter) != 0 ||
|
|
!tunnel_iter->bw_alloc_enabled)
|
|
continue;
|
|
|
|
err = drm_dp_dpcd_probe(tunnel_iter->aux, DP_DPCD_REV);
|
|
if (err) {
|
|
tun_dbg(tunnel_iter,
|
|
"Probe failed, assume disconnected (err %pe)\n",
|
|
ERR_PTR(err));
|
|
drm_dp_tunnel_set_io_error(tunnel_iter);
|
|
}
|
|
}
|
|
|
|
group_available_bw = calc_group_available_bw(tunnel);
|
|
|
|
tun_dbg(tunnel, "Updated group available BW: %d->%d\n",
|
|
DPTUN_BW_ARG(tunnel->group->available_bw),
|
|
DPTUN_BW_ARG(group_available_bw));
|
|
|
|
changed = tunnel->group->available_bw != group_available_bw;
|
|
|
|
tunnel->group->available_bw = group_available_bw;
|
|
|
|
return changed ? 1 : 0;
|
|
}
|
|
|
|
static int set_bw_alloc_mode(struct drm_dp_tunnel *tunnel, bool enable)
|
|
{
|
|
u8 mask = DP_DISPLAY_DRIVER_BW_ALLOCATION_MODE_ENABLE | DP_UNMASK_BW_ALLOCATION_IRQ;
|
|
u8 val;
|
|
|
|
if (drm_dp_dpcd_readb(tunnel->aux, DP_DPTX_BW_ALLOCATION_MODE_CONTROL, &val) < 0)
|
|
goto out_err;
|
|
|
|
if (enable)
|
|
val |= mask;
|
|
else
|
|
val &= ~mask;
|
|
|
|
if (drm_dp_dpcd_writeb(tunnel->aux, DP_DPTX_BW_ALLOCATION_MODE_CONTROL, val) < 0)
|
|
goto out_err;
|
|
|
|
tunnel->bw_alloc_enabled = enable;
|
|
|
|
return 0;
|
|
|
|
out_err:
|
|
drm_dp_tunnel_set_io_error(tunnel);
|
|
|
|
return -EIO;
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_enable_bw_alloc - Enable DP tunnel BW allocation mode
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Enable the DP tunnel BW allocation mode on @tunnel if it supports it.
|
|
*
|
|
* Returns 0 in case of success, negative error code otherwise.
|
|
*/
|
|
int drm_dp_tunnel_enable_bw_alloc(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_regs regs;
|
|
int err;
|
|
|
|
err = check_tunnel(tunnel);
|
|
if (err)
|
|
return err;
|
|
|
|
if (!tunnel->bw_alloc_supported)
|
|
return -EOPNOTSUPP;
|
|
|
|
if (!tunnel_group_id(tunnel->group->drv_group_id))
|
|
return -EINVAL;
|
|
|
|
err = set_bw_alloc_mode(tunnel, true);
|
|
if (err)
|
|
goto out;
|
|
|
|
/*
|
|
* After a BWA disable/re-enable sequence the allocated BW can either
|
|
* stay at its last requested value or, for instance after system
|
|
* suspend/resume, TBT CM can reset back the allocation to the amount
|
|
* allocated in the legacy/non-BWA mode. Accordingly allow for the
|
|
* allocation to change wrt. the last SW state.
|
|
*/
|
|
err = read_and_verify_tunnel_regs(tunnel, ®s,
|
|
ALLOW_ALLOCATED_BW_CHANGE);
|
|
if (err) {
|
|
set_bw_alloc_mode(tunnel, false);
|
|
|
|
goto out;
|
|
}
|
|
|
|
if (!tunnel->max_dprx_rate)
|
|
update_dprx_caps(tunnel, ®s);
|
|
|
|
if (tunnel->group->available_bw == -1) {
|
|
err = update_group_available_bw(tunnel, ®s);
|
|
if (err > 0)
|
|
err = 0;
|
|
}
|
|
out:
|
|
tun_dbg_stat(tunnel, err,
|
|
"Enabling BW alloc mode: DPRX:%dx%d Group alloc:%d/%d Mb/s",
|
|
tunnel->max_dprx_rate / 100, tunnel->max_dprx_lane_count,
|
|
DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
|
|
DPTUN_BW_ARG(tunnel->group->available_bw));
|
|
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_enable_bw_alloc);
|
|
|
|
/**
|
|
* drm_dp_tunnel_disable_bw_alloc - Disable DP tunnel BW allocation mode
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Disable the DP tunnel BW allocation mode on @tunnel.
|
|
*
|
|
* Returns 0 in case of success, negative error code otherwise.
|
|
*/
|
|
int drm_dp_tunnel_disable_bw_alloc(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
int err;
|
|
|
|
err = check_tunnel(tunnel);
|
|
if (err)
|
|
return err;
|
|
|
|
tunnel->allocated_bw = -1;
|
|
|
|
err = set_bw_alloc_mode(tunnel, false);
|
|
|
|
tun_dbg_stat(tunnel, err, "Disabling BW alloc mode");
|
|
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_disable_bw_alloc);
|
|
|
|
/**
|
|
* drm_dp_tunnel_bw_alloc_is_enabled - Query the BW allocation mode enabled state
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Query if the BW allocation mode is enabled for @tunnel.
|
|
*
|
|
* Returns %true if the BW allocation mode is enabled for @tunnel.
|
|
*/
|
|
bool drm_dp_tunnel_bw_alloc_is_enabled(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel && tunnel->bw_alloc_enabled;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_bw_alloc_is_enabled);
|
|
|
|
static int clear_bw_req_state(struct drm_dp_aux *aux)
|
|
{
|
|
u8 bw_req_mask = DP_BW_REQUEST_SUCCEEDED | DP_BW_REQUEST_FAILED;
|
|
|
|
if (drm_dp_dpcd_writeb(aux, DP_TUNNELING_STATUS, bw_req_mask) < 0)
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int bw_req_complete(struct drm_dp_aux *aux, bool *status_changed)
|
|
{
|
|
u8 bw_req_mask = DP_BW_REQUEST_SUCCEEDED | DP_BW_REQUEST_FAILED;
|
|
u8 status_change_mask = DP_BW_ALLOCATION_CAPABILITY_CHANGED | DP_ESTIMATED_BW_CHANGED;
|
|
u8 val;
|
|
int err;
|
|
|
|
if (drm_dp_dpcd_readb(aux, DP_TUNNELING_STATUS, &val) < 0)
|
|
return -EIO;
|
|
|
|
*status_changed = val & status_change_mask;
|
|
|
|
val &= bw_req_mask;
|
|
|
|
if (!val)
|
|
return -EAGAIN;
|
|
|
|
err = clear_bw_req_state(aux);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
return val == DP_BW_REQUEST_SUCCEEDED ? 0 : -ENOSPC;
|
|
}
|
|
|
|
static int allocate_tunnel_bw(struct drm_dp_tunnel *tunnel, int bw)
|
|
{
|
|
struct drm_dp_tunnel_mgr *mgr = tunnel->group->mgr;
|
|
int request_bw = DIV_ROUND_UP(bw, tunnel->bw_granularity);
|
|
DEFINE_WAIT_FUNC(wait, woken_wake_function);
|
|
long timeout;
|
|
int err;
|
|
|
|
if (bw < 0) {
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
if (request_bw * tunnel->bw_granularity == tunnel->allocated_bw)
|
|
return 0;
|
|
|
|
/* Atomic check should prevent the following. */
|
|
if (drm_WARN_ON(mgr->dev, request_bw > MAX_DP_REQUEST_BW)) {
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
err = clear_bw_req_state(tunnel->aux);
|
|
if (err)
|
|
goto out;
|
|
|
|
if (drm_dp_dpcd_writeb(tunnel->aux, DP_REQUEST_BW, request_bw) < 0) {
|
|
err = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
timeout = msecs_to_jiffies(3000);
|
|
add_wait_queue(&mgr->bw_req_queue, &wait);
|
|
|
|
for (;;) {
|
|
bool status_changed;
|
|
|
|
err = bw_req_complete(tunnel->aux, &status_changed);
|
|
if (err != -EAGAIN)
|
|
break;
|
|
|
|
if (status_changed) {
|
|
struct drm_dp_tunnel_regs regs;
|
|
|
|
err = read_and_verify_tunnel_regs(tunnel, ®s,
|
|
ALLOW_ALLOCATED_BW_CHANGE);
|
|
if (err)
|
|
break;
|
|
}
|
|
|
|
if (!timeout) {
|
|
err = -ETIMEDOUT;
|
|
break;
|
|
}
|
|
|
|
timeout = wait_woken(&wait, TASK_UNINTERRUPTIBLE, timeout);
|
|
};
|
|
|
|
remove_wait_queue(&mgr->bw_req_queue, &wait);
|
|
|
|
if (err)
|
|
goto out;
|
|
|
|
tunnel->allocated_bw = request_bw * tunnel->bw_granularity;
|
|
|
|
out:
|
|
tun_dbg_stat(tunnel, err, "Allocating %d/%d Mb/s for tunnel: Group alloc:%d/%d Mb/s",
|
|
DPTUN_BW_ARG(request_bw * tunnel->bw_granularity),
|
|
DPTUN_BW_ARG(get_max_tunnel_bw(tunnel)),
|
|
DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
|
|
DPTUN_BW_ARG(tunnel->group->available_bw));
|
|
|
|
if (err == -EIO)
|
|
drm_dp_tunnel_set_io_error(tunnel);
|
|
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_alloc_bw - Allocate BW for a DP tunnel
|
|
* @tunnel: Tunnel object
|
|
* @bw: BW in kB/s units
|
|
*
|
|
* Allocate @bw kB/s for @tunnel. The allocated BW must be freed after use by
|
|
* calling this function for the same tunnel setting @bw to 0.
|
|
*
|
|
* Returns 0 in case of success, a negative error code otherwise.
|
|
*/
|
|
int drm_dp_tunnel_alloc_bw(struct drm_dp_tunnel *tunnel, int bw)
|
|
{
|
|
int err;
|
|
|
|
err = check_tunnel(tunnel);
|
|
if (err)
|
|
return err;
|
|
|
|
return allocate_tunnel_bw(tunnel, bw);
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_alloc_bw);
|
|
|
|
/**
|
|
* drm_dp_tunnel_get_allocated_bw - Get the BW allocated for a DP tunnel
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Get the current BW allocated for @tunnel. After the tunnel is created /
|
|
* resumed and the BW allocation mode is enabled for it, the allocation
|
|
* becomes determined only after the first allocation request by the driver
|
|
* calling drm_dp_tunnel_alloc_bw().
|
|
*
|
|
* Return the BW allocated for the tunnel, or -1 if the allocation is
|
|
* undetermined.
|
|
*/
|
|
int drm_dp_tunnel_get_allocated_bw(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel->allocated_bw;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_get_allocated_bw);
|
|
|
|
/*
|
|
* Return 0 if the status hasn't changed, 1 if the status has changed, a
|
|
* negative error code in case of an I/O failure.
|
|
*/
|
|
static int check_and_clear_status_change(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
u8 mask = DP_BW_ALLOCATION_CAPABILITY_CHANGED | DP_ESTIMATED_BW_CHANGED;
|
|
u8 val;
|
|
|
|
if (drm_dp_dpcd_readb(tunnel->aux, DP_TUNNELING_STATUS, &val) < 0)
|
|
goto out_err;
|
|
|
|
val &= mask;
|
|
|
|
if (val) {
|
|
if (drm_dp_dpcd_writeb(tunnel->aux, DP_TUNNELING_STATUS, val) < 0)
|
|
goto out_err;
|
|
|
|
return 1;
|
|
}
|
|
|
|
if (!drm_dp_tunnel_bw_alloc_is_enabled(tunnel))
|
|
return 0;
|
|
|
|
/*
|
|
* Check for estimated BW changes explicitly to account for lost
|
|
* BW change notifications.
|
|
*/
|
|
if (drm_dp_dpcd_readb(tunnel->aux, DP_ESTIMATED_BW, &val) < 0)
|
|
goto out_err;
|
|
|
|
if (val * tunnel->bw_granularity != tunnel->estimated_bw)
|
|
return 1;
|
|
|
|
return 0;
|
|
|
|
out_err:
|
|
drm_dp_tunnel_set_io_error(tunnel);
|
|
|
|
return -EIO;
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_update_state - Update DP tunnel SW state with the HW state
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* Update the SW state of @tunnel with the HW state.
|
|
*
|
|
* Returns 0 if the state has not changed, 1 if it has changed and got updated
|
|
* successfully and a negative error code otherwise.
|
|
*/
|
|
int drm_dp_tunnel_update_state(struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_regs regs;
|
|
bool changed = false;
|
|
int ret;
|
|
|
|
ret = check_tunnel(tunnel);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
ret = check_and_clear_status_change(tunnel);
|
|
if (ret < 0)
|
|
goto out;
|
|
|
|
if (!ret)
|
|
return 0;
|
|
|
|
ret = read_and_verify_tunnel_regs(tunnel, ®s, 0);
|
|
if (ret)
|
|
goto out;
|
|
|
|
if (update_dprx_caps(tunnel, ®s))
|
|
changed = true;
|
|
|
|
ret = update_group_available_bw(tunnel, ®s);
|
|
if (ret == 1)
|
|
changed = true;
|
|
|
|
out:
|
|
tun_dbg_stat(tunnel, ret < 0 ? ret : 0,
|
|
"State update: Changed:%s DPRX:%dx%d Tunnel alloc:%d/%d Group alloc:%d/%d Mb/s",
|
|
str_yes_no(changed),
|
|
tunnel->max_dprx_rate / 100, tunnel->max_dprx_lane_count,
|
|
DPTUN_BW_ARG(tunnel->allocated_bw),
|
|
DPTUN_BW_ARG(get_max_tunnel_bw(tunnel)),
|
|
DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
|
|
DPTUN_BW_ARG(tunnel->group->available_bw));
|
|
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
if (changed)
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_update_state);
|
|
|
|
/*
|
|
* drm_dp_tunnel_handle_irq - Handle DP tunnel IRQs
|
|
*
|
|
* Handle any pending DP tunnel IRQs, waking up waiters for a completion
|
|
* event.
|
|
*
|
|
* Returns 1 if the state of the tunnel has changed which requires calling
|
|
* drm_dp_tunnel_update_state(), a negative error code in case of a failure,
|
|
* 0 otherwise.
|
|
*/
|
|
int drm_dp_tunnel_handle_irq(struct drm_dp_tunnel_mgr *mgr, struct drm_dp_aux *aux)
|
|
{
|
|
u8 val;
|
|
|
|
if (drm_dp_dpcd_readb(aux, DP_TUNNELING_STATUS, &val) < 0)
|
|
return -EIO;
|
|
|
|
if (val & (DP_BW_REQUEST_SUCCEEDED | DP_BW_REQUEST_FAILED))
|
|
wake_up_all(&mgr->bw_req_queue);
|
|
|
|
if (val & (DP_BW_ALLOCATION_CAPABILITY_CHANGED | DP_ESTIMATED_BW_CHANGED))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_handle_irq);
|
|
|
|
/**
|
|
* drm_dp_tunnel_max_dprx_rate - Query the maximum rate of the tunnel's DPRX
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* The function is used to query the maximum link rate of the DPRX connected
|
|
* to @tunnel. Note that this rate will not be limited by the BW limit of the
|
|
* tunnel, as opposed to the standard and extended DP_MAX_LINK_RATE DPCD
|
|
* registers.
|
|
*
|
|
* Returns the maximum link rate in 10 kbit/s units.
|
|
*/
|
|
int drm_dp_tunnel_max_dprx_rate(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel->max_dprx_rate;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_max_dprx_rate);
|
|
|
|
/**
|
|
* drm_dp_tunnel_max_dprx_lane_count - Query the maximum lane count of the tunnel's DPRX
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* The function is used to query the maximum lane count of the DPRX connected
|
|
* to @tunnel. Note that this lane count will not be limited by the BW limit of
|
|
* the tunnel, as opposed to the standard and extended DP_MAX_LANE_COUNT DPCD
|
|
* registers.
|
|
*
|
|
* Returns the maximum lane count.
|
|
*/
|
|
int drm_dp_tunnel_max_dprx_lane_count(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel->max_dprx_lane_count;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_max_dprx_lane_count);
|
|
|
|
/**
|
|
* drm_dp_tunnel_available_bw - Query the estimated total available BW of the tunnel
|
|
* @tunnel: Tunnel object
|
|
*
|
|
* This function is used to query the estimated total available BW of the
|
|
* tunnel. This includes the currently allocated and free BW for all the
|
|
* tunnels in @tunnel's group. The available BW is valid only after the BW
|
|
* allocation mode has been enabled for the tunnel and its state got updated
|
|
* calling drm_dp_tunnel_update_state().
|
|
*
|
|
* Returns the @tunnel group's estimated total available bandwidth in kB/s
|
|
* units, or -1 if the available BW isn't valid (the BW allocation mode is
|
|
* not enabled or the tunnel's state hasn't been updated).
|
|
*/
|
|
int drm_dp_tunnel_available_bw(const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return tunnel->group->available_bw;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_available_bw);
|
|
|
|
static struct drm_dp_tunnel_group_state *
|
|
drm_dp_tunnel_atomic_get_group_state(struct drm_atomic_state *state,
|
|
const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
return (struct drm_dp_tunnel_group_state *)
|
|
drm_atomic_get_private_obj_state(state,
|
|
&tunnel->group->base);
|
|
}
|
|
|
|
static struct drm_dp_tunnel_state *
|
|
add_tunnel_state(struct drm_dp_tunnel_group_state *group_state,
|
|
struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
|
|
tun_dbg_atomic(tunnel,
|
|
"Adding state for tunnel %p to group state %p\n",
|
|
tunnel, group_state);
|
|
|
|
tunnel_state = kzalloc(sizeof(*tunnel_state), GFP_KERNEL);
|
|
if (!tunnel_state)
|
|
return NULL;
|
|
|
|
tunnel_state->group_state = group_state;
|
|
|
|
drm_dp_tunnel_ref_get(tunnel, &tunnel_state->tunnel_ref);
|
|
|
|
INIT_LIST_HEAD(&tunnel_state->node);
|
|
list_add(&tunnel_state->node, &group_state->tunnel_states);
|
|
|
|
return tunnel_state;
|
|
}
|
|
|
|
static void free_tunnel_state(struct drm_dp_tunnel_state *tunnel_state)
|
|
{
|
|
tun_dbg_atomic(tunnel_state->tunnel_ref.tunnel,
|
|
"Freeing state for tunnel %p\n",
|
|
tunnel_state->tunnel_ref.tunnel);
|
|
|
|
list_del(&tunnel_state->node);
|
|
|
|
kfree(tunnel_state->stream_bw);
|
|
drm_dp_tunnel_ref_put(&tunnel_state->tunnel_ref);
|
|
|
|
kfree(tunnel_state);
|
|
}
|
|
|
|
static void free_group_state(struct drm_dp_tunnel_group_state *group_state)
|
|
{
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
struct drm_dp_tunnel_state *tunnel_state_tmp;
|
|
|
|
for_each_tunnel_state_safe(group_state, tunnel_state, tunnel_state_tmp)
|
|
free_tunnel_state(tunnel_state);
|
|
|
|
kfree(group_state);
|
|
}
|
|
|
|
static struct drm_dp_tunnel_state *
|
|
get_tunnel_state(struct drm_dp_tunnel_group_state *group_state,
|
|
const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
|
|
for_each_tunnel_state(group_state, tunnel_state)
|
|
if (tunnel_state->tunnel_ref.tunnel == tunnel)
|
|
return tunnel_state;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct drm_dp_tunnel_state *
|
|
get_or_add_tunnel_state(struct drm_dp_tunnel_group_state *group_state,
|
|
struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
|
|
tunnel_state = get_tunnel_state(group_state, tunnel);
|
|
if (tunnel_state)
|
|
return tunnel_state;
|
|
|
|
return add_tunnel_state(group_state, tunnel);
|
|
}
|
|
|
|
static struct drm_private_state *
|
|
tunnel_group_duplicate_state(struct drm_private_obj *obj)
|
|
{
|
|
struct drm_dp_tunnel_group_state *group_state;
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
|
|
group_state = kzalloc(sizeof(*group_state), GFP_KERNEL);
|
|
if (!group_state)
|
|
return NULL;
|
|
|
|
INIT_LIST_HEAD(&group_state->tunnel_states);
|
|
|
|
__drm_atomic_helper_private_obj_duplicate_state(obj, &group_state->base);
|
|
|
|
for_each_tunnel_state(to_group_state(obj->state), tunnel_state) {
|
|
struct drm_dp_tunnel_state *new_tunnel_state;
|
|
|
|
new_tunnel_state = get_or_add_tunnel_state(group_state,
|
|
tunnel_state->tunnel_ref.tunnel);
|
|
if (!new_tunnel_state)
|
|
goto out_free_state;
|
|
|
|
new_tunnel_state->stream_mask = tunnel_state->stream_mask;
|
|
new_tunnel_state->stream_bw = kmemdup(tunnel_state->stream_bw,
|
|
sizeof(*tunnel_state->stream_bw) *
|
|
hweight32(tunnel_state->stream_mask),
|
|
GFP_KERNEL);
|
|
|
|
if (!new_tunnel_state->stream_bw)
|
|
goto out_free_state;
|
|
}
|
|
|
|
return &group_state->base;
|
|
|
|
out_free_state:
|
|
free_group_state(group_state);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static void tunnel_group_destroy_state(struct drm_private_obj *obj, struct drm_private_state *state)
|
|
{
|
|
free_group_state(to_group_state(state));
|
|
}
|
|
|
|
static const struct drm_private_state_funcs tunnel_group_funcs = {
|
|
.atomic_duplicate_state = tunnel_group_duplicate_state,
|
|
.atomic_destroy_state = tunnel_group_destroy_state,
|
|
};
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_get_state - get/allocate the new atomic state for a tunnel
|
|
* @state: Atomic state
|
|
* @tunnel: Tunnel to get the state for
|
|
*
|
|
* Get the new atomic state for @tunnel, duplicating it from the old tunnel
|
|
* state if not yet allocated.
|
|
*
|
|
* Return the state or an ERR_PTR() error on failure.
|
|
*/
|
|
struct drm_dp_tunnel_state *
|
|
drm_dp_tunnel_atomic_get_state(struct drm_atomic_state *state,
|
|
struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_group_state *group_state;
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
|
|
group_state = drm_dp_tunnel_atomic_get_group_state(state, tunnel);
|
|
if (IS_ERR(group_state))
|
|
return ERR_CAST(group_state);
|
|
|
|
tunnel_state = get_or_add_tunnel_state(group_state, tunnel);
|
|
if (!tunnel_state)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
return tunnel_state;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_state);
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_get_old_state - get the old atomic state for a tunnel
|
|
* @state: Atomic state
|
|
* @tunnel: Tunnel to get the state for
|
|
*
|
|
* Get the old atomic state for @tunnel.
|
|
*
|
|
* Return the old state or NULL if the tunnel's atomic state is not in @state.
|
|
*/
|
|
struct drm_dp_tunnel_state *
|
|
drm_dp_tunnel_atomic_get_old_state(struct drm_atomic_state *state,
|
|
const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_group_state *old_group_state;
|
|
int i;
|
|
|
|
for_each_old_group_in_state(state, old_group_state, i)
|
|
if (to_group(old_group_state->base.obj) == tunnel->group)
|
|
return get_tunnel_state(old_group_state, tunnel);
|
|
|
|
return NULL;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_old_state);
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_get_new_state - get the new atomic state for a tunnel
|
|
* @state: Atomic state
|
|
* @tunnel: Tunnel to get the state for
|
|
*
|
|
* Get the new atomic state for @tunnel.
|
|
*
|
|
* Return the new state or NULL if the tunnel's atomic state is not in @state.
|
|
*/
|
|
struct drm_dp_tunnel_state *
|
|
drm_dp_tunnel_atomic_get_new_state(struct drm_atomic_state *state,
|
|
const struct drm_dp_tunnel *tunnel)
|
|
{
|
|
struct drm_dp_tunnel_group_state *new_group_state;
|
|
int i;
|
|
|
|
for_each_new_group_in_state(state, new_group_state, i)
|
|
if (to_group(new_group_state->base.obj) == tunnel->group)
|
|
return get_tunnel_state(new_group_state, tunnel);
|
|
|
|
return NULL;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_new_state);
|
|
|
|
static bool init_group(struct drm_dp_tunnel_mgr *mgr, struct drm_dp_tunnel_group *group)
|
|
{
|
|
struct drm_dp_tunnel_group_state *group_state;
|
|
|
|
group_state = kzalloc(sizeof(*group_state), GFP_KERNEL);
|
|
if (!group_state)
|
|
return false;
|
|
|
|
INIT_LIST_HEAD(&group_state->tunnel_states);
|
|
|
|
group->mgr = mgr;
|
|
group->available_bw = -1;
|
|
INIT_LIST_HEAD(&group->tunnels);
|
|
|
|
drm_atomic_private_obj_init(mgr->dev, &group->base, &group_state->base,
|
|
&tunnel_group_funcs);
|
|
|
|
return true;
|
|
}
|
|
|
|
static void cleanup_group(struct drm_dp_tunnel_group *group)
|
|
{
|
|
drm_atomic_private_obj_fini(&group->base);
|
|
}
|
|
|
|
#ifdef CONFIG_DRM_DISPLAY_DP_TUNNEL_STATE_DEBUG
|
|
static void check_unique_stream_ids(const struct drm_dp_tunnel_group_state *group_state)
|
|
{
|
|
const struct drm_dp_tunnel_state *tunnel_state;
|
|
u32 stream_mask = 0;
|
|
|
|
for_each_tunnel_state(group_state, tunnel_state) {
|
|
drm_WARN(to_group(group_state->base.obj)->mgr->dev,
|
|
tunnel_state->stream_mask & stream_mask,
|
|
"[DPTUN %s]: conflicting stream IDs %x (IDs in other tunnels %x)\n",
|
|
tunnel_state->tunnel_ref.tunnel->name,
|
|
tunnel_state->stream_mask,
|
|
stream_mask);
|
|
|
|
stream_mask |= tunnel_state->stream_mask;
|
|
}
|
|
}
|
|
#else
|
|
static void check_unique_stream_ids(const struct drm_dp_tunnel_group_state *group_state)
|
|
{
|
|
}
|
|
#endif
|
|
|
|
static int stream_id_to_idx(u32 stream_mask, u8 stream_id)
|
|
{
|
|
return hweight32(stream_mask & (BIT(stream_id) - 1));
|
|
}
|
|
|
|
static int resize_bw_array(struct drm_dp_tunnel_state *tunnel_state,
|
|
unsigned long old_mask, unsigned long new_mask)
|
|
{
|
|
unsigned long move_mask = old_mask & new_mask;
|
|
int *new_bws = NULL;
|
|
int id;
|
|
|
|
WARN_ON(!new_mask);
|
|
|
|
if (old_mask == new_mask)
|
|
return 0;
|
|
|
|
new_bws = kcalloc(hweight32(new_mask), sizeof(*new_bws), GFP_KERNEL);
|
|
if (!new_bws)
|
|
return -ENOMEM;
|
|
|
|
for_each_set_bit(id, &move_mask, BITS_PER_TYPE(move_mask))
|
|
new_bws[stream_id_to_idx(new_mask, id)] =
|
|
tunnel_state->stream_bw[stream_id_to_idx(old_mask, id)];
|
|
|
|
kfree(tunnel_state->stream_bw);
|
|
tunnel_state->stream_bw = new_bws;
|
|
tunnel_state->stream_mask = new_mask;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int set_stream_bw(struct drm_dp_tunnel_state *tunnel_state,
|
|
u8 stream_id, int bw)
|
|
{
|
|
int err;
|
|
|
|
err = resize_bw_array(tunnel_state,
|
|
tunnel_state->stream_mask,
|
|
tunnel_state->stream_mask | BIT(stream_id));
|
|
if (err)
|
|
return err;
|
|
|
|
tunnel_state->stream_bw[stream_id_to_idx(tunnel_state->stream_mask, stream_id)] = bw;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int clear_stream_bw(struct drm_dp_tunnel_state *tunnel_state,
|
|
u8 stream_id)
|
|
{
|
|
if (!(tunnel_state->stream_mask & ~BIT(stream_id))) {
|
|
free_tunnel_state(tunnel_state);
|
|
return 0;
|
|
}
|
|
|
|
return resize_bw_array(tunnel_state,
|
|
tunnel_state->stream_mask,
|
|
tunnel_state->stream_mask & ~BIT(stream_id));
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_set_stream_bw - Set the BW for a DP tunnel stream
|
|
* @state: Atomic state
|
|
* @tunnel: DP tunnel containing the stream
|
|
* @stream_id: Stream ID
|
|
* @bw: BW of the stream
|
|
*
|
|
* Set a DP tunnel stream's required BW in the atomic state.
|
|
*
|
|
* Returns 0 in case of success, a negative error code otherwise.
|
|
*/
|
|
int drm_dp_tunnel_atomic_set_stream_bw(struct drm_atomic_state *state,
|
|
struct drm_dp_tunnel *tunnel,
|
|
u8 stream_id, int bw)
|
|
{
|
|
struct drm_dp_tunnel_group_state *new_group_state;
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
int err;
|
|
|
|
if (drm_WARN_ON(tunnel->group->mgr->dev,
|
|
stream_id > BITS_PER_TYPE(tunnel_state->stream_mask)))
|
|
return -EINVAL;
|
|
|
|
tun_dbg(tunnel,
|
|
"Setting %d Mb/s for stream %d\n",
|
|
DPTUN_BW_ARG(bw), stream_id);
|
|
|
|
new_group_state = drm_dp_tunnel_atomic_get_group_state(state, tunnel);
|
|
if (IS_ERR(new_group_state))
|
|
return PTR_ERR(new_group_state);
|
|
|
|
if (bw == 0) {
|
|
tunnel_state = get_tunnel_state(new_group_state, tunnel);
|
|
if (!tunnel_state)
|
|
return 0;
|
|
|
|
return clear_stream_bw(tunnel_state, stream_id);
|
|
}
|
|
|
|
tunnel_state = get_or_add_tunnel_state(new_group_state, tunnel);
|
|
if (drm_WARN_ON(state->dev, !tunnel_state))
|
|
return -EINVAL;
|
|
|
|
err = set_stream_bw(tunnel_state, stream_id, bw);
|
|
if (err)
|
|
return err;
|
|
|
|
check_unique_stream_ids(new_group_state);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_set_stream_bw);
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_get_required_bw - Get the BW required by a DP tunnel
|
|
* @tunnel_state: Atomic state of the queried tunnel
|
|
*
|
|
* Calculate the BW required by a tunnel adding up the required BW of all
|
|
* the streams in the tunnel.
|
|
*
|
|
* Return the total BW required by the tunnel.
|
|
*/
|
|
int drm_dp_tunnel_atomic_get_required_bw(const struct drm_dp_tunnel_state *tunnel_state)
|
|
{
|
|
int tunnel_bw = 0;
|
|
int i;
|
|
|
|
if (!tunnel_state || !tunnel_state->stream_mask)
|
|
return 0;
|
|
|
|
for (i = 0; i < hweight32(tunnel_state->stream_mask); i++)
|
|
tunnel_bw += tunnel_state->stream_bw[i];
|
|
|
|
return tunnel_bw;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_required_bw);
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_get_group_streams_in_state - Get mask of stream IDs in a group
|
|
* @state: Atomic state
|
|
* @tunnel: Tunnel object
|
|
* @stream_mask: Mask of streams in @tunnel's group
|
|
*
|
|
* Get the mask of all the stream IDs in the tunnel group of @tunnel.
|
|
*
|
|
* Return 0 in case of success - with the stream IDs in @stream_mask - or a
|
|
* negative error code in case of failure.
|
|
*/
|
|
int drm_dp_tunnel_atomic_get_group_streams_in_state(struct drm_atomic_state *state,
|
|
const struct drm_dp_tunnel *tunnel,
|
|
u32 *stream_mask)
|
|
{
|
|
struct drm_dp_tunnel_group_state *group_state;
|
|
struct drm_dp_tunnel_state *tunnel_state;
|
|
|
|
group_state = drm_dp_tunnel_atomic_get_group_state(state, tunnel);
|
|
if (IS_ERR(group_state))
|
|
return PTR_ERR(group_state);
|
|
|
|
*stream_mask = 0;
|
|
for_each_tunnel_state(group_state, tunnel_state)
|
|
*stream_mask |= tunnel_state->stream_mask;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_group_streams_in_state);
|
|
|
|
static int
|
|
drm_dp_tunnel_atomic_check_group_bw(struct drm_dp_tunnel_group_state *new_group_state,
|
|
u32 *failed_stream_mask)
|
|
{
|
|
struct drm_dp_tunnel_group *group = to_group(new_group_state->base.obj);
|
|
struct drm_dp_tunnel_state *new_tunnel_state;
|
|
u32 group_stream_mask = 0;
|
|
int group_bw = 0;
|
|
|
|
for_each_tunnel_state(new_group_state, new_tunnel_state) {
|
|
struct drm_dp_tunnel *tunnel = new_tunnel_state->tunnel_ref.tunnel;
|
|
int max_dprx_bw = get_max_dprx_bw(tunnel);
|
|
int tunnel_bw = drm_dp_tunnel_atomic_get_required_bw(new_tunnel_state);
|
|
|
|
tun_dbg(tunnel,
|
|
"%sRequired %d/%d Mb/s total for tunnel.\n",
|
|
tunnel_bw > max_dprx_bw ? "Not enough BW: " : "",
|
|
DPTUN_BW_ARG(tunnel_bw),
|
|
DPTUN_BW_ARG(max_dprx_bw));
|
|
|
|
if (tunnel_bw > max_dprx_bw) {
|
|
*failed_stream_mask = new_tunnel_state->stream_mask;
|
|
return -ENOSPC;
|
|
}
|
|
|
|
group_bw += min(roundup(tunnel_bw, tunnel->bw_granularity),
|
|
max_dprx_bw);
|
|
group_stream_mask |= new_tunnel_state->stream_mask;
|
|
}
|
|
|
|
tun_grp_dbg(group,
|
|
"%sRequired %d/%d Mb/s total for tunnel group.\n",
|
|
group_bw > group->available_bw ? "Not enough BW: " : "",
|
|
DPTUN_BW_ARG(group_bw),
|
|
DPTUN_BW_ARG(group->available_bw));
|
|
|
|
if (group_bw > group->available_bw) {
|
|
*failed_stream_mask = group_stream_mask;
|
|
return -ENOSPC;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_atomic_check_stream_bws - Check BW limit for all streams in state
|
|
* @state: Atomic state
|
|
* @failed_stream_mask: Mask of stream IDs with a BW limit failure
|
|
*
|
|
* Check the required BW of each DP tunnel in @state against both the DPRX BW
|
|
* limit of the tunnel and the BW limit of the tunnel group. Return a mask of
|
|
* stream IDs in @failed_stream_mask once a check fails. The mask will contain
|
|
* either all the streams in a tunnel (in case a DPRX BW limit check failed) or
|
|
* all the streams in a tunnel group (in case a group BW limit check failed).
|
|
*
|
|
* Return 0 if all the BW limit checks passed, -ENOSPC in case a BW limit
|
|
* check failed - with @failed_stream_mask containing the streams failing the
|
|
* check - or a negative error code otherwise.
|
|
*/
|
|
int drm_dp_tunnel_atomic_check_stream_bws(struct drm_atomic_state *state,
|
|
u32 *failed_stream_mask)
|
|
{
|
|
struct drm_dp_tunnel_group_state *new_group_state;
|
|
int i;
|
|
|
|
for_each_new_group_in_state(state, new_group_state, i) {
|
|
int ret;
|
|
|
|
ret = drm_dp_tunnel_atomic_check_group_bw(new_group_state,
|
|
failed_stream_mask);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_atomic_check_stream_bws);
|
|
|
|
static void destroy_mgr(struct drm_dp_tunnel_mgr *mgr)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < mgr->group_count; i++) {
|
|
cleanup_group(&mgr->groups[i]);
|
|
drm_WARN_ON(mgr->dev, !list_empty(&mgr->groups[i].tunnels));
|
|
}
|
|
|
|
#ifdef CONFIG_DRM_DISPLAY_DP_TUNNEL_STATE_DEBUG
|
|
ref_tracker_dir_exit(&mgr->ref_tracker);
|
|
#endif
|
|
|
|
kfree(mgr->groups);
|
|
kfree(mgr);
|
|
}
|
|
|
|
/**
|
|
* drm_dp_tunnel_mgr_create - Create a DP tunnel manager
|
|
* @dev: DRM device object
|
|
* @max_group_count: Maximum number of tunnel groups
|
|
*
|
|
* Creates a DP tunnel manager for @dev.
|
|
*
|
|
* Returns a pointer to the tunnel manager if created successfully or NULL in
|
|
* case of an error.
|
|
*/
|
|
struct drm_dp_tunnel_mgr *
|
|
drm_dp_tunnel_mgr_create(struct drm_device *dev, int max_group_count)
|
|
{
|
|
struct drm_dp_tunnel_mgr *mgr;
|
|
int i;
|
|
|
|
mgr = kzalloc(sizeof(*mgr), GFP_KERNEL);
|
|
if (!mgr)
|
|
return NULL;
|
|
|
|
mgr->dev = dev;
|
|
init_waitqueue_head(&mgr->bw_req_queue);
|
|
|
|
mgr->groups = kcalloc(max_group_count, sizeof(*mgr->groups), GFP_KERNEL);
|
|
if (!mgr->groups) {
|
|
kfree(mgr);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
#ifdef CONFIG_DRM_DISPLAY_DP_TUNNEL_STATE_DEBUG
|
|
ref_tracker_dir_init(&mgr->ref_tracker, 16, "dptun");
|
|
#endif
|
|
|
|
for (i = 0; i < max_group_count; i++) {
|
|
if (!init_group(mgr, &mgr->groups[i])) {
|
|
destroy_mgr(mgr);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
mgr->group_count++;
|
|
}
|
|
|
|
return mgr;
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_mgr_create);
|
|
|
|
/**
|
|
* drm_dp_tunnel_mgr_destroy - Destroy DP tunnel manager
|
|
* @mgr: Tunnel manager object
|
|
*
|
|
* Destroy the tunnel manager.
|
|
*/
|
|
void drm_dp_tunnel_mgr_destroy(struct drm_dp_tunnel_mgr *mgr)
|
|
{
|
|
destroy_mgr(mgr);
|
|
}
|
|
EXPORT_SYMBOL(drm_dp_tunnel_mgr_destroy);
|