Re: [PATCH 02/19] drm/dp: Add support for DP tunneling

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Feb 07, 2024 at 10:48:43PM +0200, Imre Deak wrote:
> On Wed, Feb 07, 2024 at 10:02:18PM +0200, Ville Syrjälä wrote:
> > On Tue, Jan 23, 2024 at 12:28:33PM +0200, Imre Deak wrote:
> > > + [...]
> > > +static int group_allocated_bw(struct drm_dp_tunnel_group *group)
> > > +{
> > > +	struct drm_dp_tunnel *tunnel;
> > > +	int group_allocated_bw = 0;
> > > +
> > > +	for_each_tunnel_in_group(group, tunnel) {
> > > +		if (check_tunnel(tunnel) == 0 &&
> > > +		    tunnel->bw_alloc_enabled)
> > > +			group_allocated_bw += tunnel->allocated_bw;
> > > +	}
> > > +
> > > +	return group_allocated_bw;
> > > +}
> > > +
> > > +static int calc_group_available_bw(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return group_allocated_bw(tunnel->group) -
> > > +	       tunnel->allocated_bw +
> > > +	       tunnel->estimated_bw;
> > 
> > Hmm. So the estimated_bw=actually_free_bw + tunnel->allocated_bw?
> 
> Yes.
> 
> > Ie. how much bw might be available for this tunnel right now?
> 
> Correct.
> 
> > And here we're trying to deduce the total bandwidth available by
> > adding in the allocated_bw of all the other tunnels in the group?
> 
> Yes.
> 
> > Rather weird that we can't just get that number directly...
> 
> It is. Imo this could be simply communicated via a DPCD register
> dedicated for this. Perhaps adding this should be requested from TBT
> architects.

One reason for this design can be that a host/driver may not see all the
tunnels in the group. In that case the tunnel's current usable BW will
be only its estimated_bw (that is it can't use the BW already allocated
by other tunnels in the group, until those are released by the other
host/driver).

> I assume this could also use a code comment.
> 
> > > +}
> > > +
> > > +static int update_group_available_bw(struct drm_dp_tunnel *tunnel,
> > > +				     const struct drm_dp_tunnel_regs *regs)
> > > +{
> > > +	struct drm_dp_tunnel *tunnel_iter;
> > > +	int group_available_bw;
> > > +	bool changed;
> > > +
> > > +	tunnel->estimated_bw = tunnel_reg(regs, DP_ESTIMATED_BW) * tunnel->bw_granularity;
> > > +
> > > +	if (calc_group_available_bw(tunnel) == tunnel->group->available_bw)
> > > +		return 0;
> > > +
> > > +	for_each_tunnel_in_group(tunnel->group, tunnel_iter) {
> > > +		int err;
> > > +
> > > +		if (tunnel_iter == tunnel)
> > > +			continue;
> > > +
> > > +		if (check_tunnel(tunnel_iter) != 0 ||
> > > +		    !tunnel_iter->bw_alloc_enabled)
> > > +			continue;
> > > +
> > > +		err = drm_dp_dpcd_probe(tunnel_iter->aux, DP_DPCD_REV);
> > > +		if (err) {
> > > +			tun_dbg(tunnel_iter,
> > > +				"Probe failed, assume disconnected (err %pe)\n",
> > > +				ERR_PTR(err));
> > > +			drm_dp_tunnel_set_io_error(tunnel_iter);
> > > +		}
> > > +	}
> > > +
> > > +	group_available_bw = calc_group_available_bw(tunnel);
> > > +
> > > +	tun_dbg(tunnel, "Updated group available BW: %d->%d\n",
> > > +		DPTUN_BW_ARG(tunnel->group->available_bw),
> > > +		DPTUN_BW_ARG(group_available_bw));
> > > +
> > > +	changed = tunnel->group->available_bw != group_available_bw;
> > > +
> > > +	tunnel->group->available_bw = group_available_bw;
> > > +
> > > +	return changed ? 1 : 0;
> > > +}
> > > +
> > > +static int set_bw_alloc_mode(struct drm_dp_tunnel *tunnel, bool enable)
> > > +{
> > > +	u8 mask = DP_DISPLAY_DRIVER_BW_ALLOCATION_MODE_ENABLE | DP_UNMASK_BW_ALLOCATION_IRQ;
> > > +	u8 val;
> > > +
> > > +	if (drm_dp_dpcd_readb(tunnel->aux, DP_DPTX_BW_ALLOCATION_MODE_CONTROL, &val) < 0)
> > > +		goto out_err;
> > > +
> > > +	if (enable)
> > > +		val |= mask;
> > > +	else
> > > +		val &= ~mask;
> > > +
> > > +	if (drm_dp_dpcd_writeb(tunnel->aux, DP_DPTX_BW_ALLOCATION_MODE_CONTROL, val) < 0)
> > > +		goto out_err;
> > > +
> > > +	tunnel->bw_alloc_enabled = enable;
> > > +
> > > +	return 0;
> > > +
> > > +out_err:
> > > +	drm_dp_tunnel_set_io_error(tunnel);
> > > +
> > > +	return -EIO;
> > > +}
> > > +
> > > +/**
> > > + * drm_dp_tunnel_enable_bw_alloc: Enable DP tunnel BW allocation mode
> > > + * @tunnel: Tunnel object
> > > + *
> > > + * Enable the DP tunnel BW allocation mode on @tunnel if it supports it.
> > > + *
> > > + * Returns 0 in case of success, negative error code otherwise.
> > > + */
> > > +int drm_dp_tunnel_enable_bw_alloc(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_regs regs;
> > > +	int err = check_tunnel(tunnel);
> > > +
> > > +	if (err)
> > > +		return err;
> > > +
> > > +	if (!tunnel->bw_alloc_supported)
> > > +		return -EOPNOTSUPP;
> > > +
> > > +	if (!tunnel_group_id(tunnel->group->drv_group_id))
> > > +		return -EINVAL;
> > > +
> > > +	err = set_bw_alloc_mode(tunnel, true);
> > > +	if (err)
> > > +		goto out;
> > > +
> > > +	err = read_and_verify_tunnel_regs(tunnel, &regs, 0);
> > > +	if (err) {
> > > +		set_bw_alloc_mode(tunnel, false);
> > > +
> > > +		goto out;
> > > +	}
> > > +
> > > +	if (!tunnel->max_dprx_rate)
> > > +		update_dprx_caps(tunnel, &regs);
> > > +
> > > +	if (tunnel->group->available_bw == -1) {
> > > +		err = update_group_available_bw(tunnel, &regs);
> > > +		if (err > 0)
> > > +			err = 0;
> > > +	}
> > > +out:
> > > +	tun_dbg_stat(tunnel, err,
> > > +		     "Enabling BW alloc mode: DPRX:%dx%d Group alloc:%d/%d Mb/s",
> > > +		     tunnel->max_dprx_rate / 100, tunnel->max_dprx_lane_count,
> > > +		     DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
> > > +		     DPTUN_BW_ARG(tunnel->group->available_bw));
> > > +
> > > +	return err;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_enable_bw_alloc);
> > > +
> > > +/**
> > > + * drm_dp_tunnel_disable_bw_alloc: Disable DP tunnel BW allocation mode
> > > + * @tunnel: Tunnel object
> > > + *
> > > + * Disable the DP tunnel BW allocation mode on @tunnel.
> > > + *
> > > + * Returns 0 in case of success, negative error code otherwise.
> > > + */
> > > +int drm_dp_tunnel_disable_bw_alloc(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	int err = check_tunnel(tunnel);
> > > +
> > > +	if (err)
> > > +		return err;
> > > +
> > > +	err = set_bw_alloc_mode(tunnel, false);
> > > +
> > > +	tun_dbg_stat(tunnel, err, "Disabling BW alloc mode");
> > > +
> > > +	return err;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_disable_bw_alloc);
> > > +
> > > +bool drm_dp_tunnel_bw_alloc_is_enabled(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return tunnel->bw_alloc_enabled;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_bw_alloc_is_enabled);
> > > +
> > > +static int bw_req_complete(struct drm_dp_aux *aux, bool *status_changed)
> > > +{
> > > +	u8 bw_req_mask = DP_BW_REQUEST_SUCCEEDED | DP_BW_REQUEST_FAILED;
> > > +	u8 status_change_mask = DP_BW_ALLOCATION_CAPABILITY_CHANGED | DP_ESTIMATED_BW_CHANGED;
> > > +	u8 val;
> > > +
> > > +	if (drm_dp_dpcd_readb(aux, DP_TUNNELING_STATUS, &val) < 0)
> > > +		return -EIO;
> > > +
> > > +	*status_changed = val & status_change_mask;
> > > +
> > > +	val &= bw_req_mask;
> > > +
> > > +	if (!val)
> > > +		return -EAGAIN;
> > > +
> > > +	if (drm_dp_dpcd_writeb(aux, DP_TUNNELING_STATUS, val) < 0)
> > > +		return -EIO;
> > > +
> > > +	return val == DP_BW_REQUEST_SUCCEEDED ? 0 : -ENOSPC;
> > > +}
> > > +
> > > +static int allocate_tunnel_bw(struct drm_dp_tunnel *tunnel, int bw)
> > > +{
> > > +	struct drm_dp_tunnel_mgr *mgr = tunnel->group->mgr;
> > > +	int request_bw = DIV_ROUND_UP(bw, tunnel->bw_granularity);
> > > +	unsigned long wait_expires;
> > > +	DEFINE_WAIT(wait);
> > > +	int err;
> > > +
> > > +	/* Atomic check should prevent the following. */
> > > +	if (drm_WARN_ON(mgr->dev, request_bw > MAX_DP_REQUEST_BW)) {
> > > +		err = -EINVAL;
> > > +		goto out;
> > > +	}
> > > +
> > > +	if (drm_dp_dpcd_writeb(tunnel->aux, DP_REQUEST_BW, request_bw) < 0) {
> > > +		err = -EIO;
> > > +		goto out;
> > > +	}
> > > +
> > > +	wait_expires = jiffies + msecs_to_jiffies(3000);
> > > +
> > > +	for (;;) {
> > > +		bool status_changed;
> > > +
> > > +		err = bw_req_complete(tunnel->aux, &status_changed);
> > > +		if (err != -EAGAIN)
> > > +			break;
> > > +
> > > +		if (status_changed) {
> > > +			struct drm_dp_tunnel_regs regs;
> > > +
> > > +			err = read_and_verify_tunnel_regs(tunnel, &regs,
> > > +							  ALLOW_ALLOCATED_BW_CHANGE);
> > > +			if (err)
> > > +				break;
> > > +		}
> > > +
> > > +		if (time_after(jiffies, wait_expires)) {
> > > +			err = -ETIMEDOUT;
> > > +			break;
> > > +		}
> > > +
> > > +		prepare_to_wait(&mgr->bw_req_queue, &wait, TASK_UNINTERRUPTIBLE);
> > > +		schedule_timeout(msecs_to_jiffies(200));
> > > +	};
> > > +
> > > +	finish_wait(&mgr->bw_req_queue, &wait);
> > > +
> > > +	if (err)
> > > +		goto out;
> > > +
> > > +	tunnel->allocated_bw = request_bw * tunnel->bw_granularity;
> > > +
> > > +out:
> > > +	tun_dbg_stat(tunnel, err, "Allocating %d/%d Mb/s for tunnel: Group alloc:%d/%d Mb/s",
> > > +		     DPTUN_BW_ARG(request_bw * tunnel->bw_granularity),
> > > +		     DPTUN_BW_ARG(get_max_tunnel_bw(tunnel)),
> > > +		     DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
> > > +		     DPTUN_BW_ARG(tunnel->group->available_bw));
> > > +
> > > +	if (err == -EIO)
> > > +		drm_dp_tunnel_set_io_error(tunnel);
> > > +
> > > +	return err;
> > > +}
> > > +
> > > +int drm_dp_tunnel_alloc_bw(struct drm_dp_tunnel *tunnel, int bw)
> > > +{
> > > +	int err = check_tunnel(tunnel);
> > > +
> > > +	if (err)
> > > +		return err;
> > > +
> > > +	return allocate_tunnel_bw(tunnel, bw);
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_alloc_bw);
> > > +
> > > +static int check_and_clear_status_change(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	u8 mask = DP_BW_ALLOCATION_CAPABILITY_CHANGED | DP_ESTIMATED_BW_CHANGED;
> > > +	u8 val;
> > > +
> > > +	if (drm_dp_dpcd_readb(tunnel->aux, DP_TUNNELING_STATUS, &val) < 0)
> > > +		goto out_err;
> > > +
> > > +	val &= mask;
> > > +
> > > +	if (val) {
> > > +		if (drm_dp_dpcd_writeb(tunnel->aux, DP_TUNNELING_STATUS, val) < 0)
> > > +			goto out_err;
> > > +
> > > +		return 1;
> > > +	}
> > > +
> > > +	if (!drm_dp_tunnel_bw_alloc_is_enabled(tunnel))
> > > +		return 0;
> > > +
> > > +	/*
> > > +	 * Check for estimated BW changes explicitly to account for lost
> > > +	 * BW change notifications.
> > > +	 */
> > > +	if (drm_dp_dpcd_readb(tunnel->aux, DP_ESTIMATED_BW, &val) < 0)
> > > +		goto out_err;
> > > +
> > > +	if (val * tunnel->bw_granularity != tunnel->estimated_bw)
> > > +		return 1;
> > > +
> > > +	return 0;
> > > +
> > > +out_err:
> > > +	drm_dp_tunnel_set_io_error(tunnel);
> > > +
> > > +	return -EIO;
> > > +}
> > > +
> > > +/**
> > > + * drm_dp_tunnel_update_state: Update DP tunnel SW state with the HW state
> > > + * @tunnel: Tunnel object
> > > + *
> > > + * Update the SW state of @tunnel with the HW state.
> > > + *
> > > + * Returns 0 if the state has not changed, 1 if it has changed and got updated
> > > + * successfully and a negative error code otherwise.
> > > + */
> > > +int drm_dp_tunnel_update_state(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_regs regs;
> > > +	bool changed = false;
> > > +	int ret = check_tunnel(tunnel);
> > > +
> > > +	if (ret < 0)
> > > +		return ret;
> > > +
> > > +	ret = check_and_clear_status_change(tunnel);
> > > +	if (ret < 0)
> > > +		goto out;
> > > +
> > > +	if (!ret)
> > > +		return 0;
> > > +
> > > +	ret = read_and_verify_tunnel_regs(tunnel, &regs, 0);
> > > +	if (ret)
> > > +		goto out;
> > > +
> > > +	if (update_dprx_caps(tunnel, &regs))
> > > +		changed = true;
> > > +
> > > +	ret = update_group_available_bw(tunnel, &regs);
> > > +	if (ret == 1)
> > > +		changed = true;
> > > +
> > > +out:
> > > +	tun_dbg_stat(tunnel, ret < 0 ? ret : 0,
> > > +		     "State update: Changed:%c DPRX:%dx%d Tunnel alloc:%d/%d Group alloc:%d/%d Mb/s",
> > > +		     yes_no_chr(changed),
> > > +		     tunnel->max_dprx_rate / 100, tunnel->max_dprx_lane_count,
> > > +		     DPTUN_BW_ARG(tunnel->allocated_bw),
> > > +		     DPTUN_BW_ARG(get_max_tunnel_bw(tunnel)),
> > > +		     DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
> > > +		     DPTUN_BW_ARG(tunnel->group->available_bw));
> > > +
> > > +	if (ret < 0)
> > > +		return ret;
> > > +
> > > +	if (changed)
> > > +		return 1;
> > > +
> > > +	return 0;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_update_state);
> > > +
> > > +/*
> > > + * Returns 0 if no re-probe is needed, 1 if a re-probe is needed,
> > > + * a negative error code otherwise.
> > > + */
> > > +int drm_dp_tunnel_handle_irq(struct drm_dp_tunnel_mgr *mgr, struct drm_dp_aux *aux)
> > > +{
> > > +	u8 val;
> > > +
> > > +	if (drm_dp_dpcd_readb(aux, DP_TUNNELING_STATUS, &val) < 0)
> > > +		return -EIO;
> > > +
> > > +	if (val & (DP_BW_REQUEST_SUCCEEDED | DP_BW_REQUEST_FAILED))
> > > +		wake_up_all(&mgr->bw_req_queue);
> > > +
> > > +	if (val & (DP_BW_ALLOCATION_CAPABILITY_CHANGED | DP_ESTIMATED_BW_CHANGED))
> > > +		return 1;
> > > +
> > > +	return 0;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_handle_irq);
> > > +
> > > +/**
> > > + * drm_dp_tunnel_max_dprx_rate - Query the maximum rate of the tunnel's DPRX
> > > + * @tunnel: Tunnel object
> > > + *
> > > + * The function is used to query the maximum link rate of the DPRX connected
> > > + * to @tunnel. Note that this rate will not be limited by the BW limit of the
> > > + * tunnel, as opposed to the standard and extended DP_MAX_LINK_RATE DPCD
> > > + * registers.
> > > + *
> > > + * Returns the maximum link rate in 10 kbit/s units.
> > > + */
> > > +int drm_dp_tunnel_max_dprx_rate(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return tunnel->max_dprx_rate;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_max_dprx_rate);
> > > +
> > > +/**
> > > + * drm_dp_tunnel_max_dprx_lane_count - Query the maximum lane count of the tunnel's DPRX
> > > + * @tunnel: Tunnel object
> > > + *
> > > + * The function is used to query the maximum lane count of the DPRX connected
> > > + * to @tunnel. Note that this lane count will not be limited by the BW limit of
> > > + * the tunnel, as opposed to the standard and extended DP_MAX_LANE_COUNT DPCD
> > > + * registers.
> > > + *
> > > + * Returns the maximum lane count.
> > > + */
> > > +int drm_dp_tunnel_max_dprx_lane_count(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return tunnel->max_dprx_lane_count;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_max_dprx_lane_count);
> > > +
> > > +/**
> > > + * drm_dp_tunnel_available_bw - Query the estimated total available BW of the tunnel
> > > + * @tunnel: Tunnel object
> > > + *
> > > + * This function is used to query the estimated total available BW of the
> > > + * tunnel. This includes the currently allocated and free BW for all the
> > > + * tunnels in @tunnel's group. The available BW is valid only after the BW
> > > + * allocation mode has been enabled for the tunnel and its state got updated
> > > + * calling drm_dp_tunnel_update_state().
> > > + *
> > > + * Returns the @tunnel group's estimated total available bandwidth in kB/s
> > > + * units, or -1 if the available BW isn't valid (the BW allocation mode is
> > > + * not enabled or the tunnel's state hasn't been updated).
> > > + */
> > > +int drm_dp_tunnel_available_bw(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return tunnel->group->available_bw;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_available_bw);
> > > +
> > > +static struct drm_dp_tunnel_group_state *
> > > +drm_dp_tunnel_atomic_get_group_state(struct drm_atomic_state *state,
> > > +				     const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return (struct drm_dp_tunnel_group_state *)
> > > +		drm_atomic_get_private_obj_state(state,
> > > +						 &tunnel->group->base);
> > > +}
> > > +
> > > +static struct drm_dp_tunnel_state *
> > > +add_tunnel_state(struct drm_dp_tunnel_group_state *group_state,
> > > +		 struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +
> > > +	tun_dbg_atomic(tunnel,
> > > +		       "Adding state for tunnel %p to group state %p\n",
> > > +		       tunnel, group_state);
> > > +
> > > +	tunnel_state = kzalloc(sizeof(*tunnel_state), GFP_KERNEL);
> > > +	if (!tunnel_state)
> > > +		return NULL;
> > > +
> > > +	tunnel_state->group_state = group_state;
> > > +
> > > +	drm_dp_tunnel_ref_get(tunnel, &tunnel_state->tunnel_ref);
> > > +
> > > +	INIT_LIST_HEAD(&tunnel_state->node);
> > > +	list_add(&tunnel_state->node, &group_state->tunnel_states);
> > > +
> > > +	return tunnel_state;
> > > +}
> > > +
> > > +void drm_dp_tunnel_atomic_clear_state(struct drm_dp_tunnel_state *tunnel_state)
> > > +{
> > > +	tun_dbg_atomic(tunnel_state->tunnel_ref.tunnel,
> > > +		       "Clearing state for tunnel %p\n",
> > > +		       tunnel_state->tunnel_ref.tunnel);
> > > +
> > > +	list_del(&tunnel_state->node);
> > > +
> > > +	kfree(tunnel_state->stream_bw);
> > > +	drm_dp_tunnel_ref_put(&tunnel_state->tunnel_ref);
> > > +
> > > +	kfree(tunnel_state);
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_clear_state);
> > > +
> > > +static void clear_tunnel_group_state(struct drm_dp_tunnel_group_state *group_state)
> > > +{
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +	struct drm_dp_tunnel_state *tunnel_state_tmp;
> > > +
> > > +	for_each_tunnel_state_safe(group_state, tunnel_state, tunnel_state_tmp)
> > > +		drm_dp_tunnel_atomic_clear_state(tunnel_state);
> > > +}
> > > +
> > > +static struct drm_dp_tunnel_state *
> > > +get_tunnel_state(struct drm_dp_tunnel_group_state *group_state,
> > > +		 const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +
> > > +	for_each_tunnel_state(group_state, tunnel_state)
> > > +		if (tunnel_state->tunnel_ref.tunnel == tunnel)
> > > +			return tunnel_state;
> > > +
> > > +	return NULL;
> > > +}
> > > +
> > > +static struct drm_dp_tunnel_state *
> > > +get_or_add_tunnel_state(struct drm_dp_tunnel_group_state *group_state,
> > > +			struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +
> > > +	tunnel_state = get_tunnel_state(group_state, tunnel);
> > > +	if (tunnel_state)
> > > +		return tunnel_state;
> > > +
> > > +	return add_tunnel_state(group_state, tunnel);
> > > +}
> > > +
> > > +static struct drm_private_state *
> > > +tunnel_group_duplicate_state(struct drm_private_obj *obj)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *group_state = to_group_state(obj->state);
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +
> > > +	group_state = kzalloc(sizeof(*group_state), GFP_KERNEL);
> > > +	if (!group_state)
> > > +		return NULL;
> > > +
> > > +	INIT_LIST_HEAD(&group_state->tunnel_states);
> > > +
> > > +	__drm_atomic_helper_private_obj_duplicate_state(obj, &group_state->base);
> > > +
> > > +	for_each_tunnel_state(to_group_state(obj->state), tunnel_state) {
> > > +		struct drm_dp_tunnel_state *new_tunnel_state;
> > > +
> > > +		new_tunnel_state = get_or_add_tunnel_state(group_state,
> > > +							   tunnel_state->tunnel_ref.tunnel);
> > > +		if (!new_tunnel_state)
> > > +			goto out_free_state;
> > > +
> > > +		new_tunnel_state->stream_mask = tunnel_state->stream_mask;
> > > +		new_tunnel_state->stream_bw = kmemdup(tunnel_state->stream_bw,
> > > +						      sizeof(*tunnel_state->stream_bw) *
> > > +							hweight32(tunnel_state->stream_mask),
> > > +						      GFP_KERNEL);
> > > +
> > > +		if (!new_tunnel_state->stream_bw)
> > > +			goto out_free_state;
> > > +	}
> > > +
> > > +	return &group_state->base;
> > > +
> > > +out_free_state:
> > > +	clear_tunnel_group_state(group_state);
> > > +	kfree(group_state);
> > > +
> > > +	return NULL;
> > > +}
> > > +
> > > +static void tunnel_group_destroy_state(struct drm_private_obj *obj, struct drm_private_state *state)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *group_state = to_group_state(state);
> > > +
> > > +	clear_tunnel_group_state(group_state);
> > > +	kfree(group_state);
> > > +}
> > > +
> > > +static const struct drm_private_state_funcs tunnel_group_funcs = {
> > > +	.atomic_duplicate_state = tunnel_group_duplicate_state,
> > > +	.atomic_destroy_state = tunnel_group_destroy_state,
> > > +};
> > > +
> > > +struct drm_dp_tunnel_state *
> > > +drm_dp_tunnel_atomic_get_state(struct drm_atomic_state *state,
> > > +			       struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *group_state =
> > > +		drm_dp_tunnel_atomic_get_group_state(state, tunnel);
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +
> > > +	if (IS_ERR(group_state))
> > > +		return ERR_CAST(group_state);
> > > +
> > > +	tunnel_state = get_or_add_tunnel_state(group_state, tunnel);
> > > +	if (!tunnel_state)
> > > +		return ERR_PTR(-ENOMEM);
> > > +
> > > +	return tunnel_state;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_state);
> > > +
> > > +struct drm_dp_tunnel_state *
> > > +drm_dp_tunnel_atomic_get_new_state(struct drm_atomic_state *state,
> > > +				   const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *new_group_state;
> > > +	int i;
> > > +
> > > +	for_each_new_group_in_state(state, new_group_state, i)
> > > +		if (to_group(new_group_state->base.obj) == tunnel->group)
> > > +			return get_tunnel_state(new_group_state, tunnel);
> > > +
> > > +	return NULL;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_new_state);
> > > +
> > > +static bool init_group(struct drm_dp_tunnel_mgr *mgr, struct drm_dp_tunnel_group *group)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *group_state = kzalloc(sizeof(*group_state), GFP_KERNEL);
> > > +
> > > +	if (!group_state)
> > > +		return false;
> > > +
> > > +	INIT_LIST_HEAD(&group_state->tunnel_states);
> > > +
> > > +	group->mgr = mgr;
> > > +	group->available_bw = -1;
> > > +	INIT_LIST_HEAD(&group->tunnels);
> > > +
> > > +	drm_atomic_private_obj_init(mgr->dev, &group->base, &group_state->base,
> > > +				    &tunnel_group_funcs);
> > > +
> > > +	return true;
> > > +}
> > > +
> > > +static void cleanup_group(struct drm_dp_tunnel_group *group)
> > > +{
> > > +	drm_atomic_private_obj_fini(&group->base);
> > > +}
> > > +
> > > +#ifdef CONFIG_DRM_DISPLAY_DEBUG_DP_TUNNEL_STATE
> > > +static void check_unique_stream_ids(const struct drm_dp_tunnel_group_state *group_state)
> > > +{
> > > +	const struct drm_dp_tunnel_state *tunnel_state;
> > > +	u32 stream_mask = 0;
> > > +
> > > +	for_each_tunnel_state(group_state, tunnel_state) {
> > > +		drm_WARN(to_group(group_state->base.obj)->mgr->dev,
> > > +			 tunnel_state->stream_mask & stream_mask,
> > > +			 "[DPTUN %s]: conflicting stream IDs %x (IDs in other tunnels %x)\n",
> > > +			 tunnel_state->tunnel_ref.tunnel->name,
> > > +			 tunnel_state->stream_mask,
> > > +			 stream_mask);
> > > +
> > > +		stream_mask |= tunnel_state->stream_mask;
> > > +	}
> > > +}
> > > +#else
> > > +static void check_unique_stream_ids(const struct drm_dp_tunnel_group_state *group_state)
> > > +{
> > > +}
> > > +#endif
> > > +
> > > +static int stream_id_to_idx(u32 stream_mask, u8 stream_id)
> > > +{
> > > +	return hweight32(stream_mask & (BIT(stream_id) - 1));
> > > +}
> > > +
> > > +static int resize_bw_array(struct drm_dp_tunnel_state *tunnel_state,
> > > +			   unsigned long old_mask, unsigned long new_mask)
> > > +{
> > > +	unsigned long move_mask = old_mask & new_mask;
> > > +	int *new_bws = NULL;
> > > +	int id;
> > > +
> > > +	WARN_ON(!new_mask);
> > > +
> > > +	if (old_mask == new_mask)
> > > +		return 0;
> > > +
> > > +	new_bws = kcalloc(hweight32(new_mask), sizeof(*new_bws), GFP_KERNEL);
> > > +	if (!new_bws)
> > > +		return -ENOMEM;
> > > +
> > > +	for_each_set_bit(id, &move_mask, BITS_PER_TYPE(move_mask))
> > > +		new_bws[stream_id_to_idx(new_mask, id)] =
> > > +			tunnel_state->stream_bw[stream_id_to_idx(old_mask, id)];
> > > +
> > > +	kfree(tunnel_state->stream_bw);
> > > +	tunnel_state->stream_bw = new_bws;
> > > +	tunnel_state->stream_mask = new_mask;
> > > +
> > > +	return 0;
> > > +}
> > > +
> > > +static int set_stream_bw(struct drm_dp_tunnel_state *tunnel_state,
> > > +			 u8 stream_id, int bw)
> > > +{
> > > +	int err;
> > > +
> > > +	err = resize_bw_array(tunnel_state,
> > > +			      tunnel_state->stream_mask,
> > > +			      tunnel_state->stream_mask | BIT(stream_id));
> > > +	if (err)
> > > +		return err;
> > > +
> > > +	tunnel_state->stream_bw[stream_id_to_idx(tunnel_state->stream_mask, stream_id)] = bw;
> > > +
> > > +	return 0;
> > > +}
> > > +
> > > +static int clear_stream_bw(struct drm_dp_tunnel_state *tunnel_state,
> > > +			   u8 stream_id)
> > > +{
> > > +	if (!(tunnel_state->stream_mask & ~BIT(stream_id))) {
> > > +		drm_dp_tunnel_atomic_clear_state(tunnel_state);
> > > +		return 0;
> > > +	}
> > > +
> > > +	return resize_bw_array(tunnel_state,
> > > +			       tunnel_state->stream_mask,
> > > +			       tunnel_state->stream_mask & ~BIT(stream_id));
> > > +}
> > > +
> > > +int drm_dp_tunnel_atomic_set_stream_bw(struct drm_atomic_state *state,
> > > +					 struct drm_dp_tunnel *tunnel,
> > > +					 u8 stream_id, int bw)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *new_group_state =
> > > +		drm_dp_tunnel_atomic_get_group_state(state, tunnel);
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +	int err;
> > > +
> > > +	if (drm_WARN_ON(tunnel->group->mgr->dev,
> > > +			stream_id > BITS_PER_TYPE(tunnel_state->stream_mask)))
> > > +		return -EINVAL;
> > > +
> > > +	tun_dbg(tunnel,
> > > +		"Setting %d Mb/s for stream %d\n",
> > > +		DPTUN_BW_ARG(bw), stream_id);
> > > +
> > > +	if (bw == 0) {
> > > +		tunnel_state = get_tunnel_state(new_group_state, tunnel);
> > > +		if (!tunnel_state)
> > > +			return 0;
> > > +
> > > +		return clear_stream_bw(tunnel_state, stream_id);
> > > +	}
> > > +
> > > +	tunnel_state = get_or_add_tunnel_state(new_group_state, tunnel);
> > > +	if (drm_WARN_ON(state->dev, !tunnel_state))
> > > +		return -EINVAL;
> > > +
> > > +	err = set_stream_bw(tunnel_state, stream_id, bw);
> > > +	if (err)
> > > +		return err;
> > > +
> > > +	check_unique_stream_ids(new_group_state);
> > > +
> > > +	return 0;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_set_stream_bw);
> > > +
> > > +int drm_dp_tunnel_atomic_get_tunnel_bw(const struct drm_dp_tunnel_state *tunnel_state)
> > > +{
> > > +	int tunnel_bw = 0;
> > > +	int i;
> > > +
> > > +	for (i = 0; i < hweight32(tunnel_state->stream_mask); i++)
> > > +		tunnel_bw += tunnel_state->stream_bw[i];
> > > +
> > > +	return tunnel_bw;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_tunnel_bw);
> > > +
> > > +int drm_dp_tunnel_atomic_get_group_streams_in_state(struct drm_atomic_state *state,
> > > +						    const struct drm_dp_tunnel *tunnel,
> > > +						    u32 *stream_mask)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *group_state =
> > > +		drm_dp_tunnel_atomic_get_group_state(state, tunnel);
> > > +	struct drm_dp_tunnel_state *tunnel_state;
> > > +
> > > +	if (IS_ERR(group_state))
> > > +		return PTR_ERR(group_state);
> > > +
> > > +	*stream_mask = 0;
> > > +	for_each_tunnel_state(group_state, tunnel_state)
> > > +		*stream_mask |= tunnel_state->stream_mask;
> > > +
> > > +	return 0;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_get_group_streams_in_state);
> > > +
> > > +static int
> > > +drm_dp_tunnel_atomic_check_group_bw(struct drm_dp_tunnel_group_state *new_group_state,
> > > +				    u32 *failed_stream_mask)
> > > +{
> > > +	struct drm_dp_tunnel_group *group = to_group(new_group_state->base.obj);
> > > +	struct drm_dp_tunnel_state *new_tunnel_state;
> > > +	u32 group_stream_mask = 0;
> > > +	int group_bw = 0;
> > > +
> > > +	for_each_tunnel_state(new_group_state, new_tunnel_state) {
> > > +		struct drm_dp_tunnel *tunnel = new_tunnel_state->tunnel_ref.tunnel;
> > > +		int max_dprx_bw = get_max_dprx_bw(tunnel);
> > > +		int tunnel_bw = drm_dp_tunnel_atomic_get_tunnel_bw(new_tunnel_state);
> > > +
> > > +		tun_dbg(tunnel,
> > > +			"%sRequired %d/%d Mb/s total for tunnel.\n",
> > > +			tunnel_bw > max_dprx_bw ? "Not enough BW: " : "",
> > > +			DPTUN_BW_ARG(tunnel_bw),
> > > +			DPTUN_BW_ARG(max_dprx_bw));
> > > +
> > > +		if (tunnel_bw > max_dprx_bw) {
> > 
> > I'm a bit confused why we're checking this here. Aren't we already
> > checking this somewhere else?
> 
> Ah, yes this should be checked already by the encoder compute config +
> the MST link BW check. It can be removed, thanks.
> 
> > > +			*failed_stream_mask = new_tunnel_state->stream_mask;
> > > +			return -ENOSPC;
> > > +		}
> > > +
> > > +		group_bw += min(roundup(tunnel_bw, tunnel->bw_granularity),
> > > +				max_dprx_bw);
> > > +		group_stream_mask |= new_tunnel_state->stream_mask;
> > > +	}
> > > +
> > > +	tun_grp_dbg(group,
> > > +		    "%sRequired %d/%d Mb/s total for tunnel group.\n",
> > > +		    group_bw > group->available_bw ? "Not enough BW: " : "",
> > > +		    DPTUN_BW_ARG(group_bw),
> > > +		    DPTUN_BW_ARG(group->available_bw));
> > > +
> > > +	if (group_bw > group->available_bw) {
> > > +		*failed_stream_mask = group_stream_mask;
> > > +		return -ENOSPC;
> > > +	}
> > > +
> > > +	return 0;
> > > +}
> > > +
> > > +int drm_dp_tunnel_atomic_check_stream_bws(struct drm_atomic_state *state,
> > > +					  u32 *failed_stream_mask)
> > > +{
> > > +	struct drm_dp_tunnel_group_state *new_group_state;
> > > +	int i;
> > > +
> > > +	for_each_new_group_in_state(state, new_group_state, i) {
> > > +		int ret;
> > > +
> > > +		ret = drm_dp_tunnel_atomic_check_group_bw(new_group_state,
> > > +							  failed_stream_mask);
> > > +		if (ret)
> > > +			return ret;
> > > +	}
> > > +
> > > +	return 0;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_atomic_check_stream_bws);
> > > +
> > > +static void destroy_mgr(struct drm_dp_tunnel_mgr *mgr)
> > > +{
> > > +	int i;
> > > +
> > > +	for (i = 0; i < mgr->group_count; i++) {
> > > +		cleanup_group(&mgr->groups[i]);
> > > +		drm_WARN_ON(mgr->dev, !list_empty(&mgr->groups[i].tunnels));
> > > +	}
> > > +
> > > +#ifdef CONFIG_DRM_DISPLAY_DEBUG_DP_TUNNEL_STATE
> > > +	ref_tracker_dir_exit(&mgr->ref_tracker);
> > > +#endif
> > > +
> > > +	kfree(mgr->groups);
> > > +	kfree(mgr);
> > > +}
> > > +
> > > +/**
> > > + * drm_dp_tunnel_mgr_create - Create a DP tunnel manager
> > > + * @i915: i915 driver object
> > > + *
> > > + * Creates a DP tunnel manager.
> > > + *
> > > + * Returns a pointer to the tunnel manager if created successfully or NULL in
> > > + * case of an error.
> > > + */
> > > +struct drm_dp_tunnel_mgr *
> > > +drm_dp_tunnel_mgr_create(struct drm_device *dev, int max_group_count)
> > > +{
> > > +	struct drm_dp_tunnel_mgr *mgr = kzalloc(sizeof(*mgr), GFP_KERNEL);
> > > +	int i;
> > > +
> > > +	if (!mgr)
> > > +		return NULL;
> > > +
> > > +	mgr->dev = dev;
> > > +	init_waitqueue_head(&mgr->bw_req_queue);
> > > +
> > > +	mgr->groups = kcalloc(max_group_count, sizeof(*mgr->groups), GFP_KERNEL);
> > > +	if (!mgr->groups) {
> > > +		kfree(mgr);
> > > +
> > > +		return NULL;
> > > +	}
> > > +
> > > +#ifdef CONFIG_DRM_DISPLAY_DEBUG_DP_TUNNEL_STATE
> > > +	ref_tracker_dir_init(&mgr->ref_tracker, 16, "dptun");
> > > +#endif
> > > +
> > > +	for (i = 0; i < max_group_count; i++) {
> > > +		if (!init_group(mgr, &mgr->groups[i])) {
> > > +			destroy_mgr(mgr);
> > > +
> > > +			return NULL;
> > > +		}
> > > +
> > > +		mgr->group_count++;
> > > +	}
> > > +
> > > +	return mgr;
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_mgr_create);
> > > +
> > > +/**
> > > + * drm_dp_tunnel_mgr_destroy - Destroy DP tunnel manager
> > > + * @mgr: Tunnel manager object
> > > + *
> > > + * Destroy the tunnel manager.
> > > + */
> > > +void drm_dp_tunnel_mgr_destroy(struct drm_dp_tunnel_mgr *mgr)
> > > +{
> > > +	destroy_mgr(mgr);
> > > +}
> > > +EXPORT_SYMBOL(drm_dp_tunnel_mgr_destroy);
> > > diff --git a/include/drm/display/drm_dp.h b/include/drm/display/drm_dp.h
> > > index 281afff6ee4e5..8bfd5d007be8d 100644
> > > --- a/include/drm/display/drm_dp.h
> > > +++ b/include/drm/display/drm_dp.h
> > > @@ -1382,6 +1382,66 @@
> > >  #define DP_HDCP_2_2_REG_STREAM_TYPE_OFFSET	0x69494
> > >  #define DP_HDCP_2_2_REG_DBG_OFFSET		0x69518
> > >  
> > > +/* DP-tunneling */
> > > +#define DP_TUNNELING_OUI				0xe0000
> > > +#define  DP_TUNNELING_OUI_BYTES				3
> > > +
> > > +#define DP_TUNNELING_DEV_ID				0xe0003
> > > +#define  DP_TUNNELING_DEV_ID_BYTES			6
> > > +
> > > +#define DP_TUNNELING_HW_REV				0xe0009
> > > +#define  DP_TUNNELING_HW_REV_MAJOR_SHIFT		4
> > > +#define  DP_TUNNELING_HW_REV_MAJOR_MASK			(0xf << DP_TUNNELING_HW_REV_MAJOR_SHIFT)
> > > +#define  DP_TUNNELING_HW_REV_MINOR_SHIFT		0
> > > +#define  DP_TUNNELING_HW_REV_MINOR_MASK			(0xf << DP_TUNNELING_HW_REV_MINOR_SHIFT)
> > > +
> > > +#define DP_TUNNELING_SW_REV_MAJOR			0xe000a
> > > +#define DP_TUNNELING_SW_REV_MINOR			0xe000b
> > > +
> > > +#define DP_TUNNELING_CAPABILITIES			0xe000d
> > > +#define  DP_IN_BW_ALLOCATION_MODE_SUPPORT		(1 << 7)
> > > +#define  DP_PANEL_REPLAY_OPTIMIZATION_SUPPORT		(1 << 6)
> > > +#define  DP_TUNNELING_SUPPORT				(1 << 0)
> > > +
> > > +#define DP_IN_ADAPTER_INFO				0xe000e
> > > +#define  DP_IN_ADAPTER_NUMBER_BITS			7
> > > +#define  DP_IN_ADAPTER_NUMBER_MASK			((1 << DP_IN_ADAPTER_NUMBER_BITS) - 1)
> > > +
> > > +#define DP_USB4_DRIVER_ID				0xe000f
> > > +#define  DP_USB4_DRIVER_ID_BITS				4
> > > +#define  DP_USB4_DRIVER_ID_MASK				((1 << DP_USB4_DRIVER_ID_BITS) - 1)
> > > +
> > > +#define DP_USB4_DRIVER_BW_CAPABILITY			0xe0020
> > > +#define  DP_USB4_DRIVER_BW_ALLOCATION_MODE_SUPPORT	(1 << 7)
> > > +
> > > +#define DP_IN_ADAPTER_TUNNEL_INFORMATION		0xe0021
> > > +#define  DP_GROUP_ID_BITS				3
> > > +#define  DP_GROUP_ID_MASK				((1 << DP_GROUP_ID_BITS) - 1)
> > > +
> > > +#define DP_BW_GRANULARITY				0xe0022
> > > +#define  DP_BW_GRANULARITY_MASK				0x3
> > > +
> > > +#define DP_ESTIMATED_BW					0xe0023
> > > +#define DP_ALLOCATED_BW					0xe0024
> > > +
> > > +#define DP_TUNNELING_STATUS				0xe0025
> > > +#define  DP_BW_ALLOCATION_CAPABILITY_CHANGED		(1 << 3)
> > > +#define  DP_ESTIMATED_BW_CHANGED			(1 << 2)
> > > +#define  DP_BW_REQUEST_SUCCEEDED			(1 << 1)
> > > +#define  DP_BW_REQUEST_FAILED				(1 << 0)
> > > +
> > > +#define DP_TUNNELING_MAX_LINK_RATE			0xe0028
> > > +
> > > +#define DP_TUNNELING_MAX_LANE_COUNT			0xe0029
> > > +#define  DP_TUNNELING_MAX_LANE_COUNT_MASK		0x1f
> > > +
> > > +#define DP_DPTX_BW_ALLOCATION_MODE_CONTROL		0xe0030
> > > +#define  DP_DISPLAY_DRIVER_BW_ALLOCATION_MODE_ENABLE	(1 << 7)
> > > +#define  DP_UNMASK_BW_ALLOCATION_IRQ			(1 << 6)
> > > +
> > > +#define DP_REQUEST_BW					0xe0031
> > > +#define  MAX_DP_REQUEST_BW				255
> > > +
> > >  /* LTTPR: Link Training (LT)-tunable PHY Repeaters */
> > >  #define DP_LT_TUNABLE_PHY_REPEATER_FIELD_DATA_STRUCTURE_REV 0xf0000 /* 1.3 */
> > >  #define DP_MAX_LINK_RATE_PHY_REPEATER			    0xf0001 /* 1.4a */
> > > diff --git a/include/drm/display/drm_dp_tunnel.h b/include/drm/display/drm_dp_tunnel.h
> > > new file mode 100644
> > > index 0000000000000..f6449b1b4e6e9
> > > --- /dev/null
> > > +++ b/include/drm/display/drm_dp_tunnel.h
> > > @@ -0,0 +1,270 @@
> > > +/* SPDX-License-Identifier: MIT */
> > > +/*
> > > + * Copyright © 2023 Intel Corporation
> > > + */
> > > +
> > > +#ifndef __DRM_DP_TUNNEL_H__
> > > +#define __DRM_DP_TUNNEL_H__
> > > +
> > > +#include <linux/err.h>
> > > +#include <linux/errno.h>
> > > +#include <linux/types.h>
> > > +
> > > +struct drm_dp_aux;
> > > +
> > > +struct drm_device;
> > > +
> > > +struct drm_atomic_state;
> > > +struct drm_dp_tunnel_mgr;
> > > +struct drm_dp_tunnel_state;
> > > +
> > > +struct ref_tracker;
> > > +
> > > +struct drm_dp_tunnel_ref {
> > > +	struct drm_dp_tunnel *tunnel;
> > > +#ifdef CONFIG_DRM_DISPLAY_DEBUG_DP_TUNNEL_STATE
> > > +	struct ref_tracker *tracker;
> > > +#endif
> > > +};
> > > +
> > > +#ifdef CONFIG_DRM_DISPLAY_DP_TUNNEL
> > > +
> > > +struct drm_dp_tunnel *
> > > +drm_dp_tunnel_get_untracked(struct drm_dp_tunnel *tunnel);
> > > +void drm_dp_tunnel_put_untracked(struct drm_dp_tunnel *tunnel);
> > > +
> > > +#ifdef CONFIG_DRM_DISPLAY_DEBUG_DP_TUNNEL_STATE
> > > +struct drm_dp_tunnel *
> > > +drm_dp_tunnel_get(struct drm_dp_tunnel *tunnel, struct ref_tracker **tracker);
> > > +
> > > +void
> > > +drm_dp_tunnel_put(struct drm_dp_tunnel *tunnel, struct ref_tracker **tracker);
> > > +#else
> > > +#define drm_dp_tunnel_get(tunnel, tracker) \
> > > +	drm_dp_tunnel_get_untracked(tunnel)
> > > +
> > > +#define drm_dp_tunnel_put(tunnel, tracker) \
> > > +	drm_dp_tunnel_put_untracked(tunnel)
> > > +
> > > +#endif
> > > +
> > > +static inline void drm_dp_tunnel_ref_get(struct drm_dp_tunnel *tunnel,
> > > +					   struct drm_dp_tunnel_ref *tunnel_ref)
> > > +{
> > > +	tunnel_ref->tunnel = drm_dp_tunnel_get(tunnel, &tunnel_ref->tracker);
> > > +}
> > > +
> > > +static inline void drm_dp_tunnel_ref_put(struct drm_dp_tunnel_ref *tunnel_ref)
> > > +{
> > > +	drm_dp_tunnel_put(tunnel_ref->tunnel, &tunnel_ref->tracker);
> > > +}
> > > +
> > > +struct drm_dp_tunnel *
> > > +drm_dp_tunnel_detect(struct drm_dp_tunnel_mgr *mgr,
> > > +		     struct drm_dp_aux *aux);
> > > +int drm_dp_tunnel_destroy(struct drm_dp_tunnel *tunnel);
> > > +
> > > +int drm_dp_tunnel_enable_bw_alloc(struct drm_dp_tunnel *tunnel);
> > > +int drm_dp_tunnel_disable_bw_alloc(struct drm_dp_tunnel *tunnel);
> > > +bool drm_dp_tunnel_bw_alloc_is_enabled(const struct drm_dp_tunnel *tunnel);
> > > +int drm_dp_tunnel_alloc_bw(struct drm_dp_tunnel *tunnel, int bw);
> > > +int drm_dp_tunnel_check_state(struct drm_dp_tunnel *tunnel);
> > > +int drm_dp_tunnel_update_state(struct drm_dp_tunnel *tunnel);
> > > +
> > > +void drm_dp_tunnel_set_io_error(struct drm_dp_tunnel *tunnel);
> > > +
> > > +int drm_dp_tunnel_handle_irq(struct drm_dp_tunnel_mgr *mgr,
> > > +			     struct drm_dp_aux *aux);
> > > +
> > > +int drm_dp_tunnel_max_dprx_rate(const struct drm_dp_tunnel *tunnel);
> > > +int drm_dp_tunnel_max_dprx_lane_count(const struct drm_dp_tunnel *tunnel);
> > > +int drm_dp_tunnel_available_bw(const struct drm_dp_tunnel *tunnel);
> > > +
> > > +const char *drm_dp_tunnel_name(const struct drm_dp_tunnel *tunnel);
> > > +
> > > +struct drm_dp_tunnel_state *
> > > +drm_dp_tunnel_atomic_get_state(struct drm_atomic_state *state,
> > > +			       struct drm_dp_tunnel *tunnel);
> > > +struct drm_dp_tunnel_state *
> > > +drm_dp_tunnel_atomic_get_new_state(struct drm_atomic_state *state,
> > > +				   const struct drm_dp_tunnel *tunnel);
> > > +
> > > +void drm_dp_tunnel_atomic_clear_state(struct drm_dp_tunnel_state *tunnel_state);
> > > +
> > > +int drm_dp_tunnel_atomic_set_stream_bw(struct drm_atomic_state *state,
> > > +				       struct drm_dp_tunnel *tunnel,
> > > +				       u8 stream_id, int bw);
> > > +int drm_dp_tunnel_atomic_get_group_streams_in_state(struct drm_atomic_state *state,
> > > +						    const struct drm_dp_tunnel *tunnel,
> > > +						    u32 *stream_mask);
> > > +
> > > +int drm_dp_tunnel_atomic_check_stream_bws(struct drm_atomic_state *state,
> > > +					  u32 *failed_stream_mask);
> > > +
> > > +int drm_dp_tunnel_atomic_get_tunnel_bw(const struct drm_dp_tunnel_state *tunnel_state);
> > > +
> > > +struct drm_dp_tunnel_mgr *
> > > +drm_dp_tunnel_mgr_create(struct drm_device *dev, int max_group_count);
> > > +void drm_dp_tunnel_mgr_destroy(struct drm_dp_tunnel_mgr *mgr);
> > > +
> > > +#else
> > > +
> > > +static inline struct drm_dp_tunnel *
> > > +drm_dp_tunnel_get_untracked(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return NULL;
> > > +}
> > > +
> > > +static inline void
> > > +drm_dp_tunnel_put_untracked(struct drm_dp_tunnel *tunnel) {}
> > > +
> > > +static inline struct drm_dp_tunnel *
> > > +drm_dp_tunnel_get(struct drm_dp_tunnel *tunnel, struct ref_tracker **tracker)
> > > +{
> > > +	return NULL;
> > > +}
> > > +
> > > +static inline void
> > > +drm_dp_tunnel_put(struct drm_dp_tunnel *tunnel, struct ref_tracker **tracker) {}
> > > +
> > > +static inline void drm_dp_tunnel_ref_get(struct drm_dp_tunnel *tunnel,
> > > +					   struct drm_dp_tunnel_ref *tunnel_ref) {}
> > > +static inline void drm_dp_tunnel_ref_put(struct drm_dp_tunnel_ref *tunnel_ref) {}
> > > +
> > > +static inline struct drm_dp_tunnel *
> > > +drm_dp_tunnel_detect(struct drm_dp_tunnel_mgr *mgr,
> > > +		     struct drm_dp_aux *aux)
> > > +{
> > > +	return ERR_PTR(-EOPNOTSUPP);
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_destroy(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return 0;
> > > +}
> > > +
> > > +static inline int drm_dp_tunnel_enable_bw_alloc(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int drm_dp_tunnel_disable_bw_alloc(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline bool drm_dp_tunnel_bw_alloc_is_enabled(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return false;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_alloc_bw(struct drm_dp_tunnel *tunnel, int bw)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_check_state(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_update_state(struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline void drm_dp_tunnel_set_io_error(struct drm_dp_tunnel *tunnel) {}
> > > +static inline int
> > > +drm_dp_tunnel_handle_irq(struct drm_dp_tunnel_mgr *mgr,
> > > +			 struct drm_dp_aux *aux)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_max_dprx_rate(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return 0;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_max_dprx_lane_count(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return 0;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_available_bw(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return -1;
> > > +}
> > > +
> > > +static inline const char *
> > > +drm_dp_tunnel_name(const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return NULL;
> > > +}
> > > +
> > > +static inline struct drm_dp_tunnel_state *
> > > +drm_dp_tunnel_atomic_get_state(struct drm_atomic_state *state,
> > > +			       struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return ERR_PTR(-EOPNOTSUPP);
> > > +}
> > > +
> > > +static inline struct drm_dp_tunnel_state *
> > > +drm_dp_tunnel_atomic_get_new_state(struct drm_atomic_state *state,
> > > +				   const struct drm_dp_tunnel *tunnel)
> > > +{
> > > +	return ERR_PTR(-EOPNOTSUPP);
> > > +}
> > > +
> > > +static inline void
> > > +drm_dp_tunnel_atomic_clear_state(struct drm_dp_tunnel_state *tunnel_state) {}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_atomic_set_stream_bw(struct drm_atomic_state *state,
> > > +				   struct drm_dp_tunnel *tunnel,
> > > +				   u8 stream_id, int bw)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_atomic_get_group_streams_in_state(struct drm_atomic_state *state,
> > > +						const struct drm_dp_tunnel *tunnel,
> > > +						u32 *stream_mask)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_atomic_check_stream_bws(struct drm_atomic_state *state,
> > > +				      u32 *failed_stream_mask)
> > > +{
> > > +	return -EOPNOTSUPP;
> > > +}
> > > +
> > > +static inline int
> > > +drm_dp_tunnel_atomic_get_tunnel_bw(const struct drm_dp_tunnel_state *tunnel_state)
> > > +{
> > > +	return 0;
> > > +}
> > > +
> > > +static inline struct drm_dp_tunnel_mgr *
> > > +drm_dp_tunnel_mgr_create(struct drm_device *dev, int max_group_count)
> > > +{
> > > +	return ERR_PTR(-EOPNOTSUPP);
> > > +}
> > > +
> > > +static inline
> > > +void drm_dp_tunnel_mgr_destroy(struct drm_dp_tunnel_mgr *mgr) {}
> > > +
> > > +
> > > +#endif /* CONFIG_DRM_DISPLAY_DP_TUNNEL */
> > > +
> > > +#endif /* __DRM_DP_TUNNEL_H__ */
> > > -- 
> > > 2.39.2
> > 
> > -- 
> > Ville Syrjälä
> > Intel



[Index of Archives]     [AMD Graphics]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux