lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20160803001115.GC32025@intel.com>
Date:	Tue, 2 Aug 2016 17:11:15 -0700
From:	Matt Roper <matthew.d.roper@...el.com>
To:	Lyude <cpaul@...hat.com>
Cc:	intel-gfx@...ts.freedesktop.org,
	Ville Syrjälä 
	<ville.syrjala@...ux.intel.com>,
	Maarten Lankhorst <maarten.lankhorst@...ux.intel.com>,
	Daniel Vetter <daniel.vetter@...el.com>,
	Radhakrishna Sripada <radhakrishna.sripada@...el.com>,
	Hans de Goede <hdegoede@...hat.com>,
	Jani Nikula <jani.nikula@...ux.intel.com>,
	David Airlie <airlied@...ux.ie>,
	dri-devel@...ts.freedesktop.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH v5 6/6] drm/i915/skl: Update DDB values atomically with
 wms/plane attrs

On Tue, Aug 02, 2016 at 02:52:54PM -0400, Lyude wrote:
> Now that we can hook into update_crtcs and control the order in which we
> update CRTCs at each modeset, we can finish the final step of fixing
> Skylake's watermark handling by performing DDB updates at the same time
> as plane updates and watermark updates.
> 
> The first major change in this patch is skl_update_crtcs(), which
> handles ensuring that we order each CRTC update in our atomic commits
> properly so that they honor the DDB flush order.
> 
> The second major change in this patch is the order in which we flush the
> pipes. While the previous order may have worked, it can't be used in
> this approach since it no longer will do the right thing. For example,
> using the old ddb flush order:
> 
> We have pipes A, B, and C enabled, and we're disabling C. Initial ddb
> allocation looks like this:
> 
> |   A   |   B   |xxxxxxx|
> 
> Since we're performing the ddb updates after performing any CRTC
> disablements in intel_atomic_commit_tail(), the space to the right of
> pipe B is unallocated.
> 
> 1. Flush pipes with new allocation contained into old space. None
>    apply, so we skip this
> 2. Flush pipes having their allocation reduced, but overlapping with a
>    previous allocation. None apply, so we also skip this
> 3. Flush pipes that got more space allocated. This applies to A and B,
>    giving us the following update order: A, B
> 
> This is wrong, since updating pipe A first will cause it to overlap with
> B and potentially burst into flames. Our new order (see the code
> comments for details) would update the pipes in the proper order: B, A.
> 
> As well, we calculate the order for each DDB update during the check
> phase, and reference it later in the commit phase when we hit
> skl_update_crtcs().
> 
> This long overdue patch fixes the rest of the underruns on Skylake.
> 
> Fixes: 0e8fb7ba7ca5 ("drm/i915/skl: Flush the WM configuration")
> Fixes: 8211bd5bdf5e ("drm/i915/skl: Program the DDB allocation")
> Signed-off-by: Lyude <cpaul@...hat.com>
> [omitting CC for stable, since this patch will need to be changed for
> such backports first]
> Cc: Ville Syrjälä <ville.syrjala@...ux.intel.com>
> Cc: Daniel Vetter <daniel.vetter@...el.com>
> Cc: Radhakrishna Sripada <radhakrishna.sripada@...el.com>
> Cc: Hans de Goede <hdegoede@...hat.com>
> Cc: Matt Roper <matthew.d.roper@...el.com>
> ---
>  drivers/gpu/drm/i915/intel_display.c | 100 ++++++++++--
>  drivers/gpu/drm/i915/intel_drv.h     |  10 ++
>  drivers/gpu/drm/i915/intel_pm.c      | 287 ++++++++++++++++-------------------
>  3 files changed, 231 insertions(+), 166 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
> index d985b5b..db271d6 100644
> --- a/drivers/gpu/drm/i915/intel_display.c
> +++ b/drivers/gpu/drm/i915/intel_display.c
> @@ -12909,16 +12909,23 @@ static void verify_wm_state(struct drm_crtc *crtc,
>  			  hw_entry->start, hw_entry->end);
>  	}
>  
> -	/* cursor */
> -	hw_entry = &hw_ddb.plane[pipe][PLANE_CURSOR];
> -	sw_entry = &sw_ddb->plane[pipe][PLANE_CURSOR];
> -
> -	if (!skl_ddb_entry_equal(hw_entry, sw_entry)) {
> -		DRM_ERROR("mismatch in DDB state pipe %c cursor "
> -			  "(expected (%u,%u), found (%u,%u))\n",
> -			  pipe_name(pipe),
> -			  sw_entry->start, sw_entry->end,
> -			  hw_entry->start, hw_entry->end);
> +	/*
> +	 * cursor
> +	 * If the cursor plane isn't active, we may not have updated it's ddb
> +	 * allocation. In that case since the ddb allocation will be updated
> +	 * once the plane becomes visible, we can skip this check
> +	 */
> +	if (intel_crtc->cursor_addr) {
> +		hw_entry = &hw_ddb.plane[pipe][PLANE_CURSOR];
> +		sw_entry = &sw_ddb->plane[pipe][PLANE_CURSOR];
> +
> +		if (!skl_ddb_entry_equal(hw_entry, sw_entry)) {
> +			DRM_ERROR("mismatch in DDB state pipe %c cursor "
> +				  "(expected (%u,%u), found (%u,%u))\n",
> +				  pipe_name(pipe),
> +				  sw_entry->start, sw_entry->end,
> +				  hw_entry->start, hw_entry->end);
> +		}
>  	}
>  }
>  
> @@ -13670,6 +13677,72 @@ static void intel_update_crtcs(struct drm_atomic_state *state,
>  	}
>  }
>  
> +static inline void
> +skl_do_ddb_step(struct drm_atomic_state *state,
> +		enum skl_ddb_step step)
> +{
> +	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> +	struct drm_crtc *crtc;
> +	struct drm_crtc_state *old_crtc_state;
> +	unsigned int crtc_vblank_mask; /* unused */
> +	int i;
> +
> +	for_each_crtc_in_state(state, crtc, old_crtc_state, i) {
> +		struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
> +		struct intel_crtc_state *cstate =
> +			to_intel_crtc_state(crtc->state);
> +		bool vblank_wait = false;
> +
> +		if (cstate->wm.skl.ddb_realloc != step || !crtc->state->active)
> +			continue;
> +
> +		/*
> +		 * If we're changing the ddb allocation of this pipe to make
> +		 * room for another pipe, we have to wait for the pipe's ddb
> +		 * allocations to actually update by waiting for a vblank.
> +		 * Otherwise we risk the next pipe updating before this pipe
> +		 * finishes, resulting in the pipe fetching from ddb space for
> +		 * the wrong pipe.
> +		 *
> +		 * However, if we know we don't have any more pipes to move
> +		 * around, we can skip this wait and the new ddb allocation
> +		 * will take effect at the start of the next vblank.
> +		 */
> +		switch (step) {
> +		case SKL_DDB_STEP_NO_OVERLAP:
> +		case SKL_DDB_STEP_OVERLAP:
> +			if (step != intel_state->last_ddb_step)
> +				vblank_wait = true;
> +
> +		/* drop through */
> +		case SKL_DDB_STEP_FINAL:
> +			DRM_DEBUG_KMS(
> +			    "Updating [CRTC:%d:pipe %c] for DDB step %d\n",
> +			    crtc->base.id, pipe_name(intel_crtc->pipe),
> +			    step);
> +
> +		case SKL_DDB_STEP_NONE:
> +			break;
> +		}
> +
> +		intel_update_crtc(crtc, state, old_crtc_state,
> +				  &crtc_vblank_mask);
> +
> +		if (vblank_wait)
> +			intel_wait_for_vblank(state->dev, intel_crtc->pipe);

We only technically need to wait at the end of a step, not on each
individual CRTC within a step.  Since we only have three pipes total,
that usually winds up being the same thing, but the one case we can
optimize slightly is when we re-enable pipe B while A and C are already
on.

        AAAAAACCCCCC  (initial state)
        AAAA____CCCC  (step 1: reduce both A and C)
        AAAABBBBCCCC  (final step)

We just need to wait until both A and C have had a vblank (as
intel_atomic_wait_for_vblanks() does) rather than doing each
sequentially.


> +	}
> +}
> +
> +static void skl_update_crtcs(struct drm_atomic_state *state,
> +			     unsigned int *crtc_vblank_mask)
> +{
> +	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> +	enum skl_ddb_step step;
> +
> +	for (step = 0; step <= intel_state->last_ddb_step; step++)
> +		skl_do_ddb_step(state, step);
> +}
> +
>  static void intel_atomic_commit_tail(struct drm_atomic_state *state)
>  {
>  	struct drm_device *dev = state->dev;
> @@ -15230,8 +15303,6 @@ void intel_init_display_hooks(struct drm_i915_private *dev_priv)
>  		dev_priv->display.crtc_disable = i9xx_crtc_disable;
>  	}
>  
> -	dev_priv->display.update_crtcs = intel_update_crtcs;
> -
>  	/* Returns the core display clock speed */
>  	if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv))
>  		dev_priv->display.get_display_clock_speed =
> @@ -15321,6 +15392,11 @@ void intel_init_display_hooks(struct drm_i915_private *dev_priv)
>  			skl_modeset_calc_cdclk;
>  	}
>  
> +	if (dev_priv->info.gen >= 9)
> +		dev_priv->display.update_crtcs = skl_update_crtcs;
> +	else
> +		dev_priv->display.update_crtcs = intel_update_crtcs;
> +
>  	switch (INTEL_INFO(dev_priv)->gen) {
>  	case 2:
>  		dev_priv->display.queue_flip = intel_gen2_queue_flip;
> diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h
> index 1b444d3..cf5da83 100644
> --- a/drivers/gpu/drm/i915/intel_drv.h
> +++ b/drivers/gpu/drm/i915/intel_drv.h
> @@ -334,6 +334,7 @@ struct intel_atomic_state {
>  
>  	/* Gen9+ only */
>  	struct skl_wm_values wm_results;
> +	int last_ddb_step;
>  };
>  
>  struct intel_plane_state {
> @@ -437,6 +438,13 @@ struct skl_pipe_wm {
>  	uint32_t linetime;
>  };
>  
> +enum skl_ddb_step {
> +	SKL_DDB_STEP_NONE = 0,
> +	SKL_DDB_STEP_NO_OVERLAP,
> +	SKL_DDB_STEP_OVERLAP,
> +	SKL_DDB_STEP_FINAL
> +};
> +
>  struct intel_crtc_wm_state {
>  	union {
>  		struct {
> @@ -467,6 +475,8 @@ struct intel_crtc_wm_state {
>  			/* minimum block allocation */
>  			uint16_t minimum_blocks[I915_MAX_PLANES];
>  			uint16_t minimum_y_blocks[I915_MAX_PLANES];
> +
> +			enum skl_ddb_step ddb_realloc;
>  		} skl;
>  	};
>  
> diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c
> index 6b2452b..8596cb9 100644
> --- a/drivers/gpu/drm/i915/intel_pm.c
> +++ b/drivers/gpu/drm/i915/intel_pm.c
> @@ -3818,6 +3818,11 @@ void skl_write_plane_wm(struct intel_crtc *intel_crtc,
>  			   wm->plane[pipe][plane][level]);
>  	}
>  	I915_WRITE(PLANE_WM_TRANS(pipe, plane), wm->plane_trans[pipe][plane]);
> +
> +	skl_ddb_entry_write(dev_priv, PLANE_BUF_CFG(pipe, plane),
> +			    &wm->ddb.plane[pipe][plane]);
> +	skl_ddb_entry_write(dev_priv, PLANE_NV12_BUF_CFG(pipe, plane),
> +			    &wm->ddb.y_plane[pipe][plane]);
>  }
>  
>  void skl_write_cursor_wm(struct intel_crtc *intel_crtc,
> @@ -3839,168 +3844,46 @@ void skl_write_cursor_wm(struct intel_crtc *intel_crtc,
>  			    &wm->ddb.plane[pipe][PLANE_CURSOR]);
>  }
>  
> -static void skl_write_wm_values(struct drm_i915_private *dev_priv,
> -				const struct skl_wm_values *new)
> -{
> -	struct drm_device *dev = &dev_priv->drm;
> -	struct intel_crtc *crtc;
> -
> -	for_each_intel_crtc(dev, crtc) {
> -		int i;
> -		enum pipe pipe = crtc->pipe;
> -
> -		if ((new->dirty_pipes & drm_crtc_mask(&crtc->base)) == 0)
> -			continue;
> -		if (!crtc->active)
> -			continue;
> -
> -		for (i = 0; i < intel_num_planes(crtc); i++) {
> -			skl_ddb_entry_write(dev_priv,
> -					    PLANE_BUF_CFG(pipe, i),
> -					    &new->ddb.plane[pipe][i]);
> -			skl_ddb_entry_write(dev_priv,
> -					    PLANE_NV12_BUF_CFG(pipe, i),
> -					    &new->ddb.y_plane[pipe][i]);
> -		}
> -
> -		skl_ddb_entry_write(dev_priv, CUR_BUF_CFG(pipe),
> -				    &new->ddb.plane[pipe][PLANE_CURSOR]);
> -	}
> -}
> -
> -/*
> - * When setting up a new DDB allocation arrangement, we need to correctly
> - * sequence the times at which the new allocations for the pipes are taken into
> - * account or we'll have pipes fetching from space previously allocated to
> - * another pipe.
> - *
> - * Roughly the sequence looks like:
> - *  1. re-allocate the pipe(s) with the allocation being reduced and not
> - *     overlapping with a previous light-up pipe (another way to put it is:
> - *     pipes with their new allocation strickly included into their old ones).
> - *  2. re-allocate the other pipes that get their allocation reduced
> - *  3. allocate the pipes having their allocation increased
> - *
> - * Steps 1. and 2. are here to take care of the following case:
> - * - Initially DDB looks like this:
> - *     |   B    |   C    |
> - * - enable pipe A.
> - * - pipe B has a reduced DDB allocation that overlaps with the old pipe C
> - *   allocation
> - *     |  A  |  B  |  C  |
> - *
> - * We need to sequence the re-allocation: C, B, A (and not B, C, A).
> - */
> -
> -static void
> -skl_wm_flush_pipe(struct drm_i915_private *dev_priv, enum pipe pipe, int pass)
> +static bool
> +skl_ddb_allocation_equals(const struct skl_ddb_allocation *old,
> +			  const struct skl_ddb_allocation *new,
> +			  enum pipe pipe)
>  {
> -	int plane;
> -
> -	DRM_DEBUG_KMS("flush pipe %c (pass %d)\n", pipe_name(pipe), pass);
> -
> -	for_each_plane(dev_priv, pipe, plane) {
> -		I915_WRITE(PLANE_SURF(pipe, plane),
> -			   I915_READ(PLANE_SURF(pipe, plane)));
> -	}
> -	I915_WRITE(CURBASE(pipe), I915_READ(CURBASE(pipe)));
> +	return new->pipe[pipe].start == old->pipe[pipe].start &&
> +	       new->pipe[pipe].end == old->pipe[pipe].end;
>  }
>  
>  static bool
> -skl_ddb_allocation_included(const struct skl_ddb_allocation *old,
> +skl_ddb_allocation_overlaps(struct drm_atomic_state *state,
> +			    const struct skl_ddb_allocation *old,
>  			    const struct skl_ddb_allocation *new,
>  			    enum pipe pipe)
>  {
> -	uint16_t old_size, new_size;
> -
> -	old_size = skl_ddb_entry_size(&old->pipe[pipe]);
> -	new_size = skl_ddb_entry_size(&new->pipe[pipe]);
> -
> -	return old_size != new_size &&
> -	       new->pipe[pipe].start >= old->pipe[pipe].start &&
> -	       new->pipe[pipe].end <= old->pipe[pipe].end;
> -}
> -
> -static void skl_flush_wm_values(struct drm_i915_private *dev_priv,
> -				struct skl_wm_values *new_values)
> -{
> -	struct drm_device *dev = &dev_priv->drm;
> -	struct skl_ddb_allocation *cur_ddb, *new_ddb;
> -	bool reallocated[I915_MAX_PIPES] = {};
> -	struct intel_crtc *crtc;
> -	enum pipe pipe;
> -
> -	new_ddb = &new_values->ddb;
> -	cur_ddb = &dev_priv->wm.skl_hw.ddb;
> -
> -	/*
> -	 * First pass: flush the pipes with the new allocation contained into
> -	 * the old space.
> -	 *
> -	 * We'll wait for the vblank on those pipes to ensure we can safely
> -	 * re-allocate the freed space without this pipe fetching from it.
> -	 */
> -	for_each_intel_crtc(dev, crtc) {
> -		if (!crtc->active)
> -			continue;
> -
> -		pipe = crtc->pipe;
> -
> -		if (!skl_ddb_allocation_included(cur_ddb, new_ddb, pipe))
> -			continue;
> -
> -		skl_wm_flush_pipe(dev_priv, pipe, 1);
> -		intel_wait_for_vblank(dev, pipe);
> -
> -		reallocated[pipe] = true;
> -	}
> -
> -
> -	/*
> -	 * Second pass: flush the pipes that are having their allocation
> -	 * reduced, but overlapping with a previous allocation.
> -	 *
> -	 * Here as well we need to wait for the vblank to make sure the freed
> -	 * space is not used anymore.
> -	 */
> -	for_each_intel_crtc(dev, crtc) {
> -		if (!crtc->active)
> -			continue;
> -
> -		pipe = crtc->pipe;
> -
> -		if (reallocated[pipe])
> -			continue;
> -
> -		if (skl_ddb_entry_size(&new_ddb->pipe[pipe]) <
> -		    skl_ddb_entry_size(&cur_ddb->pipe[pipe])) {
> -			skl_wm_flush_pipe(dev_priv, pipe, 2);
> -			intel_wait_for_vblank(dev, pipe);
> -			reallocated[pipe] = true;
> -		}
> -	}
> -
> -	/*
> -	 * Third pass: flush the pipes that got more space allocated.
> -	 *
> -	 * We don't need to actively wait for the update here, next vblank
> -	 * will just get more DDB space with the correct WM values.
> -	 */
> -	for_each_intel_crtc(dev, crtc) {
> -		if (!crtc->active)
> -			continue;
> +	struct drm_device *dev = state->dev;
> +	struct intel_crtc *intel_crtc;
> +	enum pipe otherp;
>  
> -		pipe = crtc->pipe;
> +	for_each_intel_crtc(dev, intel_crtc) {
> +		otherp = intel_crtc->pipe;
>  
>  		/*
> -		 * At this point, only the pipes more space than before are
> -		 * left to re-allocate.
> +		 * When checking for overlaps, we don't want to:
> +		 *  - Compare against ourselves
> +		 *  - Compare against pipes that will be disabled in step 0
> +		 *  - Compare against pipes that won't be enabled until step 3
>  		 */
> -		if (reallocated[pipe])
> +		if (otherp == pipe || !new->pipe[otherp].end ||
> +		    !old->pipe[otherp].end)
>  			continue;
>  
> -		skl_wm_flush_pipe(dev_priv, pipe, 3);
> +		if ((new->pipe[pipe].start >= old->pipe[otherp].start &&
> +		     new->pipe[pipe].start < old->pipe[otherp].end) ||
> +		    (old->pipe[otherp].start >= new->pipe[pipe].start &&
> +		     old->pipe[otherp].start < new->pipe[pipe].end))
> +			return true;
>  	}
> +
> +	return false;
>  }
>  
>  static int skl_update_pipe_wm(struct drm_crtc_state *cstate,
> @@ -4043,8 +3926,10 @@ skl_compute_ddb(struct drm_atomic_state *state)
>  	struct drm_device *dev = state->dev;
>  	struct drm_i915_private *dev_priv = to_i915(dev);
>  	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
> +	struct intel_crtc_state *cstate;
>  	struct intel_crtc *intel_crtc;
> -	struct skl_ddb_allocation *ddb = &intel_state->wm_results.ddb;
> +	struct skl_ddb_allocation *old_ddb = &dev_priv->wm.skl_hw.ddb;
> +	struct skl_ddb_allocation *new_ddb = &intel_state->wm_results.ddb;
>  	uint32_t realloc_pipes = pipes_modified(state);
>  	int ret;
>  
> @@ -4076,13 +3961,11 @@ skl_compute_ddb(struct drm_atomic_state *state)
>  	}
>  
>  	for_each_intel_crtc_mask(dev, intel_crtc, realloc_pipes) {
> -		struct intel_crtc_state *cstate;
> -
>  		cstate = intel_atomic_get_crtc_state(state, intel_crtc);
>  		if (IS_ERR(cstate))
>  			return PTR_ERR(cstate);
>  
> -		ret = skl_allocate_pipe_ddb(cstate, ddb);
> +		ret = skl_allocate_pipe_ddb(cstate, new_ddb);
>  		if (ret)
>  			return ret;
>  
> @@ -4091,6 +3974,73 @@ skl_compute_ddb(struct drm_atomic_state *state)
>  			return ret;
>  	}
>  
> +	/*
> +	 * When setting up a new DDB allocation arrangement, we need to
> +	 * correctly sequence the times at which the new allocations for the
> +	 * pipes are taken into account or we'll have pipes fetching from space
> +	 * previously allocated to another pipe.
> +	 *
> +	 * Roughly the final sequence we want looks like this:
> +	 *  1. Disable any pipes we're not going to be using anymore
> +	 *  2. Reallocate all of the active pipes whose new ddb allocations
> +	 *  won't overlap with another active pipe's ddb allocation.
> +	 *  3. Reallocate remaining active pipes, if any.
> +	 *  4. Enable any new pipes, if any.
> +	 *
> +	 * Example:
> +	 * Initially DDB looks like this:
> +	 *   |   B    |   C    |
> +	 * And the final DDB should look like this:
> +	 *   |  B  |  C  |  A  |
> +	 *
> +	 * 1. We're not disabling any pipes, so do nothing on this step.
> +	 * 2. Pipe B's new allocation wouldn't overlap with pipe C, however
> +	 * pipe C's new allocation does overlap with pipe B's current
> +	 * allocation. Reallocate B first so the DDB looks like this:
> +	 *   |  B  |xx|   C    |
> +	 * 3. Now we can safely reallocate pipe C to it's new location:
> +	 *   |  B  |  C  |xxxxx|
> +	 * 4. Enable any remaining pipes, in this case A
> +	 *   |  B  |  C  |  A  |

It's worth noting that this sequence will work fine for the three pipe
case we have in gen9 since when adding or removing a pipe we never have
a situation where multiple pipes are processed in the 'overlap' step
(step 3).  But if some future platform decides to add more pipes while
maintaining the same general DDB programming style, this will break down
since we'll need to ensure intelligent ordering of the pipes being
processed in step 3.  E.g., with naive ordering of step 3 we could have

        AAABBBCCCDDD -> AAAABBBBCCCC:
            AAABBBCCC___ (step 1)
            AAABBB__CCCC (step 2)
            AAA!BB__CCCC (step 3a, woops)

> +	 *
> +	 * As well, between every pipe reallocation we have to wait for a
> +	 * vblank on the pipe so that we ensure it's new allocation has taken
> +	 * effect by the time we start moving the next pipe. This can be
> +	 * skipped on the last step we need to perform, which is why we keep
> +	 * track of that information here. For example, if we've reallocated
> +	 * all the pipes that need changing by the time we reach step 3, we can
> +	 * finish without waiting for the pipes we changed in step 3 to update.
> +	 */
> +	for_each_intel_crtc_mask(dev, intel_crtc, realloc_pipes) {
> +		enum pipe pipe = intel_crtc->pipe;
> +		enum skl_ddb_step step;
> +
> +		cstate = intel_atomic_get_crtc_state(state, intel_crtc);
> +		if (IS_ERR(cstate))
> +			return PTR_ERR(cstate);
> +
> +		/* Step 1: Pipes we're disabling / haven't changed */
> +		if (skl_ddb_allocation_equals(old_ddb, new_ddb, pipe) ||
> +		    new_ddb->pipe[pipe].end == 0) {
> +			step = SKL_DDB_STEP_NONE;
> +		/* Step 2-3: Active pipes we're reallocating */
> +		} else if (old_ddb->pipe[pipe].end != 0) {
> +			if (skl_ddb_allocation_overlaps(state, old_ddb, new_ddb,
> +							pipe))
> +				step = SKL_DDB_STEP_OVERLAP;
> +			else
> +				step = SKL_DDB_STEP_NO_OVERLAP;
> +		/* Step 4: Pipes we're enabling */
> +		} else {
> +			step = SKL_DDB_STEP_FINAL;
> +		}
> +
> +		cstate->wm.skl.ddb_realloc = step;
> +
> +		if (step > intel_state->last_ddb_step)
> +			intel_state->last_ddb_step = step;
> +	}
> +
>  	return 0;
>  }
>  
> @@ -4115,10 +4065,13 @@ skl_copy_wm_for_pipe(struct skl_wm_values *dst,
>  static int
>  skl_compute_wm(struct drm_atomic_state *state)
>  {
> +	struct drm_i915_private *dev_priv = to_i915(state->dev);
>  	struct drm_crtc *crtc;
>  	struct drm_crtc_state *cstate;
>  	struct intel_atomic_state *intel_state = to_intel_atomic_state(state);
>  	struct skl_wm_values *results = &intel_state->wm_results;
> +	struct skl_ddb_allocation *old_ddb = &dev_priv->wm.skl_hw.ddb;
> +	struct skl_ddb_allocation *new_ddb = &results->ddb;
>  	struct skl_pipe_wm *pipe_wm;
>  	bool changed = false;
>  	int ret, i;
> @@ -4157,7 +4110,10 @@ skl_compute_wm(struct drm_atomic_state *state)
>  		struct intel_crtc *intel_crtc = to_intel_crtc(crtc);
>  		struct intel_crtc_state *intel_cstate =
>  			to_intel_crtc_state(cstate);
> +		enum skl_ddb_step step;
> +		enum pipe pipe;
>  
> +		pipe = intel_crtc->pipe;
>  		pipe_wm = &intel_cstate->wm.skl.optimal;
>  		ret = skl_update_pipe_wm(cstate, &results->ddb, pipe_wm,
>  					 &changed);
> @@ -4172,7 +4128,18 @@ skl_compute_wm(struct drm_atomic_state *state)
>  			continue;
>  
>  		intel_cstate->update_wm_pre = true;
> +		step = intel_cstate->wm.skl.ddb_realloc;
>  		skl_compute_wm_results(crtc->dev, pipe_wm, results, intel_crtc);
> +
> +		if (!skl_ddb_entry_equal(&old_ddb->pipe[pipe],
> +					 &new_ddb->pipe[pipe])) {
> +			DRM_DEBUG_KMS(
> +			    "DDB changes for [CRTC:%d:pipe %c]: (%3d - %3d) -> (%3d - %3d) on step %d\n",
> +			    intel_crtc->base.base.id, pipe_name(pipe),
> +			    old_ddb->pipe[pipe].start, old_ddb->pipe[pipe].end,
> +			    new_ddb->pipe[pipe].start, new_ddb->pipe[pipe].end,
> +			    step);
> +		}
>  	}
>  
>  	return 0;
> @@ -4196,8 +4163,20 @@ static void skl_update_wm(struct drm_crtc *crtc)
>  
>  	mutex_lock(&dev_priv->wm.wm_mutex);
>  
> -	skl_write_wm_values(dev_priv, results);
> -	skl_flush_wm_values(dev_priv, results);
> +	/*
> +	 * If this pipe isn't active already, we're going to be enabling it
> +	 * very soon. Since it's safe to update these while the pipe's shut off,
> +	 * just do so here. Already active pipes will have their watermarks
> +	 * updated once we update their planes.
> +	 */
> +	if (!intel_crtc->active) {
> +		int plane;
> +
> +		for (plane = 0; plane < intel_num_planes(intel_crtc); plane++)
> +			skl_write_plane_wm(intel_crtc, results, plane);
> +
> +		skl_write_cursor_wm(intel_crtc, results);
> +	}

Do we actually need this?  Even if the pipe is in the process of being
enabled, all of these registers should still be programmed at the same
time as the plane registers, right?  Those register writes are based on
crtc->state->active rather than crtc->active so they should happen when
they need to.

On the other hand, if we write these when the pipe is off and isn't in
the process of being enabled, then we may not have the power well turned
on, so we'll get unclaimed register warnings.


Matt

>  
>  	/*
>  	 * Store the new configuration (but only for the pipes that have
> -- 
> 2.7.4
> 

-- 
Matt Roper
Graphics Software Engineer
IoTG Platform Enabling & Development
Intel Corporation
(916) 356-2795

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ