Recompute the state of all CRTCs on an FDI link during a modeset that may be affected by the modeset of other CRTCs on the same link. This ensures that each CRTC on the link maximizes its BW use (after another CRTC is disabled). In practice this means recomputing pipe B's config on IVB if pipe C gets disabled. v2: - Add the change recomputing affected CRTC states in a separate patch. (Ville) v3: (Ville) - Constify old and new crtc states. - Check for fused off pipe C. - Fix new vs. old crtc state mixup. - Drop check for pipe C's enabled state. Cc: Ville Syrjälä <ville.syrjala@linux.intel.com> Reviewed-by: Ville Syrjälä <ville.syrjala@linux.intel.com> Signed-off-by: Imre Deak <imre.deak@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20230921195159.2646027-12-imre.deak@intel.com
1145 lines
33 KiB
C
1145 lines
33 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2020 Intel Corporation
|
|
*/
|
|
|
|
#include <linux/string_helpers.h>
|
|
|
|
#include "i915_reg.h"
|
|
#include "intel_atomic.h"
|
|
#include "intel_crtc.h"
|
|
#include "intel_ddi.h"
|
|
#include "intel_de.h"
|
|
#include "intel_display_types.h"
|
|
#include "intel_fdi.h"
|
|
#include "intel_fdi_regs.h"
|
|
#include "intel_link_bw.h"
|
|
|
|
struct intel_fdi_funcs {
|
|
void (*fdi_link_train)(struct intel_crtc *crtc,
|
|
const struct intel_crtc_state *crtc_state);
|
|
};
|
|
|
|
static void assert_fdi_tx(struct drm_i915_private *dev_priv,
|
|
enum pipe pipe, bool state)
|
|
{
|
|
bool cur_state;
|
|
|
|
if (HAS_DDI(dev_priv)) {
|
|
/*
|
|
* DDI does not have a specific FDI_TX register.
|
|
*
|
|
* FDI is never fed from EDP transcoder
|
|
* so pipe->transcoder cast is fine here.
|
|
*/
|
|
enum transcoder cpu_transcoder = (enum transcoder)pipe;
|
|
cur_state = intel_de_read(dev_priv, TRANS_DDI_FUNC_CTL(cpu_transcoder)) & TRANS_DDI_FUNC_ENABLE;
|
|
} else {
|
|
cur_state = intel_de_read(dev_priv, FDI_TX_CTL(pipe)) & FDI_TX_ENABLE;
|
|
}
|
|
I915_STATE_WARN(dev_priv, cur_state != state,
|
|
"FDI TX state assertion failure (expected %s, current %s)\n",
|
|
str_on_off(state), str_on_off(cur_state));
|
|
}
|
|
|
|
void assert_fdi_tx_enabled(struct drm_i915_private *i915, enum pipe pipe)
|
|
{
|
|
assert_fdi_tx(i915, pipe, true);
|
|
}
|
|
|
|
void assert_fdi_tx_disabled(struct drm_i915_private *i915, enum pipe pipe)
|
|
{
|
|
assert_fdi_tx(i915, pipe, false);
|
|
}
|
|
|
|
static void assert_fdi_rx(struct drm_i915_private *dev_priv,
|
|
enum pipe pipe, bool state)
|
|
{
|
|
bool cur_state;
|
|
|
|
cur_state = intel_de_read(dev_priv, FDI_RX_CTL(pipe)) & FDI_RX_ENABLE;
|
|
I915_STATE_WARN(dev_priv, cur_state != state,
|
|
"FDI RX state assertion failure (expected %s, current %s)\n",
|
|
str_on_off(state), str_on_off(cur_state));
|
|
}
|
|
|
|
void assert_fdi_rx_enabled(struct drm_i915_private *i915, enum pipe pipe)
|
|
{
|
|
assert_fdi_rx(i915, pipe, true);
|
|
}
|
|
|
|
void assert_fdi_rx_disabled(struct drm_i915_private *i915, enum pipe pipe)
|
|
{
|
|
assert_fdi_rx(i915, pipe, false);
|
|
}
|
|
|
|
void assert_fdi_tx_pll_enabled(struct drm_i915_private *i915,
|
|
enum pipe pipe)
|
|
{
|
|
bool cur_state;
|
|
|
|
/* ILK FDI PLL is always enabled */
|
|
if (IS_IRONLAKE(i915))
|
|
return;
|
|
|
|
/* On Haswell, DDI ports are responsible for the FDI PLL setup */
|
|
if (HAS_DDI(i915))
|
|
return;
|
|
|
|
cur_state = intel_de_read(i915, FDI_TX_CTL(pipe)) & FDI_TX_PLL_ENABLE;
|
|
I915_STATE_WARN(i915, !cur_state,
|
|
"FDI TX PLL assertion failure, should be active but is disabled\n");
|
|
}
|
|
|
|
static void assert_fdi_rx_pll(struct drm_i915_private *i915,
|
|
enum pipe pipe, bool state)
|
|
{
|
|
bool cur_state;
|
|
|
|
cur_state = intel_de_read(i915, FDI_RX_CTL(pipe)) & FDI_RX_PLL_ENABLE;
|
|
I915_STATE_WARN(i915, cur_state != state,
|
|
"FDI RX PLL assertion failure (expected %s, current %s)\n",
|
|
str_on_off(state), str_on_off(cur_state));
|
|
}
|
|
|
|
void assert_fdi_rx_pll_enabled(struct drm_i915_private *i915, enum pipe pipe)
|
|
{
|
|
assert_fdi_rx_pll(i915, pipe, true);
|
|
}
|
|
|
|
void assert_fdi_rx_pll_disabled(struct drm_i915_private *i915, enum pipe pipe)
|
|
{
|
|
assert_fdi_rx_pll(i915, pipe, false);
|
|
}
|
|
|
|
void intel_fdi_link_train(struct intel_crtc *crtc,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
|
|
|
dev_priv->display.funcs.fdi->fdi_link_train(crtc, crtc_state);
|
|
}
|
|
|
|
/**
|
|
* intel_fdi_add_affected_crtcs - add CRTCs on FDI affected by other modeset CRTCs
|
|
* @state: intel atomic state
|
|
*
|
|
* Add a CRTC using FDI to @state if changing another CRTC's FDI BW usage is
|
|
* known to affect the available FDI BW for the former CRTC. In practice this
|
|
* means adding CRTC B on IVYBRIDGE if its use of FDI lanes is limited (by
|
|
* CRTC C) and CRTC C is getting disabled.
|
|
*
|
|
* Returns 0 in case of success, or a negative error code otherwise.
|
|
*/
|
|
int intel_fdi_add_affected_crtcs(struct intel_atomic_state *state)
|
|
{
|
|
struct drm_i915_private *i915 = to_i915(state->base.dev);
|
|
const struct intel_crtc_state *old_crtc_state;
|
|
const struct intel_crtc_state *new_crtc_state;
|
|
struct intel_crtc *crtc;
|
|
|
|
if (!IS_IVYBRIDGE(i915) || INTEL_NUM_PIPES(i915) != 3)
|
|
return 0;
|
|
|
|
crtc = intel_crtc_for_pipe(i915, PIPE_C);
|
|
new_crtc_state = intel_atomic_get_new_crtc_state(state, crtc);
|
|
if (!new_crtc_state)
|
|
return 0;
|
|
|
|
if (!intel_crtc_needs_modeset(new_crtc_state))
|
|
return 0;
|
|
|
|
old_crtc_state = intel_atomic_get_old_crtc_state(state, crtc);
|
|
if (!old_crtc_state->fdi_lanes)
|
|
return 0;
|
|
|
|
crtc = intel_crtc_for_pipe(i915, PIPE_B);
|
|
new_crtc_state = intel_atomic_get_crtc_state(&state->base, crtc);
|
|
if (IS_ERR(new_crtc_state))
|
|
return PTR_ERR(new_crtc_state);
|
|
|
|
old_crtc_state = intel_atomic_get_old_crtc_state(state, crtc);
|
|
if (!old_crtc_state->fdi_lanes)
|
|
return 0;
|
|
|
|
return intel_modeset_pipes_in_mask_early(state,
|
|
"FDI link BW decrease on pipe C",
|
|
BIT(PIPE_B));
|
|
}
|
|
|
|
/* units of 100MHz */
|
|
static int pipe_required_fdi_lanes(struct intel_crtc_state *crtc_state)
|
|
{
|
|
if (crtc_state->hw.enable && crtc_state->has_pch_encoder)
|
|
return crtc_state->fdi_lanes;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ilk_check_fdi_lanes(struct drm_device *dev, enum pipe pipe,
|
|
struct intel_crtc_state *pipe_config,
|
|
enum pipe *pipe_to_reduce)
|
|
{
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
|
struct drm_atomic_state *state = pipe_config->uapi.state;
|
|
struct intel_crtc *other_crtc;
|
|
struct intel_crtc_state *other_crtc_state;
|
|
|
|
*pipe_to_reduce = pipe;
|
|
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"checking fdi config on pipe %c, lanes %i\n",
|
|
pipe_name(pipe), pipe_config->fdi_lanes);
|
|
if (pipe_config->fdi_lanes > 4) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"invalid fdi lane config on pipe %c: %i lanes\n",
|
|
pipe_name(pipe), pipe_config->fdi_lanes);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) {
|
|
if (pipe_config->fdi_lanes > 2) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"only 2 lanes on haswell, required: %i lanes\n",
|
|
pipe_config->fdi_lanes);
|
|
return -EINVAL;
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
if (INTEL_NUM_PIPES(dev_priv) == 2)
|
|
return 0;
|
|
|
|
/* Ivybridge 3 pipe is really complicated */
|
|
switch (pipe) {
|
|
case PIPE_A:
|
|
return 0;
|
|
case PIPE_B:
|
|
if (pipe_config->fdi_lanes <= 2)
|
|
return 0;
|
|
|
|
other_crtc = intel_crtc_for_pipe(dev_priv, PIPE_C);
|
|
other_crtc_state =
|
|
intel_atomic_get_crtc_state(state, other_crtc);
|
|
if (IS_ERR(other_crtc_state))
|
|
return PTR_ERR(other_crtc_state);
|
|
|
|
if (pipe_required_fdi_lanes(other_crtc_state) > 0) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"invalid shared fdi lane config on pipe %c: %i lanes\n",
|
|
pipe_name(pipe), pipe_config->fdi_lanes);
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
case PIPE_C:
|
|
if (pipe_config->fdi_lanes > 2) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"only 2 lanes on pipe %c: required %i lanes\n",
|
|
pipe_name(pipe), pipe_config->fdi_lanes);
|
|
return -EINVAL;
|
|
}
|
|
|
|
other_crtc = intel_crtc_for_pipe(dev_priv, PIPE_B);
|
|
other_crtc_state =
|
|
intel_atomic_get_crtc_state(state, other_crtc);
|
|
if (IS_ERR(other_crtc_state))
|
|
return PTR_ERR(other_crtc_state);
|
|
|
|
if (pipe_required_fdi_lanes(other_crtc_state) > 2) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"fdi link B uses too many lanes to enable link C\n");
|
|
|
|
*pipe_to_reduce = PIPE_B;
|
|
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
default:
|
|
MISSING_CASE(pipe);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
void intel_fdi_pll_freq_update(struct drm_i915_private *i915)
|
|
{
|
|
if (IS_IRONLAKE(i915)) {
|
|
u32 fdi_pll_clk =
|
|
intel_de_read(i915, FDI_PLL_BIOS_0) & FDI_PLL_FB_CLOCK_MASK;
|
|
|
|
i915->display.fdi.pll_freq = (fdi_pll_clk + 2) * 10000;
|
|
} else if (IS_SANDYBRIDGE(i915) || IS_IVYBRIDGE(i915)) {
|
|
i915->display.fdi.pll_freq = 270000;
|
|
} else {
|
|
return;
|
|
}
|
|
|
|
drm_dbg(&i915->drm, "FDI PLL freq=%d\n", i915->display.fdi.pll_freq);
|
|
}
|
|
|
|
int intel_fdi_link_freq(struct drm_i915_private *i915,
|
|
const struct intel_crtc_state *pipe_config)
|
|
{
|
|
if (HAS_DDI(i915))
|
|
return pipe_config->port_clock; /* SPLL */
|
|
else
|
|
return i915->display.fdi.pll_freq;
|
|
}
|
|
|
|
/**
|
|
* intel_fdi_compute_pipe_bpp - compute pipe bpp limited by max link bpp
|
|
* @crtc_state: the crtc state
|
|
*
|
|
* Compute the pipe bpp limited by the CRTC's maximum link bpp. Encoders can
|
|
* call this function during state computation in the simple case where the
|
|
* link bpp will always match the pipe bpp. This is the case for all non-DP
|
|
* encoders, while DP encoders will use a link bpp lower than pipe bpp in case
|
|
* of DSC compression.
|
|
*
|
|
* Returns %true in case of success, %false if pipe bpp would need to be
|
|
* reduced below its valid range.
|
|
*/
|
|
bool intel_fdi_compute_pipe_bpp(struct intel_crtc_state *crtc_state)
|
|
{
|
|
int pipe_bpp = min(crtc_state->pipe_bpp,
|
|
to_bpp_int(crtc_state->max_link_bpp_x16));
|
|
|
|
pipe_bpp = rounddown(pipe_bpp, 2 * 3);
|
|
|
|
if (pipe_bpp < 6 * 3)
|
|
return false;
|
|
|
|
crtc_state->pipe_bpp = pipe_bpp;
|
|
|
|
return true;
|
|
}
|
|
|
|
int ilk_fdi_compute_config(struct intel_crtc *crtc,
|
|
struct intel_crtc_state *pipe_config)
|
|
{
|
|
struct drm_device *dev = crtc->base.dev;
|
|
struct drm_i915_private *i915 = to_i915(dev);
|
|
const struct drm_display_mode *adjusted_mode = &pipe_config->hw.adjusted_mode;
|
|
int lane, link_bw, fdi_dotclock;
|
|
|
|
/* FDI is a binary signal running at ~2.7GHz, encoding
|
|
* each output octet as 10 bits. The actual frequency
|
|
* is stored as a divider into a 100MHz clock, and the
|
|
* mode pixel clock is stored in units of 1KHz.
|
|
* Hence the bw of each lane in terms of the mode signal
|
|
* is:
|
|
*/
|
|
link_bw = intel_fdi_link_freq(i915, pipe_config);
|
|
|
|
fdi_dotclock = adjusted_mode->crtc_clock;
|
|
|
|
lane = ilk_get_lanes_required(fdi_dotclock, link_bw,
|
|
pipe_config->pipe_bpp);
|
|
|
|
pipe_config->fdi_lanes = lane;
|
|
|
|
intel_link_compute_m_n(pipe_config->pipe_bpp, lane, fdi_dotclock,
|
|
link_bw, &pipe_config->fdi_m_n, false);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int intel_fdi_atomic_check_bw(struct intel_atomic_state *state,
|
|
struct intel_crtc *crtc,
|
|
struct intel_crtc_state *pipe_config,
|
|
struct intel_link_bw_limits *limits)
|
|
{
|
|
struct drm_i915_private *i915 = to_i915(crtc->base.dev);
|
|
enum pipe pipe_to_reduce;
|
|
int ret;
|
|
|
|
ret = ilk_check_fdi_lanes(&i915->drm, crtc->pipe, pipe_config,
|
|
&pipe_to_reduce);
|
|
if (ret != -EINVAL)
|
|
return ret;
|
|
|
|
ret = intel_link_bw_reduce_bpp(state, limits,
|
|
BIT(pipe_to_reduce),
|
|
"FDI link BW");
|
|
|
|
return ret ? : -EAGAIN;
|
|
}
|
|
|
|
/**
|
|
* intel_fdi_atomic_check_link - check all modeset FDI link configuration
|
|
* @state: intel atomic state
|
|
* @limits: link BW limits
|
|
*
|
|
* Check the link configuration for all modeset FDI outputs. If the
|
|
* configuration is invalid @limits will be updated if possible to
|
|
* reduce the total BW, after which the configuration for all CRTCs in
|
|
* @state must be recomputed with the updated @limits.
|
|
*
|
|
* Returns:
|
|
* - 0 if the confugration is valid
|
|
* - %-EAGAIN, if the configuration is invalid and @limits got updated
|
|
* with fallback values with which the configuration of all CRTCs
|
|
* in @state must be recomputed
|
|
* - Other negative error, if the configuration is invalid without a
|
|
* fallback possibility, or the check failed for another reason
|
|
*/
|
|
int intel_fdi_atomic_check_link(struct intel_atomic_state *state,
|
|
struct intel_link_bw_limits *limits)
|
|
{
|
|
struct intel_crtc *crtc;
|
|
struct intel_crtc_state *crtc_state;
|
|
int i;
|
|
|
|
for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) {
|
|
int ret;
|
|
|
|
if (!crtc_state->has_pch_encoder ||
|
|
!intel_crtc_needs_modeset(crtc_state) ||
|
|
!crtc_state->hw.enable)
|
|
continue;
|
|
|
|
ret = intel_fdi_atomic_check_bw(state, crtc, crtc_state, limits);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool enable)
|
|
{
|
|
u32 temp;
|
|
|
|
temp = intel_de_read(dev_priv, SOUTH_CHICKEN1);
|
|
if (!!(temp & FDI_BC_BIFURCATION_SELECT) == enable)
|
|
return;
|
|
|
|
drm_WARN_ON(&dev_priv->drm,
|
|
intel_de_read(dev_priv, FDI_RX_CTL(PIPE_B)) &
|
|
FDI_RX_ENABLE);
|
|
drm_WARN_ON(&dev_priv->drm,
|
|
intel_de_read(dev_priv, FDI_RX_CTL(PIPE_C)) &
|
|
FDI_RX_ENABLE);
|
|
|
|
temp &= ~FDI_BC_BIFURCATION_SELECT;
|
|
if (enable)
|
|
temp |= FDI_BC_BIFURCATION_SELECT;
|
|
|
|
drm_dbg_kms(&dev_priv->drm, "%sabling fdi C rx\n",
|
|
enable ? "en" : "dis");
|
|
intel_de_write(dev_priv, SOUTH_CHICKEN1, temp);
|
|
intel_de_posting_read(dev_priv, SOUTH_CHICKEN1);
|
|
}
|
|
|
|
static void ivb_update_fdi_bc_bifurcation(const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc);
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
|
|
|
switch (crtc->pipe) {
|
|
case PIPE_A:
|
|
break;
|
|
case PIPE_B:
|
|
if (crtc_state->fdi_lanes > 2)
|
|
cpt_set_fdi_bc_bifurcation(dev_priv, false);
|
|
else
|
|
cpt_set_fdi_bc_bifurcation(dev_priv, true);
|
|
|
|
break;
|
|
case PIPE_C:
|
|
cpt_set_fdi_bc_bifurcation(dev_priv, true);
|
|
|
|
break;
|
|
default:
|
|
MISSING_CASE(crtc->pipe);
|
|
}
|
|
}
|
|
|
|
void intel_fdi_normal_train(struct intel_crtc *crtc)
|
|
{
|
|
struct drm_device *dev = crtc->base.dev;
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
i915_reg_t reg;
|
|
u32 temp;
|
|
|
|
/* enable normal train */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
if (IS_IVYBRIDGE(dev_priv)) {
|
|
temp &= ~FDI_LINK_TRAIN_NONE_IVB;
|
|
temp |= FDI_LINK_TRAIN_NONE_IVB | FDI_TX_ENHANCE_FRAME_ENABLE;
|
|
} else {
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_NONE | FDI_TX_ENHANCE_FRAME_ENABLE;
|
|
}
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
if (HAS_PCH_CPT(dev_priv)) {
|
|
temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
|
|
temp |= FDI_LINK_TRAIN_NORMAL_CPT;
|
|
} else {
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_NONE;
|
|
}
|
|
intel_de_write(dev_priv, reg, temp | FDI_RX_ENHANCE_FRAME_ENABLE);
|
|
|
|
/* wait one idle pattern time */
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(1000);
|
|
|
|
/* IVB wants error correction enabled */
|
|
if (IS_IVYBRIDGE(dev_priv))
|
|
intel_de_rmw(dev_priv, reg, 0, FDI_FS_ERRC_ENABLE | FDI_FE_ERRC_ENABLE);
|
|
}
|
|
|
|
/* The FDI link training functions for ILK/Ibexpeak. */
|
|
static void ilk_fdi_link_train(struct intel_crtc *crtc,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_device *dev = crtc->base.dev;
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
i915_reg_t reg;
|
|
u32 temp, tries;
|
|
|
|
/*
|
|
* Write the TU size bits before fdi link training, so that error
|
|
* detection works.
|
|
*/
|
|
intel_de_write(dev_priv, FDI_RX_TUSIZE1(pipe),
|
|
intel_de_read(dev_priv, PIPE_DATA_M1(pipe)) & TU_SIZE_MASK);
|
|
|
|
/* FDI needs bits from pipe first */
|
|
assert_transcoder_enabled(dev_priv, crtc_state->cpu_transcoder);
|
|
|
|
/* Train 1: umask FDI RX Interrupt symbol_lock and bit_lock bit
|
|
for train result */
|
|
reg = FDI_RX_IMR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_RX_SYMBOL_LOCK;
|
|
temp &= ~FDI_RX_BIT_LOCK;
|
|
intel_de_write(dev_priv, reg, temp);
|
|
intel_de_read(dev_priv, reg);
|
|
udelay(150);
|
|
|
|
/* enable CPU FDI TX and PCH FDI RX */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_DP_PORT_WIDTH_MASK;
|
|
temp |= FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1;
|
|
intel_de_write(dev_priv, reg, temp | FDI_TX_ENABLE);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1;
|
|
intel_de_write(dev_priv, reg, temp | FDI_RX_ENABLE);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(150);
|
|
|
|
/* Ironlake workaround, enable clock pointer after FDI enable*/
|
|
intel_de_write(dev_priv, FDI_RX_CHICKEN(pipe),
|
|
FDI_RX_PHASE_SYNC_POINTER_OVR);
|
|
intel_de_write(dev_priv, FDI_RX_CHICKEN(pipe),
|
|
FDI_RX_PHASE_SYNC_POINTER_OVR | FDI_RX_PHASE_SYNC_POINTER_EN);
|
|
|
|
reg = FDI_RX_IIR(pipe);
|
|
for (tries = 0; tries < 5; tries++) {
|
|
temp = intel_de_read(dev_priv, reg);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR 0x%x\n", temp);
|
|
|
|
if ((temp & FDI_RX_BIT_LOCK)) {
|
|
drm_dbg_kms(&dev_priv->drm, "FDI train 1 done.\n");
|
|
intel_de_write(dev_priv, reg, temp | FDI_RX_BIT_LOCK);
|
|
break;
|
|
}
|
|
}
|
|
if (tries == 5)
|
|
drm_err(&dev_priv->drm, "FDI train 1 fail!\n");
|
|
|
|
/* Train 2 */
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe),
|
|
FDI_LINK_TRAIN_NONE, FDI_LINK_TRAIN_PATTERN_2);
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(pipe),
|
|
FDI_LINK_TRAIN_NONE, FDI_LINK_TRAIN_PATTERN_2);
|
|
intel_de_posting_read(dev_priv, FDI_RX_CTL(pipe));
|
|
udelay(150);
|
|
|
|
reg = FDI_RX_IIR(pipe);
|
|
for (tries = 0; tries < 5; tries++) {
|
|
temp = intel_de_read(dev_priv, reg);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR 0x%x\n", temp);
|
|
|
|
if (temp & FDI_RX_SYMBOL_LOCK) {
|
|
intel_de_write(dev_priv, reg,
|
|
temp | FDI_RX_SYMBOL_LOCK);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI train 2 done.\n");
|
|
break;
|
|
}
|
|
}
|
|
if (tries == 5)
|
|
drm_err(&dev_priv->drm, "FDI train 2 fail!\n");
|
|
|
|
drm_dbg_kms(&dev_priv->drm, "FDI train done\n");
|
|
|
|
}
|
|
|
|
static const int snb_b_fdi_train_param[] = {
|
|
FDI_LINK_TRAIN_400MV_0DB_SNB_B,
|
|
FDI_LINK_TRAIN_400MV_6DB_SNB_B,
|
|
FDI_LINK_TRAIN_600MV_3_5DB_SNB_B,
|
|
FDI_LINK_TRAIN_800MV_0DB_SNB_B,
|
|
};
|
|
|
|
/* The FDI link training functions for SNB/Cougarpoint. */
|
|
static void gen6_fdi_link_train(struct intel_crtc *crtc,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_device *dev = crtc->base.dev;
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
i915_reg_t reg;
|
|
u32 temp, i, retry;
|
|
|
|
/*
|
|
* Write the TU size bits before fdi link training, so that error
|
|
* detection works.
|
|
*/
|
|
intel_de_write(dev_priv, FDI_RX_TUSIZE1(pipe),
|
|
intel_de_read(dev_priv, PIPE_DATA_M1(pipe)) & TU_SIZE_MASK);
|
|
|
|
/* Train 1: umask FDI RX Interrupt symbol_lock and bit_lock bit
|
|
for train result */
|
|
reg = FDI_RX_IMR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_RX_SYMBOL_LOCK;
|
|
temp &= ~FDI_RX_BIT_LOCK;
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(150);
|
|
|
|
/* enable CPU FDI TX and PCH FDI RX */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_DP_PORT_WIDTH_MASK;
|
|
temp |= FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1;
|
|
temp &= ~FDI_LINK_TRAIN_VOL_EMP_MASK;
|
|
/* SNB-B */
|
|
temp |= FDI_LINK_TRAIN_400MV_0DB_SNB_B;
|
|
intel_de_write(dev_priv, reg, temp | FDI_TX_ENABLE);
|
|
|
|
intel_de_write(dev_priv, FDI_RX_MISC(pipe),
|
|
FDI_RX_TP1_TO_TP2_48 | FDI_RX_FDI_DELAY_90);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
if (HAS_PCH_CPT(dev_priv)) {
|
|
temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
|
|
} else {
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1;
|
|
}
|
|
intel_de_write(dev_priv, reg, temp | FDI_RX_ENABLE);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(150);
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe),
|
|
FDI_LINK_TRAIN_VOL_EMP_MASK, snb_b_fdi_train_param[i]);
|
|
intel_de_posting_read(dev_priv, FDI_TX_CTL(pipe));
|
|
udelay(500);
|
|
|
|
for (retry = 0; retry < 5; retry++) {
|
|
reg = FDI_RX_IIR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR 0x%x\n", temp);
|
|
if (temp & FDI_RX_BIT_LOCK) {
|
|
intel_de_write(dev_priv, reg,
|
|
temp | FDI_RX_BIT_LOCK);
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI train 1 done.\n");
|
|
break;
|
|
}
|
|
udelay(50);
|
|
}
|
|
if (retry < 5)
|
|
break;
|
|
}
|
|
if (i == 4)
|
|
drm_err(&dev_priv->drm, "FDI train 1 fail!\n");
|
|
|
|
/* Train 2 */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_2;
|
|
if (IS_SANDYBRIDGE(dev_priv)) {
|
|
temp &= ~FDI_LINK_TRAIN_VOL_EMP_MASK;
|
|
/* SNB-B */
|
|
temp |= FDI_LINK_TRAIN_400MV_0DB_SNB_B;
|
|
}
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
if (HAS_PCH_CPT(dev_priv)) {
|
|
temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_2_CPT;
|
|
} else {
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_2;
|
|
}
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(150);
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe),
|
|
FDI_LINK_TRAIN_VOL_EMP_MASK, snb_b_fdi_train_param[i]);
|
|
intel_de_posting_read(dev_priv, FDI_TX_CTL(pipe));
|
|
udelay(500);
|
|
|
|
for (retry = 0; retry < 5; retry++) {
|
|
reg = FDI_RX_IIR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR 0x%x\n", temp);
|
|
if (temp & FDI_RX_SYMBOL_LOCK) {
|
|
intel_de_write(dev_priv, reg,
|
|
temp | FDI_RX_SYMBOL_LOCK);
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI train 2 done.\n");
|
|
break;
|
|
}
|
|
udelay(50);
|
|
}
|
|
if (retry < 5)
|
|
break;
|
|
}
|
|
if (i == 4)
|
|
drm_err(&dev_priv->drm, "FDI train 2 fail!\n");
|
|
|
|
drm_dbg_kms(&dev_priv->drm, "FDI train done.\n");
|
|
}
|
|
|
|
/* Manual link training for Ivy Bridge A0 parts */
|
|
static void ivb_manual_fdi_link_train(struct intel_crtc *crtc,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct drm_device *dev = crtc->base.dev;
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
i915_reg_t reg;
|
|
u32 temp, i, j;
|
|
|
|
ivb_update_fdi_bc_bifurcation(crtc_state);
|
|
|
|
/*
|
|
* Write the TU size bits before fdi link training, so that error
|
|
* detection works.
|
|
*/
|
|
intel_de_write(dev_priv, FDI_RX_TUSIZE1(pipe),
|
|
intel_de_read(dev_priv, PIPE_DATA_M1(pipe)) & TU_SIZE_MASK);
|
|
|
|
/* Train 1: umask FDI RX Interrupt symbol_lock and bit_lock bit
|
|
for train result */
|
|
reg = FDI_RX_IMR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_RX_SYMBOL_LOCK;
|
|
temp &= ~FDI_RX_BIT_LOCK;
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(150);
|
|
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR before link train 0x%x\n",
|
|
intel_de_read(dev_priv, FDI_RX_IIR(pipe)));
|
|
|
|
/* Try each vswing and preemphasis setting twice before moving on */
|
|
for (j = 0; j < ARRAY_SIZE(snb_b_fdi_train_param) * 2; j++) {
|
|
/* disable first in case we need to retry */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~(FDI_LINK_TRAIN_AUTO | FDI_LINK_TRAIN_NONE_IVB);
|
|
temp &= ~FDI_TX_ENABLE;
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_LINK_TRAIN_AUTO;
|
|
temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
|
|
temp &= ~FDI_RX_ENABLE;
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
/* enable CPU FDI TX and PCH FDI RX */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~FDI_DP_PORT_WIDTH_MASK;
|
|
temp |= FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1_IVB;
|
|
temp &= ~FDI_LINK_TRAIN_VOL_EMP_MASK;
|
|
temp |= snb_b_fdi_train_param[j/2];
|
|
temp |= FDI_COMPOSITE_SYNC;
|
|
intel_de_write(dev_priv, reg, temp | FDI_TX_ENABLE);
|
|
|
|
intel_de_write(dev_priv, FDI_RX_MISC(pipe),
|
|
FDI_RX_TP1_TO_TP2_48 | FDI_RX_FDI_DELAY_90);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
|
|
temp |= FDI_COMPOSITE_SYNC;
|
|
intel_de_write(dev_priv, reg, temp | FDI_RX_ENABLE);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(1); /* should be 0.5us */
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
reg = FDI_RX_IIR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR 0x%x\n", temp);
|
|
|
|
if (temp & FDI_RX_BIT_LOCK ||
|
|
(intel_de_read(dev_priv, reg) & FDI_RX_BIT_LOCK)) {
|
|
intel_de_write(dev_priv, reg,
|
|
temp | FDI_RX_BIT_LOCK);
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI train 1 done, level %i.\n",
|
|
i);
|
|
break;
|
|
}
|
|
udelay(1); /* should be 0.5us */
|
|
}
|
|
if (i == 4) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI train 1 fail on vswing %d\n", j / 2);
|
|
continue;
|
|
}
|
|
|
|
/* Train 2 */
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe),
|
|
FDI_LINK_TRAIN_NONE_IVB,
|
|
FDI_LINK_TRAIN_PATTERN_2_IVB);
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(pipe),
|
|
FDI_LINK_TRAIN_PATTERN_MASK_CPT,
|
|
FDI_LINK_TRAIN_PATTERN_2_CPT);
|
|
intel_de_posting_read(dev_priv, FDI_RX_CTL(pipe));
|
|
udelay(2); /* should be 1.5us */
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
reg = FDI_RX_IIR(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
drm_dbg_kms(&dev_priv->drm, "FDI_RX_IIR 0x%x\n", temp);
|
|
|
|
if (temp & FDI_RX_SYMBOL_LOCK ||
|
|
(intel_de_read(dev_priv, reg) & FDI_RX_SYMBOL_LOCK)) {
|
|
intel_de_write(dev_priv, reg,
|
|
temp | FDI_RX_SYMBOL_LOCK);
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI train 2 done, level %i.\n",
|
|
i);
|
|
goto train_done;
|
|
}
|
|
udelay(2); /* should be 1.5us */
|
|
}
|
|
if (i == 4)
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI train 2 fail on vswing %d\n", j / 2);
|
|
}
|
|
|
|
train_done:
|
|
drm_dbg_kms(&dev_priv->drm, "FDI train done.\n");
|
|
}
|
|
|
|
/* Starting with Haswell, different DDI ports can work in FDI mode for
|
|
* connection to the PCH-located connectors. For this, it is necessary to train
|
|
* both the DDI port and PCH receiver for the desired DDI buffer settings.
|
|
*
|
|
* The recommended port to work in FDI mode is DDI E, which we use here. Also,
|
|
* please note that when FDI mode is active on DDI E, it shares 2 lines with
|
|
* DDI A (which is used for eDP)
|
|
*/
|
|
void hsw_fdi_link_train(struct intel_encoder *encoder,
|
|
const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc);
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
|
u32 temp, i, rx_ctl_val;
|
|
int n_entries;
|
|
|
|
encoder->get_buf_trans(encoder, crtc_state, &n_entries);
|
|
|
|
hsw_prepare_dp_ddi_buffers(encoder, crtc_state);
|
|
|
|
/* Set the FDI_RX_MISC pwrdn lanes and the 2 workarounds listed at the
|
|
* mode set "sequence for CRT port" document:
|
|
* - TP1 to TP2 time with the default value
|
|
* - FDI delay to 90h
|
|
*
|
|
* WaFDIAutoLinkSetTimingOverrride:hsw
|
|
*/
|
|
intel_de_write(dev_priv, FDI_RX_MISC(PIPE_A),
|
|
FDI_RX_PWRDN_LANE1_VAL(2) |
|
|
FDI_RX_PWRDN_LANE0_VAL(2) |
|
|
FDI_RX_TP1_TO_TP2_48 |
|
|
FDI_RX_FDI_DELAY_90);
|
|
|
|
/* Enable the PCH Receiver FDI PLL */
|
|
rx_ctl_val = dev_priv->display.fdi.rx_config | FDI_RX_ENHANCE_FRAME_ENABLE |
|
|
FDI_RX_PLL_ENABLE |
|
|
FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
|
|
intel_de_write(dev_priv, FDI_RX_CTL(PIPE_A), rx_ctl_val);
|
|
intel_de_posting_read(dev_priv, FDI_RX_CTL(PIPE_A));
|
|
udelay(220);
|
|
|
|
/* Switch from Rawclk to PCDclk */
|
|
rx_ctl_val |= FDI_PCDCLK;
|
|
intel_de_write(dev_priv, FDI_RX_CTL(PIPE_A), rx_ctl_val);
|
|
|
|
/* Configure Port Clock Select */
|
|
drm_WARN_ON(&dev_priv->drm, crtc_state->shared_dpll->info->id != DPLL_ID_SPLL);
|
|
intel_ddi_enable_clock(encoder, crtc_state);
|
|
|
|
/* Start the training iterating through available voltages and emphasis,
|
|
* testing each value twice. */
|
|
for (i = 0; i < n_entries * 2; i++) {
|
|
/* Configure DP_TP_CTL with auto-training */
|
|
intel_de_write(dev_priv, DP_TP_CTL(PORT_E),
|
|
DP_TP_CTL_FDI_AUTOTRAIN |
|
|
DP_TP_CTL_ENHANCED_FRAME_ENABLE |
|
|
DP_TP_CTL_LINK_TRAIN_PAT1 |
|
|
DP_TP_CTL_ENABLE);
|
|
|
|
/* Configure and enable DDI_BUF_CTL for DDI E with next voltage.
|
|
* DDI E does not support port reversal, the functionality is
|
|
* achieved on the PCH side in FDI_RX_CTL, so no need to set the
|
|
* port reversal bit */
|
|
intel_de_write(dev_priv, DDI_BUF_CTL(PORT_E),
|
|
DDI_BUF_CTL_ENABLE |
|
|
((crtc_state->fdi_lanes - 1) << 1) |
|
|
DDI_BUF_TRANS_SELECT(i / 2));
|
|
intel_de_posting_read(dev_priv, DDI_BUF_CTL(PORT_E));
|
|
|
|
udelay(600);
|
|
|
|
/* Program PCH FDI Receiver TU */
|
|
intel_de_write(dev_priv, FDI_RX_TUSIZE1(PIPE_A), TU_SIZE(64));
|
|
|
|
/* Enable PCH FDI Receiver with auto-training */
|
|
rx_ctl_val |= FDI_RX_ENABLE | FDI_LINK_TRAIN_AUTO;
|
|
intel_de_write(dev_priv, FDI_RX_CTL(PIPE_A), rx_ctl_val);
|
|
intel_de_posting_read(dev_priv, FDI_RX_CTL(PIPE_A));
|
|
|
|
/* Wait for FDI receiver lane calibration */
|
|
udelay(30);
|
|
|
|
/* Unset FDI_RX_MISC pwrdn lanes */
|
|
intel_de_rmw(dev_priv, FDI_RX_MISC(PIPE_A),
|
|
FDI_RX_PWRDN_LANE1_MASK | FDI_RX_PWRDN_LANE0_MASK, 0);
|
|
intel_de_posting_read(dev_priv, FDI_RX_MISC(PIPE_A));
|
|
|
|
/* Wait for FDI auto training time */
|
|
udelay(5);
|
|
|
|
temp = intel_de_read(dev_priv, DP_TP_STATUS(PORT_E));
|
|
if (temp & DP_TP_STATUS_AUTOTRAIN_DONE) {
|
|
drm_dbg_kms(&dev_priv->drm,
|
|
"FDI link training done on step %d\n", i);
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* Leave things enabled even if we failed to train FDI.
|
|
* Results in less fireworks from the state checker.
|
|
*/
|
|
if (i == n_entries * 2 - 1) {
|
|
drm_err(&dev_priv->drm, "FDI link training failed!\n");
|
|
break;
|
|
}
|
|
|
|
rx_ctl_val &= ~FDI_RX_ENABLE;
|
|
intel_de_write(dev_priv, FDI_RX_CTL(PIPE_A), rx_ctl_val);
|
|
intel_de_posting_read(dev_priv, FDI_RX_CTL(PIPE_A));
|
|
|
|
intel_de_rmw(dev_priv, DDI_BUF_CTL(PORT_E), DDI_BUF_CTL_ENABLE, 0);
|
|
intel_de_posting_read(dev_priv, DDI_BUF_CTL(PORT_E));
|
|
|
|
/* Disable DP_TP_CTL and FDI_RX_CTL and retry */
|
|
intel_de_rmw(dev_priv, DP_TP_CTL(PORT_E), DP_TP_CTL_ENABLE, 0);
|
|
intel_de_posting_read(dev_priv, DP_TP_CTL(PORT_E));
|
|
|
|
intel_wait_ddi_buf_idle(dev_priv, PORT_E);
|
|
|
|
/* Reset FDI_RX_MISC pwrdn lanes */
|
|
intel_de_rmw(dev_priv, FDI_RX_MISC(PIPE_A),
|
|
FDI_RX_PWRDN_LANE1_MASK | FDI_RX_PWRDN_LANE0_MASK,
|
|
FDI_RX_PWRDN_LANE1_VAL(2) | FDI_RX_PWRDN_LANE0_VAL(2));
|
|
intel_de_posting_read(dev_priv, FDI_RX_MISC(PIPE_A));
|
|
}
|
|
|
|
/* Enable normal pixel sending for FDI */
|
|
intel_de_write(dev_priv, DP_TP_CTL(PORT_E),
|
|
DP_TP_CTL_FDI_AUTOTRAIN |
|
|
DP_TP_CTL_LINK_TRAIN_NORMAL |
|
|
DP_TP_CTL_ENHANCED_FRAME_ENABLE |
|
|
DP_TP_CTL_ENABLE);
|
|
}
|
|
|
|
void hsw_fdi_disable(struct intel_encoder *encoder)
|
|
{
|
|
struct drm_i915_private *dev_priv = to_i915(encoder->base.dev);
|
|
|
|
/*
|
|
* Bspec lists this as both step 13 (before DDI_BUF_CTL disable)
|
|
* and step 18 (after clearing PORT_CLK_SEL). Based on a BUN,
|
|
* step 13 is the correct place for it. Step 18 is where it was
|
|
* originally before the BUN.
|
|
*/
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(PIPE_A), FDI_RX_ENABLE, 0);
|
|
intel_de_rmw(dev_priv, DDI_BUF_CTL(PORT_E), DDI_BUF_CTL_ENABLE, 0);
|
|
intel_wait_ddi_buf_idle(dev_priv, PORT_E);
|
|
intel_ddi_disable_clock(encoder);
|
|
intel_de_rmw(dev_priv, FDI_RX_MISC(PIPE_A),
|
|
FDI_RX_PWRDN_LANE1_MASK | FDI_RX_PWRDN_LANE0_MASK,
|
|
FDI_RX_PWRDN_LANE1_VAL(2) | FDI_RX_PWRDN_LANE0_VAL(2));
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(PIPE_A), FDI_PCDCLK, 0);
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(PIPE_A), FDI_RX_PLL_ENABLE, 0);
|
|
}
|
|
|
|
void ilk_fdi_pll_enable(const struct intel_crtc_state *crtc_state)
|
|
{
|
|
struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc);
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
i915_reg_t reg;
|
|
u32 temp;
|
|
|
|
/* enable PCH FDI RX PLL, wait warmup plus DMI latency */
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~(FDI_DP_PORT_WIDTH_MASK | (0x7 << 16));
|
|
temp |= FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes);
|
|
temp |= (intel_de_read(dev_priv, TRANSCONF(pipe)) & TRANSCONF_BPC_MASK) << 11;
|
|
intel_de_write(dev_priv, reg, temp | FDI_RX_PLL_ENABLE);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(200);
|
|
|
|
/* Switch from Rawclk to PCDclk */
|
|
intel_de_rmw(dev_priv, reg, 0, FDI_PCDCLK);
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(200);
|
|
|
|
/* Enable CPU FDI TX PLL, always on for Ironlake */
|
|
reg = FDI_TX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
if ((temp & FDI_TX_PLL_ENABLE) == 0) {
|
|
intel_de_write(dev_priv, reg, temp | FDI_TX_PLL_ENABLE);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(100);
|
|
}
|
|
}
|
|
|
|
void ilk_fdi_pll_disable(struct intel_crtc *crtc)
|
|
{
|
|
struct drm_device *dev = crtc->base.dev;
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
|
|
/* Switch from PCDclk to Rawclk */
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(pipe), FDI_PCDCLK, 0);
|
|
|
|
/* Disable CPU FDI TX PLL */
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe), FDI_TX_PLL_ENABLE, 0);
|
|
intel_de_posting_read(dev_priv, FDI_TX_CTL(pipe));
|
|
udelay(100);
|
|
|
|
/* Wait for the clocks to turn off. */
|
|
intel_de_rmw(dev_priv, FDI_RX_CTL(pipe), FDI_RX_PLL_ENABLE, 0);
|
|
intel_de_posting_read(dev_priv, FDI_RX_CTL(pipe));
|
|
udelay(100);
|
|
}
|
|
|
|
void ilk_fdi_disable(struct intel_crtc *crtc)
|
|
{
|
|
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
|
enum pipe pipe = crtc->pipe;
|
|
i915_reg_t reg;
|
|
u32 temp;
|
|
|
|
/* disable CPU FDI tx and PCH FDI rx */
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe), FDI_TX_ENABLE, 0);
|
|
intel_de_posting_read(dev_priv, FDI_TX_CTL(pipe));
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
temp &= ~(0x7 << 16);
|
|
temp |= (intel_de_read(dev_priv, TRANSCONF(pipe)) & TRANSCONF_BPC_MASK) << 11;
|
|
intel_de_write(dev_priv, reg, temp & ~FDI_RX_ENABLE);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(100);
|
|
|
|
/* Ironlake workaround, disable clock pointer after downing FDI */
|
|
if (HAS_PCH_IBX(dev_priv))
|
|
intel_de_write(dev_priv, FDI_RX_CHICKEN(pipe),
|
|
FDI_RX_PHASE_SYNC_POINTER_OVR);
|
|
|
|
/* still set train pattern 1 */
|
|
intel_de_rmw(dev_priv, FDI_TX_CTL(pipe),
|
|
FDI_LINK_TRAIN_NONE, FDI_LINK_TRAIN_PATTERN_1);
|
|
|
|
reg = FDI_RX_CTL(pipe);
|
|
temp = intel_de_read(dev_priv, reg);
|
|
if (HAS_PCH_CPT(dev_priv)) {
|
|
temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1_CPT;
|
|
} else {
|
|
temp &= ~FDI_LINK_TRAIN_NONE;
|
|
temp |= FDI_LINK_TRAIN_PATTERN_1;
|
|
}
|
|
/* BPC in FDI rx is consistent with that in TRANSCONF */
|
|
temp &= ~(0x07 << 16);
|
|
temp |= (intel_de_read(dev_priv, TRANSCONF(pipe)) & TRANSCONF_BPC_MASK) << 11;
|
|
intel_de_write(dev_priv, reg, temp);
|
|
|
|
intel_de_posting_read(dev_priv, reg);
|
|
udelay(100);
|
|
}
|
|
|
|
static const struct intel_fdi_funcs ilk_funcs = {
|
|
.fdi_link_train = ilk_fdi_link_train,
|
|
};
|
|
|
|
static const struct intel_fdi_funcs gen6_funcs = {
|
|
.fdi_link_train = gen6_fdi_link_train,
|
|
};
|
|
|
|
static const struct intel_fdi_funcs ivb_funcs = {
|
|
.fdi_link_train = ivb_manual_fdi_link_train,
|
|
};
|
|
|
|
void
|
|
intel_fdi_init_hook(struct drm_i915_private *dev_priv)
|
|
{
|
|
if (IS_IRONLAKE(dev_priv)) {
|
|
dev_priv->display.funcs.fdi = &ilk_funcs;
|
|
} else if (IS_SANDYBRIDGE(dev_priv)) {
|
|
dev_priv->display.funcs.fdi = &gen6_funcs;
|
|
} else if (IS_IVYBRIDGE(dev_priv)) {
|
|
/* FIXME: detect B0+ stepping and use auto training */
|
|
dev_priv->display.funcs.fdi = &ivb_funcs;
|
|
}
|
|
}
|