It would be too much noise to convert the intel_de_* functions from using struct drm_i915_private to struct intel_display all at once. Add generic wrappers using __to_intel_display() to accept both. v2: Take the intel_dmc_wl_* changes into account Cc: Luca Coelho <luciano.coelho@intel.com> Reviewed-by: Rodrigo Vivi <rodrigo.vivi@intel.com> Signed-off-by: Jani Nikula <jani.nikula@intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/1b0e8e7c732535e18c8498a2e18fe1e4c123e2f5.1713358679.git.jani.nikula@intel.com
236 lines
6.4 KiB
C
236 lines
6.4 KiB
C
/* SPDX-License-Identifier: MIT */
|
|
/*
|
|
* Copyright © 2019 Intel Corporation
|
|
*/
|
|
|
|
#ifndef __INTEL_DE_H__
|
|
#define __INTEL_DE_H__
|
|
|
|
#include "i915_drv.h"
|
|
#include "i915_trace.h"
|
|
#include "intel_uncore.h"
|
|
|
|
static inline struct intel_uncore *__to_uncore(struct intel_display *display)
|
|
{
|
|
return &to_i915(display->drm)->uncore;
|
|
}
|
|
|
|
static inline u32
|
|
__intel_de_read(struct intel_display *display, i915_reg_t reg)
|
|
{
|
|
u32 val;
|
|
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
val = intel_uncore_read(__to_uncore(display), reg);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
|
|
return val;
|
|
}
|
|
#define intel_de_read(p,...) __intel_de_read(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline u8
|
|
__intel_de_read8(struct intel_display *display, i915_reg_t reg)
|
|
{
|
|
u8 val;
|
|
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
val = intel_uncore_read8(__to_uncore(display), reg);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
|
|
return val;
|
|
}
|
|
#define intel_de_read8(p,...) __intel_de_read8(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline u64
|
|
__intel_de_read64_2x32(struct intel_display *display,
|
|
i915_reg_t lower_reg, i915_reg_t upper_reg)
|
|
{
|
|
u64 val;
|
|
|
|
intel_dmc_wl_get(display, lower_reg);
|
|
intel_dmc_wl_get(display, upper_reg);
|
|
|
|
val = intel_uncore_read64_2x32(__to_uncore(display), lower_reg,
|
|
upper_reg);
|
|
|
|
intel_dmc_wl_put(display, upper_reg);
|
|
intel_dmc_wl_put(display, lower_reg);
|
|
|
|
return val;
|
|
}
|
|
#define intel_de_read64_2x32(p,...) __intel_de_read64_2x32(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline void
|
|
__intel_de_posting_read(struct intel_display *display, i915_reg_t reg)
|
|
{
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
intel_uncore_posting_read(__to_uncore(display), reg);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
}
|
|
#define intel_de_posting_read(p,...) __intel_de_posting_read(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline void
|
|
__intel_de_write(struct intel_display *display, i915_reg_t reg, u32 val)
|
|
{
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
intel_uncore_write(__to_uncore(display), reg, val);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
}
|
|
#define intel_de_write(p,...) __intel_de_write(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline u32
|
|
____intel_de_rmw_nowl(struct intel_display *display, i915_reg_t reg,
|
|
u32 clear, u32 set)
|
|
{
|
|
return intel_uncore_rmw(__to_uncore(display), reg, clear, set);
|
|
}
|
|
#define __intel_de_rmw_nowl(p,...) ____intel_de_rmw_nowl(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline u32
|
|
__intel_de_rmw(struct intel_display *display, i915_reg_t reg, u32 clear,
|
|
u32 set)
|
|
{
|
|
u32 val;
|
|
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
val = __intel_de_rmw_nowl(display, reg, clear, set);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
|
|
return val;
|
|
}
|
|
#define intel_de_rmw(p,...) __intel_de_rmw(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline int
|
|
____intel_de_wait_for_register_nowl(struct intel_display *display,
|
|
i915_reg_t reg,
|
|
u32 mask, u32 value, unsigned int timeout)
|
|
{
|
|
return intel_wait_for_register(__to_uncore(display), reg, mask,
|
|
value, timeout);
|
|
}
|
|
#define __intel_de_wait_for_register_nowl(p,...) ____intel_de_wait_for_register_nowl(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline int
|
|
__intel_de_wait(struct intel_display *display, i915_reg_t reg,
|
|
u32 mask, u32 value, unsigned int timeout)
|
|
{
|
|
int ret;
|
|
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
ret = __intel_de_wait_for_register_nowl(display, reg, mask, value,
|
|
timeout);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
|
|
return ret;
|
|
}
|
|
#define intel_de_wait(p,...) __intel_de_wait(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline int
|
|
__intel_de_wait_fw(struct intel_display *display, i915_reg_t reg,
|
|
u32 mask, u32 value, unsigned int timeout)
|
|
{
|
|
int ret;
|
|
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
ret = intel_wait_for_register_fw(__to_uncore(display), reg, mask,
|
|
value, timeout);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
|
|
return ret;
|
|
}
|
|
#define intel_de_wait_fw(p,...) __intel_de_wait_fw(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline int
|
|
__intel_de_wait_custom(struct intel_display *display, i915_reg_t reg,
|
|
u32 mask, u32 value,
|
|
unsigned int fast_timeout_us,
|
|
unsigned int slow_timeout_ms, u32 *out_value)
|
|
{
|
|
int ret;
|
|
|
|
intel_dmc_wl_get(display, reg);
|
|
|
|
ret = __intel_wait_for_register(__to_uncore(display), reg, mask,
|
|
value,
|
|
fast_timeout_us, slow_timeout_ms, out_value);
|
|
|
|
intel_dmc_wl_put(display, reg);
|
|
|
|
return ret;
|
|
}
|
|
#define intel_de_wait_custom(p,...) __intel_de_wait_custom(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline int
|
|
__intel_de_wait_for_set(struct intel_display *display, i915_reg_t reg,
|
|
u32 mask, unsigned int timeout)
|
|
{
|
|
return intel_de_wait(display, reg, mask, mask, timeout);
|
|
}
|
|
#define intel_de_wait_for_set(p,...) __intel_de_wait_for_set(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline int
|
|
__intel_de_wait_for_clear(struct intel_display *display, i915_reg_t reg,
|
|
u32 mask, unsigned int timeout)
|
|
{
|
|
return intel_de_wait(display, reg, mask, 0, timeout);
|
|
}
|
|
#define intel_de_wait_for_clear(p,...) __intel_de_wait_for_clear(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
/*
|
|
* Unlocked mmio-accessors, think carefully before using these.
|
|
*
|
|
* Certain architectures will die if the same cacheline is concurrently accessed
|
|
* by different clients (e.g. on Ivybridge). Access to registers should
|
|
* therefore generally be serialised, by either the dev_priv->uncore.lock or
|
|
* a more localised lock guarding all access to that bank of registers.
|
|
*/
|
|
static inline u32
|
|
__intel_de_read_fw(struct intel_display *display, i915_reg_t reg)
|
|
{
|
|
u32 val;
|
|
|
|
val = intel_uncore_read_fw(__to_uncore(display), reg);
|
|
trace_i915_reg_rw(false, reg, val, sizeof(val), true);
|
|
|
|
return val;
|
|
}
|
|
#define intel_de_read_fw(p,...) __intel_de_read_fw(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline void
|
|
__intel_de_write_fw(struct intel_display *display, i915_reg_t reg, u32 val)
|
|
{
|
|
trace_i915_reg_rw(true, reg, val, sizeof(val), true);
|
|
intel_uncore_write_fw(__to_uncore(display), reg, val);
|
|
}
|
|
#define intel_de_write_fw(p,...) __intel_de_write_fw(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline u32
|
|
__intel_de_read_notrace(struct intel_display *display, i915_reg_t reg)
|
|
{
|
|
return intel_uncore_read_notrace(__to_uncore(display), reg);
|
|
}
|
|
#define intel_de_read_notrace(p,...) __intel_de_read_notrace(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
static inline void
|
|
__intel_de_write_notrace(struct intel_display *display, i915_reg_t reg,
|
|
u32 val)
|
|
{
|
|
intel_uncore_write_notrace(__to_uncore(display), reg, val);
|
|
}
|
|
#define intel_de_write_notrace(p,...) __intel_de_write_notrace(__to_intel_display(p), __VA_ARGS__)
|
|
|
|
#endif /* __INTEL_DE_H__ */
|