drm/i915: group all hotplug related fields into a new struct in dev_priv
There are plenty of hotplug related fields in struct drm_i915_private scattered all around. Group them under one hotplug struct. Clean up naming while at it. No functional changes. Signed-off-by: Jani Nikula <jani.nikula@intel.com> Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
This commit is contained in:
Родитель
b0c29a33fc
Коммит
5fcece80ec
|
@ -933,8 +933,8 @@ int i915_driver_load(struct drm_device *dev, unsigned long flags)
|
|||
goto out_mtrrfree;
|
||||
}
|
||||
|
||||
dev_priv->dp_wq = alloc_ordered_workqueue("i915-dp", 0);
|
||||
if (dev_priv->dp_wq == NULL) {
|
||||
dev_priv->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0);
|
||||
if (dev_priv->hotplug.dp_wq == NULL) {
|
||||
DRM_ERROR("Failed to create our dp workqueue.\n");
|
||||
ret = -ENOMEM;
|
||||
goto out_freewq;
|
||||
|
@ -1029,7 +1029,7 @@ out_gem_unload:
|
|||
pm_qos_remove_request(&dev_priv->pm_qos);
|
||||
destroy_workqueue(dev_priv->gpu_error.hangcheck_wq);
|
||||
out_freedpwq:
|
||||
destroy_workqueue(dev_priv->dp_wq);
|
||||
destroy_workqueue(dev_priv->hotplug.dp_wq);
|
||||
out_freewq:
|
||||
destroy_workqueue(dev_priv->wq);
|
||||
out_mtrrfree:
|
||||
|
@ -1123,7 +1123,7 @@ int i915_driver_unload(struct drm_device *dev)
|
|||
intel_teardown_gmbus(dev);
|
||||
intel_teardown_mchbar(dev);
|
||||
|
||||
destroy_workqueue(dev_priv->dp_wq);
|
||||
destroy_workqueue(dev_priv->hotplug.dp_wq);
|
||||
destroy_workqueue(dev_priv->wq);
|
||||
destroy_workqueue(dev_priv->gpu_error.hangcheck_wq);
|
||||
pm_qos_remove_request(&dev_priv->pm_qos);
|
||||
|
|
|
@ -545,15 +545,15 @@ void intel_hpd_cancel_work(struct drm_i915_private *dev_priv)
|
|||
{
|
||||
spin_lock_irq(&dev_priv->irq_lock);
|
||||
|
||||
dev_priv->long_hpd_port_mask = 0;
|
||||
dev_priv->short_hpd_port_mask = 0;
|
||||
dev_priv->hpd_event_bits = 0;
|
||||
dev_priv->hotplug.long_port_mask = 0;
|
||||
dev_priv->hotplug.short_port_mask = 0;
|
||||
dev_priv->hotplug.event_bits = 0;
|
||||
|
||||
spin_unlock_irq(&dev_priv->irq_lock);
|
||||
|
||||
cancel_work_sync(&dev_priv->dig_port_work);
|
||||
cancel_work_sync(&dev_priv->hotplug_work);
|
||||
cancel_delayed_work_sync(&dev_priv->hotplug_reenable_work);
|
||||
cancel_work_sync(&dev_priv->hotplug.dig_port_work);
|
||||
cancel_work_sync(&dev_priv->hotplug.hotplug_work);
|
||||
cancel_delayed_work_sync(&dev_priv->hotplug.reenable_work);
|
||||
}
|
||||
|
||||
void i915_firmware_load_error_print(const char *fw_path, int err)
|
||||
|
|
|
@ -217,6 +217,36 @@ enum hpd_pin {
|
|||
HPD_NUM_PINS
|
||||
};
|
||||
|
||||
struct i915_hotplug {
|
||||
struct work_struct hotplug_work;
|
||||
|
||||
struct {
|
||||
unsigned long last_jiffies;
|
||||
int count;
|
||||
enum {
|
||||
HPD_ENABLED = 0,
|
||||
HPD_DISABLED = 1,
|
||||
HPD_MARK_DISABLED = 2
|
||||
} state;
|
||||
} stats[HPD_NUM_PINS];
|
||||
u32 event_bits;
|
||||
struct delayed_work reenable_work;
|
||||
|
||||
struct intel_digital_port *irq_port[I915_MAX_PORTS];
|
||||
u32 long_port_mask;
|
||||
u32 short_port_mask;
|
||||
struct work_struct dig_port_work;
|
||||
|
||||
/*
|
||||
* if we get a HPD irq from DP and a HPD irq from non-DP
|
||||
* the non-DP HPD could block the workqueue on a mode config
|
||||
* mutex getting, that userspace may have taken. However
|
||||
* userspace is waiting on the DP workqueue to run which is
|
||||
* blocked behind the non-DP one.
|
||||
*/
|
||||
struct workqueue_struct *dp_wq;
|
||||
};
|
||||
|
||||
#define I915_GEM_GPU_DOMAINS \
|
||||
(I915_GEM_DOMAIN_RENDER | \
|
||||
I915_GEM_DOMAIN_SAMPLER | \
|
||||
|
@ -1684,19 +1714,7 @@ struct drm_i915_private {
|
|||
u32 pm_rps_events;
|
||||
u32 pipestat_irq_mask[I915_MAX_PIPES];
|
||||
|
||||
struct work_struct hotplug_work;
|
||||
struct {
|
||||
unsigned long hpd_last_jiffies;
|
||||
int hpd_cnt;
|
||||
enum {
|
||||
HPD_ENABLED = 0,
|
||||
HPD_DISABLED = 1,
|
||||
HPD_MARK_DISABLED = 2
|
||||
} hpd_mark;
|
||||
} hpd_stats[HPD_NUM_PINS];
|
||||
u32 hpd_event_bits;
|
||||
struct delayed_work hotplug_reenable_work;
|
||||
|
||||
struct i915_hotplug hotplug;
|
||||
struct i915_fbc fbc;
|
||||
struct i915_drrs drrs;
|
||||
struct intel_opregion opregion;
|
||||
|
@ -1862,20 +1880,6 @@ struct drm_i915_private {
|
|||
|
||||
struct i915_runtime_pm pm;
|
||||
|
||||
struct intel_digital_port *hpd_irq_port[I915_MAX_PORTS];
|
||||
u32 long_hpd_port_mask;
|
||||
u32 short_hpd_port_mask;
|
||||
struct work_struct dig_port_work;
|
||||
|
||||
/*
|
||||
* if we get a HPD irq from DP and a HPD irq from non-DP
|
||||
* the non-DP HPD could block the workqueue on a mode config
|
||||
* mutex getting, that userspace may have taken. However
|
||||
* userspace is waiting on the DP workqueue to run which is
|
||||
* blocked behind the non-DP one.
|
||||
*/
|
||||
struct workqueue_struct *dp_wq;
|
||||
|
||||
/* Abstract the submission mechanism (legacy ringbuffer or execlists) away */
|
||||
struct {
|
||||
int (*execbuf_submit)(struct drm_device *dev, struct drm_file *file,
|
||||
|
|
|
@ -832,23 +832,23 @@ static bool intel_hpd_irq_event(struct drm_device *dev,
|
|||
static void i915_digport_work_func(struct work_struct *work)
|
||||
{
|
||||
struct drm_i915_private *dev_priv =
|
||||
container_of(work, struct drm_i915_private, dig_port_work);
|
||||
container_of(work, struct drm_i915_private, hotplug.dig_port_work);
|
||||
u32 long_port_mask, short_port_mask;
|
||||
struct intel_digital_port *intel_dig_port;
|
||||
int i;
|
||||
u32 old_bits = 0;
|
||||
|
||||
spin_lock_irq(&dev_priv->irq_lock);
|
||||
long_port_mask = dev_priv->long_hpd_port_mask;
|
||||
dev_priv->long_hpd_port_mask = 0;
|
||||
short_port_mask = dev_priv->short_hpd_port_mask;
|
||||
dev_priv->short_hpd_port_mask = 0;
|
||||
long_port_mask = dev_priv->hotplug.long_port_mask;
|
||||
dev_priv->hotplug.long_port_mask = 0;
|
||||
short_port_mask = dev_priv->hotplug.short_port_mask;
|
||||
dev_priv->hotplug.short_port_mask = 0;
|
||||
spin_unlock_irq(&dev_priv->irq_lock);
|
||||
|
||||
for (i = 0; i < I915_MAX_PORTS; i++) {
|
||||
bool valid = false;
|
||||
bool long_hpd = false;
|
||||
intel_dig_port = dev_priv->hpd_irq_port[i];
|
||||
intel_dig_port = dev_priv->hotplug.irq_port[i];
|
||||
if (!intel_dig_port || !intel_dig_port->hpd_pulse)
|
||||
continue;
|
||||
|
||||
|
@ -871,9 +871,9 @@ static void i915_digport_work_func(struct work_struct *work)
|
|||
|
||||
if (old_bits) {
|
||||
spin_lock_irq(&dev_priv->irq_lock);
|
||||
dev_priv->hpd_event_bits |= old_bits;
|
||||
dev_priv->hotplug.event_bits |= old_bits;
|
||||
spin_unlock_irq(&dev_priv->irq_lock);
|
||||
schedule_work(&dev_priv->hotplug_work);
|
||||
schedule_work(&dev_priv->hotplug.hotplug_work);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -885,7 +885,7 @@ static void i915_digport_work_func(struct work_struct *work)
|
|||
static void i915_hotplug_work_func(struct work_struct *work)
|
||||
{
|
||||
struct drm_i915_private *dev_priv =
|
||||
container_of(work, struct drm_i915_private, hotplug_work);
|
||||
container_of(work, struct drm_i915_private, hotplug.hotplug_work);
|
||||
struct drm_device *dev = dev_priv->dev;
|
||||
struct drm_mode_config *mode_config = &dev->mode_config;
|
||||
struct intel_connector *intel_connector;
|
||||
|
@ -900,20 +900,20 @@ static void i915_hotplug_work_func(struct work_struct *work)
|
|||
|
||||
spin_lock_irq(&dev_priv->irq_lock);
|
||||
|
||||
hpd_event_bits = dev_priv->hpd_event_bits;
|
||||
dev_priv->hpd_event_bits = 0;
|
||||
hpd_event_bits = dev_priv->hotplug.event_bits;
|
||||
dev_priv->hotplug.event_bits = 0;
|
||||
list_for_each_entry(connector, &mode_config->connector_list, head) {
|
||||
intel_connector = to_intel_connector(connector);
|
||||
if (!intel_connector->encoder)
|
||||
continue;
|
||||
intel_encoder = intel_connector->encoder;
|
||||
if (intel_encoder->hpd_pin > HPD_NONE &&
|
||||
dev_priv->hpd_stats[intel_encoder->hpd_pin].hpd_mark == HPD_MARK_DISABLED &&
|
||||
dev_priv->hotplug.stats[intel_encoder->hpd_pin].state == HPD_MARK_DISABLED &&
|
||||
connector->polled == DRM_CONNECTOR_POLL_HPD) {
|
||||
DRM_INFO("HPD interrupt storm detected on connector %s: "
|
||||
"switching from hotplug detection to polling\n",
|
||||
connector->name);
|
||||
dev_priv->hpd_stats[intel_encoder->hpd_pin].hpd_mark = HPD_DISABLED;
|
||||
dev_priv->hotplug.stats[intel_encoder->hpd_pin].state = HPD_DISABLED;
|
||||
connector->polled = DRM_CONNECTOR_POLL_CONNECT
|
||||
| DRM_CONNECTOR_POLL_DISCONNECT;
|
||||
hpd_disabled = true;
|
||||
|
@ -928,7 +928,7 @@ static void i915_hotplug_work_func(struct work_struct *work)
|
|||
* some connectors */
|
||||
if (hpd_disabled) {
|
||||
drm_kms_helper_poll_enable(dev);
|
||||
mod_delayed_work(system_wq, &dev_priv->hotplug_reenable_work,
|
||||
mod_delayed_work(system_wq, &dev_priv->hotplug.reenable_work,
|
||||
msecs_to_jiffies(I915_REENABLE_HOTPLUG_DELAY));
|
||||
}
|
||||
|
||||
|
@ -1448,7 +1448,7 @@ static void intel_hpd_irq_handler(struct drm_device *dev,
|
|||
continue;
|
||||
|
||||
port = get_port_from_pin(i);
|
||||
if (!port || !dev_priv->hpd_irq_port[port])
|
||||
if (!port || !dev_priv->hotplug.irq_port[port])
|
||||
continue;
|
||||
|
||||
if (!HAS_GMCH_DISPLAY(dev_priv)) {
|
||||
|
@ -1466,11 +1466,11 @@ static void intel_hpd_irq_handler(struct drm_device *dev,
|
|||
* but we still want HPD storm detection to function.
|
||||
*/
|
||||
if (long_hpd) {
|
||||
dev_priv->long_hpd_port_mask |= (1 << port);
|
||||
dev_priv->hotplug.long_port_mask |= (1 << port);
|
||||
dig_port_mask |= hpd[i];
|
||||
} else {
|
||||
/* for short HPD just trigger the digital queue */
|
||||
dev_priv->short_hpd_port_mask |= (1 << port);
|
||||
dev_priv->hotplug.short_port_mask |= (1 << port);
|
||||
hotplug_trigger &= ~hpd[i];
|
||||
}
|
||||
|
||||
|
@ -1479,7 +1479,7 @@ static void intel_hpd_irq_handler(struct drm_device *dev,
|
|||
|
||||
for (i = 1; i < HPD_NUM_PINS; i++) {
|
||||
if (hpd[i] & hotplug_trigger &&
|
||||
dev_priv->hpd_stats[i].hpd_mark == HPD_DISABLED) {
|
||||
dev_priv->hotplug.stats[i].state == HPD_DISABLED) {
|
||||
/*
|
||||
* On GMCH platforms the interrupt mask bits only
|
||||
* prevent irq generation, not the setting of the
|
||||
|
@ -1494,29 +1494,29 @@ static void intel_hpd_irq_handler(struct drm_device *dev,
|
|||
}
|
||||
|
||||
if (!(hpd[i] & hotplug_trigger) ||
|
||||
dev_priv->hpd_stats[i].hpd_mark != HPD_ENABLED)
|
||||
dev_priv->hotplug.stats[i].state != HPD_ENABLED)
|
||||
continue;
|
||||
|
||||
if (!(dig_port_mask & hpd[i])) {
|
||||
dev_priv->hpd_event_bits |= (1 << i);
|
||||
dev_priv->hotplug.event_bits |= (1 << i);
|
||||
queue_hp = true;
|
||||
}
|
||||
|
||||
if (!time_in_range(jiffies, dev_priv->hpd_stats[i].hpd_last_jiffies,
|
||||
dev_priv->hpd_stats[i].hpd_last_jiffies
|
||||
if (!time_in_range(jiffies, dev_priv->hotplug.stats[i].last_jiffies,
|
||||
dev_priv->hotplug.stats[i].last_jiffies
|
||||
+ msecs_to_jiffies(HPD_STORM_DETECT_PERIOD))) {
|
||||
dev_priv->hpd_stats[i].hpd_last_jiffies = jiffies;
|
||||
dev_priv->hpd_stats[i].hpd_cnt = 0;
|
||||
dev_priv->hotplug.stats[i].last_jiffies = jiffies;
|
||||
dev_priv->hotplug.stats[i].count = 0;
|
||||
DRM_DEBUG_KMS("Received HPD interrupt on PIN %d - cnt: 0\n", i);
|
||||
} else if (dev_priv->hpd_stats[i].hpd_cnt > HPD_STORM_THRESHOLD) {
|
||||
dev_priv->hpd_stats[i].hpd_mark = HPD_MARK_DISABLED;
|
||||
dev_priv->hpd_event_bits &= ~(1 << i);
|
||||
} else if (dev_priv->hotplug.stats[i].count > HPD_STORM_THRESHOLD) {
|
||||
dev_priv->hotplug.stats[i].state = HPD_MARK_DISABLED;
|
||||
dev_priv->hotplug.event_bits &= ~(1 << i);
|
||||
DRM_DEBUG_KMS("HPD interrupt storm detected on PIN %d\n", i);
|
||||
storm_detected = true;
|
||||
} else {
|
||||
dev_priv->hpd_stats[i].hpd_cnt++;
|
||||
dev_priv->hotplug.stats[i].count++;
|
||||
DRM_DEBUG_KMS("Received HPD interrupt on PIN %d - cnt: %d\n", i,
|
||||
dev_priv->hpd_stats[i].hpd_cnt);
|
||||
dev_priv->hotplug.stats[i].count);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1531,9 +1531,9 @@ static void intel_hpd_irq_handler(struct drm_device *dev,
|
|||
* deadlock.
|
||||
*/
|
||||
if (queue_dig)
|
||||
queue_work(dev_priv->dp_wq, &dev_priv->dig_port_work);
|
||||
queue_work(dev_priv->hotplug.dp_wq, &dev_priv->hotplug.dig_port_work);
|
||||
if (queue_hp)
|
||||
schedule_work(&dev_priv->hotplug_work);
|
||||
schedule_work(&dev_priv->hotplug.hotplug_work);
|
||||
}
|
||||
|
||||
static void gmbus_irq_handler(struct drm_device *dev)
|
||||
|
@ -3213,12 +3213,12 @@ static void ibx_hpd_irq_setup(struct drm_device *dev)
|
|||
if (HAS_PCH_IBX(dev)) {
|
||||
hotplug_irqs = SDE_HOTPLUG_MASK;
|
||||
for_each_intel_encoder(dev, intel_encoder)
|
||||
if (dev_priv->hpd_stats[intel_encoder->hpd_pin].hpd_mark == HPD_ENABLED)
|
||||
if (dev_priv->hotplug.stats[intel_encoder->hpd_pin].state == HPD_ENABLED)
|
||||
enabled_irqs |= hpd_ibx[intel_encoder->hpd_pin];
|
||||
} else {
|
||||
hotplug_irqs = SDE_HOTPLUG_MASK_CPT;
|
||||
for_each_intel_encoder(dev, intel_encoder)
|
||||
if (dev_priv->hpd_stats[intel_encoder->hpd_pin].hpd_mark == HPD_ENABLED)
|
||||
if (dev_priv->hotplug.stats[intel_encoder->hpd_pin].state == HPD_ENABLED)
|
||||
enabled_irqs |= hpd_cpt[intel_encoder->hpd_pin];
|
||||
}
|
||||
|
||||
|
@ -3247,7 +3247,7 @@ static void bxt_hpd_irq_setup(struct drm_device *dev)
|
|||
|
||||
/* Now, enable HPD */
|
||||
for_each_intel_encoder(dev, intel_encoder) {
|
||||
if (dev_priv->hpd_stats[intel_encoder->hpd_pin].hpd_mark
|
||||
if (dev_priv->hotplug.stats[intel_encoder->hpd_pin].state
|
||||
== HPD_ENABLED)
|
||||
hotplug_port |= hpd_bxt[intel_encoder->hpd_pin];
|
||||
}
|
||||
|
@ -4140,7 +4140,7 @@ static void i915_hpd_irq_setup(struct drm_device *dev)
|
|||
/* Note HDMI and DP share hotplug bits */
|
||||
/* enable bits are the same for all generations */
|
||||
for_each_intel_encoder(dev, intel_encoder)
|
||||
if (dev_priv->hpd_stats[intel_encoder->hpd_pin].hpd_mark == HPD_ENABLED)
|
||||
if (dev_priv->hotplug.stats[intel_encoder->hpd_pin].state == HPD_ENABLED)
|
||||
hotplug_en |= hpd_mask_i915[intel_encoder->hpd_pin];
|
||||
/* Programming the CRT detection parameters tends
|
||||
to generate a spurious hotplug event about three
|
||||
|
@ -4284,7 +4284,7 @@ static void intel_hpd_irq_reenable_work(struct work_struct *work)
|
|||
{
|
||||
struct drm_i915_private *dev_priv =
|
||||
container_of(work, typeof(*dev_priv),
|
||||
hotplug_reenable_work.work);
|
||||
hotplug.reenable_work.work);
|
||||
struct drm_device *dev = dev_priv->dev;
|
||||
struct drm_mode_config *mode_config = &dev->mode_config;
|
||||
int i;
|
||||
|
@ -4295,10 +4295,10 @@ static void intel_hpd_irq_reenable_work(struct work_struct *work)
|
|||
for (i = (HPD_NONE + 1); i < HPD_NUM_PINS; i++) {
|
||||
struct drm_connector *connector;
|
||||
|
||||
if (dev_priv->hpd_stats[i].hpd_mark != HPD_DISABLED)
|
||||
if (dev_priv->hotplug.stats[i].state != HPD_DISABLED)
|
||||
continue;
|
||||
|
||||
dev_priv->hpd_stats[i].hpd_mark = HPD_ENABLED;
|
||||
dev_priv->hotplug.stats[i].state = HPD_ENABLED;
|
||||
|
||||
list_for_each_entry(connector, &mode_config->connector_list, head) {
|
||||
struct intel_connector *intel_connector = to_intel_connector(connector);
|
||||
|
@ -4331,8 +4331,8 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
|
|||
{
|
||||
struct drm_device *dev = dev_priv->dev;
|
||||
|
||||
INIT_WORK(&dev_priv->hotplug_work, i915_hotplug_work_func);
|
||||
INIT_WORK(&dev_priv->dig_port_work, i915_digport_work_func);
|
||||
INIT_WORK(&dev_priv->hotplug.hotplug_work, i915_hotplug_work_func);
|
||||
INIT_WORK(&dev_priv->hotplug.dig_port_work, i915_digport_work_func);
|
||||
INIT_WORK(&dev_priv->rps.work, gen6_pm_rps_work);
|
||||
INIT_WORK(&dev_priv->l3_parity.error_work, ivybridge_parity_work);
|
||||
|
||||
|
@ -4345,7 +4345,7 @@ void intel_irq_init(struct drm_i915_private *dev_priv)
|
|||
|
||||
INIT_DELAYED_WORK(&dev_priv->gpu_error.hangcheck_work,
|
||||
i915_hangcheck_elapsed);
|
||||
INIT_DELAYED_WORK(&dev_priv->hotplug_reenable_work,
|
||||
INIT_DELAYED_WORK(&dev_priv->hotplug.reenable_work,
|
||||
intel_hpd_irq_reenable_work);
|
||||
|
||||
pm_qos_add_request(&dev_priv->pm_qos, PM_QOS_CPU_DMA_LATENCY, PM_QOS_DEFAULT_VALUE);
|
||||
|
@ -4451,8 +4451,8 @@ void intel_hpd_init(struct drm_i915_private *dev_priv)
|
|||
int i;
|
||||
|
||||
for (i = 1; i < HPD_NUM_PINS; i++) {
|
||||
dev_priv->hpd_stats[i].hpd_cnt = 0;
|
||||
dev_priv->hpd_stats[i].hpd_mark = HPD_ENABLED;
|
||||
dev_priv->hotplug.stats[i].count = 0;
|
||||
dev_priv->hotplug.stats[i].state = HPD_ENABLED;
|
||||
}
|
||||
list_for_each_entry(connector, &mode_config->connector_list, head) {
|
||||
struct intel_connector *intel_connector = to_intel_connector(connector);
|
||||
|
|
|
@ -2832,7 +2832,7 @@ void intel_ddi_init(struct drm_device *dev, enum port port)
|
|||
goto err;
|
||||
|
||||
intel_dig_port->hpd_pulse = intel_dp_hpd_pulse;
|
||||
dev_priv->hpd_irq_port[port] = intel_dig_port;
|
||||
dev_priv->hotplug.irq_port[port] = intel_dig_port;
|
||||
}
|
||||
|
||||
/* In theory we don't need the encoder->type check, but leave it just in
|
||||
|
|
|
@ -5940,7 +5940,7 @@ intel_dp_init(struct drm_device *dev, int output_reg, enum port port)
|
|||
intel_encoder->hot_plug = intel_dp_hot_plug;
|
||||
|
||||
intel_dig_port->hpd_pulse = intel_dp_hpd_pulse;
|
||||
dev_priv->hpd_irq_port[port] = intel_dig_port;
|
||||
dev_priv->hotplug.irq_port[port] = intel_dig_port;
|
||||
|
||||
if (!intel_dp_init_connector(intel_dig_port, intel_connector)) {
|
||||
drm_encoder_cleanup(encoder);
|
||||
|
@ -5956,7 +5956,7 @@ void intel_dp_mst_suspend(struct drm_device *dev)
|
|||
|
||||
/* disable MST */
|
||||
for (i = 0; i < I915_MAX_PORTS; i++) {
|
||||
struct intel_digital_port *intel_dig_port = dev_priv->hpd_irq_port[i];
|
||||
struct intel_digital_port *intel_dig_port = dev_priv->hotplug.irq_port[i];
|
||||
if (!intel_dig_port)
|
||||
continue;
|
||||
|
||||
|
@ -5975,7 +5975,7 @@ void intel_dp_mst_resume(struct drm_device *dev)
|
|||
int i;
|
||||
|
||||
for (i = 0; i < I915_MAX_PORTS; i++) {
|
||||
struct intel_digital_port *intel_dig_port = dev_priv->hpd_irq_port[i];
|
||||
struct intel_digital_port *intel_dig_port = dev_priv->hotplug.irq_port[i];
|
||||
if (!intel_dig_port)
|
||||
continue;
|
||||
if (intel_dig_port->base.type == INTEL_OUTPUT_DISPLAYPORT) {
|
||||
|
|
Загрузка…
Ссылка в новой задаче