407 строки
8.7 KiB
C
407 строки
8.7 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2013-2020, Intel Corporation. All rights reserved.
|
|
* Intel Management Engine Interface (Intel MEI) Linux driver
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/device.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/types.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/init.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/workqueue.h>
|
|
#include <linux/pm_domain.h>
|
|
#include <linux/pm_runtime.h>
|
|
|
|
#include <linux/mei.h>
|
|
|
|
|
|
#include "mei_dev.h"
|
|
#include "hw-txe.h"
|
|
|
|
static const struct pci_device_id mei_txe_pci_tbl[] = {
|
|
{PCI_VDEVICE(INTEL, 0x0F18)}, /* Baytrail */
|
|
{PCI_VDEVICE(INTEL, 0x2298)}, /* Cherrytrail */
|
|
|
|
{0, }
|
|
};
|
|
MODULE_DEVICE_TABLE(pci, mei_txe_pci_tbl);
|
|
|
|
#ifdef CONFIG_PM
|
|
static inline void mei_txe_set_pm_domain(struct mei_device *dev);
|
|
static inline void mei_txe_unset_pm_domain(struct mei_device *dev);
|
|
#else
|
|
static inline void mei_txe_set_pm_domain(struct mei_device *dev) {}
|
|
static inline void mei_txe_unset_pm_domain(struct mei_device *dev) {}
|
|
#endif /* CONFIG_PM */
|
|
|
|
/**
|
|
* mei_txe_probe - Device Initialization Routine
|
|
*
|
|
* @pdev: PCI device structure
|
|
* @ent: entry in mei_txe_pci_tbl
|
|
*
|
|
* Return: 0 on success, <0 on failure.
|
|
*/
|
|
static int mei_txe_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|
{
|
|
struct mei_device *dev;
|
|
struct mei_txe_hw *hw;
|
|
const int mask = BIT(SEC_BAR) | BIT(BRIDGE_BAR);
|
|
int err;
|
|
|
|
/* enable pci dev */
|
|
err = pcim_enable_device(pdev);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "failed to enable pci device.\n");
|
|
goto end;
|
|
}
|
|
/* set PCI host mastering */
|
|
pci_set_master(pdev);
|
|
/* pci request regions and mapping IO device memory for mei driver */
|
|
err = pcim_iomap_regions(pdev, mask, KBUILD_MODNAME);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "failed to get pci regions.\n");
|
|
goto end;
|
|
}
|
|
|
|
err = pci_set_dma_mask(pdev, DMA_BIT_MASK(36));
|
|
if (err) {
|
|
err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
|
|
if (err) {
|
|
dev_err(&pdev->dev, "No suitable DMA available.\n");
|
|
goto end;
|
|
}
|
|
}
|
|
|
|
/* allocates and initializes the mei dev structure */
|
|
dev = mei_txe_dev_init(pdev);
|
|
if (!dev) {
|
|
err = -ENOMEM;
|
|
goto end;
|
|
}
|
|
hw = to_txe_hw(dev);
|
|
hw->mem_addr = pcim_iomap_table(pdev);
|
|
|
|
pci_enable_msi(pdev);
|
|
|
|
/* clear spurious interrupts */
|
|
mei_clear_interrupts(dev);
|
|
|
|
/* request and enable interrupt */
|
|
if (pci_dev_msi_enabled(pdev))
|
|
err = request_threaded_irq(pdev->irq,
|
|
NULL,
|
|
mei_txe_irq_thread_handler,
|
|
IRQF_ONESHOT, KBUILD_MODNAME, dev);
|
|
else
|
|
err = request_threaded_irq(pdev->irq,
|
|
mei_txe_irq_quick_handler,
|
|
mei_txe_irq_thread_handler,
|
|
IRQF_SHARED, KBUILD_MODNAME, dev);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "mei: request_threaded_irq failure. irq = %d\n",
|
|
pdev->irq);
|
|
goto end;
|
|
}
|
|
|
|
if (mei_start(dev)) {
|
|
dev_err(&pdev->dev, "init hw failure.\n");
|
|
err = -ENODEV;
|
|
goto release_irq;
|
|
}
|
|
|
|
pm_runtime_set_autosuspend_delay(&pdev->dev, MEI_TXI_RPM_TIMEOUT);
|
|
pm_runtime_use_autosuspend(&pdev->dev);
|
|
|
|
err = mei_register(dev, &pdev->dev);
|
|
if (err)
|
|
goto stop;
|
|
|
|
pci_set_drvdata(pdev, dev);
|
|
|
|
/*
|
|
* MEI requires to resume from runtime suspend mode
|
|
* in order to perform link reset flow upon system suspend.
|
|
*/
|
|
dev_pm_set_driver_flags(&pdev->dev, DPM_FLAG_NO_DIRECT_COMPLETE);
|
|
|
|
/*
|
|
* TXE maps runtime suspend/resume to own power gating states,
|
|
* hence we need to go around native PCI runtime service which
|
|
* eventually brings the device into D3cold/hot state.
|
|
* But the TXE device cannot wake up from D3 unlike from own
|
|
* power gating. To get around PCI device native runtime pm,
|
|
* TXE uses runtime pm domain handlers which take precedence.
|
|
*/
|
|
mei_txe_set_pm_domain(dev);
|
|
|
|
pm_runtime_put_noidle(&pdev->dev);
|
|
|
|
return 0;
|
|
|
|
stop:
|
|
mei_stop(dev);
|
|
release_irq:
|
|
mei_cancel_work(dev);
|
|
mei_disable_interrupts(dev);
|
|
free_irq(pdev->irq, dev);
|
|
end:
|
|
dev_err(&pdev->dev, "initialization failed.\n");
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* mei_txe_remove - Device Shutdown Routine
|
|
*
|
|
* @pdev: PCI device structure
|
|
*
|
|
* mei_txe_shutdown is called from the reboot notifier
|
|
* it's a simplified version of remove so we go down
|
|
* faster.
|
|
*/
|
|
static void mei_txe_shutdown(struct pci_dev *pdev)
|
|
{
|
|
struct mei_device *dev;
|
|
|
|
dev = pci_get_drvdata(pdev);
|
|
if (!dev)
|
|
return;
|
|
|
|
dev_dbg(&pdev->dev, "shutdown\n");
|
|
mei_stop(dev);
|
|
|
|
mei_txe_unset_pm_domain(dev);
|
|
|
|
mei_disable_interrupts(dev);
|
|
free_irq(pdev->irq, dev);
|
|
}
|
|
|
|
/**
|
|
* mei_txe_remove - Device Removal Routine
|
|
*
|
|
* @pdev: PCI device structure
|
|
*
|
|
* mei_remove is called by the PCI subsystem to alert the driver
|
|
* that it should release a PCI device.
|
|
*/
|
|
static void mei_txe_remove(struct pci_dev *pdev)
|
|
{
|
|
struct mei_device *dev;
|
|
|
|
dev = pci_get_drvdata(pdev);
|
|
if (!dev) {
|
|
dev_err(&pdev->dev, "mei: dev == NULL\n");
|
|
return;
|
|
}
|
|
|
|
pm_runtime_get_noresume(&pdev->dev);
|
|
|
|
mei_stop(dev);
|
|
|
|
mei_txe_unset_pm_domain(dev);
|
|
|
|
mei_disable_interrupts(dev);
|
|
free_irq(pdev->irq, dev);
|
|
|
|
mei_deregister(dev);
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
static int mei_txe_pci_suspend(struct device *device)
|
|
{
|
|
struct pci_dev *pdev = to_pci_dev(device);
|
|
struct mei_device *dev = pci_get_drvdata(pdev);
|
|
|
|
if (!dev)
|
|
return -ENODEV;
|
|
|
|
dev_dbg(&pdev->dev, "suspend\n");
|
|
|
|
mei_stop(dev);
|
|
|
|
mei_disable_interrupts(dev);
|
|
|
|
free_irq(pdev->irq, dev);
|
|
pci_disable_msi(pdev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mei_txe_pci_resume(struct device *device)
|
|
{
|
|
struct pci_dev *pdev = to_pci_dev(device);
|
|
struct mei_device *dev;
|
|
int err;
|
|
|
|
dev = pci_get_drvdata(pdev);
|
|
if (!dev)
|
|
return -ENODEV;
|
|
|
|
pci_enable_msi(pdev);
|
|
|
|
mei_clear_interrupts(dev);
|
|
|
|
/* request and enable interrupt */
|
|
if (pci_dev_msi_enabled(pdev))
|
|
err = request_threaded_irq(pdev->irq,
|
|
NULL,
|
|
mei_txe_irq_thread_handler,
|
|
IRQF_ONESHOT, KBUILD_MODNAME, dev);
|
|
else
|
|
err = request_threaded_irq(pdev->irq,
|
|
mei_txe_irq_quick_handler,
|
|
mei_txe_irq_thread_handler,
|
|
IRQF_SHARED, KBUILD_MODNAME, dev);
|
|
if (err) {
|
|
dev_err(&pdev->dev, "request_threaded_irq failed: irq = %d.\n",
|
|
pdev->irq);
|
|
return err;
|
|
}
|
|
|
|
err = mei_restart(dev);
|
|
|
|
return err;
|
|
}
|
|
#endif /* CONFIG_PM_SLEEP */
|
|
|
|
#ifdef CONFIG_PM
|
|
static int mei_txe_pm_runtime_idle(struct device *device)
|
|
{
|
|
struct mei_device *dev;
|
|
|
|
dev_dbg(device, "rpm: txe: runtime_idle\n");
|
|
|
|
dev = dev_get_drvdata(device);
|
|
if (!dev)
|
|
return -ENODEV;
|
|
if (mei_write_is_idle(dev))
|
|
pm_runtime_autosuspend(device);
|
|
|
|
return -EBUSY;
|
|
}
|
|
static int mei_txe_pm_runtime_suspend(struct device *device)
|
|
{
|
|
struct mei_device *dev;
|
|
int ret;
|
|
|
|
dev_dbg(device, "rpm: txe: runtime suspend\n");
|
|
|
|
dev = dev_get_drvdata(device);
|
|
if (!dev)
|
|
return -ENODEV;
|
|
|
|
mutex_lock(&dev->device_lock);
|
|
|
|
if (mei_write_is_idle(dev))
|
|
ret = mei_txe_aliveness_set_sync(dev, 0);
|
|
else
|
|
ret = -EAGAIN;
|
|
|
|
/* keep irq on we are staying in D0 */
|
|
|
|
dev_dbg(device, "rpm: txe: runtime suspend ret=%d\n", ret);
|
|
|
|
mutex_unlock(&dev->device_lock);
|
|
|
|
if (ret && ret != -EAGAIN)
|
|
schedule_work(&dev->reset_work);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int mei_txe_pm_runtime_resume(struct device *device)
|
|
{
|
|
struct mei_device *dev;
|
|
int ret;
|
|
|
|
dev_dbg(device, "rpm: txe: runtime resume\n");
|
|
|
|
dev = dev_get_drvdata(device);
|
|
if (!dev)
|
|
return -ENODEV;
|
|
|
|
mutex_lock(&dev->device_lock);
|
|
|
|
mei_enable_interrupts(dev);
|
|
|
|
ret = mei_txe_aliveness_set_sync(dev, 1);
|
|
|
|
mutex_unlock(&dev->device_lock);
|
|
|
|
dev_dbg(device, "rpm: txe: runtime resume ret = %d\n", ret);
|
|
|
|
if (ret)
|
|
schedule_work(&dev->reset_work);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* mei_txe_set_pm_domain - fill and set pm domain structure for device
|
|
*
|
|
* @dev: mei_device
|
|
*/
|
|
static inline void mei_txe_set_pm_domain(struct mei_device *dev)
|
|
{
|
|
struct pci_dev *pdev = to_pci_dev(dev->dev);
|
|
|
|
if (pdev->dev.bus && pdev->dev.bus->pm) {
|
|
dev->pg_domain.ops = *pdev->dev.bus->pm;
|
|
|
|
dev->pg_domain.ops.runtime_suspend = mei_txe_pm_runtime_suspend;
|
|
dev->pg_domain.ops.runtime_resume = mei_txe_pm_runtime_resume;
|
|
dev->pg_domain.ops.runtime_idle = mei_txe_pm_runtime_idle;
|
|
|
|
dev_pm_domain_set(&pdev->dev, &dev->pg_domain);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* mei_txe_unset_pm_domain - clean pm domain structure for device
|
|
*
|
|
* @dev: mei_device
|
|
*/
|
|
static inline void mei_txe_unset_pm_domain(struct mei_device *dev)
|
|
{
|
|
/* stop using pm callbacks if any */
|
|
dev_pm_domain_set(dev->dev, NULL);
|
|
}
|
|
|
|
static const struct dev_pm_ops mei_txe_pm_ops = {
|
|
SET_SYSTEM_SLEEP_PM_OPS(mei_txe_pci_suspend,
|
|
mei_txe_pci_resume)
|
|
SET_RUNTIME_PM_OPS(
|
|
mei_txe_pm_runtime_suspend,
|
|
mei_txe_pm_runtime_resume,
|
|
mei_txe_pm_runtime_idle)
|
|
};
|
|
|
|
#define MEI_TXE_PM_OPS (&mei_txe_pm_ops)
|
|
#else
|
|
#define MEI_TXE_PM_OPS NULL
|
|
#endif /* CONFIG_PM */
|
|
|
|
/*
|
|
* PCI driver structure
|
|
*/
|
|
static struct pci_driver mei_txe_driver = {
|
|
.name = KBUILD_MODNAME,
|
|
.id_table = mei_txe_pci_tbl,
|
|
.probe = mei_txe_probe,
|
|
.remove = mei_txe_remove,
|
|
.shutdown = mei_txe_shutdown,
|
|
.driver.pm = MEI_TXE_PM_OPS,
|
|
};
|
|
|
|
module_pci_driver(mei_txe_driver);
|
|
|
|
MODULE_AUTHOR("Intel Corporation");
|
|
MODULE_DESCRIPTION("Intel(R) Trusted Execution Environment Interface");
|
|
MODULE_LICENSE("GPL v2");
|