thunderbolt: Fixes for v5.2-rc6
This includes two fixes for issues found during the current release cycle: - Fix runtime PM regression when device is authorized after the controller is runtime suspended. - Correct CIO reset flow for Titan Ridge. -----BEGIN PGP SIGNATURE----- iQJUBAABCgA+FiEEVTdhRGBbNzLrSUBaAP2fSd+ZWKAFAl0KEaIgHG1pa2Eud2Vz dGVyYmVyZ0BsaW51eC5pbnRlbC5jb20ACgkQAP2fSd+ZWKCPwA/9FGIPckiFJNK8 6eCUs7ChMeAc+hb1MJZqd0MRtwfyNGl63Ip8+kC+Gqc66C56yOs+pLNJ/p2KdR+q b7ziOLbgbEa9Iw8GMCrvDSzF+KKCgKmIXOwREf9LCKvh5AyTc2HpitLzMa8NTLP7 6fwHsmDVFIjsJj8nXBMQHhKlC+KDaGdpcEmrW6Kb/3DcoZf/5HcGVUdrq3cCQyQi vw39IR9mIs5HHLThP3l33w2QCPyIvsdYEzgQCWWvs9hyjVE8cnf7HiD998fa1p2P S5VBZDr/Ykfq1ZKFJBPV5gjuuKSh1/skTG+quCOUgdgotk8GsY9qX93w4reKFs6X TjKbQWuyzgZ0te3e5oZpM91qShypfrQvGO406u7d9gmFW4PJBL76ovme2OtxwKdK N2PK21fWOMZRpnP9MuvJz8azQ+d3x2GdpsNKWIVnNSbTbd69UJ/hDSVhjqTQ7GNf 3bv3XMxP0bFc5Xri0HzYsoeJbLLPUvKWMSGJgNURJ8l7dWho/SZiABs/j33v9x59 qhCTULugzpIjT2q1l3mYmCNa6b8NBPQZsj1MGtTUfsUOA4TZnrV0Uc0D376rKCQo JiUY+OR5fsr570KoKt7m+oQatDs3lMGRp2jghVqS/NJt60upsLxHe7ABS28zaItB R70rombROOtI/j+v8qc19aN1x8bysL8= =KHi8 -----END PGP SIGNATURE----- Merge tag 'thunderbolt-fixes-for-v5.2-rc6' of git://git.kernel.org/pub/scm/linux/kernel/git/westeri/thunderbolt into char-misc-linus Mika writes: thunderbolt: Fixes for v5.2-rc6 This includes two fixes for issues found during the current release cycle: - Fix runtime PM regression when device is authorized after the controller is runtime suspended. - Correct CIO reset flow for Titan Ridge. * tag 'thunderbolt-fixes-for-v5.2-rc6' of git://git.kernel.org/pub/scm/linux/kernel/git/westeri/thunderbolt: thunderbolt: Implement CIO reset correctly for Titan Ridge thunderbolt: Make sure device runtime resume completes before taking domain lock
This commit is contained in:
Коммит
c61802ee54
|
@ -56,6 +56,7 @@
|
|||
* @max_boot_acl: Maximum number of preboot ACL entries (%0 if not supported)
|
||||
* @rpm: Does the controller support runtime PM (RTD3)
|
||||
* @is_supported: Checks if we can support ICM on this controller
|
||||
* @cio_reset: Trigger CIO reset
|
||||
* @get_mode: Read and return the ICM firmware mode (optional)
|
||||
* @get_route: Find a route string for given switch
|
||||
* @save_devices: Ask ICM to save devices to ACL when suspending (optional)
|
||||
|
@ -74,6 +75,7 @@ struct icm {
|
|||
bool safe_mode;
|
||||
bool rpm;
|
||||
bool (*is_supported)(struct tb *tb);
|
||||
int (*cio_reset)(struct tb *tb);
|
||||
int (*get_mode)(struct tb *tb);
|
||||
int (*get_route)(struct tb *tb, u8 link, u8 depth, u64 *route);
|
||||
void (*save_devices)(struct tb *tb);
|
||||
|
@ -166,6 +168,65 @@ static inline u64 get_parent_route(u64 route)
|
|||
return depth ? route & ~(0xffULL << (depth - 1) * TB_ROUTE_SHIFT) : 0;
|
||||
}
|
||||
|
||||
static int pci2cio_wait_completion(struct icm *icm, unsigned long timeout_msec)
|
||||
{
|
||||
unsigned long end = jiffies + msecs_to_jiffies(timeout_msec);
|
||||
u32 cmd;
|
||||
|
||||
do {
|
||||
pci_read_config_dword(icm->upstream_port,
|
||||
icm->vnd_cap + PCIE2CIO_CMD, &cmd);
|
||||
if (!(cmd & PCIE2CIO_CMD_START)) {
|
||||
if (cmd & PCIE2CIO_CMD_TIMEOUT)
|
||||
break;
|
||||
return 0;
|
||||
}
|
||||
|
||||
msleep(50);
|
||||
} while (time_before(jiffies, end));
|
||||
|
||||
return -ETIMEDOUT;
|
||||
}
|
||||
|
||||
static int pcie2cio_read(struct icm *icm, enum tb_cfg_space cs,
|
||||
unsigned int port, unsigned int index, u32 *data)
|
||||
{
|
||||
struct pci_dev *pdev = icm->upstream_port;
|
||||
int ret, vnd_cap = icm->vnd_cap;
|
||||
u32 cmd;
|
||||
|
||||
cmd = index;
|
||||
cmd |= (port << PCIE2CIO_CMD_PORT_SHIFT) & PCIE2CIO_CMD_PORT_MASK;
|
||||
cmd |= (cs << PCIE2CIO_CMD_CS_SHIFT) & PCIE2CIO_CMD_CS_MASK;
|
||||
cmd |= PCIE2CIO_CMD_START;
|
||||
pci_write_config_dword(pdev, vnd_cap + PCIE2CIO_CMD, cmd);
|
||||
|
||||
ret = pci2cio_wait_completion(icm, 5000);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
pci_read_config_dword(pdev, vnd_cap + PCIE2CIO_RDDATA, data);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pcie2cio_write(struct icm *icm, enum tb_cfg_space cs,
|
||||
unsigned int port, unsigned int index, u32 data)
|
||||
{
|
||||
struct pci_dev *pdev = icm->upstream_port;
|
||||
int vnd_cap = icm->vnd_cap;
|
||||
u32 cmd;
|
||||
|
||||
pci_write_config_dword(pdev, vnd_cap + PCIE2CIO_WRDATA, data);
|
||||
|
||||
cmd = index;
|
||||
cmd |= (port << PCIE2CIO_CMD_PORT_SHIFT) & PCIE2CIO_CMD_PORT_MASK;
|
||||
cmd |= (cs << PCIE2CIO_CMD_CS_SHIFT) & PCIE2CIO_CMD_CS_MASK;
|
||||
cmd |= PCIE2CIO_CMD_WRITE | PCIE2CIO_CMD_START;
|
||||
pci_write_config_dword(pdev, vnd_cap + PCIE2CIO_CMD, cmd);
|
||||
|
||||
return pci2cio_wait_completion(icm, 5000);
|
||||
}
|
||||
|
||||
static bool icm_match(const struct tb_cfg_request *req,
|
||||
const struct ctl_pkg *pkg)
|
||||
{
|
||||
|
@ -484,6 +545,7 @@ static void add_switch(struct tb_switch *parent_sw, u64 route,
|
|||
sw->authorized = authorized;
|
||||
sw->security_level = security_level;
|
||||
sw->boot = boot;
|
||||
init_completion(&sw->rpm_complete);
|
||||
|
||||
vss = parse_intel_vss(ep_name, ep_name_size);
|
||||
if (vss)
|
||||
|
@ -523,6 +585,9 @@ static void update_switch(struct tb_switch *parent_sw, struct tb_switch *sw,
|
|||
|
||||
/* This switch still exists */
|
||||
sw->is_unplugged = false;
|
||||
|
||||
/* Runtime resume is now complete */
|
||||
complete(&sw->rpm_complete);
|
||||
}
|
||||
|
||||
static void remove_switch(struct tb_switch *sw)
|
||||
|
@ -834,6 +899,11 @@ icm_fr_xdomain_disconnected(struct tb *tb, const struct icm_pkg_header *hdr)
|
|||
}
|
||||
}
|
||||
|
||||
static int icm_tr_cio_reset(struct tb *tb)
|
||||
{
|
||||
return pcie2cio_write(tb_priv(tb), TB_CFG_SWITCH, 0, 0x777, BIT(1));
|
||||
}
|
||||
|
||||
static int
|
||||
icm_tr_driver_ready(struct tb *tb, enum tb_security_level *security_level,
|
||||
size_t *nboot_acl, bool *rpm)
|
||||
|
@ -1240,6 +1310,11 @@ static bool icm_ar_is_supported(struct tb *tb)
|
|||
return false;
|
||||
}
|
||||
|
||||
static int icm_ar_cio_reset(struct tb *tb)
|
||||
{
|
||||
return pcie2cio_write(tb_priv(tb), TB_CFG_SWITCH, 0, 0x50, BIT(9));
|
||||
}
|
||||
|
||||
static int icm_ar_get_mode(struct tb *tb)
|
||||
{
|
||||
struct tb_nhi *nhi = tb->nhi;
|
||||
|
@ -1477,65 +1552,6 @@ __icm_driver_ready(struct tb *tb, enum tb_security_level *security_level,
|
|||
return -ETIMEDOUT;
|
||||
}
|
||||
|
||||
static int pci2cio_wait_completion(struct icm *icm, unsigned long timeout_msec)
|
||||
{
|
||||
unsigned long end = jiffies + msecs_to_jiffies(timeout_msec);
|
||||
u32 cmd;
|
||||
|
||||
do {
|
||||
pci_read_config_dword(icm->upstream_port,
|
||||
icm->vnd_cap + PCIE2CIO_CMD, &cmd);
|
||||
if (!(cmd & PCIE2CIO_CMD_START)) {
|
||||
if (cmd & PCIE2CIO_CMD_TIMEOUT)
|
||||
break;
|
||||
return 0;
|
||||
}
|
||||
|
||||
msleep(50);
|
||||
} while (time_before(jiffies, end));
|
||||
|
||||
return -ETIMEDOUT;
|
||||
}
|
||||
|
||||
static int pcie2cio_read(struct icm *icm, enum tb_cfg_space cs,
|
||||
unsigned int port, unsigned int index, u32 *data)
|
||||
{
|
||||
struct pci_dev *pdev = icm->upstream_port;
|
||||
int ret, vnd_cap = icm->vnd_cap;
|
||||
u32 cmd;
|
||||
|
||||
cmd = index;
|
||||
cmd |= (port << PCIE2CIO_CMD_PORT_SHIFT) & PCIE2CIO_CMD_PORT_MASK;
|
||||
cmd |= (cs << PCIE2CIO_CMD_CS_SHIFT) & PCIE2CIO_CMD_CS_MASK;
|
||||
cmd |= PCIE2CIO_CMD_START;
|
||||
pci_write_config_dword(pdev, vnd_cap + PCIE2CIO_CMD, cmd);
|
||||
|
||||
ret = pci2cio_wait_completion(icm, 5000);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
pci_read_config_dword(pdev, vnd_cap + PCIE2CIO_RDDATA, data);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int pcie2cio_write(struct icm *icm, enum tb_cfg_space cs,
|
||||
unsigned int port, unsigned int index, u32 data)
|
||||
{
|
||||
struct pci_dev *pdev = icm->upstream_port;
|
||||
int vnd_cap = icm->vnd_cap;
|
||||
u32 cmd;
|
||||
|
||||
pci_write_config_dword(pdev, vnd_cap + PCIE2CIO_WRDATA, data);
|
||||
|
||||
cmd = index;
|
||||
cmd |= (port << PCIE2CIO_CMD_PORT_SHIFT) & PCIE2CIO_CMD_PORT_MASK;
|
||||
cmd |= (cs << PCIE2CIO_CMD_CS_SHIFT) & PCIE2CIO_CMD_CS_MASK;
|
||||
cmd |= PCIE2CIO_CMD_WRITE | PCIE2CIO_CMD_START;
|
||||
pci_write_config_dword(pdev, vnd_cap + PCIE2CIO_CMD, cmd);
|
||||
|
||||
return pci2cio_wait_completion(icm, 5000);
|
||||
}
|
||||
|
||||
static int icm_firmware_reset(struct tb *tb, struct tb_nhi *nhi)
|
||||
{
|
||||
struct icm *icm = tb_priv(tb);
|
||||
|
@ -1556,7 +1572,7 @@ static int icm_firmware_reset(struct tb *tb, struct tb_nhi *nhi)
|
|||
iowrite32(val, nhi->iobase + REG_FW_STS);
|
||||
|
||||
/* Trigger CIO reset now */
|
||||
return pcie2cio_write(icm, TB_CFG_SWITCH, 0, 0x50, BIT(9));
|
||||
return icm->cio_reset(tb);
|
||||
}
|
||||
|
||||
static int icm_firmware_start(struct tb *tb, struct tb_nhi *nhi)
|
||||
|
@ -1770,6 +1786,32 @@ static void icm_unplug_children(struct tb_switch *sw)
|
|||
}
|
||||
}
|
||||
|
||||
static int complete_rpm(struct device *dev, void *data)
|
||||
{
|
||||
struct tb_switch *sw = tb_to_switch(dev);
|
||||
|
||||
if (sw)
|
||||
complete(&sw->rpm_complete);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void remove_unplugged_switch(struct tb_switch *sw)
|
||||
{
|
||||
pm_runtime_get_sync(sw->dev.parent);
|
||||
|
||||
/*
|
||||
* Signal this and switches below for rpm_complete because
|
||||
* tb_switch_remove() calls pm_runtime_get_sync() that then waits
|
||||
* for it.
|
||||
*/
|
||||
complete_rpm(&sw->dev, NULL);
|
||||
bus_for_each_dev(&tb_bus_type, &sw->dev, NULL, complete_rpm);
|
||||
tb_switch_remove(sw);
|
||||
|
||||
pm_runtime_mark_last_busy(sw->dev.parent);
|
||||
pm_runtime_put_autosuspend(sw->dev.parent);
|
||||
}
|
||||
|
||||
static void icm_free_unplugged_children(struct tb_switch *sw)
|
||||
{
|
||||
unsigned int i;
|
||||
|
@ -1782,7 +1824,7 @@ static void icm_free_unplugged_children(struct tb_switch *sw)
|
|||
port->xdomain = NULL;
|
||||
} else if (tb_port_has_remote(port)) {
|
||||
if (port->remote->sw->is_unplugged) {
|
||||
tb_switch_remove(port->remote->sw);
|
||||
remove_unplugged_switch(port->remote->sw);
|
||||
port->remote = NULL;
|
||||
} else {
|
||||
icm_free_unplugged_children(port->remote->sw);
|
||||
|
@ -1831,6 +1873,24 @@ static int icm_runtime_suspend(struct tb *tb)
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int icm_runtime_suspend_switch(struct tb_switch *sw)
|
||||
{
|
||||
if (tb_route(sw))
|
||||
reinit_completion(&sw->rpm_complete);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int icm_runtime_resume_switch(struct tb_switch *sw)
|
||||
{
|
||||
if (tb_route(sw)) {
|
||||
if (!wait_for_completion_timeout(&sw->rpm_complete,
|
||||
msecs_to_jiffies(500))) {
|
||||
dev_dbg(&sw->dev, "runtime resuming timed out\n");
|
||||
}
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int icm_runtime_resume(struct tb *tb)
|
||||
{
|
||||
/*
|
||||
|
@ -1910,6 +1970,8 @@ static const struct tb_cm_ops icm_ar_ops = {
|
|||
.complete = icm_complete,
|
||||
.runtime_suspend = icm_runtime_suspend,
|
||||
.runtime_resume = icm_runtime_resume,
|
||||
.runtime_suspend_switch = icm_runtime_suspend_switch,
|
||||
.runtime_resume_switch = icm_runtime_resume_switch,
|
||||
.handle_event = icm_handle_event,
|
||||
.get_boot_acl = icm_ar_get_boot_acl,
|
||||
.set_boot_acl = icm_ar_set_boot_acl,
|
||||
|
@ -1930,6 +1992,8 @@ static const struct tb_cm_ops icm_tr_ops = {
|
|||
.complete = icm_complete,
|
||||
.runtime_suspend = icm_runtime_suspend,
|
||||
.runtime_resume = icm_runtime_resume,
|
||||
.runtime_suspend_switch = icm_runtime_suspend_switch,
|
||||
.runtime_resume_switch = icm_runtime_resume_switch,
|
||||
.handle_event = icm_handle_event,
|
||||
.get_boot_acl = icm_ar_get_boot_acl,
|
||||
.set_boot_acl = icm_ar_set_boot_acl,
|
||||
|
@ -1975,6 +2039,7 @@ struct tb *icm_probe(struct tb_nhi *nhi)
|
|||
case PCI_DEVICE_ID_INTEL_ALPINE_RIDGE_C_2C_NHI:
|
||||
icm->max_boot_acl = ICM_AR_PREBOOT_ACL_ENTRIES;
|
||||
icm->is_supported = icm_ar_is_supported;
|
||||
icm->cio_reset = icm_ar_cio_reset;
|
||||
icm->get_mode = icm_ar_get_mode;
|
||||
icm->get_route = icm_ar_get_route;
|
||||
icm->save_devices = icm_fr_save_devices;
|
||||
|
@ -1990,6 +2055,7 @@ struct tb *icm_probe(struct tb_nhi *nhi)
|
|||
case PCI_DEVICE_ID_INTEL_TITAN_RIDGE_4C_NHI:
|
||||
icm->max_boot_acl = ICM_AR_PREBOOT_ACL_ENTRIES;
|
||||
icm->is_supported = icm_ar_is_supported;
|
||||
icm->cio_reset = icm_tr_cio_reset;
|
||||
icm->get_mode = icm_ar_get_mode;
|
||||
icm->driver_ready = icm_tr_driver_ready;
|
||||
icm->device_connected = icm_tr_device_connected;
|
||||
|
|
|
@ -239,7 +239,16 @@ static int tb_switch_nvm_read(void *priv, unsigned int offset, void *val,
|
|||
int ret;
|
||||
|
||||
pm_runtime_get_sync(&sw->dev);
|
||||
|
||||
if (!mutex_trylock(&sw->tb->lock)) {
|
||||
ret = restart_syscall();
|
||||
goto out;
|
||||
}
|
||||
|
||||
ret = dma_port_flash_read(sw->dma_port, offset, val, bytes);
|
||||
mutex_unlock(&sw->tb->lock);
|
||||
|
||||
out:
|
||||
pm_runtime_mark_last_busy(&sw->dev);
|
||||
pm_runtime_put_autosuspend(&sw->dev);
|
||||
|
||||
|
@ -1019,7 +1028,6 @@ static int tb_switch_set_authorized(struct tb_switch *sw, unsigned int val)
|
|||
* the new tunnel too early.
|
||||
*/
|
||||
pci_lock_rescan_remove();
|
||||
pm_runtime_get_sync(&sw->dev);
|
||||
|
||||
switch (val) {
|
||||
/* Approve switch */
|
||||
|
@ -1040,8 +1048,6 @@ static int tb_switch_set_authorized(struct tb_switch *sw, unsigned int val)
|
|||
break;
|
||||
}
|
||||
|
||||
pm_runtime_mark_last_busy(&sw->dev);
|
||||
pm_runtime_put_autosuspend(&sw->dev);
|
||||
pci_unlock_rescan_remove();
|
||||
|
||||
if (!ret) {
|
||||
|
@ -1069,7 +1075,10 @@ static ssize_t authorized_store(struct device *dev,
|
|||
if (val > 2)
|
||||
return -EINVAL;
|
||||
|
||||
pm_runtime_get_sync(&sw->dev);
|
||||
ret = tb_switch_set_authorized(sw, val);
|
||||
pm_runtime_mark_last_busy(&sw->dev);
|
||||
pm_runtime_put_autosuspend(&sw->dev);
|
||||
|
||||
return ret ? ret : count;
|
||||
}
|
||||
|
@ -1195,8 +1204,12 @@ static ssize_t nvm_authenticate_store(struct device *dev,
|
|||
bool val;
|
||||
int ret;
|
||||
|
||||
if (!mutex_trylock(&sw->tb->lock))
|
||||
return restart_syscall();
|
||||
pm_runtime_get_sync(&sw->dev);
|
||||
|
||||
if (!mutex_trylock(&sw->tb->lock)) {
|
||||
ret = restart_syscall();
|
||||
goto exit_rpm;
|
||||
}
|
||||
|
||||
/* If NVMem devices are not yet added */
|
||||
if (!sw->nvm) {
|
||||
|
@ -1217,13 +1230,9 @@ static ssize_t nvm_authenticate_store(struct device *dev,
|
|||
goto exit_unlock;
|
||||
}
|
||||
|
||||
pm_runtime_get_sync(&sw->dev);
|
||||
ret = nvm_validate_and_write(sw);
|
||||
if (ret) {
|
||||
pm_runtime_mark_last_busy(&sw->dev);
|
||||
pm_runtime_put_autosuspend(&sw->dev);
|
||||
if (ret)
|
||||
goto exit_unlock;
|
||||
}
|
||||
|
||||
sw->nvm->authenticating = true;
|
||||
|
||||
|
@ -1239,12 +1248,13 @@ static ssize_t nvm_authenticate_store(struct device *dev,
|
|||
} else {
|
||||
ret = nvm_authenticate_device(sw);
|
||||
}
|
||||
pm_runtime_mark_last_busy(&sw->dev);
|
||||
pm_runtime_put_autosuspend(&sw->dev);
|
||||
}
|
||||
|
||||
exit_unlock:
|
||||
mutex_unlock(&sw->tb->lock);
|
||||
exit_rpm:
|
||||
pm_runtime_mark_last_busy(&sw->dev);
|
||||
pm_runtime_put_autosuspend(&sw->dev);
|
||||
|
||||
if (ret)
|
||||
return ret;
|
||||
|
@ -1380,11 +1390,22 @@ static void tb_switch_release(struct device *dev)
|
|||
*/
|
||||
static int __maybe_unused tb_switch_runtime_suspend(struct device *dev)
|
||||
{
|
||||
struct tb_switch *sw = tb_to_switch(dev);
|
||||
const struct tb_cm_ops *cm_ops = sw->tb->cm_ops;
|
||||
|
||||
if (cm_ops->runtime_suspend_switch)
|
||||
return cm_ops->runtime_suspend_switch(sw);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int __maybe_unused tb_switch_runtime_resume(struct device *dev)
|
||||
{
|
||||
struct tb_switch *sw = tb_to_switch(dev);
|
||||
const struct tb_cm_ops *cm_ops = sw->tb->cm_ops;
|
||||
|
||||
if (cm_ops->runtime_resume_switch)
|
||||
return cm_ops->runtime_resume_switch(sw);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
|
@ -79,6 +79,8 @@ struct tb_switch_nvm {
|
|||
* @connection_key: Connection key used with ICM messaging
|
||||
* @link: Root switch link this switch is connected (ICM only)
|
||||
* @depth: Depth in the chain this switch is connected (ICM only)
|
||||
* @rpm_complete: Completion used to wait for runtime resume to
|
||||
* complete (ICM only)
|
||||
*
|
||||
* When the switch is being added or removed to the domain (other
|
||||
* switches) you need to have domain lock held.
|
||||
|
@ -112,6 +114,7 @@ struct tb_switch {
|
|||
u8 connection_key;
|
||||
u8 link;
|
||||
u8 depth;
|
||||
struct completion rpm_complete;
|
||||
};
|
||||
|
||||
/**
|
||||
|
@ -250,6 +253,8 @@ struct tb_path {
|
|||
* @complete: Connection manager specific complete
|
||||
* @runtime_suspend: Connection manager specific runtime_suspend
|
||||
* @runtime_resume: Connection manager specific runtime_resume
|
||||
* @runtime_suspend_switch: Runtime suspend a switch
|
||||
* @runtime_resume_switch: Runtime resume a switch
|
||||
* @handle_event: Handle thunderbolt event
|
||||
* @get_boot_acl: Get boot ACL list
|
||||
* @set_boot_acl: Set boot ACL list
|
||||
|
@ -270,6 +275,8 @@ struct tb_cm_ops {
|
|||
void (*complete)(struct tb *tb);
|
||||
int (*runtime_suspend)(struct tb *tb);
|
||||
int (*runtime_resume)(struct tb *tb);
|
||||
int (*runtime_suspend_switch)(struct tb_switch *sw);
|
||||
int (*runtime_resume_switch)(struct tb_switch *sw);
|
||||
void (*handle_event)(struct tb *tb, enum tb_cfg_pkg_type,
|
||||
const void *buf, size_t size);
|
||||
int (*get_boot_acl)(struct tb *tb, uuid_t *uuids, size_t nuuids);
|
||||
|
|
Загрузка…
Ссылка в новой задаче