tpm: use tpm_msleep() value as max delay
Currently, tpm_msleep() uses delay_msec as the minimum value in usleep_range. However, that is the maximum time we want to wait. The function is modified to use the delay_msec as the maximum value, not the minimum value. After this change, performance on a TPM 1.2 with an 8 byte burstcount for 1000 extends improved from ~9sec to ~8sec. Fixes: 3b9af007869("tpm: replace msleep() with usleep_range() in TPM 1.2/ 2.0 generic drivers") Signed-off-by: Nayna Jain <nayna@linux.vnet.ibm.com> Acked-by: Mimi Zohar <zohar@linux.vnet.ibm.com> Tested-by: Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com> Reviewed-by: Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com> Signed-off-by: Jarkko Sakkinen <jarkko.sakkinen@linux.intel.com>
This commit is contained in:
Родитель
cf151a9a44
Коммит
5ef924d9e2
|
@ -515,8 +515,8 @@ int tpm_pm_resume(struct device *dev);
|
||||||
|
|
||||||
static inline void tpm_msleep(unsigned int delay_msec)
|
static inline void tpm_msleep(unsigned int delay_msec)
|
||||||
{
|
{
|
||||||
usleep_range(delay_msec * 1000,
|
usleep_range((delay_msec * 1000) - TPM_TIMEOUT_RANGE_US,
|
||||||
(delay_msec * 1000) + TPM_TIMEOUT_RANGE_US);
|
delay_msec * 1000);
|
||||||
};
|
};
|
||||||
|
|
||||||
struct tpm_chip *tpm_chip_find_get(int chip_num);
|
struct tpm_chip *tpm_chip_find_get(int chip_num);
|
||||||
|
|
Загрузка…
Ссылка в новой задаче