x86: coding style fixes to arch/x86/lib/usercopy_32.c
Before: total: 63 errors, 2 warnings, 878 lines checked After: total: 0 errors, 2 warnings, 878 lines checked Compile tested, no change in the binary output: text data bss dec hex filename 3231 0 0 3231 c9f usercopy_32.o.after 3231 0 0 3231 c9f usercopy_32.o.before md5sum: 9f9a3eb43970359ae7cecfd1c9e7cf42 usercopy_32.o.after 9f9a3eb43970359ae7cecfd1c9e7cf42 usercopy_32.o.before Signed-off-by: Paolo Ciarrocchi <paolo.ciarrocchi@gmail.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
Родитель
bdd3cee2e4
Коммит
3f50dbc1ae
|
@ -22,14 +22,14 @@ static inline int __movsl_is_ok(unsigned long a1, unsigned long a2, unsigned lon
|
|||
#endif
|
||||
return 1;
|
||||
}
|
||||
#define movsl_is_ok(a1,a2,n) \
|
||||
__movsl_is_ok((unsigned long)(a1),(unsigned long)(a2),(n))
|
||||
#define movsl_is_ok(a1, a2, n) \
|
||||
__movsl_is_ok((unsigned long)(a1), (unsigned long)(a2), (n))
|
||||
|
||||
/*
|
||||
* Copy a null terminated string from userspace.
|
||||
*/
|
||||
|
||||
#define __do_strncpy_from_user(dst,src,count,res) \
|
||||
#define __do_strncpy_from_user(dst, src, count, res) \
|
||||
do { \
|
||||
int __d0, __d1, __d2; \
|
||||
might_sleep(); \
|
||||
|
@ -120,7 +120,7 @@ EXPORT_SYMBOL(strncpy_from_user);
|
|||
do { \
|
||||
int __d0; \
|
||||
might_sleep(); \
|
||||
__asm__ __volatile__( \
|
||||
__asm__ __volatile__( \
|
||||
"0: rep; stosl\n" \
|
||||
" movl %2,%0\n" \
|
||||
"1: rep; stosb\n" \
|
||||
|
@ -429,7 +429,7 @@ __copy_user_zeroing_intel(void *to, const void __user *from, unsigned long size)
|
|||
static unsigned long __copy_user_zeroing_intel_nocache(void *to,
|
||||
const void __user *from, unsigned long size)
|
||||
{
|
||||
int d0, d1;
|
||||
int d0, d1;
|
||||
|
||||
__asm__ __volatile__(
|
||||
" .align 2,0x90\n"
|
||||
|
@ -526,7 +526,7 @@ static unsigned long __copy_user_zeroing_intel_nocache(void *to,
|
|||
static unsigned long __copy_user_intel_nocache(void *to,
|
||||
const void __user *from, unsigned long size)
|
||||
{
|
||||
int d0, d1;
|
||||
int d0, d1;
|
||||
|
||||
__asm__ __volatile__(
|
||||
" .align 2,0x90\n"
|
||||
|
@ -629,7 +629,7 @@ unsigned long __copy_user_zeroing_intel_nocache(void *to,
|
|||
#endif /* CONFIG_X86_INTEL_USERCOPY */
|
||||
|
||||
/* Generic arbitrary sized copy. */
|
||||
#define __copy_user(to,from,size) \
|
||||
#define __copy_user(to, from, size) \
|
||||
do { \
|
||||
int __d0, __d1, __d2; \
|
||||
__asm__ __volatile__( \
|
||||
|
@ -665,7 +665,7 @@ do { \
|
|||
: "memory"); \
|
||||
} while (0)
|
||||
|
||||
#define __copy_user_zeroing(to,from,size) \
|
||||
#define __copy_user_zeroing(to, from, size) \
|
||||
do { \
|
||||
int __d0, __d1, __d2; \
|
||||
__asm__ __volatile__( \
|
||||
|
@ -712,7 +712,7 @@ unsigned long __copy_to_user_ll(void __user *to, const void *from,
|
|||
{
|
||||
#ifndef CONFIG_X86_WP_WORKS_OK
|
||||
if (unlikely(boot_cpu_data.wp_works_ok == 0) &&
|
||||
((unsigned long )to) < TASK_SIZE) {
|
||||
((unsigned long)to) < TASK_SIZE) {
|
||||
/*
|
||||
* When we are in an atomic section (see
|
||||
* mm/filemap.c:file_read_actor), return the full
|
||||
|
@ -728,7 +728,7 @@ unsigned long __copy_to_user_ll(void __user *to, const void *from,
|
|||
* Do it manually. Manfred <manfred@colorfullife.com>
|
||||
*/
|
||||
while (n) {
|
||||
unsigned long offset = ((unsigned long)to)%PAGE_SIZE;
|
||||
unsigned long offset = ((unsigned long)to)%PAGE_SIZE;
|
||||
unsigned long len = PAGE_SIZE - offset;
|
||||
int retval;
|
||||
struct page *pg;
|
||||
|
@ -740,7 +740,7 @@ unsigned long __copy_to_user_ll(void __user *to, const void *from,
|
|||
survive:
|
||||
down_read(¤t->mm->mmap_sem);
|
||||
retval = get_user_pages(current, current->mm,
|
||||
(unsigned long )to, 1, 1, 0, &pg, NULL);
|
||||
(unsigned long)to, 1, 1, 0, &pg, NULL);
|
||||
|
||||
if (retval == -ENOMEM && is_global_init(current)) {
|
||||
up_read(¤t->mm->mmap_sem);
|
||||
|
@ -750,8 +750,8 @@ survive:
|
|||
|
||||
if (retval != 1) {
|
||||
up_read(¤t->mm->mmap_sem);
|
||||
break;
|
||||
}
|
||||
break;
|
||||
}
|
||||
|
||||
maddr = kmap_atomic(pg, KM_USER0);
|
||||
memcpy(maddr + offset, from, len);
|
||||
|
@ -802,12 +802,12 @@ unsigned long __copy_from_user_ll_nocache(void *to, const void __user *from,
|
|||
unsigned long n)
|
||||
{
|
||||
#ifdef CONFIG_X86_INTEL_USERCOPY
|
||||
if ( n > 64 && cpu_has_xmm2)
|
||||
n = __copy_user_zeroing_intel_nocache(to, from, n);
|
||||
if (n > 64 && cpu_has_xmm2)
|
||||
n = __copy_user_zeroing_intel_nocache(to, from, n);
|
||||
else
|
||||
__copy_user_zeroing(to, from, n);
|
||||
#else
|
||||
__copy_user_zeroing(to, from, n);
|
||||
__copy_user_zeroing(to, from, n);
|
||||
#endif
|
||||
return n;
|
||||
}
|
||||
|
@ -817,12 +817,12 @@ unsigned long __copy_from_user_ll_nocache_nozero(void *to, const void __user *fr
|
|||
unsigned long n)
|
||||
{
|
||||
#ifdef CONFIG_X86_INTEL_USERCOPY
|
||||
if ( n > 64 && cpu_has_xmm2)
|
||||
n = __copy_user_intel_nocache(to, from, n);
|
||||
if (n > 64 && cpu_has_xmm2)
|
||||
n = __copy_user_intel_nocache(to, from, n);
|
||||
else
|
||||
__copy_user(to, from, n);
|
||||
#else
|
||||
__copy_user(to, from, n);
|
||||
__copy_user(to, from, n);
|
||||
#endif
|
||||
return n;
|
||||
}
|
||||
|
|
Загрузка…
Ссылка в новой задаче