if (map_coherent)
                        vaddr = kmap_coherent(page, address);
                else
-                       vaddr = kmap_atomic(page, KM_USER0);
+                       vaddr = kmap_atomic(page);
 
                address = (unsigned long)vaddr;
        }
                if (map_coherent)
                        kunmap_coherent(vaddr);
                else
-                       kunmap_atomic(vaddr, KM_USER0);
+                       kunmap_atomic(vaddr);
        }
 }
 
 
 {
        void *vfrom, *vto;
 
-       vto = kmap_atomic(to, KM_USER1);
+       vto = kmap_atomic(to);
 
        if (boot_cpu_data.dcache.n_aliases && page_mapped(from) &&
            test_bit(PG_dcache_clean, &from->flags)) {
                copy_page(vto, vfrom);
                kunmap_coherent(vfrom);
        } else {
-               vfrom = kmap_atomic(from, KM_USER0);
+               vfrom = kmap_atomic(from);
                copy_page(vto, vfrom);
-               kunmap_atomic(vfrom, KM_USER0);
+               kunmap_atomic(vfrom);
        }
 
        if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) ||
            (vma->vm_flags & VM_EXEC))
                __flush_purge_region(vto, PAGE_SIZE);
 
-       kunmap_atomic(vto, KM_USER1);
+       kunmap_atomic(vto);
        /* Make sure this page is cleared on other CPU's too before using it */
        smp_wmb();
 }
 
 void clear_user_highpage(struct page *page, unsigned long vaddr)
 {
-       void *kaddr = kmap_atomic(page, KM_USER0);
+       void *kaddr = kmap_atomic(page);
 
        clear_page(kaddr);
 
        if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK))
                __flush_purge_region(kaddr, PAGE_SIZE);
 
-       kunmap_atomic(kaddr, KM_USER0);
+       kunmap_atomic(kaddr);
 }
 EXPORT_SYMBOL(clear_user_highpage);