}
        ASSERT_SENTINEL(zh, ZBH);
        BUG_ON(zh->size == 0 || zh->size > zbud_max_buddy_size());
-       to_va = kmap_atomic(page, KM_USER0);
+       to_va = kmap_atomic(page);
        size = zh->size;
        from_va = zbud_data(zh, size);
        ret = lzo1x_decompress_safe(from_va, size, to_va, &out_len);
        BUG_ON(ret != LZO_E_OK);
        BUG_ON(out_len != PAGE_SIZE);
-       kunmap_atomic(to_va, KM_USER0);
+       kunmap_atomic(to_va);
 out:
        spin_unlock(&zbpg->lock);
        return ret;
                goto out;
        atomic_inc(&zv_curr_dist_counts[chunks]);
        atomic_inc(&zv_cumul_dist_counts[chunks]);
-       zv = kmap_atomic(page, KM_USER0) + offset;
+       zv = kmap_atomic(page) + offset;
        zv->index = index;
        zv->oid = *oid;
        zv->pool_id = pool_id;
        SET_SENTINEL(zv, ZVH);
        memcpy((char *)zv + sizeof(struct zv_hdr), cdata, clen);
-       kunmap_atomic(zv, KM_USER0);
+       kunmap_atomic(zv);
 out:
        return zv;
 }
        ASSERT_SENTINEL(zv, ZVH);
        size = xv_get_object_size(zv) - sizeof(*zv);
        BUG_ON(size == 0);
-       to_va = kmap_atomic(page, KM_USER0);
+       to_va = kmap_atomic(page);
        ret = lzo1x_decompress_safe((char *)zv + sizeof(*zv),
                                        size, to_va, &clen);
-       kunmap_atomic(to_va, KM_USER0);
+       kunmap_atomic(to_va);
        BUG_ON(ret != LZO_E_OK);
        BUG_ON(clen != PAGE_SIZE);
 }
        BUG_ON(!irqs_disabled());
        if (unlikely(dmem == NULL || wmem == NULL))
                goto out;  /* no buffer, so can't compress */
-       from_va = kmap_atomic(from, KM_USER0);
+       from_va = kmap_atomic(from);
        mb();
        ret = lzo1x_1_compress(from_va, PAGE_SIZE, dmem, out_len, wmem);
        BUG_ON(ret != LZO_E_OK);
        *out_va = dmem;
-       kunmap_atomic(from_va, KM_USER0);
+       kunmap_atomic(from_va);
        ret = 1;
 out:
        return ret;