#endif
                          (ib->gpu_addr & 0xFFFFFFFC));
        radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFF);
-       radeon_ring_write(ring, ib->length_dw | (ib->vm_id << 24));
+       radeon_ring_write(ring, ib->length_dw | 
+                         (ib->vm ? (ib->vm->id << 24) : 0));
 
        /* flush read cache over gart for this vmid */
        radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1));
        radeon_ring_write(ring, (CP_COHER_CNTL2 - PACKET3_SET_CONFIG_REG_START) >> 2);
-       radeon_ring_write(ring, ib->vm_id);
+       radeon_ring_write(ring, ib->vm ? ib->vm->id : 0);
        radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_SYNC, 3));
        radeon_ring_write(ring, PACKET3_TC_ACTION_ENA | PACKET3_SH_ACTION_ENA);
        radeon_ring_write(ring, 0xFFFFFFFF);
 
                return r;
        }
        WREG32(scratch, 0xCAFEDEAD);
-       r = radeon_ib_get(rdev, RADEON_RING_TYPE_GFX_INDEX, &ib, 256);
+       r = radeon_ib_get(rdev, RADEON_RING_TYPE_GFX_INDEX, &ib, NULL, 256);
        if (r) {
                return r;
        }
 
                return r;
        }
        WREG32(scratch, 0xCAFEDEAD);
-       r = radeon_ib_get(rdev, ring->idx, &ib, 256);
+       r = radeon_ib_get(rdev, ring->idx, &ib, NULL, 256);
        if (r) {
                DRM_ERROR("radeon: failed to get ib (%d).\n", r);
                return r;
 
        uint32_t                        *ptr;
        int                             ring;
        struct radeon_fence             *fence;
-       unsigned                        vm_id;
+       struct radeon_vm                *vm;
        bool                            is_const_ib;
        struct radeon_fence             *sync_to[RADEON_NUM_RINGS];
        struct radeon_semaphore         *semaphore;
 };
 
 int radeon_ib_get(struct radeon_device *rdev, int ring,
-                 struct radeon_ib *ib, unsigned size);
+                 struct radeon_ib *ib, struct radeon_vm *vm,
+                 unsigned size);
 void radeon_ib_free(struct radeon_device *rdev, struct radeon_ib *ib);
 int radeon_ib_schedule(struct radeon_device *rdev, struct radeon_ib *ib,
                       struct radeon_ib *const_ib);
 
         * uncached).
         */
        r =  radeon_ib_get(rdev, parser->ring, &parser->ib,
-                          ib_chunk->length_dw * 4);
+                          NULL, ib_chunk->length_dw * 4);
        if (r) {
                DRM_ERROR("Failed to get ib !\n");
                return r;
                return r;
        }
        radeon_cs_sync_rings(parser);
-       parser->ib.vm_id = 0;
        r = radeon_ib_schedule(rdev, &parser->ib, NULL);
        if (r) {
                DRM_ERROR("Failed to schedule IB !\n");
                        return -EINVAL;
                }
                r =  radeon_ib_get(rdev, parser->ring, &parser->const_ib,
-                                  ib_chunk->length_dw * 4);
+                                  vm, ib_chunk->length_dw * 4);
                if (r) {
                        DRM_ERROR("Failed to get const ib !\n");
                        return r;
                return -EINVAL;
        }
        r =  radeon_ib_get(rdev, parser->ring, &parser->ib,
-                          ib_chunk->length_dw * 4);
+                          vm, ib_chunk->length_dw * 4);
        if (r) {
                DRM_ERROR("Failed to get ib !\n");
                return r;
        }
        radeon_cs_sync_rings(parser);
 
-       parser->ib.vm_id = vm->id;
-       /* ib pool is bind at 0 in virtual address space,
-        * so gpu_addr is the offset inside the pool bo
-        */
-       parser->ib.gpu_addr = parser->ib.sa_bo->soffset;
-
        if ((rdev->family >= CHIP_TAHITI) &&
            (parser->chunk_const_ib_idx != -1)) {
-               parser->const_ib.vm_id = vm->id;
-               /* ib pool is bind at 0 in virtual address space,
-                * so gpu_addr is the offset inside the pool bo
-                */
-               parser->const_ib.gpu_addr = parser->const_ib.sa_bo->soffset;
                r = radeon_ib_schedule(rdev, &parser->ib, &parser->const_ib);
        } else {
                r = radeon_ib_schedule(rdev, &parser->ib, NULL);
 
  * Returns 0 on success, error on failure.
  */
 int radeon_ib_get(struct radeon_device *rdev, int ring,
-                 struct radeon_ib *ib, unsigned size)
+                 struct radeon_ib *ib, struct radeon_vm *vm,
+                 unsigned size)
 {
        int i, r;
 
        ib->ring = ring;
        ib->fence = NULL;
        ib->ptr = radeon_sa_bo_cpu_addr(ib->sa_bo);
-       ib->gpu_addr = radeon_sa_bo_gpu_addr(ib->sa_bo);
-       ib->vm_id = 0;
+       ib->vm = vm;
+       if (vm) {
+               /* ib pool is bind at 0 in virtual address space,
+                * so gpu_addr is the offset inside the pool bo
+                */
+               ib->gpu_addr = ib->sa_bo->soffset;
+       } else {
+               ib->gpu_addr = radeon_sa_bo_gpu_addr(ib->sa_bo);
+       }
        ib->is_const_ib = false;
        for (i = 0; i < RADEON_NUM_RINGS; ++i)
                ib->sync_to[i] = NULL;
 
 #endif
                          (ib->gpu_addr & 0xFFFFFFFC));
        radeon_ring_write(ring, upper_32_bits(ib->gpu_addr) & 0xFFFF);
-       radeon_ring_write(ring, ib->length_dw | (ib->vm_id << 24));
+       radeon_ring_write(ring, ib->length_dw |
+                         (ib->vm ? (ib->vm->id << 24) : 0));
 
        if (!ib->is_const_ib) {
                /* flush read cache over gart for this vmid */
                radeon_ring_write(ring, PACKET3(PACKET3_SET_CONFIG_REG, 1));
                radeon_ring_write(ring, (CP_COHER_CNTL2 - PACKET3_SET_CONFIG_REG_START) >> 2);
-               radeon_ring_write(ring, ib->vm_id);
+               radeon_ring_write(ring, ib->vm ? ib->vm->id : 0);
                radeon_ring_write(ring, PACKET3(PACKET3_SURFACE_SYNC, 3));
                radeon_ring_write(ring, PACKET3_TCL1_ACTION_ENA |
                                  PACKET3_TC_ACTION_ENA |