--- /dev/null
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Copyright (C) 2019 Western Digital Corporation or its affiliates.
+ *
+ * Authors:
+ *     Atish Patra <atish.patra@wdc.com>
+ */
+
+#ifndef __KVM_VCPU_RISCV_TIMER_H
+#define __KVM_VCPU_RISCV_TIMER_H
+
+#include <linux/hrtimer.h>
+
+struct kvm_guest_timer {
+       /* Mult & Shift values to get nanoseconds from cycles */
+       u32 nsec_mult;
+       u32 nsec_shift;
+       /* Time delta value */
+       u64 time_delta;
+};
+
+struct kvm_vcpu_timer {
+       /* Flag for whether init is done */
+       bool init_done;
+       /* Flag for whether timer event is configured */
+       bool next_set;
+       /* Next timer event cycles */
+       u64 next_cycles;
+       /* Underlying hrtimer instance */
+       struct hrtimer hrt;
+};
+
+int kvm_riscv_vcpu_timer_next_event(struct kvm_vcpu *vcpu, u64 ncycles);
+int kvm_riscv_vcpu_get_reg_timer(struct kvm_vcpu *vcpu,
+                                const struct kvm_one_reg *reg);
+int kvm_riscv_vcpu_set_reg_timer(struct kvm_vcpu *vcpu,
+                                const struct kvm_one_reg *reg);
+int kvm_riscv_vcpu_timer_init(struct kvm_vcpu *vcpu);
+int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu);
+int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu);
+void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu);
+int kvm_riscv_guest_timer_init(struct kvm *kvm);
+
+#endif
 
 
        memcpy(cntx, reset_cntx, sizeof(*cntx));
 
+       kvm_riscv_vcpu_timer_reset(vcpu);
+
        WRITE_ONCE(vcpu->arch.irqs_pending, 0);
        WRITE_ONCE(vcpu->arch.irqs_pending_mask, 0);
 }
        cntx->hstatus |= HSTATUS_SPVP;
        cntx->hstatus |= HSTATUS_SPV;
 
+       /* Setup VCPU timer */
+       kvm_riscv_vcpu_timer_init(vcpu);
+
        /* Reset VCPU */
        kvm_riscv_reset_vcpu(vcpu);
 
 
 void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
 {
+       /* Cleanup VCPU timer */
+       kvm_riscv_vcpu_timer_deinit(vcpu);
+
        /* Flush the pages pre-allocated for Stage2 page table mappings */
        kvm_riscv_stage2_flush_cache(vcpu);
 }
                return kvm_riscv_vcpu_set_reg_core(vcpu, reg);
        else if ((reg->id & KVM_REG_RISCV_TYPE_MASK) == KVM_REG_RISCV_CSR)
                return kvm_riscv_vcpu_set_reg_csr(vcpu, reg);
+       else if ((reg->id & KVM_REG_RISCV_TYPE_MASK) == KVM_REG_RISCV_TIMER)
+               return kvm_riscv_vcpu_set_reg_timer(vcpu, reg);
 
        return -EINVAL;
 }
                return kvm_riscv_vcpu_get_reg_core(vcpu, reg);
        else if ((reg->id & KVM_REG_RISCV_TYPE_MASK) == KVM_REG_RISCV_CSR)
                return kvm_riscv_vcpu_get_reg_csr(vcpu, reg);
+       else if ((reg->id & KVM_REG_RISCV_TYPE_MASK) == KVM_REG_RISCV_TIMER)
+               return kvm_riscv_vcpu_get_reg_timer(vcpu, reg);
 
        return -EINVAL;
 }
 
        kvm_riscv_stage2_update_hgatp(vcpu);
 
+       kvm_riscv_vcpu_timer_restore(vcpu);
+
        vcpu->cpu = cpu;
 }
 
 
--- /dev/null
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * Copyright (C) 2019 Western Digital Corporation or its affiliates.
+ *
+ * Authors:
+ *     Atish Patra <atish.patra@wdc.com>
+ */
+
+#include <linux/errno.h>
+#include <linux/err.h>
+#include <linux/kvm_host.h>
+#include <linux/uaccess.h>
+#include <clocksource/timer-riscv.h>
+#include <asm/csr.h>
+#include <asm/delay.h>
+#include <asm/kvm_vcpu_timer.h>
+
+static u64 kvm_riscv_current_cycles(struct kvm_guest_timer *gt)
+{
+       return get_cycles64() + gt->time_delta;
+}
+
+static u64 kvm_riscv_delta_cycles2ns(u64 cycles,
+                                    struct kvm_guest_timer *gt,
+                                    struct kvm_vcpu_timer *t)
+{
+       unsigned long flags;
+       u64 cycles_now, cycles_delta, delta_ns;
+
+       local_irq_save(flags);
+       cycles_now = kvm_riscv_current_cycles(gt);
+       if (cycles_now < cycles)
+               cycles_delta = cycles - cycles_now;
+       else
+               cycles_delta = 0;
+       delta_ns = (cycles_delta * gt->nsec_mult) >> gt->nsec_shift;
+       local_irq_restore(flags);
+
+       return delta_ns;
+}
+
+static enum hrtimer_restart kvm_riscv_vcpu_hrtimer_expired(struct hrtimer *h)
+{
+       u64 delta_ns;
+       struct kvm_vcpu_timer *t = container_of(h, struct kvm_vcpu_timer, hrt);
+       struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
+       struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
+
+       if (kvm_riscv_current_cycles(gt) < t->next_cycles) {
+               delta_ns = kvm_riscv_delta_cycles2ns(t->next_cycles, gt, t);
+               hrtimer_forward_now(&t->hrt, ktime_set(0, delta_ns));
+               return HRTIMER_RESTART;
+       }
+
+       t->next_set = false;
+       kvm_riscv_vcpu_set_interrupt(vcpu, IRQ_VS_TIMER);
+
+       return HRTIMER_NORESTART;
+}
+
+static int kvm_riscv_vcpu_timer_cancel(struct kvm_vcpu_timer *t)
+{
+       if (!t->init_done || !t->next_set)
+               return -EINVAL;
+
+       hrtimer_cancel(&t->hrt);
+       t->next_set = false;
+
+       return 0;
+}
+
+int kvm_riscv_vcpu_timer_next_event(struct kvm_vcpu *vcpu, u64 ncycles)
+{
+       struct kvm_vcpu_timer *t = &vcpu->arch.timer;
+       struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
+       u64 delta_ns;
+
+       if (!t->init_done)
+               return -EINVAL;
+
+       kvm_riscv_vcpu_unset_interrupt(vcpu, IRQ_VS_TIMER);
+
+       delta_ns = kvm_riscv_delta_cycles2ns(ncycles, gt, t);
+       t->next_cycles = ncycles;
+       hrtimer_start(&t->hrt, ktime_set(0, delta_ns), HRTIMER_MODE_REL);
+       t->next_set = true;
+
+       return 0;
+}
+
+int kvm_riscv_vcpu_get_reg_timer(struct kvm_vcpu *vcpu,
+                                const struct kvm_one_reg *reg)
+{
+       struct kvm_vcpu_timer *t = &vcpu->arch.timer;
+       struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
+       u64 __user *uaddr = (u64 __user *)(unsigned long)reg->addr;
+       unsigned long reg_num = reg->id & ~(KVM_REG_ARCH_MASK |
+                                           KVM_REG_SIZE_MASK |
+                                           KVM_REG_RISCV_TIMER);
+       u64 reg_val;
+
+       if (KVM_REG_SIZE(reg->id) != sizeof(u64))
+               return -EINVAL;
+       if (reg_num >= sizeof(struct kvm_riscv_timer) / sizeof(u64))
+               return -EINVAL;
+
+       switch (reg_num) {
+       case KVM_REG_RISCV_TIMER_REG(frequency):
+               reg_val = riscv_timebase;
+               break;
+       case KVM_REG_RISCV_TIMER_REG(time):
+               reg_val = kvm_riscv_current_cycles(gt);
+               break;
+       case KVM_REG_RISCV_TIMER_REG(compare):
+               reg_val = t->next_cycles;
+               break;
+       case KVM_REG_RISCV_TIMER_REG(state):
+               reg_val = (t->next_set) ? KVM_RISCV_TIMER_STATE_ON :
+                                         KVM_RISCV_TIMER_STATE_OFF;
+               break;
+       default:
+               return -EINVAL;
+       };
+
+       if (copy_to_user(uaddr, ®_val, KVM_REG_SIZE(reg->id)))
+               return -EFAULT;
+
+       return 0;
+}
+
+int kvm_riscv_vcpu_set_reg_timer(struct kvm_vcpu *vcpu,
+                                const struct kvm_one_reg *reg)
+{
+       struct kvm_vcpu_timer *t = &vcpu->arch.timer;
+       struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
+       u64 __user *uaddr = (u64 __user *)(unsigned long)reg->addr;
+       unsigned long reg_num = reg->id & ~(KVM_REG_ARCH_MASK |
+                                           KVM_REG_SIZE_MASK |
+                                           KVM_REG_RISCV_TIMER);
+       u64 reg_val;
+       int ret = 0;
+
+       if (KVM_REG_SIZE(reg->id) != sizeof(u64))
+               return -EINVAL;
+       if (reg_num >= sizeof(struct kvm_riscv_timer) / sizeof(u64))
+               return -EINVAL;
+
+       if (copy_from_user(®_val, uaddr, KVM_REG_SIZE(reg->id)))
+               return -EFAULT;
+
+       switch (reg_num) {
+       case KVM_REG_RISCV_TIMER_REG(frequency):
+               ret = -EOPNOTSUPP;
+               break;
+       case KVM_REG_RISCV_TIMER_REG(time):
+               gt->time_delta = reg_val - get_cycles64();
+               break;
+       case KVM_REG_RISCV_TIMER_REG(compare):
+               t->next_cycles = reg_val;
+               break;
+       case KVM_REG_RISCV_TIMER_REG(state):
+               if (reg_val == KVM_RISCV_TIMER_STATE_ON)
+                       ret = kvm_riscv_vcpu_timer_next_event(vcpu, reg_val);
+               else
+                       ret = kvm_riscv_vcpu_timer_cancel(t);
+               break;
+       default:
+               ret = -EINVAL;
+               break;
+       };
+
+       return ret;
+}
+
+int kvm_riscv_vcpu_timer_init(struct kvm_vcpu *vcpu)
+{
+       struct kvm_vcpu_timer *t = &vcpu->arch.timer;
+
+       if (t->init_done)
+               return -EINVAL;
+
+       hrtimer_init(&t->hrt, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
+       t->hrt.function = kvm_riscv_vcpu_hrtimer_expired;
+       t->init_done = true;
+       t->next_set = false;
+
+       return 0;
+}
+
+int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu)
+{
+       int ret;
+
+       ret = kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
+       vcpu->arch.timer.init_done = false;
+
+       return ret;
+}
+
+int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu)
+{
+       return kvm_riscv_vcpu_timer_cancel(&vcpu->arch.timer);
+}
+
+void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu)
+{
+       struct kvm_guest_timer *gt = &vcpu->kvm->arch.timer;
+
+#ifdef CONFIG_64BIT
+       csr_write(CSR_HTIMEDELTA, gt->time_delta);
+#else
+       csr_write(CSR_HTIMEDELTA, (u32)(gt->time_delta));
+       csr_write(CSR_HTIMEDELTAH, (u32)(gt->time_delta >> 32));
+#endif
+}
+
+int kvm_riscv_guest_timer_init(struct kvm *kvm)
+{
+       struct kvm_guest_timer *gt = &kvm->arch.timer;
+
+       riscv_cs_get_mult_shift(>->nsec_mult, >->nsec_shift);
+       gt->time_delta = -get_cycles64();
+
+       return 0;
+}