hyperv.c 9.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404
  1. /*
  2. * KVM Microsoft Hyper-V emulation
  3. *
  4. * derived from arch/x86/kvm/x86.c
  5. *
  6. * Copyright (C) 2006 Qumranet, Inc.
  7. * Copyright (C) 2008 Qumranet, Inc.
  8. * Copyright IBM Corporation, 2008
  9. * Copyright 2010 Red Hat, Inc. and/or its affiliates.
  10. * Copyright (C) 2015 Andrey Smetanin <asmetanin@virtuozzo.com>
  11. *
  12. * Authors:
  13. * Avi Kivity <avi@qumranet.com>
  14. * Yaniv Kamay <yaniv@qumranet.com>
  15. * Amit Shah <amit.shah@qumranet.com>
  16. * Ben-Ami Yassour <benami@il.ibm.com>
  17. * Andrey Smetanin <asmetanin@virtuozzo.com>
  18. *
  19. * This work is licensed under the terms of the GNU GPL, version 2. See
  20. * the COPYING file in the top-level directory.
  21. *
  22. */
  23. #include "x86.h"
  24. #include "lapic.h"
  25. #include "hyperv.h"
  26. #include <linux/kvm_host.h>
  27. #include <trace/events/kvm.h>
  28. #include "trace.h"
  29. static bool kvm_hv_msr_partition_wide(u32 msr)
  30. {
  31. bool r = false;
  32. switch (msr) {
  33. case HV_X64_MSR_GUEST_OS_ID:
  34. case HV_X64_MSR_HYPERCALL:
  35. case HV_X64_MSR_REFERENCE_TSC:
  36. case HV_X64_MSR_TIME_REF_COUNT:
  37. case HV_X64_MSR_CRASH_CTL:
  38. case HV_X64_MSR_CRASH_P0 ... HV_X64_MSR_CRASH_P4:
  39. case HV_X64_MSR_RESET:
  40. r = true;
  41. break;
  42. }
  43. return r;
  44. }
  45. static int kvm_hv_msr_get_crash_data(struct kvm_vcpu *vcpu,
  46. u32 index, u64 *pdata)
  47. {
  48. struct kvm_hv *hv = &vcpu->kvm->arch.hyperv;
  49. if (WARN_ON_ONCE(index >= ARRAY_SIZE(hv->hv_crash_param)))
  50. return -EINVAL;
  51. *pdata = hv->hv_crash_param[index];
  52. return 0;
  53. }
  54. static int kvm_hv_msr_get_crash_ctl(struct kvm_vcpu *vcpu, u64 *pdata)
  55. {
  56. struct kvm_hv *hv = &vcpu->kvm->arch.hyperv;
  57. *pdata = hv->hv_crash_ctl;
  58. return 0;
  59. }
  60. static int kvm_hv_msr_set_crash_ctl(struct kvm_vcpu *vcpu, u64 data, bool host)
  61. {
  62. struct kvm_hv *hv = &vcpu->kvm->arch.hyperv;
  63. if (host)
  64. hv->hv_crash_ctl = data & HV_X64_MSR_CRASH_CTL_NOTIFY;
  65. if (!host && (data & HV_X64_MSR_CRASH_CTL_NOTIFY)) {
  66. vcpu_debug(vcpu, "hv crash (0x%llx 0x%llx 0x%llx 0x%llx 0x%llx)\n",
  67. hv->hv_crash_param[0],
  68. hv->hv_crash_param[1],
  69. hv->hv_crash_param[2],
  70. hv->hv_crash_param[3],
  71. hv->hv_crash_param[4]);
  72. /* Send notification about crash to user space */
  73. kvm_make_request(KVM_REQ_HV_CRASH, vcpu);
  74. }
  75. return 0;
  76. }
  77. static int kvm_hv_msr_set_crash_data(struct kvm_vcpu *vcpu,
  78. u32 index, u64 data)
  79. {
  80. struct kvm_hv *hv = &vcpu->kvm->arch.hyperv;
  81. if (WARN_ON_ONCE(index >= ARRAY_SIZE(hv->hv_crash_param)))
  82. return -EINVAL;
  83. hv->hv_crash_param[index] = data;
  84. return 0;
  85. }
  86. static int kvm_hv_set_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 data,
  87. bool host)
  88. {
  89. struct kvm *kvm = vcpu->kvm;
  90. struct kvm_hv *hv = &kvm->arch.hyperv;
  91. switch (msr) {
  92. case HV_X64_MSR_GUEST_OS_ID:
  93. hv->hv_guest_os_id = data;
  94. /* setting guest os id to zero disables hypercall page */
  95. if (!hv->hv_guest_os_id)
  96. hv->hv_hypercall &= ~HV_X64_MSR_HYPERCALL_ENABLE;
  97. break;
  98. case HV_X64_MSR_HYPERCALL: {
  99. u64 gfn;
  100. unsigned long addr;
  101. u8 instructions[4];
  102. /* if guest os id is not set hypercall should remain disabled */
  103. if (!hv->hv_guest_os_id)
  104. break;
  105. if (!(data & HV_X64_MSR_HYPERCALL_ENABLE)) {
  106. hv->hv_hypercall = data;
  107. break;
  108. }
  109. gfn = data >> HV_X64_MSR_HYPERCALL_PAGE_ADDRESS_SHIFT;
  110. addr = gfn_to_hva(kvm, gfn);
  111. if (kvm_is_error_hva(addr))
  112. return 1;
  113. kvm_x86_ops->patch_hypercall(vcpu, instructions);
  114. ((unsigned char *)instructions)[3] = 0xc3; /* ret */
  115. if (__copy_to_user((void __user *)addr, instructions, 4))
  116. return 1;
  117. hv->hv_hypercall = data;
  118. mark_page_dirty(kvm, gfn);
  119. break;
  120. }
  121. case HV_X64_MSR_REFERENCE_TSC: {
  122. u64 gfn;
  123. HV_REFERENCE_TSC_PAGE tsc_ref;
  124. memset(&tsc_ref, 0, sizeof(tsc_ref));
  125. hv->hv_tsc_page = data;
  126. if (!(data & HV_X64_MSR_TSC_REFERENCE_ENABLE))
  127. break;
  128. gfn = data >> HV_X64_MSR_TSC_REFERENCE_ADDRESS_SHIFT;
  129. if (kvm_write_guest(
  130. kvm,
  131. gfn << HV_X64_MSR_TSC_REFERENCE_ADDRESS_SHIFT,
  132. &tsc_ref, sizeof(tsc_ref)))
  133. return 1;
  134. mark_page_dirty(kvm, gfn);
  135. break;
  136. }
  137. case HV_X64_MSR_CRASH_P0 ... HV_X64_MSR_CRASH_P4:
  138. return kvm_hv_msr_set_crash_data(vcpu,
  139. msr - HV_X64_MSR_CRASH_P0,
  140. data);
  141. case HV_X64_MSR_CRASH_CTL:
  142. return kvm_hv_msr_set_crash_ctl(vcpu, data, host);
  143. case HV_X64_MSR_RESET:
  144. if (data == 1) {
  145. vcpu_debug(vcpu, "hyper-v reset requested\n");
  146. kvm_make_request(KVM_REQ_HV_RESET, vcpu);
  147. }
  148. break;
  149. default:
  150. vcpu_unimpl(vcpu, "Hyper-V uhandled wrmsr: 0x%x data 0x%llx\n",
  151. msr, data);
  152. return 1;
  153. }
  154. return 0;
  155. }
  156. /* Calculate cpu time spent by current task in 100ns units */
  157. static u64 current_task_runtime_100ns(void)
  158. {
  159. cputime_t utime, stime;
  160. task_cputime_adjusted(current, &utime, &stime);
  161. return div_u64(cputime_to_nsecs(utime + stime), 100);
  162. }
  163. static int kvm_hv_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host)
  164. {
  165. struct kvm_vcpu_hv *hv = &vcpu->arch.hyperv;
  166. switch (msr) {
  167. case HV_X64_MSR_APIC_ASSIST_PAGE: {
  168. u64 gfn;
  169. unsigned long addr;
  170. if (!(data & HV_X64_MSR_APIC_ASSIST_PAGE_ENABLE)) {
  171. hv->hv_vapic = data;
  172. if (kvm_lapic_enable_pv_eoi(vcpu, 0))
  173. return 1;
  174. break;
  175. }
  176. gfn = data >> HV_X64_MSR_APIC_ASSIST_PAGE_ADDRESS_SHIFT;
  177. addr = kvm_vcpu_gfn_to_hva(vcpu, gfn);
  178. if (kvm_is_error_hva(addr))
  179. return 1;
  180. if (__clear_user((void __user *)addr, PAGE_SIZE))
  181. return 1;
  182. hv->hv_vapic = data;
  183. kvm_vcpu_mark_page_dirty(vcpu, gfn);
  184. if (kvm_lapic_enable_pv_eoi(vcpu,
  185. gfn_to_gpa(gfn) | KVM_MSR_ENABLED))
  186. return 1;
  187. break;
  188. }
  189. case HV_X64_MSR_EOI:
  190. return kvm_hv_vapic_msr_write(vcpu, APIC_EOI, data);
  191. case HV_X64_MSR_ICR:
  192. return kvm_hv_vapic_msr_write(vcpu, APIC_ICR, data);
  193. case HV_X64_MSR_TPR:
  194. return kvm_hv_vapic_msr_write(vcpu, APIC_TASKPRI, data);
  195. case HV_X64_MSR_VP_RUNTIME:
  196. if (!host)
  197. return 1;
  198. hv->runtime_offset = data - current_task_runtime_100ns();
  199. break;
  200. default:
  201. vcpu_unimpl(vcpu, "Hyper-V uhandled wrmsr: 0x%x data 0x%llx\n",
  202. msr, data);
  203. return 1;
  204. }
  205. return 0;
  206. }
  207. static int kvm_hv_get_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata)
  208. {
  209. u64 data = 0;
  210. struct kvm *kvm = vcpu->kvm;
  211. struct kvm_hv *hv = &kvm->arch.hyperv;
  212. switch (msr) {
  213. case HV_X64_MSR_GUEST_OS_ID:
  214. data = hv->hv_guest_os_id;
  215. break;
  216. case HV_X64_MSR_HYPERCALL:
  217. data = hv->hv_hypercall;
  218. break;
  219. case HV_X64_MSR_TIME_REF_COUNT: {
  220. data =
  221. div_u64(get_kernel_ns() + kvm->arch.kvmclock_offset, 100);
  222. break;
  223. }
  224. case HV_X64_MSR_REFERENCE_TSC:
  225. data = hv->hv_tsc_page;
  226. break;
  227. case HV_X64_MSR_CRASH_P0 ... HV_X64_MSR_CRASH_P4:
  228. return kvm_hv_msr_get_crash_data(vcpu,
  229. msr - HV_X64_MSR_CRASH_P0,
  230. pdata);
  231. case HV_X64_MSR_CRASH_CTL:
  232. return kvm_hv_msr_get_crash_ctl(vcpu, pdata);
  233. case HV_X64_MSR_RESET:
  234. data = 0;
  235. break;
  236. default:
  237. vcpu_unimpl(vcpu, "Hyper-V unhandled rdmsr: 0x%x\n", msr);
  238. return 1;
  239. }
  240. *pdata = data;
  241. return 0;
  242. }
  243. static int kvm_hv_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata)
  244. {
  245. u64 data = 0;
  246. struct kvm_vcpu_hv *hv = &vcpu->arch.hyperv;
  247. switch (msr) {
  248. case HV_X64_MSR_VP_INDEX: {
  249. int r;
  250. struct kvm_vcpu *v;
  251. kvm_for_each_vcpu(r, v, vcpu->kvm) {
  252. if (v == vcpu) {
  253. data = r;
  254. break;
  255. }
  256. }
  257. break;
  258. }
  259. case HV_X64_MSR_EOI:
  260. return kvm_hv_vapic_msr_read(vcpu, APIC_EOI, pdata);
  261. case HV_X64_MSR_ICR:
  262. return kvm_hv_vapic_msr_read(vcpu, APIC_ICR, pdata);
  263. case HV_X64_MSR_TPR:
  264. return kvm_hv_vapic_msr_read(vcpu, APIC_TASKPRI, pdata);
  265. case HV_X64_MSR_APIC_ASSIST_PAGE:
  266. data = hv->hv_vapic;
  267. break;
  268. case HV_X64_MSR_VP_RUNTIME:
  269. data = current_task_runtime_100ns() + hv->runtime_offset;
  270. break;
  271. default:
  272. vcpu_unimpl(vcpu, "Hyper-V unhandled rdmsr: 0x%x\n", msr);
  273. return 1;
  274. }
  275. *pdata = data;
  276. return 0;
  277. }
  278. int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host)
  279. {
  280. if (kvm_hv_msr_partition_wide(msr)) {
  281. int r;
  282. mutex_lock(&vcpu->kvm->lock);
  283. r = kvm_hv_set_msr_pw(vcpu, msr, data, host);
  284. mutex_unlock(&vcpu->kvm->lock);
  285. return r;
  286. } else
  287. return kvm_hv_set_msr(vcpu, msr, data, host);
  288. }
  289. int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata)
  290. {
  291. if (kvm_hv_msr_partition_wide(msr)) {
  292. int r;
  293. mutex_lock(&vcpu->kvm->lock);
  294. r = kvm_hv_get_msr_pw(vcpu, msr, pdata);
  295. mutex_unlock(&vcpu->kvm->lock);
  296. return r;
  297. } else
  298. return kvm_hv_get_msr(vcpu, msr, pdata);
  299. }
  300. bool kvm_hv_hypercall_enabled(struct kvm *kvm)
  301. {
  302. return kvm->arch.hyperv.hv_hypercall & HV_X64_MSR_HYPERCALL_ENABLE;
  303. }
  304. int kvm_hv_hypercall(struct kvm_vcpu *vcpu)
  305. {
  306. u64 param, ingpa, outgpa, ret;
  307. uint16_t code, rep_idx, rep_cnt, res = HV_STATUS_SUCCESS, rep_done = 0;
  308. bool fast, longmode;
  309. /*
  310. * hypercall generates UD from non zero cpl and real mode
  311. * per HYPER-V spec
  312. */
  313. if (kvm_x86_ops->get_cpl(vcpu) != 0 || !is_protmode(vcpu)) {
  314. kvm_queue_exception(vcpu, UD_VECTOR);
  315. return 0;
  316. }
  317. longmode = is_64_bit_mode(vcpu);
  318. if (!longmode) {
  319. param = ((u64)kvm_register_read(vcpu, VCPU_REGS_RDX) << 32) |
  320. (kvm_register_read(vcpu, VCPU_REGS_RAX) & 0xffffffff);
  321. ingpa = ((u64)kvm_register_read(vcpu, VCPU_REGS_RBX) << 32) |
  322. (kvm_register_read(vcpu, VCPU_REGS_RCX) & 0xffffffff);
  323. outgpa = ((u64)kvm_register_read(vcpu, VCPU_REGS_RDI) << 32) |
  324. (kvm_register_read(vcpu, VCPU_REGS_RSI) & 0xffffffff);
  325. }
  326. #ifdef CONFIG_X86_64
  327. else {
  328. param = kvm_register_read(vcpu, VCPU_REGS_RCX);
  329. ingpa = kvm_register_read(vcpu, VCPU_REGS_RDX);
  330. outgpa = kvm_register_read(vcpu, VCPU_REGS_R8);
  331. }
  332. #endif
  333. code = param & 0xffff;
  334. fast = (param >> 16) & 0x1;
  335. rep_cnt = (param >> 32) & 0xfff;
  336. rep_idx = (param >> 48) & 0xfff;
  337. trace_kvm_hv_hypercall(code, fast, rep_cnt, rep_idx, ingpa, outgpa);
  338. switch (code) {
  339. case HV_X64_HV_NOTIFY_LONG_SPIN_WAIT:
  340. kvm_vcpu_on_spin(vcpu);
  341. break;
  342. default:
  343. res = HV_STATUS_INVALID_HYPERCALL_CODE;
  344. break;
  345. }
  346. ret = res | (((u64)rep_done & 0xfff) << 32);
  347. if (longmode) {
  348. kvm_register_write(vcpu, VCPU_REGS_RAX, ret);
  349. } else {
  350. kvm_register_write(vcpu, VCPU_REGS_RDX, ret >> 32);
  351. kvm_register_write(vcpu, VCPU_REGS_RAX, ret & 0xffffffff);
  352. }
  353. return 1;
  354. }