codegen.rs 51 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714
  1. use cpu::cpu::{
  2. FLAG_CARRY, FLAG_OVERFLOW, FLAG_SIGN, FLAG_ZERO, TLB_GLOBAL, TLB_HAS_CODE, TLB_NO_USER,
  3. TLB_READONLY, TLB_VALID,
  4. };
  5. use cpu::imports::mem8;
  6. use cpu::memory;
  7. use global_pointers;
  8. use jit::JitContext;
  9. use modrm;
  10. use modrm::ModrmByte;
  11. use profiler;
  12. use regs;
  13. use wasmgen::wasm_builder::{WasmBuilder, WasmLocal, WasmLocalI64};
  14. pub fn gen_add_cs_offset(ctx: &mut JitContext) {
  15. ctx.builder
  16. .load_fixed_i32(global_pointers::get_seg_offset(regs::CS));
  17. ctx.builder.add_i32();
  18. }
  19. fn gen_get_eip(builder: &mut WasmBuilder) {
  20. builder.load_fixed_i32(global_pointers::INSTRUCTION_POINTER);
  21. }
  22. pub fn gen_set_previous_eip_offset_from_eip(builder: &mut WasmBuilder, n: u32) {
  23. // previous_ip = instruction_pointer + n
  24. builder.const_i32(global_pointers::PREVIOUS_IP as i32);
  25. gen_get_eip(builder);
  26. if n != 0 {
  27. builder.const_i32(n as i32);
  28. builder.add_i32();
  29. }
  30. builder.store_aligned_i32(0);
  31. }
  32. pub fn gen_set_eip_to_after_current_instruction(ctx: &mut JitContext) {
  33. ctx.builder
  34. .const_i32(global_pointers::INSTRUCTION_POINTER as i32);
  35. gen_get_eip(ctx.builder);
  36. ctx.builder.const_i32(!0xFFF);
  37. ctx.builder.and_i32();
  38. ctx.builder.const_i32(ctx.cpu.eip as i32 & 0xFFF);
  39. ctx.builder.or_i32();
  40. ctx.builder.store_aligned_i32(0);
  41. }
  42. pub fn gen_set_previous_eip_offset_from_eip_with_low_bits(
  43. builder: &mut WasmBuilder,
  44. low_bits: i32,
  45. ) {
  46. // previous_ip = instruction_pointer & ~0xFFF | low_bits;
  47. builder.const_i32(global_pointers::PREVIOUS_IP as i32);
  48. gen_get_eip(builder);
  49. builder.const_i32(!0xFFF);
  50. builder.and_i32();
  51. builder.const_i32(low_bits);
  52. builder.or_i32();
  53. builder.store_aligned_i32(0);
  54. }
  55. pub fn gen_increment_instruction_pointer(builder: &mut WasmBuilder, n: u32) {
  56. builder.const_i32(global_pointers::INSTRUCTION_POINTER as i32);
  57. gen_get_eip(builder);
  58. builder.const_i32(n as i32);
  59. builder.add_i32();
  60. builder.store_aligned_i32(0);
  61. }
  62. pub fn gen_relative_jump(builder: &mut WasmBuilder, n: i32) {
  63. // add n to instruction_pointer (without setting the offset as above)
  64. builder.const_i32(global_pointers::INSTRUCTION_POINTER as i32);
  65. gen_get_eip(builder);
  66. builder.const_i32(n);
  67. builder.add_i32();
  68. builder.store_aligned_i32(0);
  69. }
  70. pub fn gen_absolute_indirect_jump(ctx: &mut JitContext, new_eip: WasmLocal) {
  71. ctx.builder
  72. .const_i32(global_pointers::INSTRUCTION_POINTER as i32);
  73. ctx.builder.get_local(&new_eip);
  74. ctx.builder.store_aligned_i32(0);
  75. ctx.builder.get_local(&new_eip);
  76. ctx.builder.load_fixed_i32(global_pointers::PREVIOUS_IP);
  77. ctx.builder.xor_i32();
  78. ctx.builder.const_i32(!0xFFF);
  79. ctx.builder.and_i32();
  80. ctx.builder.eqz_i32();
  81. ctx.builder.if_void();
  82. {
  83. // try staying in same page
  84. ctx.builder.get_local(&new_eip);
  85. ctx.builder.free_local(new_eip);
  86. ctx.builder
  87. .const_i32(ctx.start_of_current_instruction as i32);
  88. ctx.builder.const_i32(ctx.our_wasm_table_index as i32);
  89. ctx.builder.const_i32(ctx.state_flags.to_u32() as i32);
  90. ctx.builder.call_fn4_ret("jit_find_cache_entry_in_page");
  91. let new_basic_block_index = ctx.builder.tee_new_local();
  92. ctx.builder.const_i32(0);
  93. ctx.builder.ge_i32();
  94. ctx.builder.if_void();
  95. ctx.builder.get_local(&new_basic_block_index);
  96. ctx.builder.set_local(ctx.basic_block_index_local);
  97. ctx.builder.br(ctx.current_brtable_depth + 2); // to the loop
  98. ctx.builder.block_end();
  99. ctx.builder.free_local(new_basic_block_index);
  100. }
  101. ctx.builder.block_end();
  102. }
  103. pub fn gen_increment_timestamp_counter(builder: &mut WasmBuilder, n: i32) {
  104. builder.increment_fixed_i32(global_pointers::TIMESTAMP_COUNTER, n)
  105. }
  106. pub fn gen_get_reg8(ctx: &mut JitContext, r: u32) {
  107. match r {
  108. regs::AL | regs::CL | regs::DL | regs::BL => {
  109. ctx.builder.get_local(&ctx.register_locals[r as usize]);
  110. ctx.builder.const_i32(0xFF);
  111. ctx.builder.and_i32();
  112. },
  113. regs::AH | regs::CH | regs::DH | regs::BH => {
  114. ctx.builder
  115. .get_local(&ctx.register_locals[(r - 4) as usize]);
  116. ctx.builder.const_i32(8);
  117. ctx.builder.shr_u_i32();
  118. ctx.builder.const_i32(0xFF);
  119. ctx.builder.and_i32();
  120. },
  121. _ => assert!(false),
  122. }
  123. }
  124. /// Return a new local referencing one of the 8 bit registers or a direct reference to one of the
  125. /// register locals. Higher bits might be garbage (suitable for gen_cmp8 etc.). Must be freed with
  126. /// gen_free_reg8_or_alias.
  127. pub fn gen_get_reg8_or_alias_to_reg32(ctx: &mut JitContext, r: u32) -> WasmLocal {
  128. match r {
  129. regs::AL | regs::CL | regs::DL | regs::BL => ctx.register_locals[r as usize].unsafe_clone(),
  130. regs::AH | regs::CH | regs::DH | regs::BH => {
  131. ctx.builder
  132. .get_local(&ctx.register_locals[(r - 4) as usize]);
  133. ctx.builder.const_i32(8);
  134. ctx.builder.shr_u_i32();
  135. ctx.builder.set_new_local()
  136. },
  137. _ => panic!(),
  138. }
  139. }
  140. pub fn gen_free_reg8_or_alias(ctx: &mut JitContext, r: u32, local: WasmLocal) {
  141. match r {
  142. regs::AL | regs::CL | regs::DL | regs::BL => {},
  143. regs::AH | regs::CH | regs::DH | regs::BH => ctx.builder.free_local(local),
  144. _ => panic!(),
  145. }
  146. }
  147. pub fn gen_get_reg16(ctx: &mut JitContext, r: u32) {
  148. ctx.builder.get_local(&ctx.register_locals[r as usize]);
  149. ctx.builder.const_i32(0xFFFF);
  150. ctx.builder.and_i32();
  151. }
  152. pub fn gen_get_reg32(ctx: &mut JitContext, r: u32) {
  153. ctx.builder.get_local(&ctx.register_locals[r as usize]);
  154. }
  155. pub fn gen_set_reg8(ctx: &mut JitContext, r: u32) {
  156. match r {
  157. regs::AL | regs::CL | regs::DL | regs::BL => {
  158. // reg32[r] = stack_value & 0xFF | reg32[r] & ~0xFF
  159. ctx.builder.const_i32(0xFF);
  160. ctx.builder.and_i32();
  161. ctx.builder.get_local(&ctx.register_locals[r as usize]);
  162. ctx.builder.const_i32(!0xFF);
  163. ctx.builder.and_i32();
  164. ctx.builder.or_i32();
  165. ctx.builder.set_local(&ctx.register_locals[r as usize]);
  166. },
  167. regs::AH | regs::CH | regs::DH | regs::BH => {
  168. // reg32[r] = stack_value << 8 & 0xFF00 | reg32[r] & ~0xFF00
  169. ctx.builder.const_i32(8);
  170. ctx.builder.shl_i32();
  171. ctx.builder.const_i32(0xFF00);
  172. ctx.builder.and_i32();
  173. ctx.builder
  174. .get_local(&ctx.register_locals[(r - 4) as usize]);
  175. ctx.builder.const_i32(!0xFF00);
  176. ctx.builder.and_i32();
  177. ctx.builder.or_i32();
  178. ctx.builder
  179. .set_local(&ctx.register_locals[(r - 4) as usize]);
  180. },
  181. _ => assert!(false),
  182. }
  183. }
  184. pub fn gen_set_reg16(ctx: &mut JitContext, r: u32) {
  185. gen_set_reg16_local(ctx.builder, &ctx.register_locals[r as usize]);
  186. }
  187. pub fn gen_set_reg16_local(builder: &mut WasmBuilder, local: &WasmLocal) {
  188. // reg32[r] = v & 0xFFFF | reg32[r] & ~0xFFFF
  189. builder.const_i32(0xFFFF);
  190. builder.and_i32();
  191. builder.get_local(local);
  192. builder.const_i32(!0xFFFF);
  193. builder.and_i32();
  194. builder.or_i32();
  195. builder.set_local(local);
  196. }
  197. pub fn gen_set_reg32(ctx: &mut JitContext, r: u32) {
  198. ctx.builder.set_local(&ctx.register_locals[r as usize]);
  199. }
  200. pub fn decr_exc_asize(ctx: &mut JitContext) {
  201. gen_get_reg32(ctx, regs::ECX);
  202. ctx.builder.const_i32(1);
  203. ctx.builder.sub_i32();
  204. if ctx.cpu.asize_32() {
  205. gen_set_reg32(ctx, regs::ECX);
  206. }
  207. else {
  208. gen_set_reg16(ctx, regs::CX);
  209. }
  210. }
  211. pub fn gen_read_reg_xmm128_into_scratch(ctx: &mut JitContext, r: u32) {
  212. ctx.builder
  213. .const_i32(global_pointers::SSE_SCRATCH_REGISTER as i32);
  214. let dest = global_pointers::get_reg_xmm_offset(r);
  215. ctx.builder.const_i32(dest as i32);
  216. ctx.builder.load_aligned_i64(0);
  217. ctx.builder.store_aligned_i64(0);
  218. ctx.builder
  219. .const_i32(global_pointers::SSE_SCRATCH_REGISTER as i32 + 8);
  220. let dest = global_pointers::get_reg_xmm_offset(r) + 8;
  221. ctx.builder.const_i32(dest as i32);
  222. ctx.builder.load_aligned_i64(0);
  223. ctx.builder.store_aligned_i64(0);
  224. }
  225. pub fn gen_get_sreg(ctx: &mut JitContext, r: u32) {
  226. ctx.builder
  227. .load_fixed_u16(global_pointers::get_sreg_offset(r))
  228. }
  229. pub fn gen_get_ss_offset(ctx: &mut JitContext) {
  230. ctx.builder
  231. .load_fixed_i32(global_pointers::get_seg_offset(regs::SS));
  232. }
  233. pub fn gen_get_flags(builder: &mut WasmBuilder) { builder.load_fixed_i32(global_pointers::FLAGS); }
  234. pub fn gen_get_flags_changed(builder: &mut WasmBuilder) {
  235. builder.load_fixed_i32(global_pointers::FLAGS_CHANGED);
  236. }
  237. pub fn gen_get_last_result(builder: &mut WasmBuilder) {
  238. builder.load_fixed_i32(global_pointers::LAST_RESULT);
  239. }
  240. pub fn gen_get_last_op_size(builder: &mut WasmBuilder) {
  241. builder.load_fixed_i32(global_pointers::LAST_OP_SIZE);
  242. }
  243. pub fn gen_get_last_op1(builder: &mut WasmBuilder) {
  244. builder.load_fixed_i32(global_pointers::LAST_OP1);
  245. }
  246. pub fn gen_get_page_fault(builder: &mut WasmBuilder) {
  247. builder.load_fixed_u8(global_pointers::PAGE_FAULT);
  248. }
  249. /// sign-extend a byte value on the stack and leave it on the stack
  250. pub fn sign_extend_i8(builder: &mut WasmBuilder) {
  251. builder.const_i32(24);
  252. builder.shl_i32();
  253. builder.const_i32(24);
  254. builder.shr_s_i32();
  255. }
  256. /// sign-extend a two byte value on the stack and leave it on the stack
  257. pub fn sign_extend_i16(builder: &mut WasmBuilder) {
  258. builder.const_i32(16);
  259. builder.shl_i32();
  260. builder.const_i32(16);
  261. builder.shr_s_i32();
  262. }
  263. pub fn gen_fn0_const(builder: &mut WasmBuilder, name: &str) { builder.call_fn0(name) }
  264. pub fn gen_fn1_const(builder: &mut WasmBuilder, name: &str, arg0: u32) {
  265. builder.const_i32(arg0 as i32);
  266. builder.call_fn1(name);
  267. }
  268. pub fn gen_fn2_const(builder: &mut WasmBuilder, name: &str, arg0: u32, arg1: u32) {
  269. builder.const_i32(arg0 as i32);
  270. builder.const_i32(arg1 as i32);
  271. builder.call_fn2(name);
  272. }
  273. pub fn gen_fn3_const(builder: &mut WasmBuilder, name: &str, arg0: u32, arg1: u32, arg2: u32) {
  274. builder.const_i32(arg0 as i32);
  275. builder.const_i32(arg1 as i32);
  276. builder.const_i32(arg2 as i32);
  277. builder.call_fn3(name);
  278. }
  279. // helper functions for gen/generate_jit.js
  280. pub fn gen_modrm_fn0(builder: &mut WasmBuilder, name: &str) {
  281. // generates: fn( _ )
  282. builder.call_fn1(name);
  283. }
  284. pub fn gen_modrm_fn1(builder: &mut WasmBuilder, name: &str, arg0: u32) {
  285. // generates: fn( _, arg0 )
  286. builder.const_i32(arg0 as i32);
  287. builder.call_fn2(name);
  288. }
  289. pub fn gen_modrm_fn2(builder: &mut WasmBuilder, name: &str, arg0: u32, arg1: u32) {
  290. // generates: fn( _, arg0, arg1 )
  291. builder.const_i32(arg0 as i32);
  292. builder.const_i32(arg1 as i32);
  293. builder.call_fn3(name);
  294. }
  295. pub fn gen_modrm_resolve(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  296. modrm::gen(ctx, modrm_byte)
  297. }
  298. pub fn gen_set_reg8_r(ctx: &mut JitContext, dest: u32, src: u32) {
  299. // generates: reg8[r_dest] = reg8[r_src]
  300. gen_get_reg8(ctx, src);
  301. gen_set_reg8(ctx, dest);
  302. }
  303. pub fn gen_set_reg16_r(ctx: &mut JitContext, dest: u32, src: u32) {
  304. // generates: reg16[r_dest] = reg16[r_src]
  305. gen_get_reg16(ctx, src);
  306. gen_set_reg16(ctx, dest);
  307. }
  308. pub fn gen_set_reg32_r(ctx: &mut JitContext, dest: u32, src: u32) {
  309. // generates: reg32[r_dest] = reg32[r_src]
  310. gen_get_reg32(ctx, src);
  311. gen_set_reg32(ctx, dest);
  312. }
  313. pub fn gen_modrm_resolve_safe_read8(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  314. gen_modrm_resolve(ctx, modrm_byte);
  315. let address_local = ctx.builder.set_new_local();
  316. gen_safe_read8(ctx, &address_local);
  317. ctx.builder.free_local(address_local);
  318. }
  319. pub fn gen_modrm_resolve_safe_read16(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  320. gen_modrm_resolve(ctx, modrm_byte);
  321. let address_local = ctx.builder.set_new_local();
  322. gen_safe_read16(ctx, &address_local);
  323. ctx.builder.free_local(address_local);
  324. }
  325. pub fn gen_modrm_resolve_safe_read32(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  326. gen_modrm_resolve(ctx, modrm_byte);
  327. let address_local = ctx.builder.set_new_local();
  328. gen_safe_read32(ctx, &address_local);
  329. ctx.builder.free_local(address_local);
  330. }
  331. pub fn gen_modrm_resolve_safe_read64(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  332. gen_modrm_resolve(ctx, modrm_byte);
  333. let address_local = ctx.builder.set_new_local();
  334. gen_safe_read64(ctx, &address_local);
  335. ctx.builder.free_local(address_local);
  336. }
  337. pub fn gen_modrm_resolve_safe_read128(
  338. ctx: &mut JitContext,
  339. modrm_byte: ModrmByte,
  340. where_to_write: u32,
  341. ) {
  342. gen_modrm_resolve(ctx, modrm_byte);
  343. let address_local = ctx.builder.set_new_local();
  344. gen_safe_read128(ctx, &address_local, where_to_write);
  345. ctx.builder.free_local(address_local);
  346. }
  347. pub fn gen_safe_read8(ctx: &mut JitContext, address_local: &WasmLocal) {
  348. gen_safe_read(ctx, BitSize::BYTE, address_local, None);
  349. }
  350. pub fn gen_safe_read16(ctx: &mut JitContext, address_local: &WasmLocal) {
  351. gen_safe_read(ctx, BitSize::WORD, address_local, None);
  352. }
  353. pub fn gen_safe_read32(ctx: &mut JitContext, address_local: &WasmLocal) {
  354. gen_safe_read(ctx, BitSize::DWORD, address_local, None);
  355. }
  356. pub fn gen_safe_read64(ctx: &mut JitContext, address_local: &WasmLocal) {
  357. gen_safe_read(ctx, BitSize::QWORD, &address_local, None);
  358. }
  359. pub fn gen_safe_read128(ctx: &mut JitContext, address_local: &WasmLocal, where_to_write: u32) {
  360. gen_safe_read(ctx, BitSize::DQWORD, &address_local, Some(where_to_write));
  361. }
  362. // only used internally for gen_safe_write
  363. enum GenSafeWriteValue<'a> {
  364. I32(&'a WasmLocal),
  365. I64(&'a WasmLocalI64),
  366. TwoI64s(&'a WasmLocalI64, &'a WasmLocalI64),
  367. }
  368. #[derive(Copy, Clone, Eq, PartialEq)]
  369. pub enum BitSize {
  370. BYTE,
  371. WORD,
  372. DWORD,
  373. QWORD,
  374. DQWORD,
  375. }
  376. impl BitSize {
  377. pub fn bytes(&self) -> u32 {
  378. match self {
  379. BitSize::BYTE => 1,
  380. BitSize::WORD => 2,
  381. BitSize::DWORD => 4,
  382. BitSize::QWORD => 8,
  383. BitSize::DQWORD => 16,
  384. }
  385. }
  386. }
  387. pub fn gen_safe_write8(ctx: &mut JitContext, address_local: &WasmLocal, value_local: &WasmLocal) {
  388. gen_safe_write(
  389. ctx,
  390. BitSize::BYTE,
  391. address_local,
  392. GenSafeWriteValue::I32(value_local),
  393. )
  394. }
  395. pub fn gen_safe_write16(ctx: &mut JitContext, address_local: &WasmLocal, value_local: &WasmLocal) {
  396. gen_safe_write(
  397. ctx,
  398. BitSize::WORD,
  399. address_local,
  400. GenSafeWriteValue::I32(value_local),
  401. )
  402. }
  403. pub fn gen_safe_write32(ctx: &mut JitContext, address_local: &WasmLocal, value_local: &WasmLocal) {
  404. gen_safe_write(
  405. ctx,
  406. BitSize::DWORD,
  407. address_local,
  408. GenSafeWriteValue::I32(value_local),
  409. )
  410. }
  411. pub fn gen_safe_write64(
  412. ctx: &mut JitContext,
  413. address_local: &WasmLocal,
  414. value_local: &WasmLocalI64,
  415. ) {
  416. gen_safe_write(
  417. ctx,
  418. BitSize::QWORD,
  419. address_local,
  420. GenSafeWriteValue::I64(value_local),
  421. )
  422. }
  423. pub fn gen_safe_write128(
  424. ctx: &mut JitContext,
  425. address_local: &WasmLocal,
  426. value_local_low: &WasmLocalI64,
  427. value_local_high: &WasmLocalI64,
  428. ) {
  429. gen_safe_write(
  430. ctx,
  431. BitSize::DQWORD,
  432. address_local,
  433. GenSafeWriteValue::TwoI64s(value_local_low, value_local_high),
  434. )
  435. }
  436. fn gen_safe_read(
  437. ctx: &mut JitContext,
  438. bits: BitSize,
  439. address_local: &WasmLocal,
  440. where_to_write: Option<u32>,
  441. ) {
  442. // Execute a virtual memory read. All slow paths (memory-mapped IO, tlb miss, page fault and
  443. // read across page boundary are handled in safe_read_jit_slow
  444. // entry <- tlb_data[addr >> 12 << 2]
  445. // if entry & MASK == TLB_VALID && (addr & 0xFFF) <= 0x1000 - bytes: goto fast
  446. // entry <- safe_read_jit_slow(addr, instruction_pointer)
  447. // if page_fault: goto exit-with-pagefault
  448. // fast: mem[(entry & ~0xFFF) ^ addr]
  449. ctx.builder.block_void();
  450. ctx.builder.get_local(&address_local);
  451. ctx.builder.const_i32(12);
  452. ctx.builder.shr_u_i32();
  453. ctx.builder.const_i32(2);
  454. ctx.builder.shl_i32();
  455. ctx.builder.load_aligned_i32(global_pointers::TLB_DATA);
  456. let entry_local = ctx.builder.tee_new_local();
  457. ctx.builder.const_i32(
  458. (0xFFF
  459. & !TLB_READONLY
  460. & !TLB_GLOBAL
  461. & !TLB_HAS_CODE
  462. & !(if ctx.cpu.cpl3() { 0 } else { TLB_NO_USER })) as i32,
  463. );
  464. ctx.builder.and_i32();
  465. ctx.builder.const_i32(TLB_VALID as i32);
  466. ctx.builder.eq_i32();
  467. if bits != BitSize::BYTE {
  468. ctx.builder.get_local(&address_local);
  469. ctx.builder.const_i32(0xFFF);
  470. ctx.builder.and_i32();
  471. ctx.builder.const_i32(0x1000 - bits.bytes() as i32);
  472. ctx.builder.le_i32();
  473. ctx.builder.and_i32();
  474. }
  475. ctx.builder.br_if(0);
  476. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  477. ctx.builder.get_local(&address_local);
  478. ctx.builder.get_local(&entry_local);
  479. ctx.builder.call_fn2("report_safe_read_jit_slow");
  480. }
  481. ctx.builder.get_local(&address_local);
  482. ctx.builder
  483. .const_i32(ctx.start_of_current_instruction as i32 & 0xFFF);
  484. match bits {
  485. BitSize::BYTE => {
  486. ctx.builder.call_fn2_ret("safe_read8_slow_jit");
  487. },
  488. BitSize::WORD => {
  489. ctx.builder.call_fn2_ret("safe_read16_slow_jit");
  490. },
  491. BitSize::DWORD => {
  492. ctx.builder.call_fn2_ret("safe_read32s_slow_jit");
  493. },
  494. BitSize::QWORD => {
  495. ctx.builder.call_fn2_ret("safe_read64s_slow_jit");
  496. },
  497. BitSize::DQWORD => {
  498. ctx.builder.call_fn2_ret("safe_read128s_slow_jit");
  499. },
  500. }
  501. ctx.builder.tee_local(&entry_local);
  502. ctx.builder.const_i32(1);
  503. ctx.builder.and_i32();
  504. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  505. ctx.builder.if_void();
  506. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  507. ctx.builder.block_end();
  508. ctx.builder.get_local(&entry_local);
  509. ctx.builder.const_i32(1);
  510. ctx.builder.and_i32();
  511. }
  512. // -2 for the exit-with-pagefault block, +1 for leaving the nested if from this function
  513. ctx.builder.br_if(ctx.current_brtable_depth - 2 + 1);
  514. ctx.builder.block_end();
  515. gen_profiler_stat_increment(ctx.builder, profiler::stat::SAFE_READ_FAST); // XXX: Both fast and slow
  516. ctx.builder.get_local(&entry_local);
  517. ctx.builder.const_i32(!0xFFF);
  518. ctx.builder.and_i32();
  519. ctx.builder.get_local(&address_local);
  520. ctx.builder.xor_i32();
  521. // where_to_write is only used by dqword
  522. dbg_assert!((where_to_write != None) == (bits == BitSize::DQWORD));
  523. ctx.builder.const_i32(unsafe { mem8 } as i32);
  524. ctx.builder.add_i32();
  525. match bits {
  526. BitSize::BYTE => {
  527. ctx.builder.load_u8(0);
  528. },
  529. BitSize::WORD => {
  530. ctx.builder.load_unaligned_u16(0);
  531. },
  532. BitSize::DWORD => {
  533. ctx.builder.load_unaligned_i32(0);
  534. },
  535. BitSize::QWORD => {
  536. ctx.builder.load_unaligned_i64(0);
  537. },
  538. BitSize::DQWORD => {
  539. let where_to_write = where_to_write.unwrap();
  540. let virt_address_local = ctx.builder.set_new_local();
  541. ctx.builder.const_i32(0);
  542. ctx.builder.get_local(&virt_address_local);
  543. ctx.builder.load_unaligned_i64(0);
  544. ctx.builder.store_unaligned_i64(where_to_write);
  545. ctx.builder.const_i32(0);
  546. ctx.builder.get_local(&virt_address_local);
  547. ctx.builder.load_unaligned_i64(8);
  548. ctx.builder.store_unaligned_i64(where_to_write + 8);
  549. ctx.builder.free_local(virt_address_local);
  550. },
  551. }
  552. ctx.builder.free_local(entry_local);
  553. }
  554. fn gen_safe_write(
  555. ctx: &mut JitContext,
  556. bits: BitSize,
  557. address_local: &WasmLocal,
  558. value_local: GenSafeWriteValue,
  559. ) {
  560. // Execute a virtual memory write. All slow paths (memory-mapped IO, tlb miss, page fault,
  561. // write across page boundary and page containing jitted code are handled in safe_write_jit_slow
  562. // entry <- tlb_data[addr >> 12 << 2]
  563. // if entry & MASK == TLB_VALID && (addr & 0xFFF) <= 0x1000 - bytes: goto fast
  564. // entry <- safe_write_jit_slow(addr, value, instruction_pointer)
  565. // if page_fault: goto exit-with-pagefault
  566. // fast: mem[(entry & ~0xFFF) ^ addr] <- value
  567. ctx.builder.block_void();
  568. ctx.builder.get_local(&address_local);
  569. ctx.builder.const_i32(12);
  570. ctx.builder.shr_u_i32();
  571. ctx.builder.const_i32(2);
  572. ctx.builder.shl_i32();
  573. ctx.builder.load_aligned_i32(global_pointers::TLB_DATA);
  574. let entry_local = ctx.builder.tee_new_local();
  575. ctx.builder
  576. .const_i32((0xFFF & !TLB_GLOBAL & !(if ctx.cpu.cpl3() { 0 } else { TLB_NO_USER })) as i32);
  577. ctx.builder.and_i32();
  578. ctx.builder.const_i32(TLB_VALID as i32);
  579. ctx.builder.eq_i32();
  580. if bits != BitSize::BYTE {
  581. ctx.builder.get_local(&address_local);
  582. ctx.builder.const_i32(0xFFF);
  583. ctx.builder.and_i32();
  584. ctx.builder.const_i32(0x1000 - bits.bytes() as i32);
  585. ctx.builder.le_i32();
  586. ctx.builder.and_i32();
  587. }
  588. ctx.builder.br_if(0);
  589. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  590. ctx.builder.get_local(&address_local);
  591. ctx.builder.get_local(&entry_local);
  592. ctx.builder.call_fn2("report_safe_write_jit_slow");
  593. }
  594. ctx.builder.get_local(&address_local);
  595. match value_local {
  596. GenSafeWriteValue::I32(local) => ctx.builder.get_local(local),
  597. GenSafeWriteValue::I64(local) => ctx.builder.get_local_i64(local),
  598. GenSafeWriteValue::TwoI64s(local1, local2) => {
  599. ctx.builder.get_local_i64(local1);
  600. ctx.builder.get_local_i64(local2)
  601. },
  602. }
  603. ctx.builder
  604. .const_i32(ctx.start_of_current_instruction as i32 & 0xFFF);
  605. match bits {
  606. BitSize::BYTE => {
  607. ctx.builder.call_fn3_ret("safe_write8_slow_jit");
  608. },
  609. BitSize::WORD => {
  610. ctx.builder.call_fn3_ret("safe_write16_slow_jit");
  611. },
  612. BitSize::DWORD => {
  613. ctx.builder.call_fn3_ret("safe_write32_slow_jit");
  614. },
  615. BitSize::QWORD => {
  616. ctx.builder
  617. .call_fn3_i32_i64_i32_ret("safe_write64_slow_jit");
  618. },
  619. BitSize::DQWORD => {
  620. ctx.builder
  621. .call_fn4_i32_i64_i64_i32_ret("safe_write128_slow_jit");
  622. },
  623. }
  624. ctx.builder.tee_local(&entry_local);
  625. ctx.builder.const_i32(1);
  626. ctx.builder.and_i32();
  627. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  628. ctx.builder.if_void();
  629. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  630. ctx.builder.block_end();
  631. ctx.builder.get_local(&entry_local);
  632. ctx.builder.const_i32(1);
  633. ctx.builder.and_i32();
  634. }
  635. // -2 for the exit-with-pagefault block, +1 for leaving the nested if from this function
  636. ctx.builder.br_if(ctx.current_brtable_depth - 2 + 1);
  637. ctx.builder.block_end();
  638. gen_profiler_stat_increment(ctx.builder, profiler::stat::SAFE_WRITE_FAST); // XXX: Both fast and slow
  639. ctx.builder.get_local(&entry_local);
  640. ctx.builder.const_i32(!0xFFF);
  641. ctx.builder.and_i32();
  642. ctx.builder.get_local(&address_local);
  643. ctx.builder.xor_i32();
  644. ctx.builder.const_i32(unsafe { mem8 } as i32);
  645. ctx.builder.add_i32();
  646. match value_local {
  647. GenSafeWriteValue::I32(local) => ctx.builder.get_local(local),
  648. GenSafeWriteValue::I64(local) => ctx.builder.get_local_i64(local),
  649. GenSafeWriteValue::TwoI64s(local1, local2) => {
  650. assert!(bits == BitSize::DQWORD);
  651. let virt_address_local = ctx.builder.tee_new_local();
  652. ctx.builder.get_local_i64(local1);
  653. ctx.builder.store_unaligned_i64(0);
  654. ctx.builder.get_local(&virt_address_local);
  655. ctx.builder.get_local_i64(local2);
  656. ctx.builder.store_unaligned_i64(8);
  657. ctx.builder.free_local(virt_address_local);
  658. },
  659. }
  660. match bits {
  661. BitSize::BYTE => {
  662. ctx.builder.store_u8(0);
  663. },
  664. BitSize::WORD => {
  665. ctx.builder.store_unaligned_u16(0);
  666. },
  667. BitSize::DWORD => {
  668. ctx.builder.store_unaligned_i32(0);
  669. },
  670. BitSize::QWORD => {
  671. ctx.builder.store_unaligned_i64(0);
  672. },
  673. BitSize::DQWORD => {}, // handled above
  674. }
  675. ctx.builder.free_local(entry_local);
  676. }
  677. pub fn gen_safe_read_write(
  678. ctx: &mut JitContext,
  679. bits: BitSize,
  680. address_local: &WasmLocal,
  681. f: &dyn Fn(&mut JitContext),
  682. ) {
  683. // Execute a virtual memory read+write. All slow paths (memory-mapped IO, tlb miss, page fault,
  684. // write across page boundary and page containing jitted code are handled in
  685. // safe_read_write_jit_slow
  686. // entry <- tlb_data[addr >> 12 << 2]
  687. // can_use_fast_path <- entry & MASK == TLB_VALID && (addr & 0xFFF) <= 0x1000 - bytes
  688. // if can_use_fast_path: goto fast
  689. // entry <- safe_read_write_jit_slow(addr, instruction_pointer)
  690. // if page_fault: goto exit-with-pagefault
  691. // fast: value <- f(mem[(entry & ~0xFFF) ^ addr])
  692. // if !can_use_fast_path { safe_write_jit_slow(addr, value, instruction_pointer) }
  693. // mem[(entry & ~0xFFF) ^ addr] <- value
  694. ctx.builder.block_void();
  695. ctx.builder.get_local(address_local);
  696. ctx.builder.const_i32(12);
  697. ctx.builder.shr_u_i32();
  698. ctx.builder.const_i32(2);
  699. ctx.builder.shl_i32();
  700. ctx.builder.load_aligned_i32(global_pointers::TLB_DATA);
  701. let entry_local = ctx.builder.tee_new_local();
  702. ctx.builder
  703. .const_i32((0xFFF & !TLB_GLOBAL & !(if ctx.cpu.cpl3() { 0 } else { TLB_NO_USER })) as i32);
  704. ctx.builder.and_i32();
  705. ctx.builder.const_i32(TLB_VALID as i32);
  706. ctx.builder.eq_i32();
  707. if bits != BitSize::BYTE {
  708. ctx.builder.get_local(&address_local);
  709. ctx.builder.const_i32(0xFFF);
  710. ctx.builder.and_i32();
  711. ctx.builder.const_i32(0x1000 - bits.bytes() as i32);
  712. ctx.builder.le_i32();
  713. ctx.builder.and_i32();
  714. }
  715. let can_use_fast_path_local = ctx.builder.tee_new_local();
  716. ctx.builder.br_if(0);
  717. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  718. ctx.builder.get_local(&address_local);
  719. ctx.builder.get_local(&entry_local);
  720. ctx.builder.call_fn2("report_safe_read_write_jit_slow");
  721. }
  722. ctx.builder.get_local(&address_local);
  723. ctx.builder
  724. .const_i32(ctx.start_of_current_instruction as i32 & 0xFFF);
  725. match bits {
  726. BitSize::BYTE => {
  727. ctx.builder.call_fn2_ret("safe_read_write8_slow_jit");
  728. },
  729. BitSize::WORD => {
  730. ctx.builder.call_fn2_ret("safe_read_write16_slow_jit");
  731. },
  732. BitSize::DWORD => {
  733. ctx.builder.call_fn2_ret("safe_read_write32s_slow_jit");
  734. },
  735. BitSize::QWORD => dbg_assert!(false),
  736. BitSize::DQWORD => dbg_assert!(false),
  737. }
  738. ctx.builder.tee_local(&entry_local);
  739. ctx.builder.const_i32(1);
  740. ctx.builder.and_i32();
  741. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  742. ctx.builder.if_void();
  743. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  744. ctx.builder.block_end();
  745. ctx.builder.get_local(&entry_local);
  746. ctx.builder.const_i32(1);
  747. ctx.builder.and_i32();
  748. }
  749. // -2 for the exit-with-pagefault block, +1 for leaving the two nested ifs from this function
  750. ctx.builder.br_if(ctx.current_brtable_depth - 2 + 1);
  751. ctx.builder.block_end();
  752. gen_profiler_stat_increment(ctx.builder, profiler::stat::SAFE_READ_WRITE_FAST); // XXX: Also slow
  753. ctx.builder.get_local(&entry_local);
  754. ctx.builder.const_i32(!0xFFF);
  755. ctx.builder.and_i32();
  756. ctx.builder.get_local(&address_local);
  757. ctx.builder.xor_i32();
  758. ctx.builder.const_i32(unsafe { mem8 } as i32);
  759. ctx.builder.add_i32();
  760. ctx.builder.free_local(entry_local);
  761. let phys_addr_local = ctx.builder.tee_new_local();
  762. match bits {
  763. BitSize::BYTE => {
  764. ctx.builder.load_u8(0);
  765. },
  766. BitSize::WORD => {
  767. ctx.builder.load_unaligned_u16(0);
  768. },
  769. BitSize::DWORD => {
  770. ctx.builder.load_unaligned_i32(0);
  771. },
  772. BitSize::QWORD => assert!(false), // not used
  773. BitSize::DQWORD => assert!(false), // not used
  774. }
  775. // value is now on stack
  776. f(ctx);
  777. let value_local = ctx.builder.set_new_local(); // TODO: Could get rid of this local by returning one from f
  778. ctx.builder.get_local(&can_use_fast_path_local);
  779. ctx.builder.eqz_i32();
  780. ctx.builder.if_void();
  781. {
  782. ctx.builder.get_local(&address_local);
  783. ctx.builder.get_local(&value_local);
  784. ctx.builder
  785. .const_i32(ctx.start_of_current_instruction as i32);
  786. match bits {
  787. BitSize::BYTE => {
  788. ctx.builder.call_fn3_ret("safe_write8_slow_jit");
  789. },
  790. BitSize::WORD => {
  791. ctx.builder.call_fn3_ret("safe_write16_slow_jit");
  792. },
  793. BitSize::DWORD => {
  794. ctx.builder.call_fn3_ret("safe_write32_slow_jit");
  795. },
  796. BitSize::QWORD => dbg_assert!(false),
  797. BitSize::DQWORD => dbg_assert!(false),
  798. }
  799. ctx.builder.const_i32(1);
  800. ctx.builder.and_i32();
  801. ctx.builder.if_void();
  802. {
  803. // handled above
  804. if cfg!(debug_assertions) {
  805. ctx.builder.const_i32(match bits {
  806. BitSize::BYTE => 8,
  807. BitSize::WORD => 16,
  808. BitSize::DWORD => 32,
  809. _ => {
  810. dbg_assert!(false);
  811. 0
  812. },
  813. });
  814. ctx.builder.get_local(&address_local);
  815. ctx.builder.call_fn2("bug_gen_safe_read_write_page_fault");
  816. }
  817. else {
  818. ctx.builder.unreachable();
  819. }
  820. }
  821. ctx.builder.block_end();
  822. }
  823. ctx.builder.block_end();
  824. ctx.builder.get_local(&phys_addr_local);
  825. ctx.builder.get_local(&value_local);
  826. match bits {
  827. BitSize::BYTE => {
  828. ctx.builder.store_u8(0);
  829. },
  830. BitSize::WORD => {
  831. ctx.builder.store_unaligned_u16(0);
  832. },
  833. BitSize::DWORD => {
  834. ctx.builder.store_unaligned_i32(0);
  835. },
  836. BitSize::QWORD => dbg_assert!(false),
  837. BitSize::DQWORD => dbg_assert!(false),
  838. }
  839. ctx.builder.free_local(value_local);
  840. ctx.builder.free_local(can_use_fast_path_local);
  841. ctx.builder.free_local(phys_addr_local);
  842. }
  843. #[no_mangle]
  844. pub fn bug_gen_safe_read_write_page_fault(bits: i32, addr: u32) {
  845. dbg_log!("bug: gen_safe_read_write_page_fault {} {:x}", bits, addr);
  846. dbg_assert!(false);
  847. }
  848. pub fn gen_jmp_rel16(builder: &mut WasmBuilder, rel16: u16) {
  849. let cs_offset_addr = global_pointers::get_seg_offset(regs::CS);
  850. builder.load_fixed_i32(cs_offset_addr);
  851. let local = builder.set_new_local();
  852. // generate:
  853. // *instruction_pointer = cs_offset + ((*instruction_pointer - cs_offset + rel16) & 0xFFFF);
  854. {
  855. builder.const_i32(global_pointers::INSTRUCTION_POINTER as i32);
  856. gen_get_eip(builder);
  857. builder.get_local(&local);
  858. builder.sub_i32();
  859. builder.const_i32(rel16 as i32);
  860. builder.add_i32();
  861. builder.const_i32(0xFFFF);
  862. builder.and_i32();
  863. builder.get_local(&local);
  864. builder.add_i32();
  865. builder.store_aligned_i32(0);
  866. }
  867. builder.free_local(local);
  868. }
  869. pub fn gen_pop16_ss16(ctx: &mut JitContext) {
  870. // sp = segment_offsets[SS] + reg16[SP] (or just reg16[SP] if has_flat_segmentation)
  871. gen_get_reg16(ctx, regs::SP);
  872. if !ctx.cpu.has_flat_segmentation() {
  873. gen_get_ss_offset(ctx);
  874. ctx.builder.add_i32();
  875. }
  876. // result = safe_read16(sp)
  877. let address_local = ctx.builder.set_new_local();
  878. gen_safe_read16(ctx, &address_local);
  879. ctx.builder.free_local(address_local);
  880. // reg16[SP] += 2;
  881. gen_get_reg16(ctx, regs::SP);
  882. ctx.builder.const_i32(2);
  883. ctx.builder.add_i32();
  884. gen_set_reg16(ctx, regs::SP);
  885. // return value is already on stack
  886. }
  887. pub fn gen_pop16_ss32(ctx: &mut JitContext) {
  888. // esp = segment_offsets[SS] + reg32[ESP] (or just reg32[ESP] if has_flat_segmentation)
  889. gen_get_reg32(ctx, regs::ESP);
  890. if !ctx.cpu.has_flat_segmentation() {
  891. gen_get_ss_offset(ctx);
  892. ctx.builder.add_i32();
  893. }
  894. // result = safe_read16(esp)
  895. let address_local = ctx.builder.set_new_local();
  896. gen_safe_read16(ctx, &address_local);
  897. ctx.builder.free_local(address_local);
  898. // reg32[ESP] += 2;
  899. gen_get_reg32(ctx, regs::ESP);
  900. ctx.builder.const_i32(2);
  901. ctx.builder.add_i32();
  902. gen_set_reg32(ctx, regs::ESP);
  903. // return value is already on stack
  904. }
  905. pub fn gen_pop16(ctx: &mut JitContext) {
  906. if ctx.cpu.ssize_32() {
  907. gen_pop16_ss32(ctx);
  908. }
  909. else {
  910. gen_pop16_ss16(ctx);
  911. }
  912. }
  913. pub fn gen_pop32s_ss16(ctx: &mut JitContext) {
  914. // sp = reg16[SP]
  915. gen_get_reg16(ctx, regs::SP);
  916. // result = safe_read32s(segment_offsets[SS] + sp) (or just sp if has_flat_segmentation)
  917. if !ctx.cpu.has_flat_segmentation() {
  918. gen_get_ss_offset(ctx);
  919. ctx.builder.add_i32();
  920. }
  921. let address_local = ctx.builder.set_new_local();
  922. gen_safe_read32(ctx, &address_local);
  923. ctx.builder.free_local(address_local);
  924. // reg16[SP] = sp + 4;
  925. gen_get_reg16(ctx, regs::SP);
  926. ctx.builder.const_i32(4);
  927. ctx.builder.add_i32();
  928. gen_set_reg16(ctx, regs::SP);
  929. // return value is already on stack
  930. }
  931. pub fn gen_pop32s_ss32(ctx: &mut JitContext) {
  932. if !ctx.cpu.has_flat_segmentation() {
  933. gen_get_reg32(ctx, regs::ESP);
  934. gen_get_ss_offset(ctx);
  935. ctx.builder.add_i32();
  936. let address_local = ctx.builder.set_new_local();
  937. gen_safe_read32(ctx, &address_local);
  938. ctx.builder.free_local(address_local);
  939. }
  940. else {
  941. let reg = ctx.register_locals[regs::ESP as usize].unsafe_clone();
  942. gen_safe_read32(ctx, &reg);
  943. }
  944. gen_get_reg32(ctx, regs::ESP);
  945. ctx.builder.const_i32(4);
  946. ctx.builder.add_i32();
  947. gen_set_reg32(ctx, regs::ESP);
  948. // return value is already on stack
  949. }
  950. pub fn gen_pop32s(ctx: &mut JitContext) {
  951. if ctx.cpu.ssize_32() {
  952. gen_pop32s_ss32(ctx);
  953. }
  954. else {
  955. gen_pop32s_ss16(ctx);
  956. }
  957. }
  958. pub fn gen_adjust_stack_reg(ctx: &mut JitContext, offset: u32) {
  959. if ctx.cpu.ssize_32() {
  960. gen_get_reg32(ctx, regs::ESP);
  961. ctx.builder.const_i32(offset as i32);
  962. ctx.builder.add_i32();
  963. gen_set_reg32(ctx, regs::ESP);
  964. }
  965. else {
  966. gen_get_reg16(ctx, regs::SP);
  967. ctx.builder.const_i32(offset as i32);
  968. ctx.builder.add_i32();
  969. gen_set_reg16(ctx, regs::SP);
  970. }
  971. }
  972. pub fn gen_leave(ctx: &mut JitContext, os32: bool) {
  973. // [e]bp = safe_read{16,32}([e]bp)
  974. if ctx.cpu.ssize_32() {
  975. gen_get_reg32(ctx, regs::EBP);
  976. }
  977. else {
  978. gen_get_reg16(ctx, regs::BP);
  979. }
  980. let old_vbp = ctx.builder.tee_new_local();
  981. if !ctx.cpu.has_flat_segmentation() {
  982. gen_get_ss_offset(ctx);
  983. ctx.builder.add_i32();
  984. }
  985. if os32 {
  986. let address_local = ctx.builder.set_new_local();
  987. gen_safe_read32(ctx, &address_local);
  988. ctx.builder.free_local(address_local);
  989. gen_set_reg32(ctx, regs::EBP);
  990. }
  991. else {
  992. let address_local = ctx.builder.set_new_local();
  993. gen_safe_read16(ctx, &address_local);
  994. ctx.builder.free_local(address_local);
  995. gen_set_reg16(ctx, regs::BP);
  996. }
  997. // [e]sp = [e]bp + (os32 ? 4 : 2)
  998. if ctx.cpu.ssize_32() {
  999. ctx.builder.get_local(&old_vbp);
  1000. ctx.builder.const_i32(if os32 { 4 } else { 2 });
  1001. ctx.builder.add_i32();
  1002. gen_set_reg32(ctx, regs::ESP);
  1003. }
  1004. else {
  1005. ctx.builder.get_local(&old_vbp);
  1006. ctx.builder.const_i32(if os32 { 4 } else { 2 });
  1007. ctx.builder.add_i32();
  1008. gen_set_reg16(ctx, regs::SP);
  1009. }
  1010. ctx.builder.free_local(old_vbp);
  1011. }
  1012. pub fn gen_task_switch_test(ctx: &mut JitContext) {
  1013. // generate if(cr[0] & (CR0_EM | CR0_TS)) { task_switch_test_void(); return; }
  1014. let cr0_offset = global_pointers::get_creg_offset(0);
  1015. dbg_assert!(regs::CR0_EM | regs::CR0_TS <= 0xFF);
  1016. ctx.builder.load_fixed_u8(cr0_offset);
  1017. ctx.builder.const_i32((regs::CR0_EM | regs::CR0_TS) as i32);
  1018. ctx.builder.and_i32();
  1019. ctx.builder.if_void();
  1020. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  1021. gen_set_previous_eip_offset_from_eip_with_low_bits(
  1022. ctx.builder,
  1023. ctx.start_of_current_instruction as i32 & 0xFFF,
  1024. );
  1025. gen_move_registers_from_locals_to_memory(ctx);
  1026. gen_fn0_const(ctx.builder, "task_switch_test_jit");
  1027. ctx.builder.return_();
  1028. ctx.builder.block_end();
  1029. }
  1030. pub fn gen_task_switch_test_mmx(ctx: &mut JitContext) {
  1031. // generate if(cr[0] & (CR0_EM | CR0_TS)) { task_switch_test_mmx_void(); return; }
  1032. let cr0_offset = global_pointers::get_creg_offset(0);
  1033. dbg_assert!(regs::CR0_EM | regs::CR0_TS <= 0xFF);
  1034. ctx.builder.load_fixed_u8(cr0_offset);
  1035. ctx.builder.const_i32((regs::CR0_EM | regs::CR0_TS) as i32);
  1036. ctx.builder.and_i32();
  1037. ctx.builder.if_void();
  1038. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  1039. gen_set_previous_eip_offset_from_eip_with_low_bits(
  1040. ctx.builder,
  1041. ctx.start_of_current_instruction as i32 & 0xFFF,
  1042. );
  1043. gen_move_registers_from_locals_to_memory(ctx);
  1044. gen_fn0_const(ctx.builder, "task_switch_test_mmx_jit");
  1045. ctx.builder.return_();
  1046. ctx.builder.block_end();
  1047. }
  1048. pub fn gen_push16(ctx: &mut JitContext, value_local: &WasmLocal) {
  1049. if ctx.cpu.ssize_32() {
  1050. gen_get_reg32(ctx, regs::ESP);
  1051. }
  1052. else {
  1053. gen_get_reg16(ctx, regs::SP);
  1054. };
  1055. ctx.builder.const_i32(2);
  1056. ctx.builder.sub_i32();
  1057. let reg_updated_local = if !ctx.cpu.ssize_32() || !ctx.cpu.has_flat_segmentation() {
  1058. let reg_updated_local = ctx.builder.tee_new_local();
  1059. if !ctx.cpu.ssize_32() {
  1060. ctx.builder.const_i32(0xFFFF);
  1061. ctx.builder.and_i32();
  1062. }
  1063. if !ctx.cpu.has_flat_segmentation() {
  1064. gen_get_ss_offset(ctx);
  1065. ctx.builder.add_i32();
  1066. }
  1067. let sp_local = ctx.builder.set_new_local();
  1068. gen_safe_write16(ctx, &sp_local, &value_local);
  1069. ctx.builder.free_local(sp_local);
  1070. ctx.builder.get_local(&reg_updated_local);
  1071. reg_updated_local
  1072. }
  1073. else {
  1074. // short path: The address written to is equal to ESP/SP minus two
  1075. let reg_updated_local = ctx.builder.tee_new_local();
  1076. gen_safe_write16(ctx, &reg_updated_local, &value_local);
  1077. reg_updated_local
  1078. };
  1079. if ctx.cpu.ssize_32() {
  1080. gen_set_reg32(ctx, regs::ESP);
  1081. }
  1082. else {
  1083. gen_set_reg16(ctx, regs::SP);
  1084. };
  1085. ctx.builder.free_local(reg_updated_local);
  1086. }
  1087. pub fn gen_push32(ctx: &mut JitContext, value_local: &WasmLocal) {
  1088. if ctx.cpu.ssize_32() {
  1089. gen_get_reg32(ctx, regs::ESP);
  1090. }
  1091. else {
  1092. gen_get_reg16(ctx, regs::SP);
  1093. };
  1094. ctx.builder.const_i32(4);
  1095. ctx.builder.sub_i32();
  1096. let new_sp_local = if !ctx.cpu.ssize_32() || !ctx.cpu.has_flat_segmentation() {
  1097. let new_sp_local = ctx.builder.tee_new_local();
  1098. if !ctx.cpu.ssize_32() {
  1099. ctx.builder.const_i32(0xFFFF);
  1100. ctx.builder.and_i32();
  1101. }
  1102. if !ctx.cpu.has_flat_segmentation() {
  1103. gen_get_ss_offset(ctx);
  1104. ctx.builder.add_i32();
  1105. }
  1106. let sp_local = ctx.builder.set_new_local();
  1107. gen_safe_write32(ctx, &sp_local, &value_local);
  1108. ctx.builder.free_local(sp_local);
  1109. ctx.builder.get_local(&new_sp_local);
  1110. new_sp_local
  1111. }
  1112. else {
  1113. // short path: The address written to is equal to ESP/SP minus four
  1114. let new_sp_local = ctx.builder.tee_new_local();
  1115. gen_safe_write32(ctx, &new_sp_local, &value_local);
  1116. new_sp_local
  1117. };
  1118. if ctx.cpu.ssize_32() {
  1119. gen_set_reg32(ctx, regs::ESP);
  1120. }
  1121. else {
  1122. gen_set_reg16(ctx, regs::SP);
  1123. };
  1124. ctx.builder.free_local(new_sp_local);
  1125. }
  1126. pub fn gen_get_real_eip(ctx: &mut JitContext) {
  1127. gen_get_eip(ctx.builder);
  1128. ctx.builder
  1129. .load_fixed_i32(global_pointers::get_seg_offset(regs::CS));
  1130. ctx.builder.sub_i32();
  1131. }
  1132. pub fn gen_set_last_op1(builder: &mut WasmBuilder, source: &WasmLocal) {
  1133. builder.const_i32(global_pointers::LAST_OP1 as i32);
  1134. builder.get_local(&source);
  1135. builder.store_aligned_i32(0);
  1136. }
  1137. pub fn gen_set_last_result(builder: &mut WasmBuilder, source: &WasmLocal) {
  1138. builder.const_i32(global_pointers::LAST_RESULT as i32);
  1139. builder.get_local(&source);
  1140. builder.store_aligned_i32(0);
  1141. }
  1142. pub fn gen_set_last_op_size(builder: &mut WasmBuilder, value: i32) {
  1143. builder.const_i32(global_pointers::LAST_OP_SIZE as i32);
  1144. builder.const_i32(value);
  1145. builder.store_aligned_i32(0);
  1146. }
  1147. pub fn gen_set_flags_changed(builder: &mut WasmBuilder, value: i32) {
  1148. builder.const_i32(global_pointers::FLAGS_CHANGED as i32);
  1149. builder.const_i32(value);
  1150. builder.store_aligned_i32(0);
  1151. }
  1152. pub fn gen_set_flags_bits(builder: &mut WasmBuilder, bits_to_set: i32) {
  1153. builder.const_i32(global_pointers::FLAGS as i32);
  1154. gen_get_flags(builder);
  1155. builder.const_i32(bits_to_set);
  1156. builder.or_i32();
  1157. builder.store_aligned_i32(0);
  1158. }
  1159. pub fn gen_clear_flags_bits(builder: &mut WasmBuilder, bits_to_clear: i32) {
  1160. builder.const_i32(global_pointers::FLAGS as i32);
  1161. gen_get_flags(builder);
  1162. builder.const_i32(!bits_to_clear);
  1163. builder.and_i32();
  1164. builder.store_aligned_i32(0);
  1165. }
  1166. pub fn gen_getzf(builder: &mut WasmBuilder) {
  1167. gen_get_flags_changed(builder);
  1168. builder.const_i32(FLAG_ZERO);
  1169. builder.and_i32();
  1170. builder.if_i32();
  1171. gen_get_last_result(builder);
  1172. let last_result = builder.tee_new_local();
  1173. builder.const_i32(-1);
  1174. builder.xor_i32();
  1175. builder.get_local(&last_result);
  1176. builder.free_local(last_result);
  1177. builder.const_i32(1);
  1178. builder.sub_i32();
  1179. builder.and_i32();
  1180. gen_get_last_op_size(builder);
  1181. builder.shr_u_i32();
  1182. builder.const_i32(1);
  1183. builder.and_i32();
  1184. builder.else_();
  1185. gen_get_flags(builder);
  1186. builder.const_i32(FLAG_ZERO);
  1187. builder.and_i32();
  1188. builder.block_end();
  1189. }
  1190. pub fn gen_getcf(builder: &mut WasmBuilder) {
  1191. gen_get_flags_changed(builder);
  1192. let flags_changed = builder.tee_new_local();
  1193. builder.const_i32(FLAG_CARRY);
  1194. builder.and_i32();
  1195. builder.if_i32();
  1196. builder.get_local(&flags_changed);
  1197. builder.const_i32(31);
  1198. builder.shr_s_i32();
  1199. builder.free_local(flags_changed);
  1200. let sub_mask = builder.set_new_local();
  1201. gen_get_last_result(builder);
  1202. builder.get_local(&sub_mask);
  1203. builder.xor_i32();
  1204. gen_get_last_op1(builder);
  1205. builder.get_local(&sub_mask);
  1206. builder.xor_i32();
  1207. builder.ltu_i32();
  1208. builder.else_();
  1209. gen_get_flags(builder);
  1210. builder.const_i32(FLAG_CARRY);
  1211. builder.and_i32();
  1212. builder.block_end();
  1213. builder.free_local(sub_mask);
  1214. }
  1215. pub fn gen_getsf(builder: &mut WasmBuilder) {
  1216. gen_get_flags_changed(builder);
  1217. builder.const_i32(FLAG_SIGN);
  1218. builder.and_i32();
  1219. builder.if_i32();
  1220. {
  1221. gen_get_last_result(builder);
  1222. gen_get_last_op_size(builder);
  1223. builder.shr_u_i32();
  1224. builder.const_i32(1);
  1225. builder.and_i32();
  1226. }
  1227. builder.else_();
  1228. {
  1229. gen_get_flags(builder);
  1230. builder.const_i32(FLAG_SIGN);
  1231. builder.and_i32();
  1232. }
  1233. builder.block_end();
  1234. }
  1235. pub fn gen_getof(builder: &mut WasmBuilder) {
  1236. gen_get_flags_changed(builder);
  1237. let flags_changed = builder.tee_new_local();
  1238. builder.const_i32(FLAG_OVERFLOW);
  1239. builder.and_i32();
  1240. builder.if_i32();
  1241. {
  1242. gen_get_last_op1(builder);
  1243. let last_op1 = builder.tee_new_local();
  1244. gen_get_last_result(builder);
  1245. let last_result = builder.tee_new_local();
  1246. builder.xor_i32();
  1247. builder.get_local(&last_result);
  1248. builder.get_local(&last_op1);
  1249. builder.sub_i32();
  1250. gen_get_flags_changed(builder);
  1251. builder.const_i32(31);
  1252. builder.shr_u_i32();
  1253. builder.sub_i32();
  1254. builder.get_local(&last_result);
  1255. builder.xor_i32();
  1256. builder.and_i32();
  1257. gen_get_last_op_size(builder);
  1258. builder.shr_u_i32();
  1259. builder.const_i32(1);
  1260. builder.and_i32();
  1261. builder.free_local(last_op1);
  1262. builder.free_local(last_result);
  1263. }
  1264. builder.else_();
  1265. {
  1266. gen_get_flags(builder);
  1267. builder.const_i32(FLAG_OVERFLOW);
  1268. builder.and_i32();
  1269. }
  1270. builder.block_end();
  1271. builder.free_local(flags_changed);
  1272. }
  1273. pub fn gen_test_be(builder: &mut WasmBuilder) {
  1274. // TODO: A more efficient implementation is possible
  1275. gen_getcf(builder);
  1276. gen_getzf(builder);
  1277. builder.or_i32();
  1278. }
  1279. pub fn gen_test_l(builder: &mut WasmBuilder) {
  1280. // TODO: A more efficient implementation is possible
  1281. gen_getsf(builder);
  1282. builder.eqz_i32();
  1283. gen_getof(builder);
  1284. builder.eqz_i32();
  1285. builder.xor_i32();
  1286. }
  1287. pub fn gen_test_le(builder: &mut WasmBuilder) {
  1288. // TODO: A more efficient implementation is possible
  1289. gen_test_l(builder);
  1290. gen_getzf(builder);
  1291. builder.or_i32();
  1292. }
  1293. pub fn gen_test_loopnz(ctx: &mut JitContext, is_asize_32: bool) {
  1294. gen_test_loop(ctx, is_asize_32);
  1295. ctx.builder.eqz_i32();
  1296. gen_getzf(&mut ctx.builder);
  1297. ctx.builder.or_i32();
  1298. ctx.builder.eqz_i32();
  1299. }
  1300. pub fn gen_test_loopz(ctx: &mut JitContext, is_asize_32: bool) {
  1301. gen_test_loop(ctx, is_asize_32);
  1302. ctx.builder.eqz_i32();
  1303. gen_getzf(&mut ctx.builder);
  1304. ctx.builder.eqz_i32();
  1305. ctx.builder.or_i32();
  1306. ctx.builder.eqz_i32();
  1307. }
  1308. pub fn gen_test_loop(ctx: &mut JitContext, is_asize_32: bool) {
  1309. if is_asize_32 {
  1310. gen_get_reg32(ctx, regs::ECX);
  1311. }
  1312. else {
  1313. gen_get_reg16(ctx, regs::CX);
  1314. }
  1315. }
  1316. pub fn gen_test_jcxz(ctx: &mut JitContext, is_asize_32: bool) {
  1317. if is_asize_32 {
  1318. gen_get_reg32(ctx, regs::ECX);
  1319. }
  1320. else {
  1321. gen_get_reg16(ctx, regs::CX);
  1322. }
  1323. ctx.builder.eqz_i32();
  1324. }
  1325. pub fn gen_fpu_get_sti(ctx: &mut JitContext, i: u32) {
  1326. ctx.builder.const_i32(i as i32);
  1327. ctx.builder.call_fn1_ret_f64("fpu_get_sti");
  1328. }
  1329. pub fn gen_fpu_load_m32(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  1330. gen_modrm_resolve_safe_read32(ctx, modrm_byte);
  1331. ctx.builder.reinterpret_i32_as_f32();
  1332. ctx.builder.promote_f32_to_f64();
  1333. }
  1334. pub fn gen_fpu_load_m64(ctx: &mut JitContext, modrm_byte: ModrmByte) {
  1335. gen_modrm_resolve_safe_read64(ctx, modrm_byte);
  1336. ctx.builder.reinterpret_i64_as_f64();
  1337. }
  1338. pub fn gen_trigger_de(ctx: &mut JitContext) {
  1339. gen_move_registers_from_locals_to_memory(ctx);
  1340. gen_set_previous_eip_offset_from_eip_with_low_bits(
  1341. ctx.builder,
  1342. ctx.start_of_current_instruction as i32 & 0xFFF,
  1343. );
  1344. gen_fn0_const(ctx.builder, "trigger_de");
  1345. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  1346. ctx.builder.return_();
  1347. }
  1348. pub fn gen_trigger_ud(ctx: &mut JitContext) {
  1349. gen_move_registers_from_locals_to_memory(ctx);
  1350. gen_set_previous_eip_offset_from_eip_with_low_bits(
  1351. ctx.builder,
  1352. ctx.start_of_current_instruction as i32 & 0xFFF,
  1353. );
  1354. gen_fn0_const(ctx.builder, "trigger_ud");
  1355. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  1356. ctx.builder.return_();
  1357. }
  1358. pub fn gen_trigger_gp(ctx: &mut JitContext, error_code: u32) {
  1359. gen_move_registers_from_locals_to_memory(ctx);
  1360. gen_set_previous_eip_offset_from_eip_with_low_bits(
  1361. ctx.builder,
  1362. ctx.start_of_current_instruction as i32 & 0xFFF,
  1363. );
  1364. gen_fn1_const(ctx.builder, "trigger_gp", error_code);
  1365. gen_debug_track_jit_exit(ctx.builder, ctx.start_of_current_instruction);
  1366. ctx.builder.return_();
  1367. }
  1368. pub fn gen_condition_fn(ctx: &mut JitContext, condition: u8) {
  1369. if condition & 0xF0 == 0x00 || condition & 0xF0 == 0x70 || condition & 0xF0 == 0x80 {
  1370. match condition & 0xF {
  1371. 0x0 => {
  1372. gen_getof(ctx.builder);
  1373. },
  1374. 0x1 => {
  1375. gen_getof(ctx.builder);
  1376. ctx.builder.eqz_i32();
  1377. },
  1378. 0x2 => {
  1379. gen_getcf(ctx.builder);
  1380. },
  1381. 0x3 => {
  1382. gen_getcf(ctx.builder);
  1383. ctx.builder.eqz_i32();
  1384. },
  1385. 0x4 => {
  1386. gen_getzf(ctx.builder);
  1387. },
  1388. 0x5 => {
  1389. gen_getzf(ctx.builder);
  1390. ctx.builder.eqz_i32();
  1391. },
  1392. 0x6 => {
  1393. gen_test_be(ctx.builder);
  1394. },
  1395. 0x7 => {
  1396. gen_test_be(ctx.builder);
  1397. ctx.builder.eqz_i32();
  1398. },
  1399. 0x8 => {
  1400. gen_getsf(ctx.builder);
  1401. },
  1402. 0x9 => {
  1403. gen_getsf(ctx.builder);
  1404. ctx.builder.eqz_i32();
  1405. },
  1406. 0xA => ctx.builder.call_fn0_ret("test_p"),
  1407. 0xB => ctx.builder.call_fn0_ret("test_np"),
  1408. 0xC => {
  1409. gen_test_l(ctx.builder);
  1410. },
  1411. 0xD => {
  1412. gen_test_l(ctx.builder);
  1413. ctx.builder.eqz_i32();
  1414. },
  1415. 0xE => {
  1416. gen_test_le(ctx.builder);
  1417. },
  1418. 0xF => {
  1419. gen_test_le(ctx.builder);
  1420. ctx.builder.eqz_i32();
  1421. },
  1422. _ => dbg_assert!(false),
  1423. }
  1424. }
  1425. else {
  1426. // loop, loopnz, loopz, jcxz
  1427. dbg_assert!(condition & !0x3 == 0xE0);
  1428. if condition == 0xE0 {
  1429. gen_test_loopnz(ctx, ctx.cpu.asize_32());
  1430. }
  1431. else if condition == 0xE1 {
  1432. gen_test_loopz(ctx, ctx.cpu.asize_32());
  1433. }
  1434. else if condition == 0xE2 {
  1435. gen_test_loop(ctx, ctx.cpu.asize_32());
  1436. }
  1437. else if condition == 0xE3 {
  1438. gen_test_jcxz(ctx, ctx.cpu.asize_32());
  1439. }
  1440. }
  1441. }
  1442. const RECORD_LOCAL_MEMORY_MOVES_AT_COMPILE_TIME: bool = false;
  1443. pub fn gen_move_registers_from_locals_to_memory(ctx: &mut JitContext) {
  1444. let instruction = memory::read32s(ctx.start_of_current_instruction) as u32;
  1445. if RECORD_LOCAL_MEMORY_MOVES_AT_COMPILE_TIME {
  1446. ::opstats::record_opstat_unguarded_register(instruction);
  1447. }
  1448. else {
  1449. ::opstats::gen_opstat_unguarded_register(ctx.builder, instruction);
  1450. }
  1451. for i in 0..8 {
  1452. ctx.builder
  1453. .const_i32(global_pointers::get_reg32_offset(i as u32) as i32);
  1454. ctx.builder.get_local(&ctx.register_locals[i]);
  1455. ctx.builder.store_aligned_i32(0);
  1456. }
  1457. }
  1458. pub fn gen_move_registers_from_memory_to_locals(ctx: &mut JitContext) {
  1459. let instruction = memory::read32s(ctx.start_of_current_instruction) as u32;
  1460. if RECORD_LOCAL_MEMORY_MOVES_AT_COMPILE_TIME {
  1461. ::opstats::record_opstat_unguarded_register(instruction);
  1462. }
  1463. else {
  1464. ::opstats::gen_opstat_unguarded_register(ctx.builder, instruction);
  1465. }
  1466. for i in 0..8 {
  1467. ctx.builder
  1468. .const_i32(global_pointers::get_reg32_offset(i as u32) as i32);
  1469. ctx.builder.load_aligned_i32(0);
  1470. ctx.builder.set_local(&ctx.register_locals[i]);
  1471. }
  1472. }
  1473. pub fn gen_profiler_stat_increment(builder: &mut WasmBuilder, stat: profiler::stat) {
  1474. if !cfg!(feature = "profiler") || !cfg!(feature = "profiler_instrument") {
  1475. return;
  1476. }
  1477. let addr = unsafe { profiler::stat_array.as_mut_ptr().offset(stat as isize) } as u32;
  1478. builder.increment_fixed_i32(addr, 1)
  1479. }
  1480. pub fn gen_debug_track_jit_exit(builder: &mut WasmBuilder, address: u32) {
  1481. if cfg!(feature = "profiler") && cfg!(feature = "profiler_instrument") {
  1482. gen_fn1_const(builder, "track_jit_exit", address);
  1483. }
  1484. }