sljitNativeMIPS_common.c 62 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059
  1. /*
  2. * Stack-less Just-In-Time compiler
  3. *
  4. * Copyright Zoltan Herczeg (hzmester@freemail.hu). All rights reserved.
  5. *
  6. * Redistribution and use in source and binary forms, with or without modification, are
  7. * permitted provided that the following conditions are met:
  8. *
  9. * 1. Redistributions of source code must retain the above copyright notice, this list of
  10. * conditions and the following disclaimer.
  11. *
  12. * 2. Redistributions in binary form must reproduce the above copyright notice, this list
  13. * of conditions and the following disclaimer in the documentation and/or other materials
  14. * provided with the distribution.
  15. *
  16. * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY
  17. * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
  18. * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT
  19. * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
  20. * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
  21. * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
  22. * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
  23. * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
  24. * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
  25. */
  26. /* Latest MIPS architecture. */
  27. /* Automatically detect SLJIT_MIPS_R1 */
  28. SLJIT_API_FUNC_ATTRIBUTE const char* sljit_get_platform_name(void)
  29. {
  30. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  31. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  32. return "MIPS32-R1" SLJIT_CPUINFO;
  33. #else
  34. return "MIPS64-R1" SLJIT_CPUINFO;
  35. #endif
  36. #else /* SLJIT_MIPS_R1 */
  37. return "MIPS III" SLJIT_CPUINFO;
  38. #endif
  39. }
  40. /* Length of an instruction word
  41. Both for mips-32 and mips-64 */
  42. typedef sljit_u32 sljit_ins;
  43. #define TMP_REG1 (SLJIT_NUMBER_OF_REGISTERS + 2)
  44. #define TMP_REG2 (SLJIT_NUMBER_OF_REGISTERS + 3)
  45. #define TMP_REG3 (SLJIT_NUMBER_OF_REGISTERS + 4)
  46. /* For position independent code, t9 must contain the function address. */
  47. #define PIC_ADDR_REG TMP_REG2
  48. /* Floating point status register. */
  49. #define FCSR_REG 31
  50. /* Return address register. */
  51. #define RETURN_ADDR_REG 31
  52. /* Flags are kept in volatile registers. */
  53. #define EQUAL_FLAG 3
  54. #define OTHER_FLAG 1
  55. #define TMP_FREG1 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1)
  56. #define TMP_FREG2 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 2)
  57. static const sljit_u8 reg_map[SLJIT_NUMBER_OF_REGISTERS + 5] = {
  58. 0, 2, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 24, 23, 22, 21, 20, 19, 18, 17, 16, 29, 4, 25, 31
  59. };
  60. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  61. static const sljit_u8 freg_map[SLJIT_NUMBER_OF_FLOAT_REGISTERS + 3] = {
  62. 0, 0, 14, 2, 4, 6, 8, 12, 10
  63. };
  64. #else
  65. static const sljit_u8 freg_map[SLJIT_NUMBER_OF_FLOAT_REGISTERS + 3] = {
  66. 0, 0, 13, 14, 15, 16, 17, 12, 18
  67. };
  68. #endif
  69. /* --------------------------------------------------------------------- */
  70. /* Instrucion forms */
  71. /* --------------------------------------------------------------------- */
  72. #define S(s) (reg_map[s] << 21)
  73. #define T(t) (reg_map[t] << 16)
  74. #define D(d) (reg_map[d] << 11)
  75. #define FT(t) (freg_map[t] << 16)
  76. #define FS(s) (freg_map[s] << 11)
  77. #define FD(d) (freg_map[d] << 6)
  78. /* Absolute registers. */
  79. #define SA(s) ((s) << 21)
  80. #define TA(t) ((t) << 16)
  81. #define DA(d) ((d) << 11)
  82. #define IMM(imm) ((imm) & 0xffff)
  83. #define SH_IMM(imm) ((imm) << 6)
  84. #define DR(dr) (reg_map[dr])
  85. #define FR(dr) (freg_map[dr])
  86. #define HI(opcode) ((opcode) << 26)
  87. #define LO(opcode) (opcode)
  88. /* S = (16 << 21) D = (17 << 21) */
  89. #define FMT_S (16 << 21)
  90. #define FMT_D (17 << 21)
  91. #define ABS_S (HI(17) | FMT_S | LO(5))
  92. #define ADD_S (HI(17) | FMT_S | LO(0))
  93. #define ADDIU (HI(9))
  94. #define ADDU (HI(0) | LO(33))
  95. #define AND (HI(0) | LO(36))
  96. #define ANDI (HI(12))
  97. #define B (HI(4))
  98. #define BAL (HI(1) | (17 << 16))
  99. #define BC1F (HI(17) | (8 << 21))
  100. #define BC1T (HI(17) | (8 << 21) | (1 << 16))
  101. #define BEQ (HI(4))
  102. #define BGEZ (HI(1) | (1 << 16))
  103. #define BGTZ (HI(7))
  104. #define BLEZ (HI(6))
  105. #define BLTZ (HI(1) | (0 << 16))
  106. #define BNE (HI(5))
  107. #define BREAK (HI(0) | LO(13))
  108. #define CFC1 (HI(17) | (2 << 21))
  109. #define C_UN_S (HI(17) | FMT_S | LO(49))
  110. #define C_UEQ_S (HI(17) | FMT_S | LO(51))
  111. #define C_ULE_S (HI(17) | FMT_S | LO(55))
  112. #define C_ULT_S (HI(17) | FMT_S | LO(53))
  113. #define CVT_S_S (HI(17) | FMT_S | LO(32))
  114. #define DADDIU (HI(25))
  115. #define DADDU (HI(0) | LO(45))
  116. #define DDIV (HI(0) | LO(30))
  117. #define DDIVU (HI(0) | LO(31))
  118. #define DIV (HI(0) | LO(26))
  119. #define DIVU (HI(0) | LO(27))
  120. #define DIV_S (HI(17) | FMT_S | LO(3))
  121. #define DMULT (HI(0) | LO(28))
  122. #define DMULTU (HI(0) | LO(29))
  123. #define DSLL (HI(0) | LO(56))
  124. #define DSLL32 (HI(0) | LO(60))
  125. #define DSLLV (HI(0) | LO(20))
  126. #define DSRA (HI(0) | LO(59))
  127. #define DSRA32 (HI(0) | LO(63))
  128. #define DSRAV (HI(0) | LO(23))
  129. #define DSRL (HI(0) | LO(58))
  130. #define DSRL32 (HI(0) | LO(62))
  131. #define DSRLV (HI(0) | LO(22))
  132. #define DSUBU (HI(0) | LO(47))
  133. #define J (HI(2))
  134. #define JAL (HI(3))
  135. #define JALR (HI(0) | LO(9))
  136. #define JR (HI(0) | LO(8))
  137. #define LD (HI(55))
  138. #define LUI (HI(15))
  139. #define LW (HI(35))
  140. #define MFC1 (HI(17))
  141. #define MFHI (HI(0) | LO(16))
  142. #define MFLO (HI(0) | LO(18))
  143. #define MOV_S (HI(17) | FMT_S | LO(6))
  144. #define MTC1 (HI(17) | (4 << 21))
  145. #define MUL_S (HI(17) | FMT_S | LO(2))
  146. #define MULT (HI(0) | LO(24))
  147. #define MULTU (HI(0) | LO(25))
  148. #define NEG_S (HI(17) | FMT_S | LO(7))
  149. #define NOP (HI(0) | LO(0))
  150. #define NOR (HI(0) | LO(39))
  151. #define OR (HI(0) | LO(37))
  152. #define ORI (HI(13))
  153. #define SD (HI(63))
  154. #define SDC1 (HI(61))
  155. #define SLT (HI(0) | LO(42))
  156. #define SLTI (HI(10))
  157. #define SLTIU (HI(11))
  158. #define SLTU (HI(0) | LO(43))
  159. #define SLL (HI(0) | LO(0))
  160. #define SLLV (HI(0) | LO(4))
  161. #define SRL (HI(0) | LO(2))
  162. #define SRLV (HI(0) | LO(6))
  163. #define SRA (HI(0) | LO(3))
  164. #define SRAV (HI(0) | LO(7))
  165. #define SUB_S (HI(17) | FMT_S | LO(1))
  166. #define SUBU (HI(0) | LO(35))
  167. #define SW (HI(43))
  168. #define SWC1 (HI(57))
  169. #define TRUNC_W_S (HI(17) | FMT_S | LO(13))
  170. #define XOR (HI(0) | LO(38))
  171. #define XORI (HI(14))
  172. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  173. #define CLZ (HI(28) | LO(32))
  174. #define DCLZ (HI(28) | LO(36))
  175. #define MOVF (HI(0) | (0 << 16) | LO(1))
  176. #define MOVN (HI(0) | LO(11))
  177. #define MOVT (HI(0) | (1 << 16) | LO(1))
  178. #define MOVZ (HI(0) | LO(10))
  179. #define MUL (HI(28) | LO(2))
  180. #define PREF (HI(51))
  181. #define PREFX (HI(19) | LO(15))
  182. #define SEB (HI(31) | (16 << 6) | LO(32))
  183. #define SEH (HI(31) | (24 << 6) | LO(32))
  184. #endif
  185. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  186. #define ADDU_W ADDU
  187. #define ADDIU_W ADDIU
  188. #define SLL_W SLL
  189. #define SUBU_W SUBU
  190. #else
  191. #define ADDU_W DADDU
  192. #define ADDIU_W DADDIU
  193. #define SLL_W DSLL
  194. #define SUBU_W DSUBU
  195. #endif
  196. #define SIMM_MAX (0x7fff)
  197. #define SIMM_MIN (-0x8000)
  198. #define UIMM_MAX (0xffff)
  199. /* dest_reg is the absolute name of the register
  200. Useful for reordering instructions in the delay slot. */
  201. static sljit_s32 push_inst(struct sljit_compiler *compiler, sljit_ins ins, sljit_s32 delay_slot)
  202. {
  203. SLJIT_ASSERT(delay_slot == MOVABLE_INS || delay_slot >= UNMOVABLE_INS
  204. || delay_slot == ((ins >> 11) & 0x1f) || delay_slot == ((ins >> 16) & 0x1f));
  205. sljit_ins *ptr = (sljit_ins*)ensure_buf(compiler, sizeof(sljit_ins));
  206. FAIL_IF(!ptr);
  207. *ptr = ins;
  208. compiler->size++;
  209. compiler->delay_slot = delay_slot;
  210. return SLJIT_SUCCESS;
  211. }
  212. static SLJIT_INLINE sljit_ins invert_branch(sljit_s32 flags)
  213. {
  214. return (flags & IS_BIT26_COND) ? (1 << 26) : (1 << 16);
  215. }
  216. static SLJIT_INLINE sljit_ins* detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset)
  217. {
  218. sljit_sw diff;
  219. sljit_uw target_addr;
  220. sljit_ins *inst;
  221. sljit_ins saved_inst;
  222. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  223. if (jump->flags & (SLJIT_REWRITABLE_JUMP | IS_CALL))
  224. return code_ptr;
  225. #else
  226. if (jump->flags & SLJIT_REWRITABLE_JUMP)
  227. return code_ptr;
  228. #endif
  229. if (jump->flags & JUMP_ADDR)
  230. target_addr = jump->u.target;
  231. else {
  232. SLJIT_ASSERT(jump->flags & JUMP_LABEL);
  233. target_addr = (sljit_uw)(code + jump->u.label->size) + (sljit_uw)executable_offset;
  234. }
  235. inst = (sljit_ins *)jump->addr;
  236. if (jump->flags & IS_COND)
  237. inst--;
  238. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  239. if (jump->flags & IS_CALL)
  240. goto keep_address;
  241. #endif
  242. /* B instructions. */
  243. if (jump->flags & IS_MOVABLE) {
  244. diff = ((sljit_sw)target_addr - (sljit_sw)inst - executable_offset) >> 2;
  245. if (diff <= SIMM_MAX && diff >= SIMM_MIN) {
  246. jump->flags |= PATCH_B;
  247. if (!(jump->flags & IS_COND)) {
  248. inst[0] = inst[-1];
  249. inst[-1] = (jump->flags & IS_JAL) ? BAL : B;
  250. jump->addr -= sizeof(sljit_ins);
  251. return inst;
  252. }
  253. saved_inst = inst[0];
  254. inst[0] = inst[-1];
  255. inst[-1] = saved_inst ^ invert_branch(jump->flags);
  256. jump->addr -= 2 * sizeof(sljit_ins);
  257. return inst;
  258. }
  259. }
  260. else {
  261. diff = ((sljit_sw)target_addr - (sljit_sw)(inst + 1) - executable_offset) >> 2;
  262. if (diff <= SIMM_MAX && diff >= SIMM_MIN) {
  263. jump->flags |= PATCH_B;
  264. if (!(jump->flags & IS_COND)) {
  265. inst[0] = (jump->flags & IS_JAL) ? BAL : B;
  266. inst[1] = NOP;
  267. return inst + 1;
  268. }
  269. inst[0] = inst[0] ^ invert_branch(jump->flags);
  270. inst[1] = NOP;
  271. jump->addr -= sizeof(sljit_ins);
  272. return inst + 1;
  273. }
  274. }
  275. if (jump->flags & IS_COND) {
  276. if ((jump->flags & IS_MOVABLE) && (target_addr & ~0xfffffff) == ((jump->addr + 2 * sizeof(sljit_ins)) & ~0xfffffff)) {
  277. jump->flags |= PATCH_J;
  278. saved_inst = inst[0];
  279. inst[0] = inst[-1];
  280. inst[-1] = (saved_inst & 0xffff0000) | 3;
  281. inst[1] = J;
  282. inst[2] = NOP;
  283. return inst + 2;
  284. }
  285. else if ((target_addr & ~0xfffffff) == ((jump->addr + 3 * sizeof(sljit_ins)) & ~0xfffffff)) {
  286. jump->flags |= PATCH_J;
  287. inst[0] = (inst[0] & 0xffff0000) | 3;
  288. inst[1] = NOP;
  289. inst[2] = J;
  290. inst[3] = NOP;
  291. jump->addr += sizeof(sljit_ins);
  292. return inst + 3;
  293. }
  294. }
  295. else {
  296. /* J instuctions. */
  297. if ((jump->flags & IS_MOVABLE) && (target_addr & ~0xfffffff) == (jump->addr & ~0xfffffff)) {
  298. jump->flags |= PATCH_J;
  299. inst[0] = inst[-1];
  300. inst[-1] = (jump->flags & IS_JAL) ? JAL : J;
  301. jump->addr -= sizeof(sljit_ins);
  302. return inst;
  303. }
  304. if ((target_addr & ~0xfffffff) == ((jump->addr + sizeof(sljit_ins)) & ~0xfffffff)) {
  305. jump->flags |= PATCH_J;
  306. inst[0] = (jump->flags & IS_JAL) ? JAL : J;
  307. inst[1] = NOP;
  308. return inst + 1;
  309. }
  310. }
  311. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  312. keep_address:
  313. if (target_addr <= 0x7fffffff) {
  314. jump->flags |= PATCH_ABS32;
  315. if (jump->flags & IS_COND) {
  316. inst[0] -= 4;
  317. inst++;
  318. }
  319. inst[2] = inst[6];
  320. inst[3] = inst[7];
  321. return inst + 3;
  322. }
  323. if (target_addr <= 0x7fffffffffffl) {
  324. jump->flags |= PATCH_ABS48;
  325. if (jump->flags & IS_COND) {
  326. inst[0] -= 2;
  327. inst++;
  328. }
  329. inst[4] = inst[6];
  330. inst[5] = inst[7];
  331. return inst + 5;
  332. }
  333. #endif
  334. return code_ptr;
  335. }
  336. #ifdef __GNUC__
  337. static __attribute__ ((noinline)) void sljit_cache_flush(void* code, void* code_ptr)
  338. {
  339. SLJIT_CACHE_FLUSH(code, code_ptr);
  340. }
  341. #endif
  342. SLJIT_API_FUNC_ATTRIBUTE void* sljit_generate_code(struct sljit_compiler *compiler)
  343. {
  344. struct sljit_memory_fragment *buf;
  345. sljit_ins *code;
  346. sljit_ins *code_ptr;
  347. sljit_ins *buf_ptr;
  348. sljit_ins *buf_end;
  349. sljit_uw word_count;
  350. sljit_sw executable_offset;
  351. sljit_uw addr;
  352. struct sljit_label *label;
  353. struct sljit_jump *jump;
  354. struct sljit_const *const_;
  355. CHECK_ERROR_PTR();
  356. CHECK_PTR(check_sljit_generate_code(compiler));
  357. reverse_buf(compiler);
  358. code = (sljit_ins*)SLJIT_MALLOC_EXEC(compiler->size * sizeof(sljit_ins));
  359. PTR_FAIL_WITH_EXEC_IF(code);
  360. buf = compiler->buf;
  361. code_ptr = code;
  362. word_count = 0;
  363. executable_offset = SLJIT_EXEC_OFFSET(code);
  364. label = compiler->labels;
  365. jump = compiler->jumps;
  366. const_ = compiler->consts;
  367. do {
  368. buf_ptr = (sljit_ins*)buf->memory;
  369. buf_end = buf_ptr + (buf->used_size >> 2);
  370. do {
  371. *code_ptr = *buf_ptr++;
  372. SLJIT_ASSERT(!label || label->size >= word_count);
  373. SLJIT_ASSERT(!jump || jump->addr >= word_count);
  374. SLJIT_ASSERT(!const_ || const_->addr >= word_count);
  375. /* These structures are ordered by their address. */
  376. if (label && label->size == word_count) {
  377. label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
  378. label->size = code_ptr - code;
  379. label = label->next;
  380. }
  381. if (jump && jump->addr == word_count) {
  382. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  383. jump->addr = (sljit_uw)(code_ptr - 3);
  384. #else
  385. jump->addr = (sljit_uw)(code_ptr - 7);
  386. #endif
  387. code_ptr = detect_jump_type(jump, code_ptr, code, executable_offset);
  388. jump = jump->next;
  389. }
  390. if (const_ && const_->addr == word_count) {
  391. /* Just recording the address. */
  392. const_->addr = (sljit_uw)code_ptr;
  393. const_ = const_->next;
  394. }
  395. code_ptr ++;
  396. word_count ++;
  397. } while (buf_ptr < buf_end);
  398. buf = buf->next;
  399. } while (buf);
  400. if (label && label->size == word_count) {
  401. label->addr = (sljit_uw)code_ptr;
  402. label->size = code_ptr - code;
  403. label = label->next;
  404. }
  405. SLJIT_ASSERT(!label);
  406. SLJIT_ASSERT(!jump);
  407. SLJIT_ASSERT(!const_);
  408. SLJIT_ASSERT(code_ptr - code <= (sljit_sw)compiler->size);
  409. jump = compiler->jumps;
  410. while (jump) {
  411. do {
  412. addr = (jump->flags & JUMP_LABEL) ? jump->u.label->addr : jump->u.target;
  413. buf_ptr = (sljit_ins *)jump->addr;
  414. if (jump->flags & PATCH_B) {
  415. addr = (sljit_sw)(addr - ((sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset) + sizeof(sljit_ins))) >> 2;
  416. SLJIT_ASSERT((sljit_sw)addr <= SIMM_MAX && (sljit_sw)addr >= SIMM_MIN);
  417. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | (addr & 0xffff);
  418. break;
  419. }
  420. if (jump->flags & PATCH_J) {
  421. SLJIT_ASSERT((addr & ~0xfffffff) == (((sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset) + sizeof(sljit_ins)) & ~0xfffffff));
  422. buf_ptr[0] |= (addr >> 2) & 0x03ffffff;
  423. break;
  424. }
  425. /* Set the fields of immediate loads. */
  426. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  427. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 16) & 0xffff);
  428. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | (addr & 0xffff);
  429. #else
  430. if (jump->flags & PATCH_ABS32) {
  431. SLJIT_ASSERT(addr <= 0x7fffffff);
  432. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 16) & 0xffff);
  433. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | (addr & 0xffff);
  434. }
  435. else if (jump->flags & PATCH_ABS48) {
  436. SLJIT_ASSERT(addr <= 0x7fffffffffffl);
  437. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 32) & 0xffff);
  438. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | ((addr >> 16) & 0xffff);
  439. buf_ptr[3] = (buf_ptr[3] & 0xffff0000) | (addr & 0xffff);
  440. }
  441. else {
  442. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 48) & 0xffff);
  443. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | ((addr >> 32) & 0xffff);
  444. buf_ptr[3] = (buf_ptr[3] & 0xffff0000) | ((addr >> 16) & 0xffff);
  445. buf_ptr[5] = (buf_ptr[5] & 0xffff0000) | (addr & 0xffff);
  446. }
  447. #endif
  448. } while (0);
  449. jump = jump->next;
  450. }
  451. compiler->error = SLJIT_ERR_COMPILED;
  452. compiler->executable_offset = executable_offset;
  453. compiler->executable_size = (code_ptr - code) * sizeof(sljit_ins);
  454. code = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code, executable_offset);
  455. code_ptr = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
  456. #ifndef __GNUC__
  457. SLJIT_CACHE_FLUSH(code, code_ptr);
  458. #else
  459. /* GCC workaround for invalid code generation with -O2. */
  460. sljit_cache_flush(code, code_ptr);
  461. #endif
  462. return code;
  463. }
  464. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_has_cpu_feature(sljit_s32 feature_type)
  465. {
  466. sljit_sw fir = 0;
  467. switch (feature_type) {
  468. case SLJIT_HAS_FPU:
  469. #ifdef SLJIT_IS_FPU_AVAILABLE
  470. return SLJIT_IS_FPU_AVAILABLE;
  471. #elif defined(__GNUC__)
  472. asm ("cfc1 %0, $0" : "=r"(fir));
  473. return (fir >> 22) & 0x1;
  474. #else
  475. #error "FIR check is not implemented for this architecture"
  476. #endif
  477. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  478. case SLJIT_HAS_CLZ:
  479. case SLJIT_HAS_CMOV:
  480. return 1;
  481. #endif
  482. default:
  483. return fir;
  484. }
  485. }
  486. /* --------------------------------------------------------------------- */
  487. /* Entry, exit */
  488. /* --------------------------------------------------------------------- */
  489. /* Creates an index in data_transfer_insts array. */
  490. #define LOAD_DATA 0x01
  491. #define WORD_DATA 0x00
  492. #define BYTE_DATA 0x02
  493. #define HALF_DATA 0x04
  494. #define INT_DATA 0x06
  495. #define SIGNED_DATA 0x08
  496. /* Separates integer and floating point registers */
  497. #define GPR_REG 0x0f
  498. #define DOUBLE_DATA 0x10
  499. #define SINGLE_DATA 0x12
  500. #define MEM_MASK 0x1f
  501. #define ARG_TEST 0x00020
  502. #define ALT_KEEP_CACHE 0x00040
  503. #define CUMULATIVE_OP 0x00080
  504. #define LOGICAL_OP 0x00100
  505. #define IMM_OP 0x00200
  506. #define SRC2_IMM 0x00400
  507. #define UNUSED_DEST 0x00800
  508. #define REG_DEST 0x01000
  509. #define REG1_SOURCE 0x02000
  510. #define REG2_SOURCE 0x04000
  511. #define SLOW_SRC1 0x08000
  512. #define SLOW_SRC2 0x10000
  513. #define SLOW_DEST 0x20000
  514. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  515. #define STACK_STORE SW
  516. #define STACK_LOAD LW
  517. #else
  518. #define STACK_STORE SD
  519. #define STACK_LOAD LD
  520. #endif
  521. static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw);
  522. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  523. #include "sljitNativeMIPS_32.c"
  524. #else
  525. #include "sljitNativeMIPS_64.c"
  526. #endif
  527. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_enter(struct sljit_compiler *compiler,
  528. sljit_s32 options, sljit_s32 arg_types, sljit_s32 scratches, sljit_s32 saveds,
  529. sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
  530. {
  531. sljit_ins base;
  532. sljit_s32 args, i, tmp, offs;
  533. CHECK_ERROR();
  534. CHECK(check_sljit_emit_enter(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size));
  535. set_emit_enter(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size);
  536. local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 1) + SLJIT_LOCALS_OFFSET;
  537. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  538. local_size = (local_size + 15) & ~0xf;
  539. #else
  540. local_size = (local_size + 31) & ~0x1f;
  541. #endif
  542. compiler->local_size = local_size;
  543. if (local_size <= SIMM_MAX) {
  544. /* Frequent case. */
  545. FAIL_IF(push_inst(compiler, ADDIU_W | S(SLJIT_SP) | T(SLJIT_SP) | IMM(-local_size), DR(SLJIT_SP)));
  546. base = S(SLJIT_SP);
  547. offs = local_size - (sljit_sw)sizeof(sljit_sw);
  548. }
  549. else {
  550. FAIL_IF(load_immediate(compiler, DR(OTHER_FLAG), local_size));
  551. FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_SP) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
  552. FAIL_IF(push_inst(compiler, SUBU_W | S(SLJIT_SP) | T(OTHER_FLAG) | D(SLJIT_SP), DR(SLJIT_SP)));
  553. base = S(TMP_REG2);
  554. local_size = 0;
  555. offs = -(sljit_sw)sizeof(sljit_sw);
  556. }
  557. FAIL_IF(push_inst(compiler, STACK_STORE | base | TA(RETURN_ADDR_REG) | IMM(offs), MOVABLE_INS));
  558. tmp = saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - saveds) : SLJIT_FIRST_SAVED_REG;
  559. for (i = SLJIT_S0; i >= tmp; i--) {
  560. offs -= (sljit_s32)(sizeof(sljit_sw));
  561. FAIL_IF(push_inst(compiler, STACK_STORE | base | T(i) | IMM(offs), MOVABLE_INS));
  562. }
  563. for (i = scratches; i >= SLJIT_FIRST_SAVED_REG; i--) {
  564. offs -= (sljit_s32)(sizeof(sljit_sw));
  565. FAIL_IF(push_inst(compiler, STACK_STORE | base | T(i) | IMM(offs), MOVABLE_INS));
  566. }
  567. args = get_arg_count(arg_types);
  568. if (args >= 1)
  569. FAIL_IF(push_inst(compiler, ADDU_W | SA(4) | TA(0) | D(SLJIT_S0), DR(SLJIT_S0)));
  570. if (args >= 2)
  571. FAIL_IF(push_inst(compiler, ADDU_W | SA(5) | TA(0) | D(SLJIT_S1), DR(SLJIT_S1)));
  572. if (args >= 3)
  573. FAIL_IF(push_inst(compiler, ADDU_W | SA(6) | TA(0) | D(SLJIT_S2), DR(SLJIT_S2)));
  574. return SLJIT_SUCCESS;
  575. }
  576. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_set_context(struct sljit_compiler *compiler,
  577. sljit_s32 options, sljit_s32 arg_types, sljit_s32 scratches, sljit_s32 saveds,
  578. sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
  579. {
  580. CHECK_ERROR();
  581. CHECK(check_sljit_set_context(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size));
  582. set_set_context(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size);
  583. local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 1) + SLJIT_LOCALS_OFFSET;
  584. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  585. compiler->local_size = (local_size + 15) & ~0xf;
  586. #else
  587. compiler->local_size = (local_size + 31) & ~0x1f;
  588. #endif
  589. return SLJIT_SUCCESS;
  590. }
  591. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_return(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src, sljit_sw srcw)
  592. {
  593. sljit_s32 local_size, i, tmp, offs;
  594. sljit_ins base;
  595. CHECK_ERROR();
  596. CHECK(check_sljit_emit_return(compiler, op, src, srcw));
  597. FAIL_IF(emit_mov_before_return(compiler, op, src, srcw));
  598. local_size = compiler->local_size;
  599. if (local_size <= SIMM_MAX)
  600. base = S(SLJIT_SP);
  601. else {
  602. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), local_size));
  603. FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_SP) | T(TMP_REG1) | D(TMP_REG1), DR(TMP_REG1)));
  604. base = S(TMP_REG1);
  605. local_size = 0;
  606. }
  607. FAIL_IF(push_inst(compiler, STACK_LOAD | base | TA(RETURN_ADDR_REG) | IMM(local_size - (sljit_s32)sizeof(sljit_sw)), RETURN_ADDR_REG));
  608. offs = local_size - (sljit_s32)GET_SAVED_REGISTERS_SIZE(compiler->scratches, compiler->saveds, 1);
  609. tmp = compiler->scratches;
  610. for (i = SLJIT_FIRST_SAVED_REG; i <= tmp; i++) {
  611. FAIL_IF(push_inst(compiler, STACK_LOAD | base | T(i) | IMM(offs), DR(i)));
  612. offs += (sljit_s32)(sizeof(sljit_sw));
  613. }
  614. tmp = compiler->saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - compiler->saveds) : SLJIT_FIRST_SAVED_REG;
  615. for (i = tmp; i <= SLJIT_S0; i++) {
  616. FAIL_IF(push_inst(compiler, STACK_LOAD | base | T(i) | IMM(offs), DR(i)));
  617. offs += (sljit_s32)(sizeof(sljit_sw));
  618. }
  619. SLJIT_ASSERT(offs == local_size - (sljit_sw)(sizeof(sljit_sw)));
  620. FAIL_IF(push_inst(compiler, JR | SA(RETURN_ADDR_REG), UNMOVABLE_INS));
  621. if (compiler->local_size <= SIMM_MAX)
  622. return push_inst(compiler, ADDIU_W | S(SLJIT_SP) | T(SLJIT_SP) | IMM(compiler->local_size), UNMOVABLE_INS);
  623. else
  624. return push_inst(compiler, ADDU_W | S(TMP_REG1) | TA(0) | D(SLJIT_SP), UNMOVABLE_INS);
  625. }
  626. #undef STACK_STORE
  627. #undef STACK_LOAD
  628. /* --------------------------------------------------------------------- */
  629. /* Operators */
  630. /* --------------------------------------------------------------------- */
  631. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  632. #define ARCH_32_64(a, b) a
  633. #else
  634. #define ARCH_32_64(a, b) b
  635. #endif
  636. static const sljit_ins data_transfer_insts[16 + 4] = {
  637. /* u w s */ ARCH_32_64(HI(43) /* sw */, HI(63) /* sd */),
  638. /* u w l */ ARCH_32_64(HI(35) /* lw */, HI(55) /* ld */),
  639. /* u b s */ HI(40) /* sb */,
  640. /* u b l */ HI(36) /* lbu */,
  641. /* u h s */ HI(41) /* sh */,
  642. /* u h l */ HI(37) /* lhu */,
  643. /* u i s */ HI(43) /* sw */,
  644. /* u i l */ ARCH_32_64(HI(35) /* lw */, HI(39) /* lwu */),
  645. /* s w s */ ARCH_32_64(HI(43) /* sw */, HI(63) /* sd */),
  646. /* s w l */ ARCH_32_64(HI(35) /* lw */, HI(55) /* ld */),
  647. /* s b s */ HI(40) /* sb */,
  648. /* s b l */ HI(32) /* lb */,
  649. /* s h s */ HI(41) /* sh */,
  650. /* s h l */ HI(33) /* lh */,
  651. /* s i s */ HI(43) /* sw */,
  652. /* s i l */ HI(35) /* lw */,
  653. /* d s */ HI(61) /* sdc1 */,
  654. /* d l */ HI(53) /* ldc1 */,
  655. /* s s */ HI(57) /* swc1 */,
  656. /* s l */ HI(49) /* lwc1 */,
  657. };
  658. #undef ARCH_32_64
  659. /* reg_ar is an absoulute register! */
  660. /* Can perform an operation using at most 1 instruction. */
  661. static sljit_s32 getput_arg_fast(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw)
  662. {
  663. SLJIT_ASSERT(arg & SLJIT_MEM);
  664. if (!(arg & OFFS_REG_MASK) && argw <= SIMM_MAX && argw >= SIMM_MIN) {
  665. /* Works for both absoulte and relative addresses. */
  666. if (SLJIT_UNLIKELY(flags & ARG_TEST))
  667. return 1;
  668. FAIL_IF(push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(arg & REG_MASK)
  669. | TA(reg_ar) | IMM(argw), ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) ? reg_ar : MOVABLE_INS));
  670. return -1;
  671. }
  672. return 0;
  673. }
  674. /* See getput_arg below.
  675. Note: can_cache is called only for binary operators. Those
  676. operators always uses word arguments without write back. */
  677. static sljit_s32 can_cache(sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
  678. {
  679. SLJIT_ASSERT((arg & SLJIT_MEM) && (next_arg & SLJIT_MEM));
  680. /* Simple operation except for updates. */
  681. if (arg & OFFS_REG_MASK) {
  682. argw &= 0x3;
  683. next_argw &= 0x3;
  684. if (argw && argw == next_argw && (arg == next_arg || (arg & OFFS_REG_MASK) == (next_arg & OFFS_REG_MASK)))
  685. return 1;
  686. return 0;
  687. }
  688. if (arg == next_arg) {
  689. if (((next_argw - argw) <= SIMM_MAX && (next_argw - argw) >= SIMM_MIN))
  690. return 1;
  691. return 0;
  692. }
  693. return 0;
  694. }
  695. /* Emit the necessary instructions. See can_cache above. */
  696. static sljit_s32 getput_arg(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
  697. {
  698. sljit_s32 tmp_ar, base, delay_slot;
  699. SLJIT_ASSERT(arg & SLJIT_MEM);
  700. if (!(next_arg & SLJIT_MEM)) {
  701. next_arg = 0;
  702. next_argw = 0;
  703. }
  704. if ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) {
  705. tmp_ar = reg_ar;
  706. delay_slot = reg_ar;
  707. }
  708. else {
  709. tmp_ar = DR(TMP_REG1);
  710. delay_slot = MOVABLE_INS;
  711. }
  712. base = arg & REG_MASK;
  713. if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) {
  714. argw &= 0x3;
  715. /* Using the cache. */
  716. if (argw == compiler->cache_argw) {
  717. if (arg == compiler->cache_arg)
  718. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  719. if ((SLJIT_MEM | (arg & OFFS_REG_MASK)) == compiler->cache_arg) {
  720. if (arg == next_arg && argw == (next_argw & 0x3)) {
  721. compiler->cache_arg = arg;
  722. compiler->cache_argw = argw;
  723. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | D(TMP_REG3), DR(TMP_REG3)));
  724. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  725. }
  726. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | DA(tmp_ar), tmp_ar));
  727. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  728. }
  729. }
  730. if (SLJIT_UNLIKELY(argw)) {
  731. compiler->cache_arg = SLJIT_MEM | (arg & OFFS_REG_MASK);
  732. compiler->cache_argw = argw;
  733. FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(arg)) | D(TMP_REG3) | SH_IMM(argw), DR(TMP_REG3)));
  734. }
  735. if (arg == next_arg && argw == (next_argw & 0x3)) {
  736. compiler->cache_arg = arg;
  737. compiler->cache_argw = argw;
  738. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(!argw ? OFFS_REG(arg) : TMP_REG3) | D(TMP_REG3), DR(TMP_REG3)));
  739. tmp_ar = DR(TMP_REG3);
  740. }
  741. else
  742. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(!argw ? OFFS_REG(arg) : TMP_REG3) | DA(tmp_ar), tmp_ar));
  743. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  744. }
  745. if (compiler->cache_arg == arg && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
  746. if (argw != compiler->cache_argw) {
  747. FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
  748. compiler->cache_argw = argw;
  749. }
  750. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  751. }
  752. if (compiler->cache_arg == SLJIT_MEM && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
  753. if (argw != compiler->cache_argw)
  754. FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
  755. }
  756. else {
  757. compiler->cache_arg = SLJIT_MEM;
  758. FAIL_IF(load_immediate(compiler, DR(TMP_REG3), argw));
  759. }
  760. compiler->cache_argw = argw;
  761. if (!base)
  762. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  763. if (arg == next_arg && next_argw - argw <= SIMM_MAX && next_argw - argw >= SIMM_MIN) {
  764. compiler->cache_arg = arg;
  765. FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | T(base) | D(TMP_REG3), DR(TMP_REG3)));
  766. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  767. }
  768. FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | T(base) | DA(tmp_ar), tmp_ar));
  769. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  770. }
  771. static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw)
  772. {
  773. sljit_s32 tmp_ar, base, delay_slot;
  774. if (getput_arg_fast(compiler, flags, reg_ar, arg, argw))
  775. return compiler->error;
  776. if ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) {
  777. tmp_ar = reg_ar;
  778. delay_slot = reg_ar;
  779. }
  780. else {
  781. tmp_ar = DR(TMP_REG1);
  782. delay_slot = MOVABLE_INS;
  783. }
  784. base = arg & REG_MASK;
  785. if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) {
  786. argw &= 0x3;
  787. if (SLJIT_UNLIKELY(argw)) {
  788. FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(arg)) | DA(tmp_ar) | SH_IMM(argw), tmp_ar));
  789. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | TA(tmp_ar) | DA(tmp_ar), tmp_ar));
  790. }
  791. else
  792. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(OFFS_REG(arg)) | DA(tmp_ar), tmp_ar));
  793. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  794. }
  795. FAIL_IF(load_immediate(compiler, tmp_ar, argw));
  796. if (base != 0)
  797. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | TA(tmp_ar) | DA(tmp_ar), tmp_ar));
  798. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  799. }
  800. static SLJIT_INLINE sljit_s32 emit_op_mem2(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg1, sljit_sw arg1w, sljit_s32 arg2, sljit_sw arg2w)
  801. {
  802. if (getput_arg_fast(compiler, flags, reg, arg1, arg1w))
  803. return compiler->error;
  804. return getput_arg(compiler, flags, reg, arg1, arg1w, arg2, arg2w);
  805. }
  806. static sljit_s32 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags,
  807. sljit_s32 dst, sljit_sw dstw,
  808. sljit_s32 src1, sljit_sw src1w,
  809. sljit_s32 src2, sljit_sw src2w)
  810. {
  811. /* arg1 goes to TMP_REG1 or src reg
  812. arg2 goes to TMP_REG2, imm or src reg
  813. TMP_REG3 can be used for caching
  814. result goes to TMP_REG2, so put result can use TMP_REG1 and TMP_REG3. */
  815. sljit_s32 dst_r = TMP_REG2;
  816. sljit_s32 src1_r;
  817. sljit_sw src2_r = 0;
  818. sljit_s32 sugg_src2_r = TMP_REG2;
  819. if (!(flags & ALT_KEEP_CACHE)) {
  820. compiler->cache_arg = 0;
  821. compiler->cache_argw = 0;
  822. }
  823. if (SLJIT_UNLIKELY(dst == SLJIT_UNUSED)) {
  824. SLJIT_ASSERT(HAS_FLAGS(op));
  825. flags |= UNUSED_DEST;
  826. }
  827. else if (FAST_IS_REG(dst)) {
  828. dst_r = dst;
  829. flags |= REG_DEST;
  830. if (op >= SLJIT_MOV && op <= SLJIT_MOV_P)
  831. sugg_src2_r = dst_r;
  832. }
  833. else if ((dst & SLJIT_MEM) && !getput_arg_fast(compiler, flags | ARG_TEST, DR(TMP_REG1), dst, dstw))
  834. flags |= SLOW_DEST;
  835. if (flags & IMM_OP) {
  836. if ((src2 & SLJIT_IMM) && src2w) {
  837. if ((!(flags & LOGICAL_OP) && (src2w <= SIMM_MAX && src2w >= SIMM_MIN))
  838. || ((flags & LOGICAL_OP) && !(src2w & ~UIMM_MAX))) {
  839. flags |= SRC2_IMM;
  840. src2_r = src2w;
  841. }
  842. }
  843. if (!(flags & SRC2_IMM) && (flags & CUMULATIVE_OP) && (src1 & SLJIT_IMM) && src1w) {
  844. if ((!(flags & LOGICAL_OP) && (src1w <= SIMM_MAX && src1w >= SIMM_MIN))
  845. || ((flags & LOGICAL_OP) && !(src1w & ~UIMM_MAX))) {
  846. flags |= SRC2_IMM;
  847. src2_r = src1w;
  848. /* And swap arguments. */
  849. src1 = src2;
  850. src1w = src2w;
  851. src2 = SLJIT_IMM;
  852. /* src2w = src2_r unneeded. */
  853. }
  854. }
  855. }
  856. /* Source 1. */
  857. if (FAST_IS_REG(src1)) {
  858. src1_r = src1;
  859. flags |= REG1_SOURCE;
  860. }
  861. else if (src1 & SLJIT_IMM) {
  862. if (src1w) {
  863. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), src1w));
  864. src1_r = TMP_REG1;
  865. }
  866. else
  867. src1_r = 0;
  868. }
  869. else {
  870. if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w))
  871. FAIL_IF(compiler->error);
  872. else
  873. flags |= SLOW_SRC1;
  874. src1_r = TMP_REG1;
  875. }
  876. /* Source 2. */
  877. if (FAST_IS_REG(src2)) {
  878. src2_r = src2;
  879. flags |= REG2_SOURCE;
  880. if (!(flags & REG_DEST) && op >= SLJIT_MOV && op <= SLJIT_MOV_P)
  881. dst_r = src2_r;
  882. }
  883. else if (src2 & SLJIT_IMM) {
  884. if (!(flags & SRC2_IMM)) {
  885. if (src2w) {
  886. FAIL_IF(load_immediate(compiler, DR(sugg_src2_r), src2w));
  887. src2_r = sugg_src2_r;
  888. }
  889. else {
  890. src2_r = 0;
  891. if ((op >= SLJIT_MOV && op <= SLJIT_MOV_P) && (dst & SLJIT_MEM))
  892. dst_r = 0;
  893. }
  894. }
  895. }
  896. else {
  897. if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(sugg_src2_r), src2, src2w))
  898. FAIL_IF(compiler->error);
  899. else
  900. flags |= SLOW_SRC2;
  901. src2_r = sugg_src2_r;
  902. }
  903. if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
  904. SLJIT_ASSERT(src2_r == TMP_REG2);
  905. if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
  906. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, src1, src1w));
  907. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, dst, dstw));
  908. }
  909. else {
  910. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, src2, src2w));
  911. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, dst, dstw));
  912. }
  913. }
  914. else if (flags & SLOW_SRC1)
  915. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, dst, dstw));
  916. else if (flags & SLOW_SRC2)
  917. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(sugg_src2_r), src2, src2w, dst, dstw));
  918. FAIL_IF(emit_single_op(compiler, op, flags, dst_r, src1_r, src2_r));
  919. if (dst & SLJIT_MEM) {
  920. if (!(flags & SLOW_DEST)) {
  921. getput_arg_fast(compiler, flags, DR(dst_r), dst, dstw);
  922. return compiler->error;
  923. }
  924. return getput_arg(compiler, flags, DR(dst_r), dst, dstw, 0, 0);
  925. }
  926. return SLJIT_SUCCESS;
  927. }
  928. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op0(struct sljit_compiler *compiler, sljit_s32 op)
  929. {
  930. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  931. sljit_s32 int_op = op & SLJIT_I32_OP;
  932. #endif
  933. CHECK_ERROR();
  934. CHECK(check_sljit_emit_op0(compiler, op));
  935. op = GET_OPCODE(op);
  936. switch (op) {
  937. case SLJIT_BREAKPOINT:
  938. return push_inst(compiler, BREAK, UNMOVABLE_INS);
  939. case SLJIT_NOP:
  940. return push_inst(compiler, NOP, UNMOVABLE_INS);
  941. case SLJIT_LMUL_UW:
  942. case SLJIT_LMUL_SW:
  943. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  944. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? DMULTU : DMULT) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  945. #else
  946. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? MULTU : MULT) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  947. #endif
  948. FAIL_IF(push_inst(compiler, MFLO | D(SLJIT_R0), DR(SLJIT_R0)));
  949. return push_inst(compiler, MFHI | D(SLJIT_R1), DR(SLJIT_R1));
  950. case SLJIT_DIVMOD_UW:
  951. case SLJIT_DIVMOD_SW:
  952. case SLJIT_DIV_UW:
  953. case SLJIT_DIV_SW:
  954. SLJIT_COMPILE_ASSERT((SLJIT_DIVMOD_UW & 0x2) == 0 && SLJIT_DIV_UW - 0x2 == SLJIT_DIVMOD_UW, bad_div_opcode_assignments);
  955. #if !(defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  956. FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  957. FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  958. #endif
  959. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  960. if (int_op)
  961. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  962. else
  963. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DDIVU : DDIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  964. #else
  965. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  966. #endif
  967. FAIL_IF(push_inst(compiler, MFLO | D(SLJIT_R0), DR(SLJIT_R0)));
  968. return (op >= SLJIT_DIV_UW) ? SLJIT_SUCCESS : push_inst(compiler, MFHI | D(SLJIT_R1), DR(SLJIT_R1));
  969. }
  970. return SLJIT_SUCCESS;
  971. }
  972. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  973. static sljit_s32 emit_prefetch(struct sljit_compiler *compiler,
  974. sljit_s32 src, sljit_sw srcw)
  975. {
  976. if (!(src & OFFS_REG_MASK)) {
  977. if (srcw <= SIMM_MAX && srcw >= SIMM_MIN)
  978. return push_inst(compiler, PREF | S(src & REG_MASK) | IMM(srcw), MOVABLE_INS);
  979. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
  980. return push_inst(compiler, PREFX | S(src & REG_MASK) | T(TMP_REG1), MOVABLE_INS);
  981. }
  982. srcw &= 0x3;
  983. if (SLJIT_UNLIKELY(srcw != 0)) {
  984. FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(src)) | D(TMP_REG1) | SH_IMM(srcw), DR(TMP_REG1)));
  985. return push_inst(compiler, PREFX | S(src & REG_MASK) | T(TMP_REG1), MOVABLE_INS);
  986. }
  987. return push_inst(compiler, PREFX | S(src & REG_MASK) | T(OFFS_REG(src)), MOVABLE_INS);
  988. }
  989. #endif
  990. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op1(struct sljit_compiler *compiler, sljit_s32 op,
  991. sljit_s32 dst, sljit_sw dstw,
  992. sljit_s32 src, sljit_sw srcw)
  993. {
  994. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  995. # define flags 0
  996. #else
  997. sljit_s32 flags = 0;
  998. #endif
  999. CHECK_ERROR();
  1000. CHECK(check_sljit_emit_op1(compiler, op, dst, dstw, src, srcw));
  1001. ADJUST_LOCAL_OFFSET(dst, dstw);
  1002. ADJUST_LOCAL_OFFSET(src, srcw);
  1003. if (dst == SLJIT_UNUSED && !HAS_FLAGS(op)) {
  1004. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  1005. if (op <= SLJIT_MOV_P && (src & SLJIT_MEM))
  1006. return emit_prefetch(compiler, src, srcw);
  1007. #endif
  1008. return SLJIT_SUCCESS;
  1009. }
  1010. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1011. if ((op & SLJIT_I32_OP) && GET_OPCODE(op) >= SLJIT_NOT)
  1012. flags |= INT_DATA | SIGNED_DATA;
  1013. #endif
  1014. switch (GET_OPCODE(op)) {
  1015. case SLJIT_MOV:
  1016. case SLJIT_MOV_P:
  1017. return emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
  1018. case SLJIT_MOV_U32:
  1019. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1020. return emit_op(compiler, SLJIT_MOV_U32, INT_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
  1021. #else
  1022. return emit_op(compiler, SLJIT_MOV_U32, INT_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u32)srcw : srcw);
  1023. #endif
  1024. case SLJIT_MOV_S32:
  1025. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1026. return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
  1027. #else
  1028. return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s32)srcw : srcw);
  1029. #endif
  1030. case SLJIT_MOV_U8:
  1031. return emit_op(compiler, SLJIT_MOV_U8, BYTE_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u8)srcw : srcw);
  1032. case SLJIT_MOV_S8:
  1033. return emit_op(compiler, SLJIT_MOV_S8, BYTE_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s8)srcw : srcw);
  1034. case SLJIT_MOV_U16:
  1035. return emit_op(compiler, SLJIT_MOV_U16, HALF_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u16)srcw : srcw);
  1036. case SLJIT_MOV_S16:
  1037. return emit_op(compiler, SLJIT_MOV_S16, HALF_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s16)srcw : srcw);
  1038. case SLJIT_NOT:
  1039. return emit_op(compiler, op, flags, dst, dstw, TMP_REG1, 0, src, srcw);
  1040. case SLJIT_NEG:
  1041. return emit_op(compiler, SLJIT_SUB | GET_ALL_FLAGS(op), flags | IMM_OP, dst, dstw, SLJIT_IMM, 0, src, srcw);
  1042. case SLJIT_CLZ:
  1043. return emit_op(compiler, op, flags, dst, dstw, TMP_REG1, 0, src, srcw);
  1044. }
  1045. SLJIT_UNREACHABLE();
  1046. return SLJIT_SUCCESS;
  1047. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1048. # undef flags
  1049. #endif
  1050. }
  1051. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op,
  1052. sljit_s32 dst, sljit_sw dstw,
  1053. sljit_s32 src1, sljit_sw src1w,
  1054. sljit_s32 src2, sljit_sw src2w)
  1055. {
  1056. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1057. # define flags 0
  1058. #else
  1059. sljit_s32 flags = 0;
  1060. #endif
  1061. CHECK_ERROR();
  1062. CHECK(check_sljit_emit_op2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
  1063. ADJUST_LOCAL_OFFSET(dst, dstw);
  1064. ADJUST_LOCAL_OFFSET(src1, src1w);
  1065. ADJUST_LOCAL_OFFSET(src2, src2w);
  1066. if (dst == SLJIT_UNUSED && !HAS_FLAGS(op))
  1067. return SLJIT_SUCCESS;
  1068. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1069. if (op & SLJIT_I32_OP) {
  1070. flags |= INT_DATA | SIGNED_DATA;
  1071. if (src1 & SLJIT_IMM)
  1072. src1w = (sljit_s32)src1w;
  1073. if (src2 & SLJIT_IMM)
  1074. src2w = (sljit_s32)src2w;
  1075. }
  1076. #endif
  1077. switch (GET_OPCODE(op)) {
  1078. case SLJIT_ADD:
  1079. case SLJIT_ADDC:
  1080. return emit_op(compiler, op, flags | CUMULATIVE_OP | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1081. case SLJIT_SUB:
  1082. case SLJIT_SUBC:
  1083. return emit_op(compiler, op, flags | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1084. case SLJIT_MUL:
  1085. return emit_op(compiler, op, flags | CUMULATIVE_OP, dst, dstw, src1, src1w, src2, src2w);
  1086. case SLJIT_AND:
  1087. case SLJIT_OR:
  1088. case SLJIT_XOR:
  1089. return emit_op(compiler, op, flags | CUMULATIVE_OP | LOGICAL_OP | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1090. case SLJIT_SHL:
  1091. case SLJIT_LSHR:
  1092. case SLJIT_ASHR:
  1093. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1094. if (src2 & SLJIT_IMM)
  1095. src2w &= 0x1f;
  1096. #else
  1097. if (src2 & SLJIT_IMM) {
  1098. if (op & SLJIT_I32_OP)
  1099. src2w &= 0x1f;
  1100. else
  1101. src2w &= 0x3f;
  1102. }
  1103. #endif
  1104. return emit_op(compiler, op, flags | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1105. }
  1106. SLJIT_UNREACHABLE();
  1107. return SLJIT_SUCCESS;
  1108. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1109. # undef flags
  1110. #endif
  1111. }
  1112. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_register_index(sljit_s32 reg)
  1113. {
  1114. CHECK_REG_INDEX(check_sljit_get_register_index(reg));
  1115. return reg_map[reg];
  1116. }
  1117. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_float_register_index(sljit_s32 reg)
  1118. {
  1119. CHECK_REG_INDEX(check_sljit_get_float_register_index(reg));
  1120. return FR(reg);
  1121. }
  1122. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_custom(struct sljit_compiler *compiler,
  1123. void *instruction, sljit_s32 size)
  1124. {
  1125. CHECK_ERROR();
  1126. CHECK(check_sljit_emit_op_custom(compiler, instruction, size));
  1127. return push_inst(compiler, *(sljit_ins*)instruction, UNMOVABLE_INS);
  1128. }
  1129. /* --------------------------------------------------------------------- */
  1130. /* Floating point operators */
  1131. /* --------------------------------------------------------------------- */
  1132. #define FLOAT_DATA(op) (DOUBLE_DATA | ((op & SLJIT_F32_OP) >> 7))
  1133. #define FMT(op) (((op & SLJIT_F32_OP) ^ SLJIT_F32_OP) << (21 - 8))
  1134. static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler *compiler, sljit_s32 op,
  1135. sljit_s32 dst, sljit_sw dstw,
  1136. sljit_s32 src, sljit_sw srcw)
  1137. {
  1138. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1139. # define flags 0
  1140. #else
  1141. sljit_s32 flags = (GET_OPCODE(op) == SLJIT_CONV_SW_FROM_F64) << 21;
  1142. #endif
  1143. if (src & SLJIT_MEM) {
  1144. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src, srcw, dst, dstw));
  1145. src = TMP_FREG1;
  1146. }
  1147. FAIL_IF(push_inst(compiler, (TRUNC_W_S ^ (flags >> 19)) | FMT(op) | FS(src) | FD(TMP_FREG1), MOVABLE_INS));
  1148. if (FAST_IS_REG(dst))
  1149. return push_inst(compiler, MFC1 | flags | T(dst) | FS(TMP_FREG1), MOVABLE_INS);
  1150. /* Store the integer value from a VFP register. */
  1151. return emit_op_mem2(compiler, flags ? DOUBLE_DATA : SINGLE_DATA, FR(TMP_FREG1), dst, dstw, 0, 0);
  1152. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1153. # undef is_long
  1154. #endif
  1155. }
  1156. static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler *compiler, sljit_s32 op,
  1157. sljit_s32 dst, sljit_sw dstw,
  1158. sljit_s32 src, sljit_sw srcw)
  1159. {
  1160. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1161. # define flags 0
  1162. #else
  1163. sljit_s32 flags = (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_SW) << 21;
  1164. #endif
  1165. sljit_s32 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
  1166. if (FAST_IS_REG(src))
  1167. FAIL_IF(push_inst(compiler, MTC1 | flags | T(src) | FS(TMP_FREG1), MOVABLE_INS));
  1168. else if (src & SLJIT_MEM) {
  1169. /* Load the integer value into a VFP register. */
  1170. FAIL_IF(emit_op_mem2(compiler, ((flags) ? DOUBLE_DATA : SINGLE_DATA) | LOAD_DATA, FR(TMP_FREG1), src, srcw, dst, dstw));
  1171. }
  1172. else {
  1173. #if (defined SLJIT_CONFIG_X86_64 && SLJIT_CONFIG_X86_64)
  1174. if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32)
  1175. srcw = (sljit_s32)srcw;
  1176. #endif
  1177. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
  1178. FAIL_IF(push_inst(compiler, MTC1 | flags | T(TMP_REG1) | FS(TMP_FREG1), MOVABLE_INS));
  1179. }
  1180. FAIL_IF(push_inst(compiler, CVT_S_S | flags | (4 << 21) | (((op & SLJIT_F32_OP) ^ SLJIT_F32_OP) >> 8) | FS(TMP_FREG1) | FD(dst_r), MOVABLE_INS));
  1181. if (dst & SLJIT_MEM)
  1182. return emit_op_mem2(compiler, FLOAT_DATA(op), FR(TMP_FREG1), dst, dstw, 0, 0);
  1183. return SLJIT_SUCCESS;
  1184. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1185. # undef flags
  1186. #endif
  1187. }
  1188. static SLJIT_INLINE sljit_s32 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op,
  1189. sljit_s32 src1, sljit_sw src1w,
  1190. sljit_s32 src2, sljit_sw src2w)
  1191. {
  1192. sljit_ins inst;
  1193. if (src1 & SLJIT_MEM) {
  1194. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, src2, src2w));
  1195. src1 = TMP_FREG1;
  1196. }
  1197. if (src2 & SLJIT_MEM) {
  1198. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, 0, 0));
  1199. src2 = TMP_FREG2;
  1200. }
  1201. switch (GET_FLAG_TYPE(op)) {
  1202. case SLJIT_EQUAL_F64:
  1203. case SLJIT_NOT_EQUAL_F64:
  1204. inst = C_UEQ_S;
  1205. break;
  1206. case SLJIT_LESS_F64:
  1207. case SLJIT_GREATER_EQUAL_F64:
  1208. inst = C_ULT_S;
  1209. break;
  1210. case SLJIT_GREATER_F64:
  1211. case SLJIT_LESS_EQUAL_F64:
  1212. inst = C_ULE_S;
  1213. break;
  1214. default:
  1215. SLJIT_ASSERT(GET_FLAG_TYPE(op) == SLJIT_UNORDERED_F64 || GET_FLAG_TYPE(op) == SLJIT_ORDERED_F64);
  1216. inst = C_UN_S;
  1217. break;
  1218. }
  1219. return push_inst(compiler, inst | FMT(op) | FT(src2) | FS(src1), UNMOVABLE_INS);
  1220. }
  1221. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop1(struct sljit_compiler *compiler, sljit_s32 op,
  1222. sljit_s32 dst, sljit_sw dstw,
  1223. sljit_s32 src, sljit_sw srcw)
  1224. {
  1225. sljit_s32 dst_r;
  1226. CHECK_ERROR();
  1227. compiler->cache_arg = 0;
  1228. compiler->cache_argw = 0;
  1229. SLJIT_COMPILE_ASSERT((SLJIT_F32_OP == 0x100) && !(DOUBLE_DATA & 0x2), float_transfer_bit_error);
  1230. SELECT_FOP1_OPERATION_WITH_CHECKS(compiler, op, dst, dstw, src, srcw);
  1231. if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_F32)
  1232. op ^= SLJIT_F32_OP;
  1233. dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
  1234. if (src & SLJIT_MEM) {
  1235. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(dst_r), src, srcw, dst, dstw));
  1236. src = dst_r;
  1237. }
  1238. switch (GET_OPCODE(op)) {
  1239. case SLJIT_MOV_F64:
  1240. if (src != dst_r) {
  1241. if (dst_r != TMP_FREG1)
  1242. FAIL_IF(push_inst(compiler, MOV_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
  1243. else
  1244. dst_r = src;
  1245. }
  1246. break;
  1247. case SLJIT_NEG_F64:
  1248. FAIL_IF(push_inst(compiler, NEG_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
  1249. break;
  1250. case SLJIT_ABS_F64:
  1251. FAIL_IF(push_inst(compiler, ABS_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
  1252. break;
  1253. case SLJIT_CONV_F64_FROM_F32:
  1254. FAIL_IF(push_inst(compiler, CVT_S_S | ((op & SLJIT_F32_OP) ? 1 : (1 << 21)) | FS(src) | FD(dst_r), MOVABLE_INS));
  1255. op ^= SLJIT_F32_OP;
  1256. break;
  1257. }
  1258. if (dst & SLJIT_MEM)
  1259. return emit_op_mem2(compiler, FLOAT_DATA(op), FR(dst_r), dst, dstw, 0, 0);
  1260. return SLJIT_SUCCESS;
  1261. }
  1262. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op,
  1263. sljit_s32 dst, sljit_sw dstw,
  1264. sljit_s32 src1, sljit_sw src1w,
  1265. sljit_s32 src2, sljit_sw src2w)
  1266. {
  1267. sljit_s32 dst_r, flags = 0;
  1268. CHECK_ERROR();
  1269. CHECK(check_sljit_emit_fop2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
  1270. ADJUST_LOCAL_OFFSET(dst, dstw);
  1271. ADJUST_LOCAL_OFFSET(src1, src1w);
  1272. ADJUST_LOCAL_OFFSET(src2, src2w);
  1273. compiler->cache_arg = 0;
  1274. compiler->cache_argw = 0;
  1275. dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG2;
  1276. if (src1 & SLJIT_MEM) {
  1277. if (getput_arg_fast(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w)) {
  1278. FAIL_IF(compiler->error);
  1279. src1 = TMP_FREG1;
  1280. } else
  1281. flags |= SLOW_SRC1;
  1282. }
  1283. if (src2 & SLJIT_MEM) {
  1284. if (getput_arg_fast(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w)) {
  1285. FAIL_IF(compiler->error);
  1286. src2 = TMP_FREG2;
  1287. } else
  1288. flags |= SLOW_SRC2;
  1289. }
  1290. if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
  1291. if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
  1292. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, src1, src1w));
  1293. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, dst, dstw));
  1294. }
  1295. else {
  1296. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, src2, src2w));
  1297. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, dst, dstw));
  1298. }
  1299. }
  1300. else if (flags & SLOW_SRC1)
  1301. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, dst, dstw));
  1302. else if (flags & SLOW_SRC2)
  1303. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, dst, dstw));
  1304. if (flags & SLOW_SRC1)
  1305. src1 = TMP_FREG1;
  1306. if (flags & SLOW_SRC2)
  1307. src2 = TMP_FREG2;
  1308. switch (GET_OPCODE(op)) {
  1309. case SLJIT_ADD_F64:
  1310. FAIL_IF(push_inst(compiler, ADD_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1311. break;
  1312. case SLJIT_SUB_F64:
  1313. FAIL_IF(push_inst(compiler, SUB_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1314. break;
  1315. case SLJIT_MUL_F64:
  1316. FAIL_IF(push_inst(compiler, MUL_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1317. break;
  1318. case SLJIT_DIV_F64:
  1319. FAIL_IF(push_inst(compiler, DIV_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1320. break;
  1321. }
  1322. if (dst_r == TMP_FREG2)
  1323. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op), FR(TMP_FREG2), dst, dstw, 0, 0));
  1324. return SLJIT_SUCCESS;
  1325. }
  1326. /* --------------------------------------------------------------------- */
  1327. /* Other instructions */
  1328. /* --------------------------------------------------------------------- */
  1329. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_enter(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw)
  1330. {
  1331. CHECK_ERROR();
  1332. CHECK(check_sljit_emit_fast_enter(compiler, dst, dstw));
  1333. ADJUST_LOCAL_OFFSET(dst, dstw);
  1334. if (FAST_IS_REG(dst))
  1335. return push_inst(compiler, ADDU_W | SA(RETURN_ADDR_REG) | TA(0) | D(dst), DR(dst));
  1336. /* Memory. */
  1337. return emit_op_mem(compiler, WORD_DATA, RETURN_ADDR_REG, dst, dstw);
  1338. }
  1339. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_return(struct sljit_compiler *compiler, sljit_s32 src, sljit_sw srcw)
  1340. {
  1341. CHECK_ERROR();
  1342. CHECK(check_sljit_emit_fast_return(compiler, src, srcw));
  1343. ADJUST_LOCAL_OFFSET(src, srcw);
  1344. if (FAST_IS_REG(src))
  1345. FAIL_IF(push_inst(compiler, ADDU_W | S(src) | TA(0) | DA(RETURN_ADDR_REG), RETURN_ADDR_REG));
  1346. else
  1347. FAIL_IF(emit_op_mem(compiler, WORD_DATA | LOAD_DATA, RETURN_ADDR_REG, src, srcw));
  1348. FAIL_IF(push_inst(compiler, JR | SA(RETURN_ADDR_REG), UNMOVABLE_INS));
  1349. return push_inst(compiler, NOP, UNMOVABLE_INS);
  1350. }
  1351. /* --------------------------------------------------------------------- */
  1352. /* Conditional instructions */
  1353. /* --------------------------------------------------------------------- */
  1354. SLJIT_API_FUNC_ATTRIBUTE struct sljit_label* sljit_emit_label(struct sljit_compiler *compiler)
  1355. {
  1356. struct sljit_label *label;
  1357. CHECK_ERROR_PTR();
  1358. CHECK_PTR(check_sljit_emit_label(compiler));
  1359. if (compiler->last_label && compiler->last_label->size == compiler->size)
  1360. return compiler->last_label;
  1361. label = (struct sljit_label*)ensure_abuf(compiler, sizeof(struct sljit_label));
  1362. PTR_FAIL_IF(!label);
  1363. set_label(label, compiler);
  1364. compiler->delay_slot = UNMOVABLE_INS;
  1365. return label;
  1366. }
  1367. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1368. #define JUMP_LENGTH 4
  1369. #else
  1370. #define JUMP_LENGTH 8
  1371. #endif
  1372. #define BR_Z(src) \
  1373. inst = BEQ | SA(src) | TA(0) | JUMP_LENGTH; \
  1374. flags = IS_BIT26_COND; \
  1375. delay_check = src;
  1376. #define BR_NZ(src) \
  1377. inst = BNE | SA(src) | TA(0) | JUMP_LENGTH; \
  1378. flags = IS_BIT26_COND; \
  1379. delay_check = src;
  1380. #define BR_T() \
  1381. inst = BC1T | JUMP_LENGTH; \
  1382. flags = IS_BIT16_COND; \
  1383. delay_check = FCSR_FCC;
  1384. #define BR_F() \
  1385. inst = BC1F | JUMP_LENGTH; \
  1386. flags = IS_BIT16_COND; \
  1387. delay_check = FCSR_FCC;
  1388. SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_jump(struct sljit_compiler *compiler, sljit_s32 type)
  1389. {
  1390. struct sljit_jump *jump;
  1391. sljit_ins inst;
  1392. sljit_s32 flags = 0;
  1393. sljit_s32 delay_check = UNMOVABLE_INS;
  1394. CHECK_ERROR_PTR();
  1395. CHECK_PTR(check_sljit_emit_jump(compiler, type));
  1396. jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
  1397. PTR_FAIL_IF(!jump);
  1398. set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
  1399. type &= 0xff;
  1400. switch (type) {
  1401. case SLJIT_EQUAL:
  1402. BR_NZ(EQUAL_FLAG);
  1403. break;
  1404. case SLJIT_NOT_EQUAL:
  1405. BR_Z(EQUAL_FLAG);
  1406. break;
  1407. case SLJIT_LESS:
  1408. case SLJIT_GREATER:
  1409. case SLJIT_SIG_LESS:
  1410. case SLJIT_SIG_GREATER:
  1411. case SLJIT_OVERFLOW:
  1412. case SLJIT_MUL_OVERFLOW:
  1413. BR_Z(OTHER_FLAG);
  1414. break;
  1415. case SLJIT_GREATER_EQUAL:
  1416. case SLJIT_LESS_EQUAL:
  1417. case SLJIT_SIG_GREATER_EQUAL:
  1418. case SLJIT_SIG_LESS_EQUAL:
  1419. case SLJIT_NOT_OVERFLOW:
  1420. case SLJIT_MUL_NOT_OVERFLOW:
  1421. BR_NZ(OTHER_FLAG);
  1422. break;
  1423. case SLJIT_NOT_EQUAL_F64:
  1424. case SLJIT_GREATER_EQUAL_F64:
  1425. case SLJIT_GREATER_F64:
  1426. case SLJIT_ORDERED_F64:
  1427. BR_T();
  1428. break;
  1429. case SLJIT_EQUAL_F64:
  1430. case SLJIT_LESS_F64:
  1431. case SLJIT_LESS_EQUAL_F64:
  1432. case SLJIT_UNORDERED_F64:
  1433. BR_F();
  1434. break;
  1435. default:
  1436. /* Not conditional branch. */
  1437. inst = 0;
  1438. break;
  1439. }
  1440. jump->flags |= flags;
  1441. if (compiler->delay_slot == MOVABLE_INS || (compiler->delay_slot != UNMOVABLE_INS && compiler->delay_slot != delay_check))
  1442. jump->flags |= IS_MOVABLE;
  1443. if (inst)
  1444. PTR_FAIL_IF(push_inst(compiler, inst, UNMOVABLE_INS));
  1445. PTR_FAIL_IF(emit_const(compiler, TMP_REG2, 0));
  1446. if (type <= SLJIT_JUMP)
  1447. PTR_FAIL_IF(push_inst(compiler, JR | S(TMP_REG2), UNMOVABLE_INS));
  1448. else {
  1449. jump->flags |= IS_JAL;
  1450. PTR_FAIL_IF(push_inst(compiler, JALR | S(TMP_REG2) | DA(RETURN_ADDR_REG), UNMOVABLE_INS));
  1451. }
  1452. jump->addr = compiler->size;
  1453. PTR_FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1454. return jump;
  1455. }
  1456. #define RESOLVE_IMM1() \
  1457. if (src1 & SLJIT_IMM) { \
  1458. if (src1w) { \
  1459. PTR_FAIL_IF(load_immediate(compiler, DR(TMP_REG1), src1w)); \
  1460. src1 = TMP_REG1; \
  1461. } \
  1462. else \
  1463. src1 = 0; \
  1464. }
  1465. #define RESOLVE_IMM2() \
  1466. if (src2 & SLJIT_IMM) { \
  1467. if (src2w) { \
  1468. PTR_FAIL_IF(load_immediate(compiler, DR(TMP_REG2), src2w)); \
  1469. src2 = TMP_REG2; \
  1470. } \
  1471. else \
  1472. src2 = 0; \
  1473. }
  1474. SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_cmp(struct sljit_compiler *compiler, sljit_s32 type,
  1475. sljit_s32 src1, sljit_sw src1w,
  1476. sljit_s32 src2, sljit_sw src2w)
  1477. {
  1478. struct sljit_jump *jump;
  1479. sljit_s32 flags;
  1480. sljit_ins inst;
  1481. CHECK_ERROR_PTR();
  1482. CHECK_PTR(check_sljit_emit_cmp(compiler, type, src1, src1w, src2, src2w));
  1483. ADJUST_LOCAL_OFFSET(src1, src1w);
  1484. ADJUST_LOCAL_OFFSET(src2, src2w);
  1485. compiler->cache_arg = 0;
  1486. compiler->cache_argw = 0;
  1487. flags = ((type & SLJIT_I32_OP) ? INT_DATA : WORD_DATA) | LOAD_DATA;
  1488. if (src1 & SLJIT_MEM) {
  1489. PTR_FAIL_IF(emit_op_mem2(compiler, flags, DR(TMP_REG1), src1, src1w, src2, src2w));
  1490. src1 = TMP_REG1;
  1491. }
  1492. if (src2 & SLJIT_MEM) {
  1493. PTR_FAIL_IF(emit_op_mem2(compiler, flags, DR(TMP_REG2), src2, src2w, 0, 0));
  1494. src2 = TMP_REG2;
  1495. }
  1496. jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
  1497. PTR_FAIL_IF(!jump);
  1498. set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
  1499. type &= 0xff;
  1500. if (type <= SLJIT_NOT_EQUAL) {
  1501. RESOLVE_IMM1();
  1502. RESOLVE_IMM2();
  1503. jump->flags |= IS_BIT26_COND;
  1504. if (compiler->delay_slot == MOVABLE_INS || (compiler->delay_slot != UNMOVABLE_INS && compiler->delay_slot != DR(src1) && compiler->delay_slot != DR(src2)))
  1505. jump->flags |= IS_MOVABLE;
  1506. PTR_FAIL_IF(push_inst(compiler, (type == SLJIT_EQUAL ? BNE : BEQ) | S(src1) | T(src2) | JUMP_LENGTH, UNMOVABLE_INS));
  1507. }
  1508. else if (type >= SLJIT_SIG_LESS && (((src1 & SLJIT_IMM) && (src1w == 0)) || ((src2 & SLJIT_IMM) && (src2w == 0)))) {
  1509. inst = NOP;
  1510. if ((src1 & SLJIT_IMM) && (src1w == 0)) {
  1511. RESOLVE_IMM2();
  1512. switch (type) {
  1513. case SLJIT_SIG_LESS:
  1514. inst = BLEZ;
  1515. jump->flags |= IS_BIT26_COND;
  1516. break;
  1517. case SLJIT_SIG_GREATER_EQUAL:
  1518. inst = BGTZ;
  1519. jump->flags |= IS_BIT26_COND;
  1520. break;
  1521. case SLJIT_SIG_GREATER:
  1522. inst = BGEZ;
  1523. jump->flags |= IS_BIT16_COND;
  1524. break;
  1525. case SLJIT_SIG_LESS_EQUAL:
  1526. inst = BLTZ;
  1527. jump->flags |= IS_BIT16_COND;
  1528. break;
  1529. }
  1530. src1 = src2;
  1531. }
  1532. else {
  1533. RESOLVE_IMM1();
  1534. switch (type) {
  1535. case SLJIT_SIG_LESS:
  1536. inst = BGEZ;
  1537. jump->flags |= IS_BIT16_COND;
  1538. break;
  1539. case SLJIT_SIG_GREATER_EQUAL:
  1540. inst = BLTZ;
  1541. jump->flags |= IS_BIT16_COND;
  1542. break;
  1543. case SLJIT_SIG_GREATER:
  1544. inst = BLEZ;
  1545. jump->flags |= IS_BIT26_COND;
  1546. break;
  1547. case SLJIT_SIG_LESS_EQUAL:
  1548. inst = BGTZ;
  1549. jump->flags |= IS_BIT26_COND;
  1550. break;
  1551. }
  1552. }
  1553. PTR_FAIL_IF(push_inst(compiler, inst | S(src1) | JUMP_LENGTH, UNMOVABLE_INS));
  1554. }
  1555. else {
  1556. if (type == SLJIT_LESS || type == SLJIT_GREATER_EQUAL || type == SLJIT_SIG_LESS || type == SLJIT_SIG_GREATER_EQUAL) {
  1557. RESOLVE_IMM1();
  1558. if ((src2 & SLJIT_IMM) && src2w <= SIMM_MAX && src2w >= SIMM_MIN)
  1559. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTIU : SLTI) | S(src1) | T(TMP_REG1) | IMM(src2w), DR(TMP_REG1)));
  1560. else {
  1561. RESOLVE_IMM2();
  1562. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTU : SLT) | S(src1) | T(src2) | D(TMP_REG1), DR(TMP_REG1)));
  1563. }
  1564. type = (type == SLJIT_LESS || type == SLJIT_SIG_LESS) ? SLJIT_NOT_EQUAL : SLJIT_EQUAL;
  1565. }
  1566. else {
  1567. RESOLVE_IMM2();
  1568. if ((src1 & SLJIT_IMM) && src1w <= SIMM_MAX && src1w >= SIMM_MIN)
  1569. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTIU : SLTI) | S(src2) | T(TMP_REG1) | IMM(src1w), DR(TMP_REG1)));
  1570. else {
  1571. RESOLVE_IMM1();
  1572. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTU : SLT) | S(src2) | T(src1) | D(TMP_REG1), DR(TMP_REG1)));
  1573. }
  1574. type = (type == SLJIT_GREATER || type == SLJIT_SIG_GREATER) ? SLJIT_NOT_EQUAL : SLJIT_EQUAL;
  1575. }
  1576. jump->flags |= IS_BIT26_COND;
  1577. PTR_FAIL_IF(push_inst(compiler, (type == SLJIT_EQUAL ? BNE : BEQ) | S(TMP_REG1) | TA(0) | JUMP_LENGTH, UNMOVABLE_INS));
  1578. }
  1579. PTR_FAIL_IF(emit_const(compiler, TMP_REG2, 0));
  1580. PTR_FAIL_IF(push_inst(compiler, JR | S(TMP_REG2), UNMOVABLE_INS));
  1581. jump->addr = compiler->size;
  1582. PTR_FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1583. return jump;
  1584. }
  1585. #undef RESOLVE_IMM1
  1586. #undef RESOLVE_IMM2
  1587. #undef JUMP_LENGTH
  1588. #undef BR_Z
  1589. #undef BR_NZ
  1590. #undef BR_T
  1591. #undef BR_F
  1592. #undef FLOAT_DATA
  1593. #undef FMT
  1594. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_ijump(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src, sljit_sw srcw)
  1595. {
  1596. struct sljit_jump *jump = NULL;
  1597. CHECK_ERROR();
  1598. CHECK(check_sljit_emit_ijump(compiler, type, src, srcw));
  1599. ADJUST_LOCAL_OFFSET(src, srcw);
  1600. if (src & SLJIT_IMM) {
  1601. jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
  1602. FAIL_IF(!jump);
  1603. set_jump(jump, compiler, JUMP_ADDR | ((type >= SLJIT_FAST_CALL) ? IS_JAL : 0));
  1604. jump->u.target = srcw;
  1605. if (compiler->delay_slot != UNMOVABLE_INS)
  1606. jump->flags |= IS_MOVABLE;
  1607. FAIL_IF(emit_const(compiler, TMP_REG2, 0));
  1608. src = TMP_REG2;
  1609. }
  1610. else if (src & SLJIT_MEM) {
  1611. FAIL_IF(emit_op_mem(compiler, WORD_DATA | LOAD_DATA, DR(TMP_REG2), src, srcw));
  1612. src = TMP_REG2;
  1613. }
  1614. FAIL_IF(push_inst(compiler, JR | S(src), UNMOVABLE_INS));
  1615. if (jump)
  1616. jump->addr = compiler->size;
  1617. FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1618. return SLJIT_SUCCESS;
  1619. }
  1620. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_flags(struct sljit_compiler *compiler, sljit_s32 op,
  1621. sljit_s32 dst, sljit_sw dstw,
  1622. sljit_s32 type)
  1623. {
  1624. sljit_s32 src_ar, dst_ar;
  1625. sljit_s32 saved_op = op;
  1626. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1627. sljit_s32 mem_type = WORD_DATA;
  1628. #else
  1629. sljit_s32 mem_type = (op & SLJIT_I32_OP) ? (INT_DATA | SIGNED_DATA) : WORD_DATA;
  1630. #endif
  1631. CHECK_ERROR();
  1632. CHECK(check_sljit_emit_op_flags(compiler, op, dst, dstw, type));
  1633. ADJUST_LOCAL_OFFSET(dst, dstw);
  1634. op = GET_OPCODE(op);
  1635. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1636. if (op == SLJIT_MOV_S32)
  1637. mem_type = INT_DATA | SIGNED_DATA;
  1638. #endif
  1639. dst_ar = DR((op < SLJIT_ADD && FAST_IS_REG(dst)) ? dst : TMP_REG2);
  1640. compiler->cache_arg = 0;
  1641. compiler->cache_argw = 0;
  1642. if (op >= SLJIT_ADD && (dst & SLJIT_MEM))
  1643. FAIL_IF(emit_op_mem2(compiler, mem_type | LOAD_DATA, DR(TMP_REG1), dst, dstw, dst, dstw));
  1644. switch (type & 0xff) {
  1645. case SLJIT_EQUAL:
  1646. case SLJIT_NOT_EQUAL:
  1647. FAIL_IF(push_inst(compiler, SLTIU | SA(EQUAL_FLAG) | TA(dst_ar) | IMM(1), dst_ar));
  1648. src_ar = dst_ar;
  1649. break;
  1650. case SLJIT_MUL_OVERFLOW:
  1651. case SLJIT_MUL_NOT_OVERFLOW:
  1652. FAIL_IF(push_inst(compiler, SLTIU | SA(OTHER_FLAG) | TA(dst_ar) | IMM(1), dst_ar));
  1653. src_ar = dst_ar;
  1654. type ^= 0x1; /* Flip type bit for the XORI below. */
  1655. break;
  1656. case SLJIT_GREATER_F64:
  1657. case SLJIT_LESS_EQUAL_F64:
  1658. type ^= 0x1; /* Flip type bit for the XORI below. */
  1659. case SLJIT_EQUAL_F64:
  1660. case SLJIT_NOT_EQUAL_F64:
  1661. case SLJIT_LESS_F64:
  1662. case SLJIT_GREATER_EQUAL_F64:
  1663. case SLJIT_UNORDERED_F64:
  1664. case SLJIT_ORDERED_F64:
  1665. FAIL_IF(push_inst(compiler, CFC1 | TA(dst_ar) | DA(FCSR_REG), dst_ar));
  1666. FAIL_IF(push_inst(compiler, SRL | TA(dst_ar) | DA(dst_ar) | SH_IMM(23), dst_ar));
  1667. FAIL_IF(push_inst(compiler, ANDI | SA(dst_ar) | TA(dst_ar) | IMM(1), dst_ar));
  1668. src_ar = dst_ar;
  1669. break;
  1670. default:
  1671. src_ar = OTHER_FLAG;
  1672. break;
  1673. }
  1674. if (type & 0x1) {
  1675. FAIL_IF(push_inst(compiler, XORI | SA(src_ar) | TA(dst_ar) | IMM(1), dst_ar));
  1676. src_ar = dst_ar;
  1677. }
  1678. if (op < SLJIT_ADD) {
  1679. if (dst & SLJIT_MEM)
  1680. return emit_op_mem(compiler, mem_type, src_ar, dst, dstw);
  1681. if (src_ar != dst_ar)
  1682. return push_inst(compiler, ADDU_W | SA(src_ar) | TA(0) | DA(dst_ar), dst_ar);
  1683. return SLJIT_SUCCESS;
  1684. }
  1685. /* OTHER_FLAG cannot be specified as src2 argument at the moment. */
  1686. if (DR(TMP_REG2) != src_ar)
  1687. FAIL_IF(push_inst(compiler, ADDU_W | SA(src_ar) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
  1688. mem_type |= CUMULATIVE_OP | LOGICAL_OP | IMM_OP | ALT_KEEP_CACHE;
  1689. if (dst & SLJIT_MEM)
  1690. return emit_op(compiler, saved_op, mem_type, dst, dstw, TMP_REG1, 0, TMP_REG2, 0);
  1691. return emit_op(compiler, saved_op, mem_type, dst, dstw, dst, dstw, TMP_REG2, 0);
  1692. }
  1693. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_cmov(struct sljit_compiler *compiler, sljit_s32 type,
  1694. sljit_s32 dst_reg,
  1695. sljit_s32 src, sljit_sw srcw)
  1696. {
  1697. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  1698. sljit_ins ins;
  1699. #endif
  1700. CHECK_ERROR();
  1701. CHECK(check_sljit_emit_cmov(compiler, type, dst_reg, src, srcw));
  1702. #if (defined SLJIT_MIPS_R1 && SLJIT_MIPS_R1)
  1703. if (SLJIT_UNLIKELY(src & SLJIT_IMM)) {
  1704. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1705. if (dst_reg & SLJIT_I32_OP)
  1706. srcw = (sljit_s32)srcw;
  1707. #endif
  1708. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
  1709. src = TMP_REG1;
  1710. srcw = 0;
  1711. }
  1712. dst_reg &= ~SLJIT_I32_OP;
  1713. switch (type & 0xff) {
  1714. case SLJIT_EQUAL:
  1715. ins = MOVZ | TA(EQUAL_FLAG);
  1716. break;
  1717. case SLJIT_NOT_EQUAL:
  1718. ins = MOVN | TA(EQUAL_FLAG);
  1719. break;
  1720. case SLJIT_LESS:
  1721. case SLJIT_GREATER:
  1722. case SLJIT_SIG_LESS:
  1723. case SLJIT_SIG_GREATER:
  1724. case SLJIT_OVERFLOW:
  1725. case SLJIT_MUL_OVERFLOW:
  1726. ins = MOVN | TA(OTHER_FLAG);
  1727. break;
  1728. case SLJIT_GREATER_EQUAL:
  1729. case SLJIT_LESS_EQUAL:
  1730. case SLJIT_SIG_GREATER_EQUAL:
  1731. case SLJIT_SIG_LESS_EQUAL:
  1732. case SLJIT_NOT_OVERFLOW:
  1733. case SLJIT_MUL_NOT_OVERFLOW:
  1734. ins = MOVZ | TA(OTHER_FLAG);
  1735. break;
  1736. case SLJIT_EQUAL_F64:
  1737. case SLJIT_LESS_F64:
  1738. case SLJIT_LESS_EQUAL_F64:
  1739. case SLJIT_UNORDERED_F64:
  1740. ins = MOVT;
  1741. break;
  1742. case SLJIT_NOT_EQUAL_F64:
  1743. case SLJIT_GREATER_EQUAL_F64:
  1744. case SLJIT_GREATER_F64:
  1745. case SLJIT_ORDERED_F64:
  1746. ins = MOVF;
  1747. break;
  1748. default:
  1749. ins = MOVZ | TA(OTHER_FLAG);
  1750. SLJIT_UNREACHABLE();
  1751. break;
  1752. }
  1753. return push_inst(compiler, ins | S(src) | D(dst_reg), DR(dst_reg));
  1754. #else
  1755. return sljit_emit_cmov_generic(compiler, type, dst_reg, src, srcw);
  1756. #endif
  1757. }
  1758. SLJIT_API_FUNC_ATTRIBUTE struct sljit_const* sljit_emit_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw init_value)
  1759. {
  1760. struct sljit_const *const_;
  1761. sljit_s32 reg;
  1762. CHECK_ERROR_PTR();
  1763. CHECK_PTR(check_sljit_emit_const(compiler, dst, dstw, init_value));
  1764. ADJUST_LOCAL_OFFSET(dst, dstw);
  1765. const_ = (struct sljit_const*)ensure_abuf(compiler, sizeof(struct sljit_const));
  1766. PTR_FAIL_IF(!const_);
  1767. set_const(const_, compiler);
  1768. reg = FAST_IS_REG(dst) ? dst : TMP_REG2;
  1769. PTR_FAIL_IF(emit_const(compiler, reg, init_value));
  1770. if (dst & SLJIT_MEM)
  1771. PTR_FAIL_IF(emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, TMP_REG2, 0));
  1772. return const_;
  1773. }