sljitNativeMIPS_common.c 72 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835183618371838183918401841184218431844184518461847184818491850185118521853185418551856185718581859186018611862186318641865186618671868186918701871187218731874187518761877187818791880188118821883188418851886188718881889189018911892189318941895189618971898189919001901190219031904190519061907190819091910191119121913191419151916191719181919192019211922192319241925192619271928192919301931193219331934193519361937193819391940194119421943194419451946194719481949195019511952195319541955195619571958195919601961196219631964196519661967196819691970197119721973197419751976197719781979198019811982198319841985198619871988198919901991199219931994199519961997199819992000200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024202520262027202820292030203120322033203420352036203720382039204020412042204320442045204620472048204920502051205220532054205520562057205820592060206120622063206420652066206720682069207020712072207320742075207620772078207920802081208220832084208520862087208820892090209120922093209420952096209720982099210021012102210321042105210621072108210921102111211221132114211521162117211821192120212121222123212421252126212721282129213021312132213321342135213621372138213921402141214221432144214521462147214821492150215121522153215421552156215721582159216021612162216321642165216621672168216921702171217221732174217521762177217821792180218121822183218421852186218721882189219021912192219321942195219621972198219922002201220222032204220522062207220822092210221122122213221422152216221722182219222022212222222322242225222622272228222922302231223222332234223522362237223822392240224122422243224422452246224722482249225022512252225322542255225622572258225922602261226222632264226522662267226822692270227122722273227422752276227722782279228022812282228322842285228622872288228922902291229222932294229522962297229822992300230123022303230423052306230723082309
  1. /*
  2. * Stack-less Just-In-Time compiler
  3. *
  4. * Copyright Zoltan Herczeg (hzmester@freemail.hu). All rights reserved.
  5. *
  6. * Redistribution and use in source and binary forms, with or without modification, are
  7. * permitted provided that the following conditions are met:
  8. *
  9. * 1. Redistributions of source code must retain the above copyright notice, this list of
  10. * conditions and the following disclaimer.
  11. *
  12. * 2. Redistributions in binary form must reproduce the above copyright notice, this list
  13. * of conditions and the following disclaimer in the documentation and/or other materials
  14. * provided with the distribution.
  15. *
  16. * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY
  17. * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
  18. * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT
  19. * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
  20. * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
  21. * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
  22. * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
  23. * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
  24. * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
  25. */
  26. /* Latest MIPS architecture. */
  27. #ifndef __mips_hard_float
  28. /* Disable automatic detection, covers both -msoft-float and -mno-float */
  29. #undef SLJIT_IS_FPU_AVAILABLE
  30. #define SLJIT_IS_FPU_AVAILABLE 0
  31. #endif
  32. SLJIT_API_FUNC_ATTRIBUTE const char* sljit_get_platform_name(void)
  33. {
  34. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  35. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  36. return "MIPS32-R6" SLJIT_CPUINFO;
  37. #else /* !SLJIT_CONFIG_MIPS_32 */
  38. return "MIPS64-R6" SLJIT_CPUINFO;
  39. #endif /* SLJIT_CONFIG_MIPS_32 */
  40. #elif (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1)
  41. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  42. return "MIPS32-R1" SLJIT_CPUINFO;
  43. #else /* !SLJIT_CONFIG_MIPS_32 */
  44. return "MIPS64-R1" SLJIT_CPUINFO;
  45. #endif /* SLJIT_CONFIG_MIPS_32 */
  46. #else /* SLJIT_MIPS_REV < 1 */
  47. return "MIPS III" SLJIT_CPUINFO;
  48. #endif /* SLJIT_MIPS_REV >= 6 */
  49. }
  50. /* Length of an instruction word
  51. Both for mips-32 and mips-64 */
  52. typedef sljit_u32 sljit_ins;
  53. #define TMP_REG1 (SLJIT_NUMBER_OF_REGISTERS + 2)
  54. #define TMP_REG2 (SLJIT_NUMBER_OF_REGISTERS + 3)
  55. #define TMP_REG3 (SLJIT_NUMBER_OF_REGISTERS + 4)
  56. /* For position independent code, t9 must contain the function address. */
  57. #define PIC_ADDR_REG TMP_REG2
  58. /* Floating point status register. */
  59. #define FCSR_REG 31
  60. /* Return address register. */
  61. #define RETURN_ADDR_REG 31
  62. /* Flags are kept in volatile registers. */
  63. #define EQUAL_FLAG 3
  64. #define OTHER_FLAG 1
  65. #define TMP_FREG1 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 1)
  66. #define TMP_FREG2 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 2)
  67. #define TMP_FREG3 (SLJIT_NUMBER_OF_FLOAT_REGISTERS + 3)
  68. static const sljit_u8 reg_map[SLJIT_NUMBER_OF_REGISTERS + 5] = {
  69. 0, 2, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 24, 23, 22, 21, 20, 19, 18, 17, 16, 29, 4, 25, 31
  70. };
  71. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  72. static const sljit_u8 freg_map[SLJIT_NUMBER_OF_FLOAT_REGISTERS + 4] = {
  73. 0, 0, 14, 2, 4, 6, 8, 12, 10, 16
  74. };
  75. #else
  76. static const sljit_u8 freg_map[SLJIT_NUMBER_OF_FLOAT_REGISTERS + 4] = {
  77. 0, 0, 13, 14, 15, 16, 17, 12, 18, 10
  78. };
  79. #endif
  80. /* --------------------------------------------------------------------- */
  81. /* Instrucion forms */
  82. /* --------------------------------------------------------------------- */
  83. #define S(s) (reg_map[s] << 21)
  84. #define T(t) (reg_map[t] << 16)
  85. #define D(d) (reg_map[d] << 11)
  86. #define FT(t) (freg_map[t] << 16)
  87. #define FS(s) (freg_map[s] << 11)
  88. #define FD(d) (freg_map[d] << 6)
  89. /* Absolute registers. */
  90. #define SA(s) ((s) << 21)
  91. #define TA(t) ((t) << 16)
  92. #define DA(d) ((d) << 11)
  93. #define IMM(imm) ((imm) & 0xffff)
  94. #define SH_IMM(imm) ((imm) << 6)
  95. #define DR(dr) (reg_map[dr])
  96. #define FR(dr) (freg_map[dr])
  97. #define HI(opcode) ((opcode) << 26)
  98. #define LO(opcode) (opcode)
  99. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  100. /* CMP.cond.fmt */
  101. /* S = (20 << 21) D = (21 << 21) */
  102. #define CMP_FMT_S (20 << 21)
  103. #endif /* SLJIT_MIPS_REV >= 6 */
  104. /* S = (16 << 21) D = (17 << 21) */
  105. #define FMT_S (16 << 21)
  106. #define FMT_D (17 << 21)
  107. #define ABS_S (HI(17) | FMT_S | LO(5))
  108. #define ADD_S (HI(17) | FMT_S | LO(0))
  109. #define ADDIU (HI(9))
  110. #define ADDU (HI(0) | LO(33))
  111. #define AND (HI(0) | LO(36))
  112. #define ANDI (HI(12))
  113. #define B (HI(4))
  114. #define BAL (HI(1) | (17 << 16))
  115. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  116. #define BC1EQZ (HI(17) | (9 << 21) | FT(TMP_FREG3))
  117. #define BC1NEZ (HI(17) | (13 << 21) | FT(TMP_FREG3))
  118. #else /* SLJIT_MIPS_REV < 6 */
  119. #define BC1F (HI(17) | (8 << 21))
  120. #define BC1T (HI(17) | (8 << 21) | (1 << 16))
  121. #endif /* SLJIT_MIPS_REV >= 6 */
  122. #define BEQ (HI(4))
  123. #define BGEZ (HI(1) | (1 << 16))
  124. #define BGTZ (HI(7))
  125. #define BLEZ (HI(6))
  126. #define BLTZ (HI(1) | (0 << 16))
  127. #define BNE (HI(5))
  128. #define BREAK (HI(0) | LO(13))
  129. #define CFC1 (HI(17) | (2 << 21))
  130. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  131. #define C_UEQ_S (HI(17) | CMP_FMT_S | LO(3))
  132. #define C_ULE_S (HI(17) | CMP_FMT_S | LO(7))
  133. #define C_ULT_S (HI(17) | CMP_FMT_S | LO(5))
  134. #define C_UN_S (HI(17) | CMP_FMT_S | LO(1))
  135. #define C_FD (FD(TMP_FREG3))
  136. #else /* SLJIT_MIPS_REV < 6 */
  137. #define C_UEQ_S (HI(17) | FMT_S | LO(51))
  138. #define C_ULE_S (HI(17) | FMT_S | LO(55))
  139. #define C_ULT_S (HI(17) | FMT_S | LO(53))
  140. #define C_UN_S (HI(17) | FMT_S | LO(49))
  141. #define C_FD (0)
  142. #endif /* SLJIT_MIPS_REV >= 6 */
  143. #define CVT_S_S (HI(17) | FMT_S | LO(32))
  144. #define DADDIU (HI(25))
  145. #define DADDU (HI(0) | LO(45))
  146. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  147. #define DDIV (HI(0) | (2 << 6) | LO(30))
  148. #define DDIVU (HI(0) | (2 << 6) | LO(31))
  149. #define DMOD (HI(0) | (3 << 6) | LO(30))
  150. #define DMODU (HI(0) | (3 << 6) | LO(31))
  151. #define DIV (HI(0) | (2 << 6) | LO(26))
  152. #define DIVU (HI(0) | (2 << 6) | LO(27))
  153. #define DMUH (HI(0) | (3 << 6) | LO(28))
  154. #define DMUHU (HI(0) | (3 << 6) | LO(29))
  155. #define DMUL (HI(0) | (2 << 6) | LO(28))
  156. #define DMULU (HI(0) | (2 << 6) | LO(29))
  157. #else /* SLJIT_MIPS_REV < 6 */
  158. #define DDIV (HI(0) | LO(30))
  159. #define DDIVU (HI(0) | LO(31))
  160. #define DIV (HI(0) | LO(26))
  161. #define DIVU (HI(0) | LO(27))
  162. #define DMULT (HI(0) | LO(28))
  163. #define DMULTU (HI(0) | LO(29))
  164. #endif /* SLJIT_MIPS_REV >= 6 */
  165. #define DIV_S (HI(17) | FMT_S | LO(3))
  166. #define DSLL (HI(0) | LO(56))
  167. #define DSLL32 (HI(0) | LO(60))
  168. #define DSLLV (HI(0) | LO(20))
  169. #define DSRA (HI(0) | LO(59))
  170. #define DSRA32 (HI(0) | LO(63))
  171. #define DSRAV (HI(0) | LO(23))
  172. #define DSRL (HI(0) | LO(58))
  173. #define DSRL32 (HI(0) | LO(62))
  174. #define DSRLV (HI(0) | LO(22))
  175. #define DSUBU (HI(0) | LO(47))
  176. #define J (HI(2))
  177. #define JAL (HI(3))
  178. #define JALR (HI(0) | LO(9))
  179. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  180. #define JR (HI(0) | LO(9))
  181. #else /* SLJIT_MIPS_REV < 6 */
  182. #define JR (HI(0) | LO(8))
  183. #endif /* SLJIT_MIPS_REV >= 6 */
  184. #define LD (HI(55))
  185. #define LUI (HI(15))
  186. #define LW (HI(35))
  187. #define MFC1 (HI(17))
  188. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  189. #define MOD (HI(0) | (3 << 6) | LO(26))
  190. #define MODU (HI(0) | (3 << 6) | LO(27))
  191. #else /* SLJIT_MIPS_REV < 6 */
  192. #define MFHI (HI(0) | LO(16))
  193. #define MFLO (HI(0) | LO(18))
  194. #endif /* SLJIT_MIPS_REV >= 6 */
  195. #define MOV_S (HI(17) | FMT_S | LO(6))
  196. #define MTC1 (HI(17) | (4 << 21))
  197. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  198. #define MUH (HI(0) | (3 << 6) | LO(24))
  199. #define MUHU (HI(0) | (3 << 6) | LO(25))
  200. #define MUL (HI(0) | (2 << 6) | LO(24))
  201. #define MULU (HI(0) | (2 << 6) | LO(25))
  202. #else /* SLJIT_MIPS_REV < 6 */
  203. #define MULT (HI(0) | LO(24))
  204. #define MULTU (HI(0) | LO(25))
  205. #endif /* SLJIT_MIPS_REV >= 6 */
  206. #define MUL_S (HI(17) | FMT_S | LO(2))
  207. #define NEG_S (HI(17) | FMT_S | LO(7))
  208. #define NOP (HI(0) | LO(0))
  209. #define NOR (HI(0) | LO(39))
  210. #define OR (HI(0) | LO(37))
  211. #define ORI (HI(13))
  212. #define SD (HI(63))
  213. #define SDC1 (HI(61))
  214. #define SLT (HI(0) | LO(42))
  215. #define SLTI (HI(10))
  216. #define SLTIU (HI(11))
  217. #define SLTU (HI(0) | LO(43))
  218. #define SLL (HI(0) | LO(0))
  219. #define SLLV (HI(0) | LO(4))
  220. #define SRL (HI(0) | LO(2))
  221. #define SRLV (HI(0) | LO(6))
  222. #define SRA (HI(0) | LO(3))
  223. #define SRAV (HI(0) | LO(7))
  224. #define SUB_S (HI(17) | FMT_S | LO(1))
  225. #define SUBU (HI(0) | LO(35))
  226. #define SW (HI(43))
  227. #define SWC1 (HI(57))
  228. #define TRUNC_W_S (HI(17) | FMT_S | LO(13))
  229. #define XOR (HI(0) | LO(38))
  230. #define XORI (HI(14))
  231. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1)
  232. #define CLZ (HI(28) | LO(32))
  233. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  234. #define DCLZ (LO(18))
  235. #else /* SLJIT_MIPS_REV < 6 */
  236. #define DCLZ (HI(28) | LO(36))
  237. #define MOVF (HI(0) | (0 << 16) | LO(1))
  238. #define MOVN (HI(0) | LO(11))
  239. #define MOVT (HI(0) | (1 << 16) | LO(1))
  240. #define MOVZ (HI(0) | LO(10))
  241. #define MUL (HI(28) | LO(2))
  242. #endif /* SLJIT_MIPS_REV >= 6 */
  243. #define PREF (HI(51))
  244. #define PREFX (HI(19) | LO(15))
  245. #define SEB (HI(31) | (16 << 6) | LO(32))
  246. #define SEH (HI(31) | (24 << 6) | LO(32))
  247. #endif /* SLJIT_MIPS_REV >= 1 */
  248. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  249. #define ADDU_W ADDU
  250. #define ADDIU_W ADDIU
  251. #define SLL_W SLL
  252. #define SUBU_W SUBU
  253. #else
  254. #define ADDU_W DADDU
  255. #define ADDIU_W DADDIU
  256. #define SLL_W DSLL
  257. #define SUBU_W DSUBU
  258. #endif
  259. #define SIMM_MAX (0x7fff)
  260. #define SIMM_MIN (-0x8000)
  261. #define UIMM_MAX (0xffff)
  262. /* dest_reg is the absolute name of the register
  263. Useful for reordering instructions in the delay slot. */
  264. static sljit_s32 push_inst(struct sljit_compiler *compiler, sljit_ins ins, sljit_s32 delay_slot)
  265. {
  266. sljit_ins *ptr = (sljit_ins*)ensure_buf(compiler, sizeof(sljit_ins));
  267. SLJIT_ASSERT(delay_slot == MOVABLE_INS || delay_slot >= UNMOVABLE_INS
  268. || delay_slot == ((ins >> 11) & 0x1f) || delay_slot == ((ins >> 16) & 0x1f));
  269. FAIL_IF(!ptr);
  270. *ptr = ins;
  271. compiler->size++;
  272. compiler->delay_slot = delay_slot;
  273. return SLJIT_SUCCESS;
  274. }
  275. static SLJIT_INLINE sljit_ins invert_branch(sljit_s32 flags)
  276. {
  277. if (flags & IS_BIT26_COND)
  278. return (1 << 26);
  279. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  280. if (flags & IS_BIT23_COND)
  281. return (1 << 23);
  282. #endif /* SLJIT_MIPS_REV >= 6 */
  283. return (1 << 16);
  284. }
  285. static SLJIT_INLINE sljit_ins* detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset)
  286. {
  287. sljit_sw diff;
  288. sljit_uw target_addr;
  289. sljit_ins *inst;
  290. sljit_ins saved_inst;
  291. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  292. if (jump->flags & (SLJIT_REWRITABLE_JUMP | IS_CALL))
  293. return code_ptr;
  294. #else
  295. if (jump->flags & SLJIT_REWRITABLE_JUMP)
  296. return code_ptr;
  297. #endif
  298. if (jump->flags & JUMP_ADDR)
  299. target_addr = jump->u.target;
  300. else {
  301. SLJIT_ASSERT(jump->flags & JUMP_LABEL);
  302. target_addr = (sljit_uw)(code + jump->u.label->size) + (sljit_uw)executable_offset;
  303. }
  304. inst = (sljit_ins *)jump->addr;
  305. if (jump->flags & IS_COND)
  306. inst--;
  307. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  308. if (jump->flags & IS_CALL)
  309. goto keep_address;
  310. #endif
  311. /* B instructions. */
  312. if (jump->flags & IS_MOVABLE) {
  313. diff = ((sljit_sw)target_addr - (sljit_sw)inst - executable_offset) >> 2;
  314. if (diff <= SIMM_MAX && diff >= SIMM_MIN) {
  315. jump->flags |= PATCH_B;
  316. if (!(jump->flags & IS_COND)) {
  317. inst[0] = inst[-1];
  318. inst[-1] = (jump->flags & IS_JAL) ? BAL : B;
  319. jump->addr -= sizeof(sljit_ins);
  320. return inst;
  321. }
  322. saved_inst = inst[0];
  323. inst[0] = inst[-1];
  324. inst[-1] = saved_inst ^ invert_branch(jump->flags);
  325. jump->addr -= 2 * sizeof(sljit_ins);
  326. return inst;
  327. }
  328. }
  329. else {
  330. diff = ((sljit_sw)target_addr - (sljit_sw)(inst + 1) - executable_offset) >> 2;
  331. if (diff <= SIMM_MAX && diff >= SIMM_MIN) {
  332. jump->flags |= PATCH_B;
  333. if (!(jump->flags & IS_COND)) {
  334. inst[0] = (jump->flags & IS_JAL) ? BAL : B;
  335. inst[1] = NOP;
  336. return inst + 1;
  337. }
  338. inst[0] = inst[0] ^ invert_branch(jump->flags);
  339. inst[1] = NOP;
  340. jump->addr -= sizeof(sljit_ins);
  341. return inst + 1;
  342. }
  343. }
  344. if (jump->flags & IS_COND) {
  345. if ((jump->flags & IS_MOVABLE) && (target_addr & ~0xfffffff) == ((jump->addr + 2 * sizeof(sljit_ins)) & ~0xfffffff)) {
  346. jump->flags |= PATCH_J;
  347. saved_inst = inst[0];
  348. inst[0] = inst[-1];
  349. inst[-1] = (saved_inst & 0xffff0000) | 3;
  350. inst[1] = J;
  351. inst[2] = NOP;
  352. return inst + 2;
  353. }
  354. else if ((target_addr & ~0xfffffff) == ((jump->addr + 3 * sizeof(sljit_ins)) & ~0xfffffff)) {
  355. jump->flags |= PATCH_J;
  356. inst[0] = (inst[0] & 0xffff0000) | 3;
  357. inst[1] = NOP;
  358. inst[2] = J;
  359. inst[3] = NOP;
  360. jump->addr += sizeof(sljit_ins);
  361. return inst + 3;
  362. }
  363. }
  364. else {
  365. /* J instuctions. */
  366. if ((jump->flags & IS_MOVABLE) && (target_addr & ~0xfffffff) == (jump->addr & ~0xfffffff)) {
  367. jump->flags |= PATCH_J;
  368. inst[0] = inst[-1];
  369. inst[-1] = (jump->flags & IS_JAL) ? JAL : J;
  370. jump->addr -= sizeof(sljit_ins);
  371. return inst;
  372. }
  373. if ((target_addr & ~0xfffffff) == ((jump->addr + sizeof(sljit_ins)) & ~0xfffffff)) {
  374. jump->flags |= PATCH_J;
  375. inst[0] = (jump->flags & IS_JAL) ? JAL : J;
  376. inst[1] = NOP;
  377. return inst + 1;
  378. }
  379. }
  380. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  381. keep_address:
  382. if (target_addr <= 0x7fffffff) {
  383. jump->flags |= PATCH_ABS32;
  384. if (jump->flags & IS_COND) {
  385. inst[0] -= 4;
  386. inst++;
  387. }
  388. inst[2] = inst[6];
  389. inst[3] = inst[7];
  390. return inst + 3;
  391. }
  392. if (target_addr <= 0x7fffffffffffl) {
  393. jump->flags |= PATCH_ABS48;
  394. if (jump->flags & IS_COND) {
  395. inst[0] -= 2;
  396. inst++;
  397. }
  398. inst[4] = inst[6];
  399. inst[5] = inst[7];
  400. return inst + 5;
  401. }
  402. #endif
  403. return code_ptr;
  404. }
  405. #ifdef __GNUC__
  406. static __attribute__ ((noinline)) void sljit_cache_flush(void* code, void* code_ptr)
  407. {
  408. SLJIT_CACHE_FLUSH(code, code_ptr);
  409. }
  410. #endif
  411. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  412. static SLJIT_INLINE sljit_sw put_label_get_length(struct sljit_put_label *put_label, sljit_uw max_label)
  413. {
  414. if (max_label < 0x80000000l) {
  415. put_label->flags = 0;
  416. return 1;
  417. }
  418. if (max_label < 0x800000000000l) {
  419. put_label->flags = 1;
  420. return 3;
  421. }
  422. put_label->flags = 2;
  423. return 5;
  424. }
  425. static SLJIT_INLINE void put_label_set(struct sljit_put_label *put_label)
  426. {
  427. sljit_uw addr = put_label->label->addr;
  428. sljit_ins *inst = (sljit_ins *)put_label->addr;
  429. sljit_s32 reg = *inst;
  430. if (put_label->flags == 0) {
  431. SLJIT_ASSERT(addr < 0x80000000l);
  432. inst[0] = LUI | T(reg) | IMM(addr >> 16);
  433. }
  434. else if (put_label->flags == 1) {
  435. SLJIT_ASSERT(addr < 0x800000000000l);
  436. inst[0] = LUI | T(reg) | IMM(addr >> 32);
  437. inst[1] = ORI | S(reg) | T(reg) | IMM((addr >> 16) & 0xffff);
  438. inst[2] = DSLL | T(reg) | D(reg) | SH_IMM(16);
  439. inst += 2;
  440. }
  441. else {
  442. inst[0] = LUI | T(reg) | IMM(addr >> 48);
  443. inst[1] = ORI | S(reg) | T(reg) | IMM((addr >> 32) & 0xffff);
  444. inst[2] = DSLL | T(reg) | D(reg) | SH_IMM(16);
  445. inst[3] = ORI | S(reg) | T(reg) | IMM((addr >> 16) & 0xffff);
  446. inst[4] = DSLL | T(reg) | D(reg) | SH_IMM(16);
  447. inst += 4;
  448. }
  449. inst[1] = ORI | S(reg) | T(reg) | IMM(addr & 0xffff);
  450. }
  451. #endif
  452. SLJIT_API_FUNC_ATTRIBUTE void* sljit_generate_code(struct sljit_compiler *compiler)
  453. {
  454. struct sljit_memory_fragment *buf;
  455. sljit_ins *code;
  456. sljit_ins *code_ptr;
  457. sljit_ins *buf_ptr;
  458. sljit_ins *buf_end;
  459. sljit_uw word_count;
  460. sljit_uw next_addr;
  461. sljit_sw executable_offset;
  462. sljit_uw addr;
  463. struct sljit_label *label;
  464. struct sljit_jump *jump;
  465. struct sljit_const *const_;
  466. struct sljit_put_label *put_label;
  467. CHECK_ERROR_PTR();
  468. CHECK_PTR(check_sljit_generate_code(compiler));
  469. reverse_buf(compiler);
  470. code = (sljit_ins*)SLJIT_MALLOC_EXEC(compiler->size * sizeof(sljit_ins), compiler->exec_allocator_data);
  471. PTR_FAIL_WITH_EXEC_IF(code);
  472. buf = compiler->buf;
  473. code_ptr = code;
  474. word_count = 0;
  475. next_addr = 0;
  476. executable_offset = SLJIT_EXEC_OFFSET(code);
  477. label = compiler->labels;
  478. jump = compiler->jumps;
  479. const_ = compiler->consts;
  480. put_label = compiler->put_labels;
  481. do {
  482. buf_ptr = (sljit_ins*)buf->memory;
  483. buf_end = buf_ptr + (buf->used_size >> 2);
  484. do {
  485. *code_ptr = *buf_ptr++;
  486. if (next_addr == word_count) {
  487. SLJIT_ASSERT(!label || label->size >= word_count);
  488. SLJIT_ASSERT(!jump || jump->addr >= word_count);
  489. SLJIT_ASSERT(!const_ || const_->addr >= word_count);
  490. SLJIT_ASSERT(!put_label || put_label->addr >= word_count);
  491. /* These structures are ordered by their address. */
  492. if (label && label->size == word_count) {
  493. label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
  494. label->size = code_ptr - code;
  495. label = label->next;
  496. }
  497. if (jump && jump->addr == word_count) {
  498. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  499. jump->addr = (sljit_uw)(code_ptr - 3);
  500. #else
  501. jump->addr = (sljit_uw)(code_ptr - 7);
  502. #endif
  503. code_ptr = detect_jump_type(jump, code_ptr, code, executable_offset);
  504. jump = jump->next;
  505. }
  506. if (const_ && const_->addr == word_count) {
  507. const_->addr = (sljit_uw)code_ptr;
  508. const_ = const_->next;
  509. }
  510. if (put_label && put_label->addr == word_count) {
  511. SLJIT_ASSERT(put_label->label);
  512. put_label->addr = (sljit_uw)code_ptr;
  513. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  514. code_ptr += put_label_get_length(put_label, (sljit_uw)(SLJIT_ADD_EXEC_OFFSET(code, executable_offset) + put_label->label->size));
  515. word_count += 5;
  516. #endif
  517. put_label = put_label->next;
  518. }
  519. next_addr = compute_next_addr(label, jump, const_, put_label);
  520. }
  521. code_ptr ++;
  522. word_count ++;
  523. } while (buf_ptr < buf_end);
  524. buf = buf->next;
  525. } while (buf);
  526. if (label && label->size == word_count) {
  527. label->addr = (sljit_uw)code_ptr;
  528. label->size = code_ptr - code;
  529. label = label->next;
  530. }
  531. SLJIT_ASSERT(!label);
  532. SLJIT_ASSERT(!jump);
  533. SLJIT_ASSERT(!const_);
  534. SLJIT_ASSERT(!put_label);
  535. SLJIT_ASSERT(code_ptr - code <= (sljit_sw)compiler->size);
  536. jump = compiler->jumps;
  537. while (jump) {
  538. do {
  539. addr = (jump->flags & JUMP_LABEL) ? jump->u.label->addr : jump->u.target;
  540. buf_ptr = (sljit_ins *)jump->addr;
  541. if (jump->flags & PATCH_B) {
  542. addr = (sljit_sw)(addr - ((sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset) + sizeof(sljit_ins))) >> 2;
  543. SLJIT_ASSERT((sljit_sw)addr <= SIMM_MAX && (sljit_sw)addr >= SIMM_MIN);
  544. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | (addr & 0xffff);
  545. break;
  546. }
  547. if (jump->flags & PATCH_J) {
  548. SLJIT_ASSERT((addr & ~0xfffffff) == (((sljit_uw)SLJIT_ADD_EXEC_OFFSET(buf_ptr, executable_offset) + sizeof(sljit_ins)) & ~0xfffffff));
  549. buf_ptr[0] |= (addr >> 2) & 0x03ffffff;
  550. break;
  551. }
  552. /* Set the fields of immediate loads. */
  553. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  554. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 16) & 0xffff);
  555. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | (addr & 0xffff);
  556. #else
  557. if (jump->flags & PATCH_ABS32) {
  558. SLJIT_ASSERT(addr <= 0x7fffffff);
  559. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 16) & 0xffff);
  560. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | (addr & 0xffff);
  561. }
  562. else if (jump->flags & PATCH_ABS48) {
  563. SLJIT_ASSERT(addr <= 0x7fffffffffffl);
  564. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 32) & 0xffff);
  565. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | ((addr >> 16) & 0xffff);
  566. buf_ptr[3] = (buf_ptr[3] & 0xffff0000) | (addr & 0xffff);
  567. }
  568. else {
  569. buf_ptr[0] = (buf_ptr[0] & 0xffff0000) | ((addr >> 48) & 0xffff);
  570. buf_ptr[1] = (buf_ptr[1] & 0xffff0000) | ((addr >> 32) & 0xffff);
  571. buf_ptr[3] = (buf_ptr[3] & 0xffff0000) | ((addr >> 16) & 0xffff);
  572. buf_ptr[5] = (buf_ptr[5] & 0xffff0000) | (addr & 0xffff);
  573. }
  574. #endif
  575. } while (0);
  576. jump = jump->next;
  577. }
  578. put_label = compiler->put_labels;
  579. while (put_label) {
  580. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  581. addr = put_label->label->addr;
  582. buf_ptr = (sljit_ins *)put_label->addr;
  583. SLJIT_ASSERT((buf_ptr[0] & 0xffe00000) == LUI && (buf_ptr[1] & 0xfc000000) == ORI);
  584. buf_ptr[0] |= (addr >> 16) & 0xffff;
  585. buf_ptr[1] |= addr & 0xffff;
  586. #else
  587. put_label_set(put_label);
  588. #endif
  589. put_label = put_label->next;
  590. }
  591. compiler->error = SLJIT_ERR_COMPILED;
  592. compiler->executable_offset = executable_offset;
  593. compiler->executable_size = (code_ptr - code) * sizeof(sljit_ins);
  594. code = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code, executable_offset);
  595. code_ptr = (sljit_ins *)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset);
  596. #ifndef __GNUC__
  597. SLJIT_CACHE_FLUSH(code, code_ptr);
  598. #else
  599. /* GCC workaround for invalid code generation with -O2. */
  600. sljit_cache_flush(code, code_ptr);
  601. #endif
  602. SLJIT_UPDATE_WX_FLAGS(code, code_ptr, 1);
  603. return code;
  604. }
  605. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_has_cpu_feature(sljit_s32 feature_type)
  606. {
  607. sljit_sw fir = 0;
  608. switch (feature_type) {
  609. case SLJIT_HAS_FPU:
  610. #ifdef SLJIT_IS_FPU_AVAILABLE
  611. return SLJIT_IS_FPU_AVAILABLE;
  612. #elif defined(__GNUC__)
  613. __asm__ ("cfc1 %0, $0" : "=r"(fir));
  614. return (fir >> 22) & 0x1;
  615. #else
  616. #error "FIR check is not implemented for this architecture"
  617. #endif
  618. case SLJIT_HAS_ZERO_REGISTER:
  619. return 1;
  620. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1)
  621. case SLJIT_HAS_CLZ:
  622. case SLJIT_HAS_CMOV:
  623. case SLJIT_HAS_PREFETCH:
  624. return 1;
  625. #endif /* SLJIT_MIPS_REV >= 1 */
  626. default:
  627. return fir;
  628. }
  629. }
  630. /* --------------------------------------------------------------------- */
  631. /* Entry, exit */
  632. /* --------------------------------------------------------------------- */
  633. /* Creates an index in data_transfer_insts array. */
  634. #define LOAD_DATA 0x01
  635. #define WORD_DATA 0x00
  636. #define BYTE_DATA 0x02
  637. #define HALF_DATA 0x04
  638. #define INT_DATA 0x06
  639. #define SIGNED_DATA 0x08
  640. /* Separates integer and floating point registers */
  641. #define GPR_REG 0x0f
  642. #define DOUBLE_DATA 0x10
  643. #define SINGLE_DATA 0x12
  644. #define MEM_MASK 0x1f
  645. #define ARG_TEST 0x00020
  646. #define ALT_KEEP_CACHE 0x00040
  647. #define CUMULATIVE_OP 0x00080
  648. #define LOGICAL_OP 0x00100
  649. #define IMM_OP 0x00200
  650. #define SRC2_IMM 0x00400
  651. #define UNUSED_DEST 0x00800
  652. #define REG_DEST 0x01000
  653. #define REG1_SOURCE 0x02000
  654. #define REG2_SOURCE 0x04000
  655. #define SLOW_SRC1 0x08000
  656. #define SLOW_SRC2 0x10000
  657. #define SLOW_DEST 0x20000
  658. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  659. #define STACK_STORE SW
  660. #define STACK_LOAD LW
  661. #else
  662. #define STACK_STORE SD
  663. #define STACK_LOAD LD
  664. #endif
  665. static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw);
  666. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  667. #include "sljitNativeMIPS_32.c"
  668. #else
  669. #include "sljitNativeMIPS_64.c"
  670. #endif
  671. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_enter(struct sljit_compiler *compiler,
  672. sljit_s32 options, sljit_s32 arg_types, sljit_s32 scratches, sljit_s32 saveds,
  673. sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
  674. {
  675. sljit_ins base;
  676. sljit_s32 args, i, tmp, offs;
  677. CHECK_ERROR();
  678. CHECK(check_sljit_emit_enter(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size));
  679. set_emit_enter(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size);
  680. local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 1) + SLJIT_LOCALS_OFFSET;
  681. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  682. local_size = (local_size + 15) & ~0xf;
  683. #else
  684. local_size = (local_size + 31) & ~0x1f;
  685. #endif
  686. compiler->local_size = local_size;
  687. if (local_size <= SIMM_MAX) {
  688. /* Frequent case. */
  689. FAIL_IF(push_inst(compiler, ADDIU_W | S(SLJIT_SP) | T(SLJIT_SP) | IMM(-local_size), DR(SLJIT_SP)));
  690. base = S(SLJIT_SP);
  691. offs = local_size - (sljit_sw)sizeof(sljit_sw);
  692. }
  693. else {
  694. FAIL_IF(load_immediate(compiler, DR(OTHER_FLAG), local_size));
  695. FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_SP) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
  696. FAIL_IF(push_inst(compiler, SUBU_W | S(SLJIT_SP) | T(OTHER_FLAG) | D(SLJIT_SP), DR(SLJIT_SP)));
  697. base = S(TMP_REG2);
  698. local_size = 0;
  699. offs = -(sljit_sw)sizeof(sljit_sw);
  700. }
  701. FAIL_IF(push_inst(compiler, STACK_STORE | base | TA(RETURN_ADDR_REG) | IMM(offs), MOVABLE_INS));
  702. tmp = saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - saveds) : SLJIT_FIRST_SAVED_REG;
  703. for (i = SLJIT_S0; i >= tmp; i--) {
  704. offs -= (sljit_s32)(sizeof(sljit_sw));
  705. FAIL_IF(push_inst(compiler, STACK_STORE | base | T(i) | IMM(offs), MOVABLE_INS));
  706. }
  707. for (i = scratches; i >= SLJIT_FIRST_SAVED_REG; i--) {
  708. offs -= (sljit_s32)(sizeof(sljit_sw));
  709. FAIL_IF(push_inst(compiler, STACK_STORE | base | T(i) | IMM(offs), MOVABLE_INS));
  710. }
  711. args = get_arg_count(arg_types);
  712. if (args >= 1)
  713. FAIL_IF(push_inst(compiler, ADDU_W | SA(4) | TA(0) | D(SLJIT_S0), DR(SLJIT_S0)));
  714. if (args >= 2)
  715. FAIL_IF(push_inst(compiler, ADDU_W | SA(5) | TA(0) | D(SLJIT_S1), DR(SLJIT_S1)));
  716. if (args >= 3)
  717. FAIL_IF(push_inst(compiler, ADDU_W | SA(6) | TA(0) | D(SLJIT_S2), DR(SLJIT_S2)));
  718. return SLJIT_SUCCESS;
  719. }
  720. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_set_context(struct sljit_compiler *compiler,
  721. sljit_s32 options, sljit_s32 arg_types, sljit_s32 scratches, sljit_s32 saveds,
  722. sljit_s32 fscratches, sljit_s32 fsaveds, sljit_s32 local_size)
  723. {
  724. CHECK_ERROR();
  725. CHECK(check_sljit_set_context(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size));
  726. set_set_context(compiler, options, arg_types, scratches, saveds, fscratches, fsaveds, local_size);
  727. local_size += GET_SAVED_REGISTERS_SIZE(scratches, saveds, 1) + SLJIT_LOCALS_OFFSET;
  728. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  729. compiler->local_size = (local_size + 15) & ~0xf;
  730. #else
  731. compiler->local_size = (local_size + 31) & ~0x1f;
  732. #endif
  733. return SLJIT_SUCCESS;
  734. }
  735. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_return(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 src, sljit_sw srcw)
  736. {
  737. sljit_s32 local_size, i, tmp, offs;
  738. sljit_ins base;
  739. CHECK_ERROR();
  740. CHECK(check_sljit_emit_return(compiler, op, src, srcw));
  741. FAIL_IF(emit_mov_before_return(compiler, op, src, srcw));
  742. local_size = compiler->local_size;
  743. if (local_size <= SIMM_MAX)
  744. base = S(SLJIT_SP);
  745. else {
  746. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), local_size));
  747. FAIL_IF(push_inst(compiler, ADDU_W | S(SLJIT_SP) | T(TMP_REG1) | D(TMP_REG1), DR(TMP_REG1)));
  748. base = S(TMP_REG1);
  749. local_size = 0;
  750. }
  751. FAIL_IF(push_inst(compiler, STACK_LOAD | base | TA(RETURN_ADDR_REG) | IMM(local_size - (sljit_s32)sizeof(sljit_sw)), RETURN_ADDR_REG));
  752. offs = local_size - (sljit_s32)GET_SAVED_REGISTERS_SIZE(compiler->scratches, compiler->saveds, 1);
  753. tmp = compiler->scratches;
  754. for (i = SLJIT_FIRST_SAVED_REG; i <= tmp; i++) {
  755. FAIL_IF(push_inst(compiler, STACK_LOAD | base | T(i) | IMM(offs), DR(i)));
  756. offs += (sljit_s32)(sizeof(sljit_sw));
  757. }
  758. tmp = compiler->saveds < SLJIT_NUMBER_OF_SAVED_REGISTERS ? (SLJIT_S0 + 1 - compiler->saveds) : SLJIT_FIRST_SAVED_REG;
  759. for (i = tmp; i <= SLJIT_S0; i++) {
  760. FAIL_IF(push_inst(compiler, STACK_LOAD | base | T(i) | IMM(offs), DR(i)));
  761. offs += (sljit_s32)(sizeof(sljit_sw));
  762. }
  763. SLJIT_ASSERT(offs == local_size - (sljit_sw)(sizeof(sljit_sw)));
  764. FAIL_IF(push_inst(compiler, JR | SA(RETURN_ADDR_REG), UNMOVABLE_INS));
  765. if (compiler->local_size <= SIMM_MAX)
  766. return push_inst(compiler, ADDIU_W | S(SLJIT_SP) | T(SLJIT_SP) | IMM(compiler->local_size), UNMOVABLE_INS);
  767. else
  768. return push_inst(compiler, ADDU_W | S(TMP_REG1) | TA(0) | D(SLJIT_SP), UNMOVABLE_INS);
  769. }
  770. #undef STACK_STORE
  771. #undef STACK_LOAD
  772. /* --------------------------------------------------------------------- */
  773. /* Operators */
  774. /* --------------------------------------------------------------------- */
  775. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  776. #define ARCH_32_64(a, b) a
  777. #else
  778. #define ARCH_32_64(a, b) b
  779. #endif
  780. static const sljit_ins data_transfer_insts[16 + 4] = {
  781. /* u w s */ ARCH_32_64(HI(43) /* sw */, HI(63) /* sd */),
  782. /* u w l */ ARCH_32_64(HI(35) /* lw */, HI(55) /* ld */),
  783. /* u b s */ HI(40) /* sb */,
  784. /* u b l */ HI(36) /* lbu */,
  785. /* u h s */ HI(41) /* sh */,
  786. /* u h l */ HI(37) /* lhu */,
  787. /* u i s */ HI(43) /* sw */,
  788. /* u i l */ ARCH_32_64(HI(35) /* lw */, HI(39) /* lwu */),
  789. /* s w s */ ARCH_32_64(HI(43) /* sw */, HI(63) /* sd */),
  790. /* s w l */ ARCH_32_64(HI(35) /* lw */, HI(55) /* ld */),
  791. /* s b s */ HI(40) /* sb */,
  792. /* s b l */ HI(32) /* lb */,
  793. /* s h s */ HI(41) /* sh */,
  794. /* s h l */ HI(33) /* lh */,
  795. /* s i s */ HI(43) /* sw */,
  796. /* s i l */ HI(35) /* lw */,
  797. /* d s */ HI(61) /* sdc1 */,
  798. /* d l */ HI(53) /* ldc1 */,
  799. /* s s */ HI(57) /* swc1 */,
  800. /* s l */ HI(49) /* lwc1 */,
  801. };
  802. #undef ARCH_32_64
  803. /* reg_ar is an absoulute register! */
  804. /* Can perform an operation using at most 1 instruction. */
  805. static sljit_s32 getput_arg_fast(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw)
  806. {
  807. SLJIT_ASSERT(arg & SLJIT_MEM);
  808. if (!(arg & OFFS_REG_MASK) && argw <= SIMM_MAX && argw >= SIMM_MIN) {
  809. /* Works for both absoulte and relative addresses. */
  810. if (SLJIT_UNLIKELY(flags & ARG_TEST))
  811. return 1;
  812. FAIL_IF(push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(arg & REG_MASK)
  813. | TA(reg_ar) | IMM(argw), ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) ? reg_ar : MOVABLE_INS));
  814. return -1;
  815. }
  816. return 0;
  817. }
  818. /* See getput_arg below.
  819. Note: can_cache is called only for binary operators. Those
  820. operators always uses word arguments without write back. */
  821. static sljit_s32 can_cache(sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
  822. {
  823. SLJIT_ASSERT((arg & SLJIT_MEM) && (next_arg & SLJIT_MEM));
  824. /* Simple operation except for updates. */
  825. if (arg & OFFS_REG_MASK) {
  826. argw &= 0x3;
  827. next_argw &= 0x3;
  828. if (argw && argw == next_argw && (arg == next_arg || (arg & OFFS_REG_MASK) == (next_arg & OFFS_REG_MASK)))
  829. return 1;
  830. return 0;
  831. }
  832. if (arg == next_arg) {
  833. if (((next_argw - argw) <= SIMM_MAX && (next_argw - argw) >= SIMM_MIN))
  834. return 1;
  835. return 0;
  836. }
  837. return 0;
  838. }
  839. /* Emit the necessary instructions. See can_cache above. */
  840. static sljit_s32 getput_arg(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw, sljit_s32 next_arg, sljit_sw next_argw)
  841. {
  842. sljit_s32 tmp_ar, base, delay_slot;
  843. SLJIT_ASSERT(arg & SLJIT_MEM);
  844. if (!(next_arg & SLJIT_MEM)) {
  845. next_arg = 0;
  846. next_argw = 0;
  847. }
  848. if ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) {
  849. tmp_ar = reg_ar;
  850. delay_slot = reg_ar;
  851. }
  852. else {
  853. tmp_ar = DR(TMP_REG1);
  854. delay_slot = MOVABLE_INS;
  855. }
  856. base = arg & REG_MASK;
  857. if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) {
  858. argw &= 0x3;
  859. /* Using the cache. */
  860. if (argw == compiler->cache_argw) {
  861. if (arg == compiler->cache_arg)
  862. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  863. if ((SLJIT_MEM | (arg & OFFS_REG_MASK)) == compiler->cache_arg) {
  864. if (arg == next_arg && argw == (next_argw & 0x3)) {
  865. compiler->cache_arg = arg;
  866. compiler->cache_argw = argw;
  867. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | D(TMP_REG3), DR(TMP_REG3)));
  868. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  869. }
  870. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(TMP_REG3) | DA(tmp_ar), tmp_ar));
  871. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  872. }
  873. }
  874. if (SLJIT_UNLIKELY(argw)) {
  875. compiler->cache_arg = SLJIT_MEM | (arg & OFFS_REG_MASK);
  876. compiler->cache_argw = argw;
  877. FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(arg)) | D(TMP_REG3) | SH_IMM(argw), DR(TMP_REG3)));
  878. }
  879. if (arg == next_arg && argw == (next_argw & 0x3)) {
  880. compiler->cache_arg = arg;
  881. compiler->cache_argw = argw;
  882. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(!argw ? OFFS_REG(arg) : TMP_REG3) | D(TMP_REG3), DR(TMP_REG3)));
  883. tmp_ar = DR(TMP_REG3);
  884. }
  885. else
  886. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(!argw ? OFFS_REG(arg) : TMP_REG3) | DA(tmp_ar), tmp_ar));
  887. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  888. }
  889. if (compiler->cache_arg == arg && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
  890. if (argw != compiler->cache_argw) {
  891. FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
  892. compiler->cache_argw = argw;
  893. }
  894. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  895. }
  896. if (compiler->cache_arg == SLJIT_MEM && argw - compiler->cache_argw <= SIMM_MAX && argw - compiler->cache_argw >= SIMM_MIN) {
  897. if (argw != compiler->cache_argw)
  898. FAIL_IF(push_inst(compiler, ADDIU_W | S(TMP_REG3) | T(TMP_REG3) | IMM(argw - compiler->cache_argw), DR(TMP_REG3)));
  899. }
  900. else {
  901. compiler->cache_arg = SLJIT_MEM;
  902. FAIL_IF(load_immediate(compiler, DR(TMP_REG3), argw));
  903. }
  904. compiler->cache_argw = argw;
  905. if (!base)
  906. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  907. if (arg == next_arg && next_argw - argw <= SIMM_MAX && next_argw - argw >= SIMM_MIN) {
  908. compiler->cache_arg = arg;
  909. FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | T(base) | D(TMP_REG3), DR(TMP_REG3)));
  910. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | S(TMP_REG3) | TA(reg_ar), delay_slot);
  911. }
  912. FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | T(base) | DA(tmp_ar), tmp_ar));
  913. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  914. }
  915. static SLJIT_INLINE sljit_s32 emit_op_mem(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg_ar, sljit_s32 arg, sljit_sw argw)
  916. {
  917. sljit_s32 tmp_ar, base, delay_slot;
  918. if (getput_arg_fast(compiler, flags, reg_ar, arg, argw))
  919. return compiler->error;
  920. if ((flags & MEM_MASK) <= GPR_REG && (flags & LOAD_DATA)) {
  921. tmp_ar = reg_ar;
  922. delay_slot = reg_ar;
  923. }
  924. else {
  925. tmp_ar = DR(TMP_REG1);
  926. delay_slot = MOVABLE_INS;
  927. }
  928. base = arg & REG_MASK;
  929. if (SLJIT_UNLIKELY(arg & OFFS_REG_MASK)) {
  930. argw &= 0x3;
  931. if (SLJIT_UNLIKELY(argw)) {
  932. FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(arg)) | DA(tmp_ar) | SH_IMM(argw), tmp_ar));
  933. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | TA(tmp_ar) | DA(tmp_ar), tmp_ar));
  934. }
  935. else
  936. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | T(OFFS_REG(arg)) | DA(tmp_ar), tmp_ar));
  937. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  938. }
  939. FAIL_IF(load_immediate(compiler, tmp_ar, argw));
  940. if (base != 0)
  941. FAIL_IF(push_inst(compiler, ADDU_W | S(base) | TA(tmp_ar) | DA(tmp_ar), tmp_ar));
  942. return push_inst(compiler, data_transfer_insts[flags & MEM_MASK] | SA(tmp_ar) | TA(reg_ar), delay_slot);
  943. }
  944. static SLJIT_INLINE sljit_s32 emit_op_mem2(struct sljit_compiler *compiler, sljit_s32 flags, sljit_s32 reg, sljit_s32 arg1, sljit_sw arg1w, sljit_s32 arg2, sljit_sw arg2w)
  945. {
  946. if (getput_arg_fast(compiler, flags, reg, arg1, arg1w))
  947. return compiler->error;
  948. return getput_arg(compiler, flags, reg, arg1, arg1w, arg2, arg2w);
  949. }
  950. static sljit_s32 emit_op(struct sljit_compiler *compiler, sljit_s32 op, sljit_s32 flags,
  951. sljit_s32 dst, sljit_sw dstw,
  952. sljit_s32 src1, sljit_sw src1w,
  953. sljit_s32 src2, sljit_sw src2w)
  954. {
  955. /* arg1 goes to TMP_REG1 or src reg
  956. arg2 goes to TMP_REG2, imm or src reg
  957. TMP_REG3 can be used for caching
  958. result goes to TMP_REG2, so put result can use TMP_REG1 and TMP_REG3. */
  959. sljit_s32 dst_r = TMP_REG2;
  960. sljit_s32 src1_r;
  961. sljit_sw src2_r = 0;
  962. sljit_s32 sugg_src2_r = TMP_REG2;
  963. if (!(flags & ALT_KEEP_CACHE)) {
  964. compiler->cache_arg = 0;
  965. compiler->cache_argw = 0;
  966. }
  967. if (SLJIT_UNLIKELY(dst == SLJIT_UNUSED)) {
  968. SLJIT_ASSERT(HAS_FLAGS(op));
  969. flags |= UNUSED_DEST;
  970. }
  971. else if (FAST_IS_REG(dst)) {
  972. dst_r = dst;
  973. flags |= REG_DEST;
  974. if (op >= SLJIT_MOV && op <= SLJIT_MOV_P)
  975. sugg_src2_r = dst_r;
  976. }
  977. else if ((dst & SLJIT_MEM) && !getput_arg_fast(compiler, flags | ARG_TEST, DR(TMP_REG1), dst, dstw))
  978. flags |= SLOW_DEST;
  979. if (flags & IMM_OP) {
  980. if ((src2 & SLJIT_IMM) && src2w) {
  981. if ((!(flags & LOGICAL_OP) && (src2w <= SIMM_MAX && src2w >= SIMM_MIN))
  982. || ((flags & LOGICAL_OP) && !(src2w & ~UIMM_MAX))) {
  983. flags |= SRC2_IMM;
  984. src2_r = src2w;
  985. }
  986. }
  987. if (!(flags & SRC2_IMM) && (flags & CUMULATIVE_OP) && (src1 & SLJIT_IMM) && src1w) {
  988. if ((!(flags & LOGICAL_OP) && (src1w <= SIMM_MAX && src1w >= SIMM_MIN))
  989. || ((flags & LOGICAL_OP) && !(src1w & ~UIMM_MAX))) {
  990. flags |= SRC2_IMM;
  991. src2_r = src1w;
  992. /* And swap arguments. */
  993. src1 = src2;
  994. src1w = src2w;
  995. src2 = SLJIT_IMM;
  996. /* src2w = src2_r unneeded. */
  997. }
  998. }
  999. }
  1000. /* Source 1. */
  1001. if (FAST_IS_REG(src1)) {
  1002. src1_r = src1;
  1003. flags |= REG1_SOURCE;
  1004. }
  1005. else if (src1 & SLJIT_IMM) {
  1006. if (src1w) {
  1007. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), src1w));
  1008. src1_r = TMP_REG1;
  1009. }
  1010. else
  1011. src1_r = 0;
  1012. }
  1013. else {
  1014. if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w))
  1015. FAIL_IF(compiler->error);
  1016. else
  1017. flags |= SLOW_SRC1;
  1018. src1_r = TMP_REG1;
  1019. }
  1020. /* Source 2. */
  1021. if (FAST_IS_REG(src2)) {
  1022. src2_r = src2;
  1023. flags |= REG2_SOURCE;
  1024. if (!(flags & REG_DEST) && op >= SLJIT_MOV && op <= SLJIT_MOV_P)
  1025. dst_r = src2_r;
  1026. }
  1027. else if (src2 & SLJIT_IMM) {
  1028. if (!(flags & SRC2_IMM)) {
  1029. if (src2w) {
  1030. FAIL_IF(load_immediate(compiler, DR(sugg_src2_r), src2w));
  1031. src2_r = sugg_src2_r;
  1032. }
  1033. else {
  1034. src2_r = 0;
  1035. if ((op >= SLJIT_MOV && op <= SLJIT_MOV_P) && (dst & SLJIT_MEM))
  1036. dst_r = 0;
  1037. }
  1038. }
  1039. }
  1040. else {
  1041. if (getput_arg_fast(compiler, flags | LOAD_DATA, DR(sugg_src2_r), src2, src2w))
  1042. FAIL_IF(compiler->error);
  1043. else
  1044. flags |= SLOW_SRC2;
  1045. src2_r = sugg_src2_r;
  1046. }
  1047. if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
  1048. SLJIT_ASSERT(src2_r == TMP_REG2);
  1049. if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
  1050. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, src1, src1w));
  1051. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, dst, dstw));
  1052. }
  1053. else {
  1054. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, src2, src2w));
  1055. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG2), src2, src2w, dst, dstw));
  1056. }
  1057. }
  1058. else if (flags & SLOW_SRC1)
  1059. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(TMP_REG1), src1, src1w, dst, dstw));
  1060. else if (flags & SLOW_SRC2)
  1061. FAIL_IF(getput_arg(compiler, flags | LOAD_DATA, DR(sugg_src2_r), src2, src2w, dst, dstw));
  1062. FAIL_IF(emit_single_op(compiler, op, flags, dst_r, src1_r, src2_r));
  1063. if (dst & SLJIT_MEM) {
  1064. if (!(flags & SLOW_DEST)) {
  1065. getput_arg_fast(compiler, flags, DR(dst_r), dst, dstw);
  1066. return compiler->error;
  1067. }
  1068. return getput_arg(compiler, flags, DR(dst_r), dst, dstw, 0, 0);
  1069. }
  1070. return SLJIT_SUCCESS;
  1071. }
  1072. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op0(struct sljit_compiler *compiler, sljit_s32 op)
  1073. {
  1074. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1075. sljit_s32 int_op = op & SLJIT_I32_OP;
  1076. #endif
  1077. CHECK_ERROR();
  1078. CHECK(check_sljit_emit_op0(compiler, op));
  1079. op = GET_OPCODE(op);
  1080. switch (op) {
  1081. case SLJIT_BREAKPOINT:
  1082. return push_inst(compiler, BREAK, UNMOVABLE_INS);
  1083. case SLJIT_NOP:
  1084. return push_inst(compiler, NOP, UNMOVABLE_INS);
  1085. case SLJIT_LMUL_UW:
  1086. case SLJIT_LMUL_SW:
  1087. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  1088. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1089. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? DMULU : DMUL) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG3), DR(TMP_REG3)));
  1090. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? DMUHU : DMUH) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG1), DR(TMP_REG1)));
  1091. #else /* !SLJIT_CONFIG_MIPS_64 */
  1092. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? MULU : MUL) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG3), DR(TMP_REG3)));
  1093. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? MUHU : MUH) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG1), DR(TMP_REG1)));
  1094. #endif /* SLJIT_CONFIG_MIPS_64 */
  1095. FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | TA(0) | D(SLJIT_R0), DR(SLJIT_R0)));
  1096. return push_inst(compiler, ADDU_W | S(TMP_REG1) | TA(0) | D(SLJIT_R1), DR(SLJIT_R1));
  1097. #else /* SLJIT_MIPS_REV < 6 */
  1098. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1099. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? DMULTU : DMULT) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  1100. #else /* !SLJIT_CONFIG_MIPS_64 */
  1101. FAIL_IF(push_inst(compiler, (op == SLJIT_LMUL_UW ? MULTU : MULT) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  1102. #endif /* SLJIT_CONFIG_MIPS_64 */
  1103. FAIL_IF(push_inst(compiler, MFLO | D(SLJIT_R0), DR(SLJIT_R0)));
  1104. return push_inst(compiler, MFHI | D(SLJIT_R1), DR(SLJIT_R1));
  1105. #endif /* SLJIT_MIPS_REV >= 6 */
  1106. case SLJIT_DIVMOD_UW:
  1107. case SLJIT_DIVMOD_SW:
  1108. case SLJIT_DIV_UW:
  1109. case SLJIT_DIV_SW:
  1110. SLJIT_COMPILE_ASSERT((SLJIT_DIVMOD_UW & 0x2) == 0 && SLJIT_DIV_UW - 0x2 == SLJIT_DIVMOD_UW, bad_div_opcode_assignments);
  1111. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  1112. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1113. if (int_op) {
  1114. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG3), DR(TMP_REG3)));
  1115. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? MODU : MOD) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG1), DR(TMP_REG1)));
  1116. }
  1117. else {
  1118. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DDIVU : DDIV) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG3), DR(TMP_REG3)));
  1119. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DMODU : DMOD) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG1), DR(TMP_REG1)));
  1120. }
  1121. #else /* !SLJIT_CONFIG_MIPS_64 */
  1122. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG3), DR(TMP_REG3)));
  1123. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? MODU : MOD) | S(SLJIT_R0) | T(SLJIT_R1) | D(TMP_REG1), DR(TMP_REG1)));
  1124. #endif /* SLJIT_CONFIG_MIPS_64 */
  1125. FAIL_IF(push_inst(compiler, ADDU_W | S(TMP_REG3) | TA(0) | D(SLJIT_R0), DR(SLJIT_R0)));
  1126. return (op >= SLJIT_DIV_UW) ? SLJIT_SUCCESS : push_inst(compiler, ADDU_W | S(TMP_REG1) | TA(0) | D(SLJIT_R1), DR(SLJIT_R1));
  1127. #else /* SLJIT_MIPS_REV < 6 */
  1128. #if !(defined SLJIT_MIPS_REV)
  1129. FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1130. FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1131. #endif /* !SLJIT_MIPS_REV */
  1132. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1133. if (int_op)
  1134. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  1135. else
  1136. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DDIVU : DDIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  1137. #else /* !SLJIT_CONFIG_MIPS_64 */
  1138. FAIL_IF(push_inst(compiler, ((op | 0x2) == SLJIT_DIV_UW ? DIVU : DIV) | S(SLJIT_R0) | T(SLJIT_R1), MOVABLE_INS));
  1139. #endif /* SLJIT_CONFIG_MIPS_64 */
  1140. FAIL_IF(push_inst(compiler, MFLO | D(SLJIT_R0), DR(SLJIT_R0)));
  1141. return (op >= SLJIT_DIV_UW) ? SLJIT_SUCCESS : push_inst(compiler, MFHI | D(SLJIT_R1), DR(SLJIT_R1));
  1142. #endif /* SLJIT_MIPS_REV >= 6 */
  1143. case SLJIT_ENDBR:
  1144. case SLJIT_SKIP_FRAMES_BEFORE_RETURN:
  1145. return SLJIT_SUCCESS;
  1146. }
  1147. return SLJIT_SUCCESS;
  1148. }
  1149. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1)
  1150. static sljit_s32 emit_prefetch(struct sljit_compiler *compiler,
  1151. sljit_s32 src, sljit_sw srcw)
  1152. {
  1153. if (!(src & OFFS_REG_MASK)) {
  1154. if (srcw <= SIMM_MAX && srcw >= SIMM_MIN)
  1155. return push_inst(compiler, PREF | S(src & REG_MASK) | IMM(srcw), MOVABLE_INS);
  1156. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
  1157. return push_inst(compiler, PREFX | S(src & REG_MASK) | T(TMP_REG1), MOVABLE_INS);
  1158. }
  1159. srcw &= 0x3;
  1160. if (SLJIT_UNLIKELY(srcw != 0)) {
  1161. FAIL_IF(push_inst(compiler, SLL_W | T(OFFS_REG(src)) | D(TMP_REG1) | SH_IMM(srcw), DR(TMP_REG1)));
  1162. return push_inst(compiler, PREFX | S(src & REG_MASK) | T(TMP_REG1), MOVABLE_INS);
  1163. }
  1164. return push_inst(compiler, PREFX | S(src & REG_MASK) | T(OFFS_REG(src)), MOVABLE_INS);
  1165. }
  1166. #endif /* SLJIT_MIPS_REV >= 1 */
  1167. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op1(struct sljit_compiler *compiler, sljit_s32 op,
  1168. sljit_s32 dst, sljit_sw dstw,
  1169. sljit_s32 src, sljit_sw srcw)
  1170. {
  1171. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1172. # define flags 0
  1173. #else
  1174. sljit_s32 flags = 0;
  1175. #endif
  1176. CHECK_ERROR();
  1177. CHECK(check_sljit_emit_op1(compiler, op, dst, dstw, src, srcw));
  1178. ADJUST_LOCAL_OFFSET(dst, dstw);
  1179. ADJUST_LOCAL_OFFSET(src, srcw);
  1180. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1181. if ((op & SLJIT_I32_OP) && GET_OPCODE(op) >= SLJIT_NOT)
  1182. flags |= INT_DATA | SIGNED_DATA;
  1183. #endif
  1184. switch (GET_OPCODE(op)) {
  1185. case SLJIT_MOV:
  1186. case SLJIT_MOV_P:
  1187. return emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
  1188. case SLJIT_MOV_U32:
  1189. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1190. return emit_op(compiler, SLJIT_MOV_U32, INT_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
  1191. #else
  1192. return emit_op(compiler, SLJIT_MOV_U32, INT_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u32)srcw : srcw);
  1193. #endif
  1194. case SLJIT_MOV_S32:
  1195. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1196. return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, srcw);
  1197. #else
  1198. return emit_op(compiler, SLJIT_MOV_S32, INT_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s32)srcw : srcw);
  1199. #endif
  1200. case SLJIT_MOV_U8:
  1201. return emit_op(compiler, SLJIT_MOV_U8, BYTE_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u8)srcw : srcw);
  1202. case SLJIT_MOV_S8:
  1203. return emit_op(compiler, SLJIT_MOV_S8, BYTE_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s8)srcw : srcw);
  1204. case SLJIT_MOV_U16:
  1205. return emit_op(compiler, SLJIT_MOV_U16, HALF_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_u16)srcw : srcw);
  1206. case SLJIT_MOV_S16:
  1207. return emit_op(compiler, SLJIT_MOV_S16, HALF_DATA | SIGNED_DATA, dst, dstw, TMP_REG1, 0, src, (src & SLJIT_IMM) ? (sljit_s16)srcw : srcw);
  1208. case SLJIT_NOT:
  1209. return emit_op(compiler, op, flags, dst, dstw, TMP_REG1, 0, src, srcw);
  1210. case SLJIT_NEG:
  1211. compiler->status_flags_state = SLJIT_CURRENT_FLAGS_ADD_SUB;
  1212. return emit_op(compiler, SLJIT_SUB | GET_ALL_FLAGS(op), flags | IMM_OP, dst, dstw, SLJIT_IMM, 0, src, srcw);
  1213. case SLJIT_CLZ:
  1214. return emit_op(compiler, op, flags, dst, dstw, TMP_REG1, 0, src, srcw);
  1215. }
  1216. SLJIT_UNREACHABLE();
  1217. return SLJIT_SUCCESS;
  1218. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1219. # undef flags
  1220. #endif
  1221. }
  1222. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op2(struct sljit_compiler *compiler, sljit_s32 op,
  1223. sljit_s32 dst, sljit_sw dstw,
  1224. sljit_s32 src1, sljit_sw src1w,
  1225. sljit_s32 src2, sljit_sw src2w)
  1226. {
  1227. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1228. # define flags 0
  1229. #else
  1230. sljit_s32 flags = 0;
  1231. #endif
  1232. CHECK_ERROR();
  1233. CHECK(check_sljit_emit_op2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
  1234. ADJUST_LOCAL_OFFSET(dst, dstw);
  1235. ADJUST_LOCAL_OFFSET(src1, src1w);
  1236. ADJUST_LOCAL_OFFSET(src2, src2w);
  1237. if (dst == SLJIT_UNUSED && !HAS_FLAGS(op))
  1238. return SLJIT_SUCCESS;
  1239. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1240. if (op & SLJIT_I32_OP) {
  1241. flags |= INT_DATA | SIGNED_DATA;
  1242. if (src1 & SLJIT_IMM)
  1243. src1w = (sljit_s32)src1w;
  1244. if (src2 & SLJIT_IMM)
  1245. src2w = (sljit_s32)src2w;
  1246. }
  1247. #endif
  1248. switch (GET_OPCODE(op)) {
  1249. case SLJIT_ADD:
  1250. case SLJIT_ADDC:
  1251. compiler->status_flags_state = SLJIT_CURRENT_FLAGS_ADD_SUB;
  1252. return emit_op(compiler, op, flags | CUMULATIVE_OP | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1253. case SLJIT_SUB:
  1254. case SLJIT_SUBC:
  1255. compiler->status_flags_state = SLJIT_CURRENT_FLAGS_ADD_SUB;
  1256. return emit_op(compiler, op, flags | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1257. case SLJIT_MUL:
  1258. compiler->status_flags_state = 0;
  1259. return emit_op(compiler, op, flags | CUMULATIVE_OP, dst, dstw, src1, src1w, src2, src2w);
  1260. case SLJIT_AND:
  1261. case SLJIT_OR:
  1262. case SLJIT_XOR:
  1263. return emit_op(compiler, op, flags | CUMULATIVE_OP | LOGICAL_OP | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1264. case SLJIT_SHL:
  1265. case SLJIT_LSHR:
  1266. case SLJIT_ASHR:
  1267. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1268. if (src2 & SLJIT_IMM)
  1269. src2w &= 0x1f;
  1270. #else
  1271. if (src2 & SLJIT_IMM) {
  1272. if (op & SLJIT_I32_OP)
  1273. src2w &= 0x1f;
  1274. else
  1275. src2w &= 0x3f;
  1276. }
  1277. #endif
  1278. return emit_op(compiler, op, flags | IMM_OP, dst, dstw, src1, src1w, src2, src2w);
  1279. }
  1280. SLJIT_UNREACHABLE();
  1281. return SLJIT_SUCCESS;
  1282. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1283. # undef flags
  1284. #endif
  1285. }
  1286. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_src(struct sljit_compiler *compiler, sljit_s32 op,
  1287. sljit_s32 src, sljit_sw srcw)
  1288. {
  1289. CHECK_ERROR();
  1290. CHECK(check_sljit_emit_op_src(compiler, op, src, srcw));
  1291. ADJUST_LOCAL_OFFSET(src, srcw);
  1292. switch (op) {
  1293. case SLJIT_FAST_RETURN:
  1294. if (FAST_IS_REG(src))
  1295. FAIL_IF(push_inst(compiler, ADDU_W | S(src) | TA(0) | DA(RETURN_ADDR_REG), RETURN_ADDR_REG));
  1296. else
  1297. FAIL_IF(emit_op_mem(compiler, WORD_DATA | LOAD_DATA, RETURN_ADDR_REG, src, srcw));
  1298. FAIL_IF(push_inst(compiler, JR | SA(RETURN_ADDR_REG), UNMOVABLE_INS));
  1299. return push_inst(compiler, NOP, UNMOVABLE_INS);
  1300. case SLJIT_SKIP_FRAMES_BEFORE_FAST_RETURN:
  1301. return SLJIT_SUCCESS;
  1302. case SLJIT_PREFETCH_L1:
  1303. case SLJIT_PREFETCH_L2:
  1304. case SLJIT_PREFETCH_L3:
  1305. case SLJIT_PREFETCH_ONCE:
  1306. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1)
  1307. return emit_prefetch(compiler, src, srcw);
  1308. #else /* SLJIT_MIPS_REV < 1 */
  1309. return SLJIT_SUCCESS;
  1310. #endif /* SLJIT_MIPS_REV >= 1 */
  1311. }
  1312. return SLJIT_SUCCESS;
  1313. }
  1314. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_register_index(sljit_s32 reg)
  1315. {
  1316. CHECK_REG_INDEX(check_sljit_get_register_index(reg));
  1317. return reg_map[reg];
  1318. }
  1319. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_get_float_register_index(sljit_s32 reg)
  1320. {
  1321. CHECK_REG_INDEX(check_sljit_get_float_register_index(reg));
  1322. return FR(reg);
  1323. }
  1324. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_custom(struct sljit_compiler *compiler,
  1325. void *instruction, sljit_s32 size)
  1326. {
  1327. CHECK_ERROR();
  1328. CHECK(check_sljit_emit_op_custom(compiler, instruction, size));
  1329. return push_inst(compiler, *(sljit_ins*)instruction, UNMOVABLE_INS);
  1330. }
  1331. /* --------------------------------------------------------------------- */
  1332. /* Floating point operators */
  1333. /* --------------------------------------------------------------------- */
  1334. #define FLOAT_DATA(op) (DOUBLE_DATA | ((op & SLJIT_F32_OP) >> 7))
  1335. #define FMT(op) (((op & SLJIT_F32_OP) ^ SLJIT_F32_OP) << (21 - 8))
  1336. static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_sw_from_f64(struct sljit_compiler *compiler, sljit_s32 op,
  1337. sljit_s32 dst, sljit_sw dstw,
  1338. sljit_s32 src, sljit_sw srcw)
  1339. {
  1340. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1341. # define flags 0
  1342. #else
  1343. sljit_s32 flags = (GET_OPCODE(op) == SLJIT_CONV_SW_FROM_F64) << 21;
  1344. #endif
  1345. if (src & SLJIT_MEM) {
  1346. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src, srcw, dst, dstw));
  1347. src = TMP_FREG1;
  1348. }
  1349. FAIL_IF(push_inst(compiler, (TRUNC_W_S ^ (flags >> 19)) | FMT(op) | FS(src) | FD(TMP_FREG1), MOVABLE_INS));
  1350. if (FAST_IS_REG(dst))
  1351. return push_inst(compiler, MFC1 | flags | T(dst) | FS(TMP_FREG1), MOVABLE_INS);
  1352. /* Store the integer value from a VFP register. */
  1353. return emit_op_mem2(compiler, flags ? DOUBLE_DATA : SINGLE_DATA, FR(TMP_FREG1), dst, dstw, 0, 0);
  1354. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1355. # undef is_long
  1356. #endif
  1357. }
  1358. static SLJIT_INLINE sljit_s32 sljit_emit_fop1_conv_f64_from_sw(struct sljit_compiler *compiler, sljit_s32 op,
  1359. sljit_s32 dst, sljit_sw dstw,
  1360. sljit_s32 src, sljit_sw srcw)
  1361. {
  1362. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1363. # define flags 0
  1364. #else
  1365. sljit_s32 flags = (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_SW) << 21;
  1366. #endif
  1367. sljit_s32 dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
  1368. if (FAST_IS_REG(src))
  1369. FAIL_IF(push_inst(compiler, MTC1 | flags | T(src) | FS(TMP_FREG1), MOVABLE_INS));
  1370. else if (src & SLJIT_MEM) {
  1371. /* Load the integer value into a VFP register. */
  1372. FAIL_IF(emit_op_mem2(compiler, ((flags) ? DOUBLE_DATA : SINGLE_DATA) | LOAD_DATA, FR(TMP_FREG1), src, srcw, dst, dstw));
  1373. }
  1374. else {
  1375. #if (defined SLJIT_CONFIG_X86_64 && SLJIT_CONFIG_X86_64)
  1376. if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_S32)
  1377. srcw = (sljit_s32)srcw;
  1378. #endif
  1379. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
  1380. FAIL_IF(push_inst(compiler, MTC1 | flags | T(TMP_REG1) | FS(TMP_FREG1), MOVABLE_INS));
  1381. }
  1382. FAIL_IF(push_inst(compiler, CVT_S_S | flags | (4 << 21) | (((op & SLJIT_F32_OP) ^ SLJIT_F32_OP) >> 8) | FS(TMP_FREG1) | FD(dst_r), MOVABLE_INS));
  1383. if (dst & SLJIT_MEM)
  1384. return emit_op_mem2(compiler, FLOAT_DATA(op), FR(TMP_FREG1), dst, dstw, 0, 0);
  1385. return SLJIT_SUCCESS;
  1386. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1387. # undef flags
  1388. #endif
  1389. }
  1390. static SLJIT_INLINE sljit_s32 sljit_emit_fop1_cmp(struct sljit_compiler *compiler, sljit_s32 op,
  1391. sljit_s32 src1, sljit_sw src1w,
  1392. sljit_s32 src2, sljit_sw src2w)
  1393. {
  1394. sljit_ins inst;
  1395. if (src1 & SLJIT_MEM) {
  1396. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, src2, src2w));
  1397. src1 = TMP_FREG1;
  1398. }
  1399. if (src2 & SLJIT_MEM) {
  1400. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, 0, 0));
  1401. src2 = TMP_FREG2;
  1402. }
  1403. switch (GET_FLAG_TYPE(op)) {
  1404. case SLJIT_EQUAL_F64:
  1405. case SLJIT_NOT_EQUAL_F64:
  1406. inst = C_UEQ_S;
  1407. break;
  1408. case SLJIT_LESS_F64:
  1409. case SLJIT_GREATER_EQUAL_F64:
  1410. inst = C_ULT_S;
  1411. break;
  1412. case SLJIT_GREATER_F64:
  1413. case SLJIT_LESS_EQUAL_F64:
  1414. inst = C_ULE_S;
  1415. break;
  1416. default:
  1417. SLJIT_ASSERT(GET_FLAG_TYPE(op) == SLJIT_UNORDERED_F64 || GET_FLAG_TYPE(op) == SLJIT_ORDERED_F64);
  1418. inst = C_UN_S;
  1419. break;
  1420. }
  1421. return push_inst(compiler, inst | FMT(op) | FT(src2) | FS(src1) | C_FD, UNMOVABLE_INS);
  1422. }
  1423. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop1(struct sljit_compiler *compiler, sljit_s32 op,
  1424. sljit_s32 dst, sljit_sw dstw,
  1425. sljit_s32 src, sljit_sw srcw)
  1426. {
  1427. sljit_s32 dst_r;
  1428. CHECK_ERROR();
  1429. compiler->cache_arg = 0;
  1430. compiler->cache_argw = 0;
  1431. SLJIT_COMPILE_ASSERT((SLJIT_F32_OP == 0x100) && !(DOUBLE_DATA & 0x2), float_transfer_bit_error);
  1432. SELECT_FOP1_OPERATION_WITH_CHECKS(compiler, op, dst, dstw, src, srcw);
  1433. if (GET_OPCODE(op) == SLJIT_CONV_F64_FROM_F32)
  1434. op ^= SLJIT_F32_OP;
  1435. dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG1;
  1436. if (src & SLJIT_MEM) {
  1437. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(dst_r), src, srcw, dst, dstw));
  1438. src = dst_r;
  1439. }
  1440. switch (GET_OPCODE(op)) {
  1441. case SLJIT_MOV_F64:
  1442. if (src != dst_r) {
  1443. if (dst_r != TMP_FREG1)
  1444. FAIL_IF(push_inst(compiler, MOV_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
  1445. else
  1446. dst_r = src;
  1447. }
  1448. break;
  1449. case SLJIT_NEG_F64:
  1450. FAIL_IF(push_inst(compiler, NEG_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
  1451. break;
  1452. case SLJIT_ABS_F64:
  1453. FAIL_IF(push_inst(compiler, ABS_S | FMT(op) | FS(src) | FD(dst_r), MOVABLE_INS));
  1454. break;
  1455. case SLJIT_CONV_F64_FROM_F32:
  1456. FAIL_IF(push_inst(compiler, CVT_S_S | ((op & SLJIT_F32_OP) ? 1 : (1 << 21)) | FS(src) | FD(dst_r), MOVABLE_INS));
  1457. op ^= SLJIT_F32_OP;
  1458. break;
  1459. }
  1460. if (dst & SLJIT_MEM)
  1461. return emit_op_mem2(compiler, FLOAT_DATA(op), FR(dst_r), dst, dstw, 0, 0);
  1462. return SLJIT_SUCCESS;
  1463. }
  1464. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fop2(struct sljit_compiler *compiler, sljit_s32 op,
  1465. sljit_s32 dst, sljit_sw dstw,
  1466. sljit_s32 src1, sljit_sw src1w,
  1467. sljit_s32 src2, sljit_sw src2w)
  1468. {
  1469. sljit_s32 dst_r, flags = 0;
  1470. CHECK_ERROR();
  1471. CHECK(check_sljit_emit_fop2(compiler, op, dst, dstw, src1, src1w, src2, src2w));
  1472. ADJUST_LOCAL_OFFSET(dst, dstw);
  1473. ADJUST_LOCAL_OFFSET(src1, src1w);
  1474. ADJUST_LOCAL_OFFSET(src2, src2w);
  1475. compiler->cache_arg = 0;
  1476. compiler->cache_argw = 0;
  1477. dst_r = FAST_IS_REG(dst) ? dst : TMP_FREG2;
  1478. if (src1 & SLJIT_MEM) {
  1479. if (getput_arg_fast(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w)) {
  1480. FAIL_IF(compiler->error);
  1481. src1 = TMP_FREG1;
  1482. } else
  1483. flags |= SLOW_SRC1;
  1484. }
  1485. if (src2 & SLJIT_MEM) {
  1486. if (getput_arg_fast(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w)) {
  1487. FAIL_IF(compiler->error);
  1488. src2 = TMP_FREG2;
  1489. } else
  1490. flags |= SLOW_SRC2;
  1491. }
  1492. if ((flags & (SLOW_SRC1 | SLOW_SRC2)) == (SLOW_SRC1 | SLOW_SRC2)) {
  1493. if (!can_cache(src1, src1w, src2, src2w) && can_cache(src1, src1w, dst, dstw)) {
  1494. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, src1, src1w));
  1495. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, dst, dstw));
  1496. }
  1497. else {
  1498. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, src2, src2w));
  1499. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, dst, dstw));
  1500. }
  1501. }
  1502. else if (flags & SLOW_SRC1)
  1503. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG1), src1, src1w, dst, dstw));
  1504. else if (flags & SLOW_SRC2)
  1505. FAIL_IF(getput_arg(compiler, FLOAT_DATA(op) | LOAD_DATA, FR(TMP_FREG2), src2, src2w, dst, dstw));
  1506. if (flags & SLOW_SRC1)
  1507. src1 = TMP_FREG1;
  1508. if (flags & SLOW_SRC2)
  1509. src2 = TMP_FREG2;
  1510. switch (GET_OPCODE(op)) {
  1511. case SLJIT_ADD_F64:
  1512. FAIL_IF(push_inst(compiler, ADD_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1513. break;
  1514. case SLJIT_SUB_F64:
  1515. FAIL_IF(push_inst(compiler, SUB_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1516. break;
  1517. case SLJIT_MUL_F64:
  1518. FAIL_IF(push_inst(compiler, MUL_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1519. break;
  1520. case SLJIT_DIV_F64:
  1521. FAIL_IF(push_inst(compiler, DIV_S | FMT(op) | FT(src2) | FS(src1) | FD(dst_r), MOVABLE_INS));
  1522. break;
  1523. }
  1524. if (dst_r == TMP_FREG2)
  1525. FAIL_IF(emit_op_mem2(compiler, FLOAT_DATA(op), FR(TMP_FREG2), dst, dstw, 0, 0));
  1526. return SLJIT_SUCCESS;
  1527. }
  1528. /* --------------------------------------------------------------------- */
  1529. /* Other instructions */
  1530. /* --------------------------------------------------------------------- */
  1531. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_fast_enter(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw)
  1532. {
  1533. CHECK_ERROR();
  1534. CHECK(check_sljit_emit_fast_enter(compiler, dst, dstw));
  1535. ADJUST_LOCAL_OFFSET(dst, dstw);
  1536. if (FAST_IS_REG(dst))
  1537. return push_inst(compiler, ADDU_W | SA(RETURN_ADDR_REG) | TA(0) | D(dst), UNMOVABLE_INS);
  1538. /* Memory. */
  1539. FAIL_IF(emit_op_mem(compiler, WORD_DATA, RETURN_ADDR_REG, dst, dstw));
  1540. compiler->delay_slot = UNMOVABLE_INS;
  1541. return SLJIT_SUCCESS;
  1542. }
  1543. /* --------------------------------------------------------------------- */
  1544. /* Conditional instructions */
  1545. /* --------------------------------------------------------------------- */
  1546. SLJIT_API_FUNC_ATTRIBUTE struct sljit_label* sljit_emit_label(struct sljit_compiler *compiler)
  1547. {
  1548. struct sljit_label *label;
  1549. CHECK_ERROR_PTR();
  1550. CHECK_PTR(check_sljit_emit_label(compiler));
  1551. if (compiler->last_label && compiler->last_label->size == compiler->size)
  1552. return compiler->last_label;
  1553. label = (struct sljit_label*)ensure_abuf(compiler, sizeof(struct sljit_label));
  1554. PTR_FAIL_IF(!label);
  1555. set_label(label, compiler);
  1556. compiler->delay_slot = UNMOVABLE_INS;
  1557. return label;
  1558. }
  1559. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1560. #define JUMP_LENGTH 4
  1561. #else
  1562. #define JUMP_LENGTH 8
  1563. #endif
  1564. #define BR_Z(src) \
  1565. inst = BEQ | SA(src) | TA(0) | JUMP_LENGTH; \
  1566. flags = IS_BIT26_COND; \
  1567. delay_check = src;
  1568. #define BR_NZ(src) \
  1569. inst = BNE | SA(src) | TA(0) | JUMP_LENGTH; \
  1570. flags = IS_BIT26_COND; \
  1571. delay_check = src;
  1572. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  1573. #define BR_T() \
  1574. inst = BC1NEZ; \
  1575. flags = IS_BIT23_COND; \
  1576. delay_check = FCSR_FCC;
  1577. #define BR_F() \
  1578. inst = BC1EQZ; \
  1579. flags = IS_BIT23_COND; \
  1580. delay_check = FCSR_FCC;
  1581. #else /* SLJIT_MIPS_REV < 6 */
  1582. #define BR_T() \
  1583. inst = BC1T | JUMP_LENGTH; \
  1584. flags = IS_BIT16_COND; \
  1585. delay_check = FCSR_FCC;
  1586. #define BR_F() \
  1587. inst = BC1F | JUMP_LENGTH; \
  1588. flags = IS_BIT16_COND; \
  1589. delay_check = FCSR_FCC;
  1590. #endif /* SLJIT_MIPS_REV >= 6 */
  1591. SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_jump(struct sljit_compiler *compiler, sljit_s32 type)
  1592. {
  1593. struct sljit_jump *jump;
  1594. sljit_ins inst;
  1595. sljit_s32 flags = 0;
  1596. sljit_s32 delay_check = UNMOVABLE_INS;
  1597. CHECK_ERROR_PTR();
  1598. CHECK_PTR(check_sljit_emit_jump(compiler, type));
  1599. jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
  1600. PTR_FAIL_IF(!jump);
  1601. set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
  1602. type &= 0xff;
  1603. switch (type) {
  1604. case SLJIT_EQUAL:
  1605. BR_NZ(EQUAL_FLAG);
  1606. break;
  1607. case SLJIT_NOT_EQUAL:
  1608. BR_Z(EQUAL_FLAG);
  1609. break;
  1610. case SLJIT_LESS:
  1611. case SLJIT_GREATER:
  1612. case SLJIT_SIG_LESS:
  1613. case SLJIT_SIG_GREATER:
  1614. case SLJIT_OVERFLOW:
  1615. BR_Z(OTHER_FLAG);
  1616. break;
  1617. case SLJIT_GREATER_EQUAL:
  1618. case SLJIT_LESS_EQUAL:
  1619. case SLJIT_SIG_GREATER_EQUAL:
  1620. case SLJIT_SIG_LESS_EQUAL:
  1621. case SLJIT_NOT_OVERFLOW:
  1622. BR_NZ(OTHER_FLAG);
  1623. break;
  1624. case SLJIT_NOT_EQUAL_F64:
  1625. case SLJIT_GREATER_EQUAL_F64:
  1626. case SLJIT_GREATER_F64:
  1627. case SLJIT_ORDERED_F64:
  1628. BR_T();
  1629. break;
  1630. case SLJIT_EQUAL_F64:
  1631. case SLJIT_LESS_F64:
  1632. case SLJIT_LESS_EQUAL_F64:
  1633. case SLJIT_UNORDERED_F64:
  1634. BR_F();
  1635. break;
  1636. default:
  1637. /* Not conditional branch. */
  1638. inst = 0;
  1639. break;
  1640. }
  1641. jump->flags |= flags;
  1642. if (compiler->delay_slot == MOVABLE_INS || (compiler->delay_slot != UNMOVABLE_INS && compiler->delay_slot != delay_check))
  1643. jump->flags |= IS_MOVABLE;
  1644. if (inst)
  1645. PTR_FAIL_IF(push_inst(compiler, inst, UNMOVABLE_INS));
  1646. PTR_FAIL_IF(emit_const(compiler, TMP_REG2, 0));
  1647. if (type <= SLJIT_JUMP)
  1648. PTR_FAIL_IF(push_inst(compiler, JR | S(TMP_REG2), UNMOVABLE_INS));
  1649. else {
  1650. jump->flags |= IS_JAL;
  1651. PTR_FAIL_IF(push_inst(compiler, JALR | S(TMP_REG2) | DA(RETURN_ADDR_REG), UNMOVABLE_INS));
  1652. }
  1653. jump->addr = compiler->size;
  1654. PTR_FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1655. return jump;
  1656. }
  1657. #define RESOLVE_IMM1() \
  1658. if (src1 & SLJIT_IMM) { \
  1659. if (src1w) { \
  1660. PTR_FAIL_IF(load_immediate(compiler, DR(TMP_REG1), src1w)); \
  1661. src1 = TMP_REG1; \
  1662. } \
  1663. else \
  1664. src1 = 0; \
  1665. }
  1666. #define RESOLVE_IMM2() \
  1667. if (src2 & SLJIT_IMM) { \
  1668. if (src2w) { \
  1669. PTR_FAIL_IF(load_immediate(compiler, DR(TMP_REG2), src2w)); \
  1670. src2 = TMP_REG2; \
  1671. } \
  1672. else \
  1673. src2 = 0; \
  1674. }
  1675. SLJIT_API_FUNC_ATTRIBUTE struct sljit_jump* sljit_emit_cmp(struct sljit_compiler *compiler, sljit_s32 type,
  1676. sljit_s32 src1, sljit_sw src1w,
  1677. sljit_s32 src2, sljit_sw src2w)
  1678. {
  1679. struct sljit_jump *jump;
  1680. sljit_s32 flags;
  1681. sljit_ins inst;
  1682. CHECK_ERROR_PTR();
  1683. CHECK_PTR(check_sljit_emit_cmp(compiler, type, src1, src1w, src2, src2w));
  1684. ADJUST_LOCAL_OFFSET(src1, src1w);
  1685. ADJUST_LOCAL_OFFSET(src2, src2w);
  1686. compiler->cache_arg = 0;
  1687. compiler->cache_argw = 0;
  1688. flags = ((type & SLJIT_I32_OP) ? INT_DATA : WORD_DATA) | LOAD_DATA;
  1689. if (src1 & SLJIT_MEM) {
  1690. PTR_FAIL_IF(emit_op_mem2(compiler, flags, DR(TMP_REG1), src1, src1w, src2, src2w));
  1691. src1 = TMP_REG1;
  1692. }
  1693. if (src2 & SLJIT_MEM) {
  1694. PTR_FAIL_IF(emit_op_mem2(compiler, flags, DR(TMP_REG2), src2, src2w, 0, 0));
  1695. src2 = TMP_REG2;
  1696. }
  1697. jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
  1698. PTR_FAIL_IF(!jump);
  1699. set_jump(jump, compiler, type & SLJIT_REWRITABLE_JUMP);
  1700. type &= 0xff;
  1701. if (type <= SLJIT_NOT_EQUAL) {
  1702. RESOLVE_IMM1();
  1703. RESOLVE_IMM2();
  1704. jump->flags |= IS_BIT26_COND;
  1705. if (compiler->delay_slot == MOVABLE_INS || (compiler->delay_slot != UNMOVABLE_INS && compiler->delay_slot != DR(src1) && compiler->delay_slot != DR(src2)))
  1706. jump->flags |= IS_MOVABLE;
  1707. PTR_FAIL_IF(push_inst(compiler, (type == SLJIT_EQUAL ? BNE : BEQ) | S(src1) | T(src2) | JUMP_LENGTH, UNMOVABLE_INS));
  1708. }
  1709. else if (type >= SLJIT_SIG_LESS && (((src1 & SLJIT_IMM) && (src1w == 0)) || ((src2 & SLJIT_IMM) && (src2w == 0)))) {
  1710. inst = NOP;
  1711. if ((src1 & SLJIT_IMM) && (src1w == 0)) {
  1712. RESOLVE_IMM2();
  1713. switch (type) {
  1714. case SLJIT_SIG_LESS:
  1715. inst = BLEZ;
  1716. jump->flags |= IS_BIT26_COND;
  1717. break;
  1718. case SLJIT_SIG_GREATER_EQUAL:
  1719. inst = BGTZ;
  1720. jump->flags |= IS_BIT26_COND;
  1721. break;
  1722. case SLJIT_SIG_GREATER:
  1723. inst = BGEZ;
  1724. jump->flags |= IS_BIT16_COND;
  1725. break;
  1726. case SLJIT_SIG_LESS_EQUAL:
  1727. inst = BLTZ;
  1728. jump->flags |= IS_BIT16_COND;
  1729. break;
  1730. }
  1731. src1 = src2;
  1732. }
  1733. else {
  1734. RESOLVE_IMM1();
  1735. switch (type) {
  1736. case SLJIT_SIG_LESS:
  1737. inst = BGEZ;
  1738. jump->flags |= IS_BIT16_COND;
  1739. break;
  1740. case SLJIT_SIG_GREATER_EQUAL:
  1741. inst = BLTZ;
  1742. jump->flags |= IS_BIT16_COND;
  1743. break;
  1744. case SLJIT_SIG_GREATER:
  1745. inst = BLEZ;
  1746. jump->flags |= IS_BIT26_COND;
  1747. break;
  1748. case SLJIT_SIG_LESS_EQUAL:
  1749. inst = BGTZ;
  1750. jump->flags |= IS_BIT26_COND;
  1751. break;
  1752. }
  1753. }
  1754. PTR_FAIL_IF(push_inst(compiler, inst | S(src1) | JUMP_LENGTH, UNMOVABLE_INS));
  1755. }
  1756. else {
  1757. if (type == SLJIT_LESS || type == SLJIT_GREATER_EQUAL || type == SLJIT_SIG_LESS || type == SLJIT_SIG_GREATER_EQUAL) {
  1758. RESOLVE_IMM1();
  1759. if ((src2 & SLJIT_IMM) && src2w <= SIMM_MAX && src2w >= SIMM_MIN)
  1760. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTIU : SLTI) | S(src1) | T(TMP_REG1) | IMM(src2w), DR(TMP_REG1)));
  1761. else {
  1762. RESOLVE_IMM2();
  1763. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTU : SLT) | S(src1) | T(src2) | D(TMP_REG1), DR(TMP_REG1)));
  1764. }
  1765. type = (type == SLJIT_LESS || type == SLJIT_SIG_LESS) ? SLJIT_NOT_EQUAL : SLJIT_EQUAL;
  1766. }
  1767. else {
  1768. RESOLVE_IMM2();
  1769. if ((src1 & SLJIT_IMM) && src1w <= SIMM_MAX && src1w >= SIMM_MIN)
  1770. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTIU : SLTI) | S(src2) | T(TMP_REG1) | IMM(src1w), DR(TMP_REG1)));
  1771. else {
  1772. RESOLVE_IMM1();
  1773. PTR_FAIL_IF(push_inst(compiler, (type <= SLJIT_LESS_EQUAL ? SLTU : SLT) | S(src2) | T(src1) | D(TMP_REG1), DR(TMP_REG1)));
  1774. }
  1775. type = (type == SLJIT_GREATER || type == SLJIT_SIG_GREATER) ? SLJIT_NOT_EQUAL : SLJIT_EQUAL;
  1776. }
  1777. jump->flags |= IS_BIT26_COND;
  1778. PTR_FAIL_IF(push_inst(compiler, (type == SLJIT_EQUAL ? BNE : BEQ) | S(TMP_REG1) | TA(0) | JUMP_LENGTH, UNMOVABLE_INS));
  1779. }
  1780. PTR_FAIL_IF(emit_const(compiler, TMP_REG2, 0));
  1781. PTR_FAIL_IF(push_inst(compiler, JR | S(TMP_REG2), UNMOVABLE_INS));
  1782. jump->addr = compiler->size;
  1783. PTR_FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1784. return jump;
  1785. }
  1786. #undef RESOLVE_IMM1
  1787. #undef RESOLVE_IMM2
  1788. #undef JUMP_LENGTH
  1789. #undef BR_Z
  1790. #undef BR_NZ
  1791. #undef BR_T
  1792. #undef BR_F
  1793. #undef FLOAT_DATA
  1794. #undef FMT
  1795. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_ijump(struct sljit_compiler *compiler, sljit_s32 type, sljit_s32 src, sljit_sw srcw)
  1796. {
  1797. struct sljit_jump *jump = NULL;
  1798. CHECK_ERROR();
  1799. CHECK(check_sljit_emit_ijump(compiler, type, src, srcw));
  1800. ADJUST_LOCAL_OFFSET(src, srcw);
  1801. if (src & SLJIT_IMM) {
  1802. jump = (struct sljit_jump*)ensure_abuf(compiler, sizeof(struct sljit_jump));
  1803. FAIL_IF(!jump);
  1804. set_jump(jump, compiler, JUMP_ADDR | ((type >= SLJIT_FAST_CALL) ? IS_JAL : 0));
  1805. jump->u.target = srcw;
  1806. if (compiler->delay_slot != UNMOVABLE_INS)
  1807. jump->flags |= IS_MOVABLE;
  1808. FAIL_IF(emit_const(compiler, TMP_REG2, 0));
  1809. src = TMP_REG2;
  1810. }
  1811. else if (src & SLJIT_MEM) {
  1812. FAIL_IF(emit_op_mem(compiler, WORD_DATA | LOAD_DATA, DR(TMP_REG2), src, srcw));
  1813. src = TMP_REG2;
  1814. }
  1815. FAIL_IF(push_inst(compiler, JR | S(src), UNMOVABLE_INS));
  1816. if (jump)
  1817. jump->addr = compiler->size;
  1818. FAIL_IF(push_inst(compiler, NOP, UNMOVABLE_INS));
  1819. return SLJIT_SUCCESS;
  1820. }
  1821. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_op_flags(struct sljit_compiler *compiler, sljit_s32 op,
  1822. sljit_s32 dst, sljit_sw dstw,
  1823. sljit_s32 type)
  1824. {
  1825. sljit_s32 src_ar, dst_ar;
  1826. sljit_s32 saved_op = op;
  1827. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1828. sljit_s32 mem_type = WORD_DATA;
  1829. #else
  1830. sljit_s32 mem_type = (op & SLJIT_I32_OP) ? (INT_DATA | SIGNED_DATA) : WORD_DATA;
  1831. #endif
  1832. CHECK_ERROR();
  1833. CHECK(check_sljit_emit_op_flags(compiler, op, dst, dstw, type));
  1834. ADJUST_LOCAL_OFFSET(dst, dstw);
  1835. op = GET_OPCODE(op);
  1836. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1837. if (op == SLJIT_MOV_S32)
  1838. mem_type = INT_DATA | SIGNED_DATA;
  1839. #endif
  1840. dst_ar = DR((op < SLJIT_ADD && FAST_IS_REG(dst)) ? dst : TMP_REG2);
  1841. compiler->cache_arg = 0;
  1842. compiler->cache_argw = 0;
  1843. if (op >= SLJIT_ADD && (dst & SLJIT_MEM))
  1844. FAIL_IF(emit_op_mem2(compiler, mem_type | LOAD_DATA, DR(TMP_REG1), dst, dstw, dst, dstw));
  1845. switch (type & 0xff) {
  1846. case SLJIT_EQUAL:
  1847. case SLJIT_NOT_EQUAL:
  1848. FAIL_IF(push_inst(compiler, SLTIU | SA(EQUAL_FLAG) | TA(dst_ar) | IMM(1), dst_ar));
  1849. src_ar = dst_ar;
  1850. break;
  1851. case SLJIT_OVERFLOW:
  1852. case SLJIT_NOT_OVERFLOW:
  1853. if (compiler->status_flags_state & SLJIT_CURRENT_FLAGS_ADD_SUB) {
  1854. src_ar = OTHER_FLAG;
  1855. break;
  1856. }
  1857. FAIL_IF(push_inst(compiler, SLTIU | SA(OTHER_FLAG) | TA(dst_ar) | IMM(1), dst_ar));
  1858. src_ar = dst_ar;
  1859. type ^= 0x1; /* Flip type bit for the XORI below. */
  1860. break;
  1861. case SLJIT_GREATER_F64:
  1862. case SLJIT_LESS_EQUAL_F64:
  1863. type ^= 0x1; /* Flip type bit for the XORI below. */
  1864. case SLJIT_EQUAL_F64:
  1865. case SLJIT_NOT_EQUAL_F64:
  1866. case SLJIT_LESS_F64:
  1867. case SLJIT_GREATER_EQUAL_F64:
  1868. case SLJIT_UNORDERED_F64:
  1869. case SLJIT_ORDERED_F64:
  1870. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 6)
  1871. FAIL_IF(push_inst(compiler, MFC1 | TA(dst_ar) | FS(TMP_FREG3), dst_ar));
  1872. #else /* SLJIT_MIPS_REV < 6 */
  1873. FAIL_IF(push_inst(compiler, CFC1 | TA(dst_ar) | DA(FCSR_REG), dst_ar));
  1874. #endif /* SLJIT_MIPS_REV >= 6 */
  1875. FAIL_IF(push_inst(compiler, SRL | TA(dst_ar) | DA(dst_ar) | SH_IMM(23), dst_ar));
  1876. FAIL_IF(push_inst(compiler, ANDI | SA(dst_ar) | TA(dst_ar) | IMM(1), dst_ar));
  1877. src_ar = dst_ar;
  1878. break;
  1879. default:
  1880. src_ar = OTHER_FLAG;
  1881. break;
  1882. }
  1883. if (type & 0x1) {
  1884. FAIL_IF(push_inst(compiler, XORI | SA(src_ar) | TA(dst_ar) | IMM(1), dst_ar));
  1885. src_ar = dst_ar;
  1886. }
  1887. if (op < SLJIT_ADD) {
  1888. if (dst & SLJIT_MEM)
  1889. return emit_op_mem(compiler, mem_type, src_ar, dst, dstw);
  1890. if (src_ar != dst_ar)
  1891. return push_inst(compiler, ADDU_W | SA(src_ar) | TA(0) | DA(dst_ar), dst_ar);
  1892. return SLJIT_SUCCESS;
  1893. }
  1894. /* OTHER_FLAG cannot be specified as src2 argument at the moment. */
  1895. if (DR(TMP_REG2) != src_ar)
  1896. FAIL_IF(push_inst(compiler, ADDU_W | SA(src_ar) | TA(0) | D(TMP_REG2), DR(TMP_REG2)));
  1897. mem_type |= CUMULATIVE_OP | LOGICAL_OP | IMM_OP | ALT_KEEP_CACHE;
  1898. if (dst & SLJIT_MEM)
  1899. return emit_op(compiler, saved_op, mem_type, dst, dstw, TMP_REG1, 0, TMP_REG2, 0);
  1900. return emit_op(compiler, saved_op, mem_type, dst, dstw, dst, dstw, TMP_REG2, 0);
  1901. }
  1902. SLJIT_API_FUNC_ATTRIBUTE sljit_s32 sljit_emit_cmov(struct sljit_compiler *compiler, sljit_s32 type,
  1903. sljit_s32 dst_reg,
  1904. sljit_s32 src, sljit_sw srcw)
  1905. {
  1906. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1 && SLJIT_MIPS_REV < 6)
  1907. sljit_ins ins;
  1908. #endif /* SLJIT_MIPS_REV >= 1 && SLJIT_MIPS_REV < 6 */
  1909. CHECK_ERROR();
  1910. CHECK(check_sljit_emit_cmov(compiler, type, dst_reg, src, srcw));
  1911. #if (defined SLJIT_MIPS_REV && SLJIT_MIPS_REV >= 1 && SLJIT_MIPS_REV < 6)
  1912. if (SLJIT_UNLIKELY(src & SLJIT_IMM)) {
  1913. #if (defined SLJIT_CONFIG_MIPS_64 && SLJIT_CONFIG_MIPS_64)
  1914. if (dst_reg & SLJIT_I32_OP)
  1915. srcw = (sljit_s32)srcw;
  1916. #endif
  1917. FAIL_IF(load_immediate(compiler, DR(TMP_REG1), srcw));
  1918. src = TMP_REG1;
  1919. srcw = 0;
  1920. }
  1921. dst_reg &= ~SLJIT_I32_OP;
  1922. switch (type & 0xff) {
  1923. case SLJIT_EQUAL:
  1924. ins = MOVZ | TA(EQUAL_FLAG);
  1925. break;
  1926. case SLJIT_NOT_EQUAL:
  1927. ins = MOVN | TA(EQUAL_FLAG);
  1928. break;
  1929. case SLJIT_LESS:
  1930. case SLJIT_GREATER:
  1931. case SLJIT_SIG_LESS:
  1932. case SLJIT_SIG_GREATER:
  1933. case SLJIT_OVERFLOW:
  1934. ins = MOVN | TA(OTHER_FLAG);
  1935. break;
  1936. case SLJIT_GREATER_EQUAL:
  1937. case SLJIT_LESS_EQUAL:
  1938. case SLJIT_SIG_GREATER_EQUAL:
  1939. case SLJIT_SIG_LESS_EQUAL:
  1940. case SLJIT_NOT_OVERFLOW:
  1941. ins = MOVZ | TA(OTHER_FLAG);
  1942. break;
  1943. case SLJIT_EQUAL_F64:
  1944. case SLJIT_LESS_F64:
  1945. case SLJIT_LESS_EQUAL_F64:
  1946. case SLJIT_UNORDERED_F64:
  1947. ins = MOVT;
  1948. break;
  1949. case SLJIT_NOT_EQUAL_F64:
  1950. case SLJIT_GREATER_EQUAL_F64:
  1951. case SLJIT_GREATER_F64:
  1952. case SLJIT_ORDERED_F64:
  1953. ins = MOVF;
  1954. break;
  1955. default:
  1956. ins = MOVZ | TA(OTHER_FLAG);
  1957. SLJIT_UNREACHABLE();
  1958. break;
  1959. }
  1960. return push_inst(compiler, ins | S(src) | D(dst_reg), DR(dst_reg));
  1961. #else /* SLJIT_MIPS_REV < 1 || SLJIT_MIPS_REV >= 6 */
  1962. return sljit_emit_cmov_generic(compiler, type, dst_reg, src, srcw);
  1963. #endif /* SLJIT_MIPS_REV >= 1 */
  1964. }
  1965. SLJIT_API_FUNC_ATTRIBUTE struct sljit_const* sljit_emit_const(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw, sljit_sw init_value)
  1966. {
  1967. struct sljit_const *const_;
  1968. sljit_s32 dst_r;
  1969. CHECK_ERROR_PTR();
  1970. CHECK_PTR(check_sljit_emit_const(compiler, dst, dstw, init_value));
  1971. ADJUST_LOCAL_OFFSET(dst, dstw);
  1972. const_ = (struct sljit_const*)ensure_abuf(compiler, sizeof(struct sljit_const));
  1973. PTR_FAIL_IF(!const_);
  1974. set_const(const_, compiler);
  1975. dst_r = FAST_IS_REG(dst) ? dst : TMP_REG2;
  1976. PTR_FAIL_IF(emit_const(compiler, dst_r, init_value));
  1977. if (dst & SLJIT_MEM)
  1978. PTR_FAIL_IF(emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, TMP_REG2, 0));
  1979. return const_;
  1980. }
  1981. SLJIT_API_FUNC_ATTRIBUTE struct sljit_put_label* sljit_emit_put_label(struct sljit_compiler *compiler, sljit_s32 dst, sljit_sw dstw)
  1982. {
  1983. struct sljit_put_label *put_label;
  1984. sljit_s32 dst_r;
  1985. CHECK_ERROR_PTR();
  1986. CHECK_PTR(check_sljit_emit_put_label(compiler, dst, dstw));
  1987. ADJUST_LOCAL_OFFSET(dst, dstw);
  1988. put_label = (struct sljit_put_label*)ensure_abuf(compiler, sizeof(struct sljit_put_label));
  1989. PTR_FAIL_IF(!put_label);
  1990. set_put_label(put_label, compiler, 0);
  1991. dst_r = FAST_IS_REG(dst) ? dst : TMP_REG2;
  1992. #if (defined SLJIT_CONFIG_MIPS_32 && SLJIT_CONFIG_MIPS_32)
  1993. PTR_FAIL_IF(emit_const(compiler, dst_r, 0));
  1994. #else
  1995. PTR_FAIL_IF(push_inst(compiler, dst_r, UNMOVABLE_INS));
  1996. compiler->size += 5;
  1997. #endif
  1998. if (dst & SLJIT_MEM)
  1999. PTR_FAIL_IF(emit_op(compiler, SLJIT_MOV, WORD_DATA, dst, dstw, TMP_REG1, 0, TMP_REG2, 0));
  2000. return put_label;
  2001. }