Tarantool development patches archive
 help / color / mirror / Atom feed
From: Sergey Bronnikov via Tarantool-patches <tarantool-patches@dev.tarantool.org>
To: Sergey Kaplun <skaplun@tarantool.org>, Igor Munkin <imun@tarantool.org>
Cc: tarantool-patches@dev.tarantool.org
Subject: Re: [Tarantool-patches] [PATCH luajit 06/19] PPC: Add soft-float support to JIT compiler backend.
Date: Thu, 17 Aug 2023 17:33:30 +0300	[thread overview]
Message-ID: <da27ca67-822d-1c9a-1120-811400d53b27@tarantool.org> (raw)
In-Reply-To: <cfd1879d0b52c929e83ae060e462b5ca20c8f415.1691592488.git.skaplun@tarantool.org>

Hi, Sergey!


thanks for the patch! LGTM

On 8/9/23 18:35, Sergey Kaplun wrote:
> From: Mike Pall <mike>
>
> Contributed by Djordje Kovacevic and Stefan Pejic from RT-RK.com.
> Sponsored by Cisco Systems, Inc.
>
> (cherry-picked from commit 71b7bc88341945f13f3951e2bb5fd247b639ff7a)
>
> The software floating point library is used on machines which do not
> have hardware support for floating point [1]. This patch enables
> support for such machines in the JIT compiler for powerpc.
> This includes:
> * All fp-depending paths are instrumented with `LJ_SOFTFP` macro.
> * `asm_sfpmin_max()` is introduced for min/max operations on soft-float
>    point.
> * `asm_sfpcomp()` is introduced for soft-float point comparisons.
>
> [1]: https://gcc.gnu.org/onlinedocs/gccint/Soft-float-library-routines.html
>
> Sergey Kaplun:
> * added the description for the feature
>
> Part of tarantool/tarantool#8825
> ---
>   src/lj_arch.h    |   1 -
>   src/lj_asm_ppc.h | 321 ++++++++++++++++++++++++++++++++++++++++-------
>   2 files changed, 278 insertions(+), 44 deletions(-)
>
> diff --git a/src/lj_arch.h b/src/lj_arch.h
> index 8bb8757d..7397492e 100644
> --- a/src/lj_arch.h
> +++ b/src/lj_arch.h
> @@ -281,7 +281,6 @@
>   #endif
>   
>   #if LJ_ABI_SOFTFP
> -#define LJ_ARCH_NOJIT		1  /* NYI */
>   #define LJ_ARCH_NUMMODE		LJ_NUMMODE_DUAL
>   #else
>   #define LJ_ARCH_NUMMODE		LJ_NUMMODE_DUAL_SINGLE
> diff --git a/src/lj_asm_ppc.h b/src/lj_asm_ppc.h
> index aa2d45c0..6cb608f7 100644
> --- a/src/lj_asm_ppc.h
> +++ b/src/lj_asm_ppc.h
> @@ -226,6 +226,7 @@ static void asm_fusexrefx(ASMState *as, PPCIns pi, Reg rt, IRRef ref,
>     emit_tab(as, pi, rt, left, right);
>   }
>   
> +#if !LJ_SOFTFP
>   /* Fuse to multiply-add/sub instruction. */
>   static int asm_fusemadd(ASMState *as, IRIns *ir, PPCIns pi, PPCIns pir)
>   {
> @@ -245,6 +246,7 @@ static int asm_fusemadd(ASMState *as, IRIns *ir, PPCIns pi, PPCIns pir)
>     }
>     return 0;
>   }
> +#endif
>   
>   /* -- Calls --------------------------------------------------------------- */
>   
> @@ -253,13 +255,17 @@ static void asm_gencall(ASMState *as, const CCallInfo *ci, IRRef *args)
>   {
>     uint32_t n, nargs = CCI_XNARGS(ci);
>     int32_t ofs = 8;
> -  Reg gpr = REGARG_FIRSTGPR, fpr = REGARG_FIRSTFPR;
> +  Reg gpr = REGARG_FIRSTGPR;
> +#if !LJ_SOFTFP
> +  Reg fpr = REGARG_FIRSTFPR;
> +#endif
>     if ((void *)ci->func)
>       emit_call(as, (void *)ci->func);
>     for (n = 0; n < nargs; n++) {  /* Setup args. */
>       IRRef ref = args[n];
>       if (ref) {
>         IRIns *ir = IR(ref);
> +#if !LJ_SOFTFP
>         if (irt_isfp(ir->t)) {
>   	if (fpr <= REGARG_LASTFPR) {
>   	  lua_assert(rset_test(as->freeset, fpr));  /* Already evicted. */
> @@ -271,7 +277,9 @@ static void asm_gencall(ASMState *as, const CCallInfo *ci, IRRef *args)
>   	  emit_spstore(as, ir, r, ofs);
>   	  ofs += irt_isnum(ir->t) ? 8 : 4;
>   	}
> -      } else {
> +      } else
> +#endif
> +      {
>   	if (gpr <= REGARG_LASTGPR) {
>   	  lua_assert(rset_test(as->freeset, gpr));  /* Already evicted. */
>   	  ra_leftov(as, gpr, ref);
> @@ -290,8 +298,10 @@ static void asm_gencall(ASMState *as, const CCallInfo *ci, IRRef *args)
>       }
>       checkmclim(as);
>     }
> +#if !LJ_SOFTFP
>     if ((ci->flags & CCI_VARARG))  /* Vararg calls need to know about FPR use. */
>       emit_tab(as, fpr == REGARG_FIRSTFPR ? PPCI_CRXOR : PPCI_CREQV, 6, 6, 6);
> +#endif
>   }
>   
>   /* Setup result reg/sp for call. Evict scratch regs. */
> @@ -299,8 +309,10 @@ static void asm_setupresult(ASMState *as, IRIns *ir, const CCallInfo *ci)
>   {
>     RegSet drop = RSET_SCRATCH;
>     int hiop = ((ir+1)->o == IR_HIOP && !irt_isnil((ir+1)->t));
> +#if !LJ_SOFTFP
>     if ((ci->flags & CCI_NOFPRCLOBBER))
>       drop &= ~RSET_FPR;
> +#endif
>     if (ra_hasreg(ir->r))
>       rset_clear(drop, ir->r);  /* Dest reg handled below. */
>     if (hiop && ra_hasreg((ir+1)->r))
> @@ -308,7 +320,7 @@ static void asm_setupresult(ASMState *as, IRIns *ir, const CCallInfo *ci)
>     ra_evictset(as, drop);  /* Evictions must be performed first. */
>     if (ra_used(ir)) {
>       lua_assert(!irt_ispri(ir->t));
> -    if (irt_isfp(ir->t)) {
> +    if (!LJ_SOFTFP && irt_isfp(ir->t)) {
>         if ((ci->flags & CCI_CASTU64)) {
>   	/* Use spill slot or temp slots. */
>   	int32_t ofs = ir->s ? sps_scale(ir->s) : SPOFS_TMP;
> @@ -377,6 +389,7 @@ static void asm_retf(ASMState *as, IRIns *ir)
>   
>   /* -- Type conversions ---------------------------------------------------- */
>   
> +#if !LJ_SOFTFP
>   static void asm_tointg(ASMState *as, IRIns *ir, Reg left)
>   {
>     RegSet allow = RSET_FPR;
> @@ -409,15 +422,23 @@ static void asm_tobit(ASMState *as, IRIns *ir)
>     emit_fai(as, PPCI_STFD, tmp, RID_SP, SPOFS_TMP);
>     emit_fab(as, PPCI_FADD, tmp, left, right);
>   }
> +#endif
>   
>   static void asm_conv(ASMState *as, IRIns *ir)
>   {
>     IRType st = (IRType)(ir->op2 & IRCONV_SRCMASK);
> +#if !LJ_SOFTFP
>     int stfp = (st == IRT_NUM || st == IRT_FLOAT);
> +#endif
>     IRRef lref = ir->op1;
> -  lua_assert(irt_type(ir->t) != st);
>     lua_assert(!(irt_isint64(ir->t) ||
>   	       (st == IRT_I64 || st == IRT_U64))); /* Handled by SPLIT. */
> +#if LJ_SOFTFP
> +  /* FP conversions are handled by SPLIT. */
> +  lua_assert(!irt_isfp(ir->t) && !(st == IRT_NUM || st == IRT_FLOAT));
> +  /* Can't check for same types: SPLIT uses CONV int.int + BXOR for sfp NEG. */
> +#else
> +  lua_assert(irt_type(ir->t) != st);
>     if (irt_isfp(ir->t)) {
>       Reg dest = ra_dest(as, ir, RSET_FPR);
>       if (stfp) {  /* FP to FP conversion. */
> @@ -476,7 +497,9 @@ static void asm_conv(ASMState *as, IRIns *ir)
>   	emit_fb(as, PPCI_FCTIWZ, tmp, left);
>         }
>       }
> -  } else {
> +  } else
> +#endif
> +  {
>       Reg dest = ra_dest(as, ir, RSET_GPR);
>       if (st >= IRT_I8 && st <= IRT_U16) {  /* Extend to 32 bit integer. */
>         Reg left = ra_alloc1(as, ir->op1, RSET_GPR);
> @@ -496,17 +519,41 @@ static void asm_strto(ASMState *as, IRIns *ir)
>   {
>     const CCallInfo *ci = &lj_ir_callinfo[IRCALL_lj_strscan_num];
>     IRRef args[2];
> -  int32_t ofs;
> +  int32_t ofs = SPOFS_TMP;
> +#if LJ_SOFTFP
> +  ra_evictset(as, RSET_SCRATCH);
> +  if (ra_used(ir)) {
> +    if (ra_hasspill(ir->s) && ra_hasspill((ir+1)->s) &&
> +	(ir->s & 1) == LJ_BE && (ir->s ^ 1) == (ir+1)->s) {
> +      int i;
> +      for (i = 0; i < 2; i++) {
> +	Reg r = (ir+i)->r;
> +	if (ra_hasreg(r)) {
> +	  ra_free(as, r);
> +	  ra_modified(as, r);
> +	  emit_spload(as, ir+i, r, sps_scale((ir+i)->s));
> +	}
> +      }
> +      ofs = sps_scale(ir->s & ~1);
> +    } else {
> +      Reg rhi = ra_dest(as, ir+1, RSET_GPR);
> +      Reg rlo = ra_dest(as, ir, rset_exclude(RSET_GPR, rhi));
> +      emit_tai(as, PPCI_LWZ, rhi, RID_SP, ofs);
> +      emit_tai(as, PPCI_LWZ, rlo, RID_SP, ofs+4);
> +    }
> +  }
> +#else
>     RegSet drop = RSET_SCRATCH;
>     if (ra_hasreg(ir->r)) rset_set(drop, ir->r);  /* Spill dest reg (if any). */
>     ra_evictset(as, drop);
> +  if (ir->s) ofs = sps_scale(ir->s);
> +#endif
>     asm_guardcc(as, CC_EQ);
>     emit_ai(as, PPCI_CMPWI, RID_RET, 0);  /* Test return status. */
>     args[0] = ir->op1;      /* GCstr *str */
>     args[1] = ASMREF_TMP1;  /* TValue *n  */
>     asm_gencall(as, ci, args);
>     /* Store the result to the spill slot or temp slots. */
> -  ofs = ir->s ? sps_scale(ir->s) : SPOFS_TMP;
>     emit_tai(as, PPCI_ADDI, ra_releasetmp(as, ASMREF_TMP1), RID_SP, ofs);
>   }
>   
> @@ -530,7 +577,10 @@ static void asm_tvptr(ASMState *as, Reg dest, IRRef ref)
>         Reg src = ra_alloc1(as, ref, allow);
>         emit_setgl(as, src, tmptv.gcr);
>       }
> -    type = ra_allock(as, irt_toitype(ir->t), allow);
> +    if (LJ_SOFTFP && (ir+1)->o == IR_HIOP)
> +      type = ra_alloc1(as, ref+1, allow);
> +    else
> +      type = ra_allock(as, irt_toitype(ir->t), allow);
>       emit_setgl(as, type, tmptv.it);
>     }
>   }
> @@ -574,11 +624,27 @@ static void asm_href(ASMState *as, IRIns *ir, IROp merge)
>     Reg tisnum = RID_NONE, tmpnum = RID_NONE;
>     IRRef refkey = ir->op2;
>     IRIns *irkey = IR(refkey);
> +  int isk = irref_isk(refkey);
>     IRType1 kt = irkey->t;
>     uint32_t khash;
>     MCLabel l_end, l_loop, l_next;
>   
>     rset_clear(allow, tab);
> +#if LJ_SOFTFP
> +  if (!isk) {
> +    key = ra_alloc1(as, refkey, allow);
> +    rset_clear(allow, key);
> +    if (irkey[1].o == IR_HIOP) {
> +      if (ra_hasreg((irkey+1)->r)) {
> +	tmpnum = (irkey+1)->r;
> +	ra_noweak(as, tmpnum);
> +      } else {
> +	tmpnum = ra_allocref(as, refkey+1, allow);
> +      }
> +      rset_clear(allow, tmpnum);
> +    }
> +  }
> +#else
>     if (irt_isnum(kt)) {
>       key = ra_alloc1(as, refkey, RSET_FPR);
>       tmpnum = ra_scratch(as, rset_exclude(RSET_FPR, key));
> @@ -588,6 +654,7 @@ static void asm_href(ASMState *as, IRIns *ir, IROp merge)
>       key = ra_alloc1(as, refkey, allow);
>       rset_clear(allow, key);
>     }
> +#endif
>     tmp2 = ra_scratch(as, allow);
>     rset_clear(allow, tmp2);
>   
> @@ -610,7 +677,7 @@ static void asm_href(ASMState *as, IRIns *ir, IROp merge)
>       asm_guardcc(as, CC_EQ);
>     else
>       emit_condbranch(as, PPCI_BC|PPCF_Y, CC_EQ, l_end);
> -  if (irt_isnum(kt)) {
> +  if (!LJ_SOFTFP && irt_isnum(kt)) {
>       emit_fab(as, PPCI_FCMPU, 0, tmpnum, key);
>       emit_condbranch(as, PPCI_BC, CC_GE, l_next);
>       emit_ab(as, PPCI_CMPLW, tmp1, tisnum);
> @@ -620,7 +687,10 @@ static void asm_href(ASMState *as, IRIns *ir, IROp merge)
>         emit_ab(as, PPCI_CMPW, tmp2, key);
>         emit_condbranch(as, PPCI_BC, CC_NE, l_next);
>       }
> -    emit_ai(as, PPCI_CMPWI, tmp1, irt_toitype(irkey->t));
> +    if (LJ_SOFTFP && ra_hasreg(tmpnum))
> +      emit_ab(as, PPCI_CMPW, tmp1, tmpnum);
> +    else
> +      emit_ai(as, PPCI_CMPWI, tmp1, irt_toitype(irkey->t));
>       if (!irt_ispri(kt))
>         emit_tai(as, PPCI_LWZ, tmp2, dest, (int32_t)offsetof(Node, key.gcr));
>     }
> @@ -629,19 +699,19 @@ static void asm_href(ASMState *as, IRIns *ir, IROp merge)
>   	    (((char *)as->mcp-(char *)l_loop) & 0xffffu);
>   
>     /* Load main position relative to tab->node into dest. */
> -  khash = irref_isk(refkey) ? ir_khash(irkey) : 1;
> +  khash = isk ? ir_khash(irkey) : 1;
>     if (khash == 0) {
>       emit_tai(as, PPCI_LWZ, dest, tab, (int32_t)offsetof(GCtab, node));
>     } else {
>       Reg tmphash = tmp1;
> -    if (irref_isk(refkey))
> +    if (isk)
>         tmphash = ra_allock(as, khash, allow);
>       emit_tab(as, PPCI_ADD, dest, dest, tmp1);
>       emit_tai(as, PPCI_MULLI, tmp1, tmp1, sizeof(Node));
>       emit_asb(as, PPCI_AND, tmp1, tmp2, tmphash);
>       emit_tai(as, PPCI_LWZ, dest, tab, (int32_t)offsetof(GCtab, node));
>       emit_tai(as, PPCI_LWZ, tmp2, tab, (int32_t)offsetof(GCtab, hmask));
> -    if (irref_isk(refkey)) {
> +    if (isk) {
>         /* Nothing to do. */
>       } else if (irt_isstr(kt)) {
>         emit_tai(as, PPCI_LWZ, tmp1, key, (int32_t)offsetof(GCstr, hash));
> @@ -651,13 +721,19 @@ static void asm_href(ASMState *as, IRIns *ir, IROp merge)
>         emit_asb(as, PPCI_XOR, tmp1, tmp1, tmp2);
>         emit_rotlwi(as, tmp1, tmp1, (HASH_ROT2+HASH_ROT1)&31);
>         emit_tab(as, PPCI_SUBF, tmp2, dest, tmp2);
> -      if (irt_isnum(kt)) {
> +      if (LJ_SOFTFP ? (irkey[1].o == IR_HIOP) : irt_isnum(kt)) {
> +#if LJ_SOFTFP
> +	emit_asb(as, PPCI_XOR, tmp2, key, tmp1);
> +	emit_rotlwi(as, dest, tmp1, HASH_ROT1);
> +	emit_tab(as, PPCI_ADD, tmp1, tmpnum, tmpnum);
> +#else
>   	int32_t ofs = ra_spill(as, irkey);
>   	emit_asb(as, PPCI_XOR, tmp2, tmp2, tmp1);
>   	emit_rotlwi(as, dest, tmp1, HASH_ROT1);
>   	emit_tab(as, PPCI_ADD, tmp1, tmp1, tmp1);
>   	emit_tai(as, PPCI_LWZ, tmp2, RID_SP, ofs+4);
>   	emit_tai(as, PPCI_LWZ, tmp1, RID_SP, ofs);
> +#endif
>         } else {
>   	emit_asb(as, PPCI_XOR, tmp2, key, tmp1);
>   	emit_rotlwi(as, dest, tmp1, HASH_ROT1);
> @@ -784,8 +860,8 @@ static PPCIns asm_fxloadins(IRIns *ir)
>     case IRT_U8: return PPCI_LBZ;
>     case IRT_I16: return PPCI_LHA;
>     case IRT_U16: return PPCI_LHZ;
> -  case IRT_NUM: return PPCI_LFD;
> -  case IRT_FLOAT: return PPCI_LFS;
> +  case IRT_NUM: lua_assert(!LJ_SOFTFP); return PPCI_LFD;
> +  case IRT_FLOAT: if (!LJ_SOFTFP) return PPCI_LFS;
>     default: return PPCI_LWZ;
>     }
>   }
> @@ -795,8 +871,8 @@ static PPCIns asm_fxstoreins(IRIns *ir)
>     switch (irt_type(ir->t)) {
>     case IRT_I8: case IRT_U8: return PPCI_STB;
>     case IRT_I16: case IRT_U16: return PPCI_STH;
> -  case IRT_NUM: return PPCI_STFD;
> -  case IRT_FLOAT: return PPCI_STFS;
> +  case IRT_NUM: lua_assert(!LJ_SOFTFP); return PPCI_STFD;
> +  case IRT_FLOAT: if (!LJ_SOFTFP) return PPCI_STFS;
>     default: return PPCI_STW;
>     }
>   }
> @@ -839,7 +915,8 @@ static void asm_fstore(ASMState *as, IRIns *ir)
>   
>   static void asm_xload(ASMState *as, IRIns *ir)
>   {
> -  Reg dest = ra_dest(as, ir, irt_isfp(ir->t) ? RSET_FPR : RSET_GPR);
> +  Reg dest = ra_dest(as, ir,
> +    (!LJ_SOFTFP && irt_isfp(ir->t)) ? RSET_FPR : RSET_GPR);
>     lua_assert(!(ir->op2 & IRXLOAD_UNALIGNED));
>     if (irt_isi8(ir->t))
>       emit_as(as, PPCI_EXTSB, dest, dest);
> @@ -857,7 +934,8 @@ static void asm_xstore_(ASMState *as, IRIns *ir, int32_t ofs)
>       Reg src = ra_alloc1(as, irb->op1, RSET_GPR);
>       asm_fusexrefx(as, PPCI_STWBRX, src, ir->op1, rset_exclude(RSET_GPR, src));
>     } else {
> -    Reg src = ra_alloc1(as, ir->op2, irt_isfp(ir->t) ? RSET_FPR : RSET_GPR);
> +    Reg src = ra_alloc1(as, ir->op2,
> +      (!LJ_SOFTFP && irt_isfp(ir->t)) ? RSET_FPR : RSET_GPR);
>       asm_fusexref(as, asm_fxstoreins(ir), src, ir->op1,
>   		 rset_exclude(RSET_GPR, src), ofs);
>     }
> @@ -871,10 +949,19 @@ static void asm_ahuvload(ASMState *as, IRIns *ir)
>     Reg dest = RID_NONE, type = RID_TMP, tmp = RID_TMP, idx;
>     RegSet allow = RSET_GPR;
>     int32_t ofs = AHUREF_LSX;
> +  if (LJ_SOFTFP && (ir+1)->o == IR_HIOP) {
> +    t.irt = IRT_NUM;
> +    if (ra_used(ir+1)) {
> +      type = ra_dest(as, ir+1, allow);
> +      rset_clear(allow, type);
> +    }
> +    ofs = 0;
> +  }
>     if (ra_used(ir)) {
> -    lua_assert(irt_isnum(t) || irt_isint(t) || irt_isaddr(t));
> -    if (!irt_isnum(t)) ofs = 0;
> -    dest = ra_dest(as, ir, irt_isnum(t) ? RSET_FPR : RSET_GPR);
> +    lua_assert((LJ_SOFTFP ? 0 : irt_isnum(ir->t)) ||
> +	       irt_isint(ir->t) || irt_isaddr(ir->t));
> +    if (LJ_SOFTFP || !irt_isnum(t)) ofs = 0;
> +    dest = ra_dest(as, ir, (!LJ_SOFTFP && irt_isnum(t)) ? RSET_FPR : allow);
>       rset_clear(allow, dest);
>     }
>     idx = asm_fuseahuref(as, ir->op1, &ofs, allow);
> @@ -883,12 +970,13 @@ static void asm_ahuvload(ASMState *as, IRIns *ir)
>       asm_guardcc(as, CC_GE);
>       emit_ab(as, PPCI_CMPLW, type, tisnum);
>       if (ra_hasreg(dest)) {
> -      if (ofs == AHUREF_LSX) {
> +      if (!LJ_SOFTFP && ofs == AHUREF_LSX) {
>   	tmp = ra_scratch(as, rset_exclude(rset_exclude(RSET_GPR,
>   						       (idx&255)), (idx>>8)));
>   	emit_fab(as, PPCI_LFDX, dest, (idx&255), tmp);
>         } else {
> -	emit_fai(as, PPCI_LFD, dest, idx, ofs);
> +	emit_fai(as, LJ_SOFTFP ? PPCI_LWZ : PPCI_LFD, dest, idx,
> +		 ofs+4*LJ_SOFTFP);
>         }
>       }
>     } else {
> @@ -911,7 +999,7 @@ static void asm_ahustore(ASMState *as, IRIns *ir)
>     int32_t ofs = AHUREF_LSX;
>     if (ir->r == RID_SINK)
>       return;
> -  if (irt_isnum(ir->t)) {
> +  if (!LJ_SOFTFP && irt_isnum(ir->t)) {
>       src = ra_alloc1(as, ir->op2, RSET_FPR);
>     } else {
>       if (!irt_ispri(ir->t)) {
> @@ -919,11 +1007,14 @@ static void asm_ahustore(ASMState *as, IRIns *ir)
>         rset_clear(allow, src);
>         ofs = 0;
>       }
> -    type = ra_allock(as, (int32_t)irt_toitype(ir->t), allow);
> +    if (LJ_SOFTFP && (ir+1)->o == IR_HIOP)
> +      type = ra_alloc1(as, (ir+1)->op2, allow);
> +    else
> +      type = ra_allock(as, (int32_t)irt_toitype(ir->t), allow);
>       rset_clear(allow, type);
>     }
>     idx = asm_fuseahuref(as, ir->op1, &ofs, allow);
> -  if (irt_isnum(ir->t)) {
> +  if (!LJ_SOFTFP && irt_isnum(ir->t)) {
>       if (ofs == AHUREF_LSX) {
>         emit_fab(as, PPCI_STFDX, src, (idx&255), RID_TMP);
>         emit_slwi(as, RID_TMP, (idx>>8), 3);
> @@ -948,21 +1039,33 @@ static void asm_sload(ASMState *as, IRIns *ir)
>     IRType1 t = ir->t;
>     Reg dest = RID_NONE, type = RID_NONE, base;
>     RegSet allow = RSET_GPR;
> +  int hiop = (LJ_SOFTFP && (ir+1)->o == IR_HIOP);
> +  if (hiop)
> +    t.irt = IRT_NUM;
>     lua_assert(!(ir->op2 & IRSLOAD_PARENT));  /* Handled by asm_head_side(). */
> -  lua_assert(irt_isguard(t) || !(ir->op2 & IRSLOAD_TYPECHECK));
> +  lua_assert(irt_isguard(ir->t) || !(ir->op2 & IRSLOAD_TYPECHECK));
>     lua_assert(LJ_DUALNUM ||
>   	     !irt_isint(t) || (ir->op2 & (IRSLOAD_CONVERT|IRSLOAD_FRAME)));
> +#if LJ_SOFTFP
> +  lua_assert(!(ir->op2 & IRSLOAD_CONVERT));  /* Handled by LJ_SOFTFP SPLIT. */
> +  if (hiop && ra_used(ir+1)) {
> +    type = ra_dest(as, ir+1, allow);
> +    rset_clear(allow, type);
> +  }
> +#else
>     if ((ir->op2 & IRSLOAD_CONVERT) && irt_isguard(t) && irt_isint(t)) {
>       dest = ra_scratch(as, RSET_FPR);
>       asm_tointg(as, ir, dest);
>       t.irt = IRT_NUM;  /* Continue with a regular number type check. */
> -  } else if (ra_used(ir)) {
> +  } else
> +#endif
> +  if (ra_used(ir)) {
>       lua_assert(irt_isnum(t) || irt_isint(t) || irt_isaddr(t));
> -    dest = ra_dest(as, ir, irt_isnum(t) ? RSET_FPR : RSET_GPR);
> +    dest = ra_dest(as, ir, (!LJ_SOFTFP && irt_isnum(t)) ? RSET_FPR : allow);
>       rset_clear(allow, dest);
>       base = ra_alloc1(as, REF_BASE, allow);
>       rset_clear(allow, base);
> -    if ((ir->op2 & IRSLOAD_CONVERT)) {
> +    if (!LJ_SOFTFP && (ir->op2 & IRSLOAD_CONVERT)) {
>         if (irt_isint(t)) {
>   	emit_tai(as, PPCI_LWZ, dest, RID_SP, SPOFS_TMPLO);
>   	dest = ra_scratch(as, RSET_FPR);
> @@ -994,10 +1097,13 @@ dotypecheck:
>       if ((ir->op2 & IRSLOAD_TYPECHECK)) {
>         Reg tisnum = ra_allock(as, (int32_t)LJ_TISNUM, allow);
>         asm_guardcc(as, CC_GE);
> -      emit_ab(as, PPCI_CMPLW, RID_TMP, tisnum);
> +#if !LJ_SOFTFP
>         type = RID_TMP;
> +#endif
> +      emit_ab(as, PPCI_CMPLW, type, tisnum);
>       }
> -    if (ra_hasreg(dest)) emit_fai(as, PPCI_LFD, dest, base, ofs-4);
> +    if (ra_hasreg(dest)) emit_fai(as, LJ_SOFTFP ? PPCI_LWZ : PPCI_LFD, dest,
> +				  base, ofs-(LJ_SOFTFP?0:4));
>     } else {
>       if ((ir->op2 & IRSLOAD_TYPECHECK)) {
>         asm_guardcc(as, CC_NE);
> @@ -1122,6 +1228,7 @@ static void asm_obar(ASMState *as, IRIns *ir)
>   
>   /* -- Arithmetic and logic operations ------------------------------------- */
>   
> +#if !LJ_SOFTFP
>   static void asm_fparith(ASMState *as, IRIns *ir, PPCIns pi)
>   {
>     Reg dest = ra_dest(as, ir, RSET_FPR);
> @@ -1149,13 +1256,17 @@ static void asm_fpmath(ASMState *as, IRIns *ir)
>     else
>       asm_callid(as, ir, IRCALL_lj_vm_floor + ir->op2);
>   }
> +#endif
>   
>   static void asm_add(ASMState *as, IRIns *ir)
>   {
> +#if !LJ_SOFTFP
>     if (irt_isnum(ir->t)) {
>       if (!asm_fusemadd(as, ir, PPCI_FMADD, PPCI_FMADD))
>         asm_fparith(as, ir, PPCI_FADD);
> -  } else {
> +  } else
> +#endif
> +  {
>       Reg dest = ra_dest(as, ir, RSET_GPR);
>       Reg right, left = ra_hintalloc(as, ir->op1, dest, RSET_GPR);
>       PPCIns pi;
> @@ -1194,10 +1305,13 @@ static void asm_add(ASMState *as, IRIns *ir)
>   
>   static void asm_sub(ASMState *as, IRIns *ir)
>   {
> +#if !LJ_SOFTFP
>     if (irt_isnum(ir->t)) {
>       if (!asm_fusemadd(as, ir, PPCI_FMSUB, PPCI_FNMSUB))
>         asm_fparith(as, ir, PPCI_FSUB);
> -  } else {
> +  } else
> +#endif
> +  {
>       PPCIns pi = PPCI_SUBF;
>       Reg dest = ra_dest(as, ir, RSET_GPR);
>       Reg left, right;
> @@ -1223,9 +1337,12 @@ static void asm_sub(ASMState *as, IRIns *ir)
>   
>   static void asm_mul(ASMState *as, IRIns *ir)
>   {
> +#if !LJ_SOFTFP
>     if (irt_isnum(ir->t)) {
>       asm_fparith(as, ir, PPCI_FMUL);
> -  } else {
> +  } else
> +#endif
> +  {
>       PPCIns pi = PPCI_MULLW;
>       Reg dest = ra_dest(as, ir, RSET_GPR);
>       Reg right, left = ra_hintalloc(as, ir->op1, dest, RSET_GPR);
> @@ -1253,9 +1370,12 @@ static void asm_mul(ASMState *as, IRIns *ir)
>   
>   static void asm_neg(ASMState *as, IRIns *ir)
>   {
> +#if !LJ_SOFTFP
>     if (irt_isnum(ir->t)) {
>       asm_fpunary(as, ir, PPCI_FNEG);
> -  } else {
> +  } else
> +#endif
> +  {
>       Reg dest, left;
>       PPCIns pi = PPCI_NEG;
>       if (as->flagmcp == as->mcp) {
> @@ -1566,9 +1686,40 @@ static void asm_bitshift(ASMState *as, IRIns *ir, PPCIns pi, PPCIns pik)
>   		       PPCI_RLWINM|PPCF_MB(0)|PPCF_ME(31))
>   #define asm_bror(as, ir)	lua_assert(0)
>   
> +#if LJ_SOFTFP
> +static void asm_sfpmin_max(ASMState *as, IRIns *ir)
> +{
> +  CCallInfo ci = lj_ir_callinfo[IRCALL_softfp_cmp];
> +  IRRef args[4];
> +  MCLabel l_right, l_end;
> +  Reg desthi = ra_dest(as, ir, RSET_GPR), destlo = ra_dest(as, ir+1, RSET_GPR);
> +  Reg righthi, lefthi = ra_alloc2(as, ir, RSET_GPR);
> +  Reg rightlo, leftlo = ra_alloc2(as, ir+1, RSET_GPR);
> +  PPCCC cond = (IROp)ir->o == IR_MIN ? CC_EQ : CC_NE;
> +  righthi = (lefthi >> 8); lefthi &= 255;
> +  rightlo = (leftlo >> 8); leftlo &= 255;
> +  args[0^LJ_BE] = ir->op1; args[1^LJ_BE] = (ir+1)->op1;
> +  args[2^LJ_BE] = ir->op2; args[3^LJ_BE] = (ir+1)->op2;
> +  l_end = emit_label(as);
> +  if (desthi != righthi) emit_mr(as, desthi, righthi);
> +  if (destlo != rightlo) emit_mr(as, destlo, rightlo);
> +  l_right = emit_label(as);
> +  if (l_end != l_right) emit_jmp(as, l_end);
> +  if (desthi != lefthi) emit_mr(as, desthi, lefthi);
> +  if (destlo != leftlo) emit_mr(as, destlo, leftlo);
> +  if (l_right == as->mcp+1) {
> +    cond ^= 4; l_right = l_end; ++as->mcp;
> +  }
> +  emit_condbranch(as, PPCI_BC, cond, l_right);
> +  ra_evictset(as, RSET_SCRATCH);
> +  emit_cmpi(as, RID_RET, 1);
> +  asm_gencall(as, &ci, args);
> +}
> +#endif
> +
>   static void asm_min_max(ASMState *as, IRIns *ir, int ismax)
>   {
> -  if (irt_isnum(ir->t)) {
> +  if (!LJ_SOFTFP && irt_isnum(ir->t)) {
>       Reg dest = ra_dest(as, ir, RSET_FPR);
>       Reg tmp = dest;
>       Reg right, left = ra_alloc2(as, ir, RSET_FPR);
> @@ -1656,7 +1807,7 @@ static void asm_intcomp_(ASMState *as, IRRef lref, IRRef rref, Reg cr, PPCCC cc)
>   static void asm_comp(ASMState *as, IRIns *ir)
>   {
>     PPCCC cc = asm_compmap[ir->o];
> -  if (irt_isnum(ir->t)) {
> +  if (!LJ_SOFTFP && irt_isnum(ir->t)) {
>       Reg right, left = ra_alloc2(as, ir, RSET_FPR);
>       right = (left >> 8); left &= 255;
>       asm_guardcc(as, (cc >> 4));
> @@ -1677,6 +1828,44 @@ static void asm_comp(ASMState *as, IRIns *ir)
>   
>   #define asm_equal(as, ir)	asm_comp(as, ir)
>   
> +#if LJ_SOFTFP
> +/* SFP comparisons. */
> +static void asm_sfpcomp(ASMState *as, IRIns *ir)
> +{
> +  const CCallInfo *ci = &lj_ir_callinfo[IRCALL_softfp_cmp];
> +  RegSet drop = RSET_SCRATCH;
> +  Reg r;
> +  IRRef args[4];
> +  args[0^LJ_BE] = ir->op1; args[1^LJ_BE] = (ir+1)->op1;
> +  args[2^LJ_BE] = ir->op2; args[3^LJ_BE] = (ir+1)->op2;
> +
> +  for (r = REGARG_FIRSTGPR; r <= REGARG_FIRSTGPR+3; r++) {
> +    if (!rset_test(as->freeset, r) &&
> +	regcost_ref(as->cost[r]) == args[r-REGARG_FIRSTGPR])
> +      rset_clear(drop, r);
> +  }
> +  ra_evictset(as, drop);
> +  asm_setupresult(as, ir, ci);
> +  switch ((IROp)ir->o) {
> +  case IR_ULT:
> +    asm_guardcc(as, CC_EQ);
> +    emit_ai(as, PPCI_CMPWI, RID_RET, 0);
> +  case IR_ULE:
> +    asm_guardcc(as, CC_EQ);
> +    emit_ai(as, PPCI_CMPWI, RID_RET, 1);
> +    break;
> +  case IR_GE: case IR_GT:
> +    asm_guardcc(as, CC_EQ);
> +    emit_ai(as, PPCI_CMPWI, RID_RET, 2);
> +  default:
> +    asm_guardcc(as, (asm_compmap[ir->o] & 0xf));
> +    emit_ai(as, PPCI_CMPWI, RID_RET, 0);
> +    break;
> +  }
> +  asm_gencall(as, ci, args);
> +}
> +#endif
> +
>   #if LJ_HASFFI
>   /* 64 bit integer comparisons. */
>   static void asm_comp64(ASMState *as, IRIns *ir)
> @@ -1706,19 +1895,36 @@ static void asm_comp64(ASMState *as, IRIns *ir)
>   /* Hiword op of a split 64 bit op. Previous op must be the loword op. */
>   static void asm_hiop(ASMState *as, IRIns *ir)
>   {
> -#if LJ_HASFFI
> +#if LJ_HASFFI || LJ_SOFTFP
>     /* HIOP is marked as a store because it needs its own DCE logic. */
>     int uselo = ra_used(ir-1), usehi = ra_used(ir);  /* Loword/hiword used? */
>     if (LJ_UNLIKELY(!(as->flags & JIT_F_OPT_DCE))) uselo = usehi = 1;
>     if ((ir-1)->o == IR_CONV) {  /* Conversions to/from 64 bit. */
>       as->curins--;  /* Always skip the CONV. */
> +#if LJ_HASFFI && !LJ_SOFTFP
>       if (usehi || uselo)
>         asm_conv64(as, ir);
>       return;
> +#endif
>     } else if ((ir-1)->o <= IR_NE) {  /* 64 bit integer comparisons. ORDER IR. */
>       as->curins--;  /* Always skip the loword comparison. */
> +#if LJ_SOFTFP
> +    if (!irt_isint(ir->t)) {
> +      asm_sfpcomp(as, ir-1);
> +      return;
> +    }
> +#endif
> +#if LJ_HASFFI
>       asm_comp64(as, ir);
> +#endif
> +    return;
> +#if LJ_SOFTFP
> +  } else if ((ir-1)->o == IR_MIN || (ir-1)->o == IR_MAX) {
> +      as->curins--;  /* Always skip the loword min/max. */
> +    if (uselo || usehi)
> +      asm_sfpmin_max(as, ir-1);
>       return;
> +#endif
>     } else if ((ir-1)->o == IR_XSTORE) {
>       as->curins--;  /* Handle both stores here. */
>       if ((ir-1)->r != RID_SINK) {
> @@ -1729,14 +1935,27 @@ static void asm_hiop(ASMState *as, IRIns *ir)
>     }
>     if (!usehi) return;  /* Skip unused hiword op for all remaining ops. */
>     switch ((ir-1)->o) {
> +#if LJ_HASFFI
>     case IR_ADD: as->curins--; asm_add64(as, ir); break;
>     case IR_SUB: as->curins--; asm_sub64(as, ir); break;
>     case IR_NEG: as->curins--; asm_neg64(as, ir); break;
> +#endif
> +#if LJ_SOFTFP
> +  case IR_SLOAD: case IR_ALOAD: case IR_HLOAD: case IR_ULOAD: case IR_VLOAD:
> +  case IR_STRTO:
> +    if (!uselo)
> +      ra_allocref(as, ir->op1, RSET_GPR);  /* Mark lo op as used. */
> +    break;
> +#endif
>     case IR_CALLN:
> +  case IR_CALLS:
>     case IR_CALLXS:
>       if (!uselo)
>         ra_allocref(as, ir->op1, RID2RSET(RID_RETLO));  /* Mark lo op as used. */
>       break;
> +#if LJ_SOFTFP
> +  case IR_ASTORE: case IR_HSTORE: case IR_USTORE: case IR_TOSTR:
> +#endif
>     case IR_CNEWI:
>       /* Nothing to do here. Handled by lo op itself. */
>       break;
> @@ -1800,8 +2019,19 @@ static void asm_stack_restore(ASMState *as, SnapShot *snap)
>       if ((sn & SNAP_NORESTORE))
>         continue;
>       if (irt_isnum(ir->t)) {
> +#if LJ_SOFTFP
> +      Reg tmp;
> +      RegSet allow = rset_exclude(RSET_GPR, RID_BASE);
> +      lua_assert(irref_isk(ref));  /* LJ_SOFTFP: must be a number constant. */
> +      tmp = ra_allock(as, (int32_t)ir_knum(ir)->u32.lo, allow);
> +      emit_tai(as, PPCI_STW, tmp, RID_BASE, ofs+(LJ_BE?4:0));
> +      if (rset_test(as->freeset, tmp+1)) allow = RID2RSET(tmp+1);
> +      tmp = ra_allock(as, (int32_t)ir_knum(ir)->u32.hi, allow);
> +      emit_tai(as, PPCI_STW, tmp, RID_BASE, ofs+(LJ_BE?0:4));
> +#else
>         Reg src = ra_alloc1(as, ref, RSET_FPR);
>         emit_fai(as, PPCI_STFD, src, RID_BASE, ofs);
> +#endif
>       } else {
>         Reg type;
>         RegSet allow = rset_exclude(RSET_GPR, RID_BASE);
> @@ -1814,6 +2044,10 @@ static void asm_stack_restore(ASMState *as, SnapShot *snap)
>         if ((sn & (SNAP_CONT|SNAP_FRAME))) {
>   	if (s == 0) continue;  /* Do not overwrite link to previous frame. */
>   	type = ra_allock(as, (int32_t)(*flinks--), allow);
> +#if LJ_SOFTFP
> +      } else if ((sn & SNAP_SOFTFPNUM)) {
> +	type = ra_alloc1(as, ref+1, rset_exclude(RSET_GPR, RID_BASE));
> +#endif
>         } else {
>   	type = ra_allock(as, (int32_t)irt_toitype(ir->t), allow);
>         }
> @@ -1950,14 +2184,15 @@ static Reg asm_setup_call_slots(ASMState *as, IRIns *ir, const CCallInfo *ci)
>     int nslots = 2, ngpr = REGARG_NUMGPR, nfpr = REGARG_NUMFPR;
>     asm_collectargs(as, ir, ci, args);
>     for (i = 0; i < nargs; i++)
> -    if (args[i] && irt_isfp(IR(args[i])->t)) {
> +    if (!LJ_SOFTFP && args[i] && irt_isfp(IR(args[i])->t)) {
>         if (nfpr > 0) nfpr--; else nslots = (nslots+3) & ~1;
>       } else {
>         if (ngpr > 0) ngpr--; else nslots++;
>       }
>     if (nslots > as->evenspill)  /* Leave room for args in stack slots. */
>       as->evenspill = nslots;
> -  return irt_isfp(ir->t) ? REGSP_HINT(RID_FPRET) : REGSP_HINT(RID_RET);
> +  return (!LJ_SOFTFP && irt_isfp(ir->t)) ? REGSP_HINT(RID_FPRET) :
> +					   REGSP_HINT(RID_RET);
>   }
>   
>   static void asm_setup_target(ASMState *as)

  parent reply	other threads:[~2023-08-17 14:33 UTC|newest]

Thread overview: 97+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-08-09 15:35 [Tarantool-patches] [PATCH luajit 00/19] Prerequisites for improve assertions Sergey Kaplun via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 01/19] MIPS: Use precise search for exit jump patching Sergey Kaplun via Tarantool-patches
2023-08-15  9:36   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 12:40     ` Sergey Kaplun via Tarantool-patches
2023-08-16 13:25   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 02/19] test: introduce mcode generator for tests Sergey Kaplun via Tarantool-patches
2023-08-15 10:14   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 12:55     ` Sergey Kaplun via Tarantool-patches
2023-08-16 13:06       ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 14:32   ` Sergey Bronnikov via Tarantool-patches
2023-08-16 15:20     ` Sergey Kaplun via Tarantool-patches
2023-08-16 16:08       ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 03/19] MIPS: Fix handling of spare long-range jump slots Sergey Kaplun via Tarantool-patches
2023-08-15 11:13   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:05     ` Sergey Kaplun via Tarantool-patches
2023-08-16 15:02   ` Sergey Bronnikov via Tarantool-patches
2023-08-16 15:32     ` Sergey Kaplun via Tarantool-patches
2023-08-16 16:08       ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 04/19] MIPS64: Add soft-float support to JIT compiler backend Sergey Kaplun via Tarantool-patches
2023-08-15 11:27   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:10     ` Sergey Kaplun via Tarantool-patches
2023-08-16 16:07   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 05/19] PPC: Add soft-float support to interpreter Sergey Kaplun via Tarantool-patches
2023-08-15 11:40   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:13     ` Sergey Kaplun via Tarantool-patches
2023-08-17 14:53   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 06/19] PPC: Add soft-float support to JIT compiler backend Sergey Kaplun via Tarantool-patches
2023-08-15 11:46   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:21     ` Sergey Kaplun via Tarantool-patches
2023-08-17 14:33   ` Sergey Bronnikov via Tarantool-patches [this message]
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 07/19] build: fix non-Linux/macOS builds Sergey Kaplun via Tarantool-patches
2023-08-15 11:58   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:40     ` Sergey Kaplun via Tarantool-patches
2023-08-17 14:31   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 08/19] Windows: Add UWP support, part 1 Sergey Kaplun via Tarantool-patches
2023-08-15 12:09   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:50     ` Sergey Kaplun via Tarantool-patches
2023-08-16 16:40   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 09/19] FFI: Eliminate hardcoded string hashes Sergey Kaplun via Tarantool-patches
2023-08-15 13:07   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:52     ` Sergey Kaplun via Tarantool-patches
2023-08-16 17:04     ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:35 ` [Tarantool-patches] [PATCH luajit 10/19] Cleanup math function compilation and fix inconsistencies Sergey Kaplun via Tarantool-patches
2023-08-11  8:06   ` Sergey Kaplun via Tarantool-patches
2023-08-15 13:10   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 17:15   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 11/19] Fix GCC 7 -Wimplicit-fallthrough warnings Sergey Kaplun via Tarantool-patches
2023-08-15 13:17   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 13:59     ` Sergey Kaplun via Tarantool-patches
2023-08-17  7:37   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 12/19] DynASM: Fix warning Sergey Kaplun via Tarantool-patches
2023-08-15 13:21   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 14:01     ` Sergey Kaplun via Tarantool-patches
2023-08-17  7:39   ` Sergey Bronnikov via Tarantool-patches
2023-08-17  7:51     ` Sergey Bronnikov via Tarantool-patches
2023-08-17  7:58       ` Sergey Kaplun via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 13/19] ARM: Fix GCC 7 -Wimplicit-fallthrough warnings Sergey Kaplun via Tarantool-patches
2023-08-15 13:25   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 14:08     ` Sergey Kaplun via Tarantool-patches
2023-08-17  7:44   ` Sergey Bronnikov via Tarantool-patches
2023-08-17  8:01     ` Sergey Kaplun via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 14/19] Fix debug.getinfo() argument check Sergey Kaplun via Tarantool-patches
2023-08-15 13:35   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 14:20     ` Sergey Kaplun via Tarantool-patches
2023-08-16 20:13       ` Maxim Kokryashkin via Tarantool-patches
2023-08-17  8:29   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 15/19] Fix LJ_MAX_JSLOTS assertion in rec_check_slots() Sergey Kaplun via Tarantool-patches
2023-08-15 14:07   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 14:22     ` Sergey Kaplun via Tarantool-patches
2023-08-17  8:57   ` Sergey Bronnikov via Tarantool-patches
2023-08-17  8:57     ` Sergey Kaplun via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 16/19] Prevent integer overflow while parsing long strings Sergey Kaplun via Tarantool-patches
2023-08-15 14:38   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 14:52     ` Sergey Kaplun via Tarantool-patches
2023-08-17 10:53   ` Sergey Bronnikov via Tarantool-patches
2023-08-17 13:57     ` Sergey Kaplun via Tarantool-patches
2023-08-17 14:28       ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 17/19] MIPS64: Fix register allocation in assembly of HREF Sergey Kaplun via Tarantool-patches
2023-08-16  9:01   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 15:17     ` Sergey Kaplun via Tarantool-patches
2023-08-16 20:14       ` Maxim Kokryashkin via Tarantool-patches
2023-08-17 11:06   ` Sergey Bronnikov via Tarantool-patches
2023-08-17 13:50     ` Sergey Kaplun via Tarantool-patches
2023-08-17 14:30       ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 18/19] DynASM/MIPS: Fix shadowed variable Sergey Kaplun via Tarantool-patches
2023-08-16  9:03   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 15:22     ` Sergey Kaplun via Tarantool-patches
2023-08-17 12:01   ` Sergey Bronnikov via Tarantool-patches
2023-08-09 15:36 ` [Tarantool-patches] [PATCH luajit 19/19] MIPS: Add MIPS64 R6 port Sergey Kaplun via Tarantool-patches
2023-08-16  9:16   ` Maxim Kokryashkin via Tarantool-patches
2023-08-16 15:24     ` Sergey Kaplun via Tarantool-patches
2023-08-17 13:03   ` Sergey Bronnikov via Tarantool-patches
2023-08-17 13:59     ` Sergey Kaplun via Tarantool-patches
2023-08-16 15:35 ` [Tarantool-patches] [PATCH luajit 00/19] Prerequisites for improve assertions Sergey Kaplun via Tarantool-patches
2023-08-17 14:06   ` Maxim Kokryashkin via Tarantool-patches
2023-08-17 14:38 ` Sergey Bronnikov via Tarantool-patches
2023-08-31 15:17 ` Igor Munkin via Tarantool-patches

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=da27ca67-822d-1c9a-1120-811400d53b27@tarantool.org \
    --to=tarantool-patches@dev.tarantool.org \
    --cc=imun@tarantool.org \
    --cc=sergeyb@tarantool.org \
    --cc=skaplun@tarantool.org \
    --subject='Re: [Tarantool-patches] [PATCH luajit 06/19] PPC: Add soft-float support to JIT compiler backend.' \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox