diff options
| author | Jakub Konka <kubkon@jakubkonka.com> | 2022-05-10 20:45:57 +0200 |
|---|---|---|
| committer | Jakub Konka <kubkon@jakubkonka.com> | 2022-05-10 20:45:57 +0200 |
| commit | 2a738599a0180b6f12439d33722a52182b11c21e (patch) | |
| tree | 08aa21fa0e3d12b8a5977b4c0e78c0c3733e75d6 | |
| parent | f131e41db9f8b1b1bbd678b52ef377821d83bddc (diff) | |
| download | zig-2a738599a0180b6f12439d33722a52182b11c21e.tar.gz zig-2a738599a0180b6f12439d33722a52182b11c21e.zip | |
x64: implement missing bits in add_with_overflow and sub_with_overflow
| -rw-r--r-- | src/arch/x86_64/CodeGen.zig | 173 | ||||
| -rw-r--r-- | test/behavior/math.zig | 1 |
2 files changed, 101 insertions, 73 deletions
diff --git a/src/arch/x86_64/CodeGen.zig b/src/arch/x86_64/CodeGen.zig index 2ed61006f2..15610f1bd2 100644 --- a/src/arch/x86_64/CodeGen.zig +++ b/src/arch/x86_64/CodeGen.zig @@ -1311,18 +1311,15 @@ fn airAddSubWithOverflow(self: *Self, inst: Air.Inst.Index) !void { const bin_op = self.air.extraData(Air.Bin, ty_pl.payload).data; const result = if (self.liveness.isUnused(inst)) .dead else result: { const ty = self.air.typeOf(bin_op.lhs); - + const abi_size = ty.abiSize(self.target.*); switch (ty.zigTypeTag()) { .Vector => return self.fail("TODO implement add_with_overflow for Vector type", .{}), .Int => { - const int_info = ty.intInfo(self.target.*); - - if (int_info.bits > 64) { + if (abi_size > 8) { return self.fail("TODO implement add_with_overflow for Ints larger than 64bits", .{}); } try self.spillCompareFlagsIfOccupied(); - self.compare_flags_inst = inst; const lhs = try self.resolveInst(bin_op.lhs); const rhs = try self.resolveInst(bin_op.rhs); @@ -1333,11 +1330,30 @@ fn airAddSubWithOverflow(self: *Self, inst: Air.Inst.Index) !void { else => unreachable, }; const partial = try self.genBinOp(base_tag, null, lhs, rhs, ty, ty); - const result: MCValue = switch (int_info.signedness) { - .signed => .{ .register_overflow_signed = partial.register }, - .unsigned => .{ .register_overflow_unsigned = partial.register }, - }; - break :result result; + + const int_info = ty.intInfo(self.target.*); + + if (math.isPowerOfTwo(int_info.bits) and int_info.bits >= 8) { + self.compare_flags_inst = inst; + + const result: MCValue = switch (int_info.signedness) { + .signed => .{ .register_overflow_signed = partial.register }, + .unsigned => .{ .register_overflow_unsigned = partial.register }, + }; + break :result result; + } + + self.compare_flags_inst = null; + + const tuple_ty = self.air.typeOfIndex(inst); + const tuple_size = @intCast(u32, tuple_ty.abiSize(self.target.*)); + const tuple_align = tuple_ty.abiAlignment(self.target.*); + const overflow_bit_offset = @intCast(i32, tuple_ty.structFieldOffset(1, self.target.*)); + const stack_offset = @intCast(i32, try self.allocMem(inst, tuple_size, tuple_align)); + + try self.genSetStackTruncatedOverflowCompare(ty, stack_offset, overflow_bit_offset, partial.register); + + break :result MCValue{ .stack_offset = stack_offset }; }, else => unreachable, } @@ -1346,6 +1362,75 @@ fn airAddSubWithOverflow(self: *Self, inst: Air.Inst.Index) !void { return self.finishAir(inst, result, .{ bin_op.lhs, bin_op.rhs, .none }); } +fn genSetStackTruncatedOverflowCompare( + self: *Self, + ty: Type, + stack_offset: i32, + overflow_bit_offset: i32, + reg: Register, +) !void { + const reg_lock = self.register_manager.lockReg(reg); + defer if (reg_lock) |lock| self.register_manager.unlockReg(lock); + + const int_info = ty.intInfo(self.target.*); + const extended_ty = switch (int_info.signedness) { + .signed => Type.isize, + .unsigned => ty, + }; + + const temp_regs = try self.register_manager.allocRegs(3, .{ null, null, null }); + const temp_regs_locks = self.register_manager.lockRegsAssumeUnused(3, temp_regs); + defer for (temp_regs_locks) |rreg| { + self.register_manager.unlockReg(rreg); + }; + + const overflow_reg = temp_regs[0]; + const flags: u2 = switch (int_info.signedness) { + .signed => 0b00, + .unsigned => 0b10, + }; + _ = try self.addInst(.{ + .tag = .cond_set_byte_overflow, + .ops = (Mir.Ops{ + .reg1 = overflow_reg.to8(), + .flags = flags, + }).encode(), + .data = undefined, + }); + + const scratch_reg = temp_regs[1]; + try self.genSetReg(extended_ty, scratch_reg, .{ .register = reg }); + try self.truncateRegister(ty, scratch_reg); + try self.genBinOpMir( + .cmp, + extended_ty, + .{ .register = reg }, + .{ .register = scratch_reg }, + ); + + const eq_reg = temp_regs[2]; + _ = try self.addInst(.{ + .tag = .cond_set_byte_eq_ne, + .ops = (Mir.Ops{ + .reg1 = eq_reg.to8(), + .flags = 0b00, + }).encode(), + .data = undefined, + }); + + try self.genBinOpMir( + .@"or", + Type.u8, + .{ .register = overflow_reg }, + .{ .register = eq_reg }, + ); + + try self.genSetStack(ty, stack_offset, .{ .register = scratch_reg }, .{}); + try self.genSetStack(Type.initTag(.u1), stack_offset - overflow_bit_offset, .{ + .register = overflow_reg.to8(), + }, .{}); +} + fn airMulWithOverflow(self: *Self, inst: Air.Inst.Index) !void { const ty_pl = self.air.instructions.items(.data)[inst].ty_pl; const bin_op = self.air.extraData(Air.Bin, ty_pl.payload).data; @@ -1355,17 +1440,18 @@ fn airMulWithOverflow(self: *Self, inst: Air.Inst.Index) !void { } const ty = self.air.typeOf(bin_op.lhs); + const abi_size = ty.abiSize(self.target.*); const result: MCValue = result: { switch (ty.zigTypeTag()) { .Vector => return self.fail("TODO implement mul_with_overflow for Vector type", .{}), .Int => { - const int_info = ty.intInfo(self.target.*); - - if (int_info.bits > 64) { + if (abi_size > 8) { return self.fail("TODO implement mul_with_overflow for Ints larger than 64bits", .{}); } - if (math.isPowerOfTwo(int_info.bits)) { + const int_info = ty.intInfo(self.target.*); + + if (math.isPowerOfTwo(int_info.bits) and int_info.bits >= 8) { try self.spillCompareFlagsIfOccupied(); self.compare_flags_inst = inst; @@ -1428,71 +1514,14 @@ fn airMulWithOverflow(self: *Self, inst: Air.Inst.Index) !void { }, } }; - const dst_reg_lock = self.register_manager.lockRegAssumeUnused(dst_reg); - defer self.register_manager.unlockReg(dst_reg_lock); const tuple_ty = self.air.typeOfIndex(inst); const tuple_size = @intCast(u32, tuple_ty.abiSize(self.target.*)); const tuple_align = tuple_ty.abiAlignment(self.target.*); const overflow_bit_offset = @intCast(i32, tuple_ty.structFieldOffset(1, self.target.*)); - const stack_offset = @intCast(i32, try self.allocMem(inst, tuple_size, tuple_align)); - const extended_ty = switch (int_info.signedness) { - .signed => Type.isize, - .unsigned => ty, - }; - - const temp_regs = try self.register_manager.allocRegs(3, .{ null, null, null }); - const temp_regs_locks = self.register_manager.lockRegsAssumeUnused(3, temp_regs); - defer for (temp_regs_locks) |reg| { - self.register_manager.unlockReg(reg); - }; - - const overflow_reg = temp_regs[0]; - const flags: u2 = switch (int_info.signedness) { - .signed => 0b00, - .unsigned => 0b10, - }; - _ = try self.addInst(.{ - .tag = .cond_set_byte_overflow, - .ops = (Mir.Ops{ - .reg1 = overflow_reg.to8(), - .flags = flags, - }).encode(), - .data = undefined, - }); - - const scratch_reg = temp_regs[1]; - try self.genSetReg(extended_ty, scratch_reg, .{ .register = dst_reg }); - try self.truncateRegister(ty, scratch_reg); - try self.genBinOpMir( - .cmp, - extended_ty, - .{ .register = dst_reg }, - .{ .register = scratch_reg }, - ); - - const eq_reg = temp_regs[2]; - _ = try self.addInst(.{ - .tag = .cond_set_byte_eq_ne, - .ops = (Mir.Ops{ - .reg1 = eq_reg.to8(), - .flags = 0b00, - }).encode(), - .data = undefined, - }); - - try self.genBinOpMir( - .@"or", - Type.u8, - .{ .register = overflow_reg }, - .{ .register = eq_reg }, - ); - try self.genSetStack(ty, stack_offset, .{ .register = scratch_reg }, .{}); - try self.genSetStack(Type.initTag(.u1), stack_offset - overflow_bit_offset, .{ - .register = overflow_reg.to8(), - }, .{}); + try self.genSetStackTruncatedOverflowCompare(ty, stack_offset, overflow_bit_offset, dst_reg); break :result MCValue{ .stack_offset = stack_offset }; }, diff --git a/test/behavior/math.zig b/test/behavior/math.zig index 53b11d1d9f..201fccb8ae 100644 --- a/test/behavior/math.zig +++ b/test/behavior/math.zig @@ -640,7 +640,6 @@ test "@addWithOverflow" { test "small int addition" { if (builtin.zig_backend == .stage2_c) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_x86_64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO |
