aboutsummaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorAndrew Kelley <andrew@ziglang.org>2020-10-29 18:29:24 -0400
committerGitHub <noreply@github.com>2020-10-29 18:29:24 -0400
commitf4bb8be9fc8766fec93618f89552c28d3f0a201b (patch)
treece990f0559cdb2f10a315986ee1711b3d3964c15 /src
parentaaff66b8edc549de79b997328ebad668838c5433 (diff)
parent7b4f3c7cfce202117f62b12e9036e5c8ca6a0e56 (diff)
downloadzig-f4bb8be9fc8766fec93618f89552c28d3f0a201b.tar.gz
zig-f4bb8be9fc8766fec93618f89552c28d3f0a201b.zip
Merge pull request #6654 from joachimschmidt557/stage2-arm
stage2 ARM: more stuff
Diffstat (limited to 'src')
-rw-r--r--src/codegen.zig183
-rw-r--r--src/codegen/arm.zig226
2 files changed, 360 insertions, 49 deletions
diff --git a/src/codegen.zig b/src/codegen.zig
index 70c97ca531..5e5215c992 100644
--- a/src/codegen.zig
+++ b/src/codegen.zig
@@ -573,25 +573,54 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
// sub sp, sp, #reloc
mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.push(.al, .{ .fp, .lr }).toU32());
mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, .fp, Instruction.Operand.reg(.sp, Instruction.Operand.Shift.none)).toU32());
- // TODO: prepare stack for local variables
- // const backpatch_reloc = try self.code.addManyAsArray(4);
+ const backpatch_reloc = self.code.items.len;
+ try self.code.resize(backpatch_reloc + 4);
try self.dbgSetPrologueEnd();
try self.genBody(self.mod_fn.analysis.success);
// Backpatch stack offset
- // const stack_end = self.max_end_stack;
- // const aligned_stack_end = mem.alignForward(stack_end, self.stack_align);
- // mem.writeIntLittle(u32, backpatch_reloc, Instruction.sub(.al, .sp, .sp, Instruction.Operand.imm()));
+ const stack_end = self.max_end_stack;
+ const aligned_stack_end = mem.alignForward(stack_end, self.stack_align);
+ if (Instruction.Operand.fromU32(@intCast(u32, aligned_stack_end))) |op| {
+ mem.writeIntLittle(u32, self.code.items[backpatch_reloc..][0..4], Instruction.sub(.al, .sp, .sp, op).toU32());
+ } else {
+ return self.fail(self.src, "TODO ARM: allow larger stacks", .{});
+ }
try self.dbgSetEpilogueBegin();
+ // exitlude jumps
+ if (self.exitlude_jump_relocs.items.len == 1) {
+ // There is only one relocation. Hence,
+ // this relocation must be at the end of
+ // the code. Therefore, we can just delete
+ // the space initially reserved for the
+ // jump
+ self.code.items.len -= 4;
+ } else for (self.exitlude_jump_relocs.items) |jmp_reloc| {
+ const amt = self.code.items.len - (jmp_reloc + 4);
+ if (amt == 0) {
+ // This return is at the end of the
+ // code block. We can't just delete
+ // the space because there may be
+ // other jumps we already relocated to
+ // the address. Instead, insert a nop
+ mem.writeIntLittle(u32, self.code.items[jmp_reloc..][0..4], Instruction.nop().toU32());
+ } else {
+ if (math.cast(i26, amt)) |offset| {
+ mem.writeIntLittle(u32, self.code.items[jmp_reloc..][0..4], Instruction.b(.al, offset).toU32());
+ } else |err| {
+ return self.fail(self.src, "exitlude jump is too large", .{});
+ }
+ }
+ }
+
// mov sp, fp
// pop {fp, pc}
- // TODO: return by jumping to this code, use relocations
- // mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, .sp, Instruction.Operand.reg(.fp, Instruction.Operand.Shift.none)).toU32());
- // mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.pop(.al, .{ .fp, .pc }).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, .sp, Instruction.Operand.reg(.fp, Instruction.Operand.Shift.none)).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.pop(.al, .{ .fp, .pc }).toU32());
} else {
try self.dbgSetPrologueEnd();
try self.genBody(self.mod_fn.analysis.success);
@@ -1661,12 +1690,9 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.jalr(.zero, 0, .ra).toU32());
},
.arm => {
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, .sp, Instruction.Operand.reg(.fp, Instruction.Operand.Shift.none)).toU32());
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.pop(.al, .{ .fp, .pc }).toU32());
- // TODO: jump to the end with relocation
- // // Just add space for an instruction, patch this later
- // try self.code.resize(self.code.items.len + 4);
- // try self.exitlude_jump_relocs.append(self.gpa, self.code.items.len - 4);
+ // Just add space for an instruction, patch this later
+ try self.code.resize(self.code.items.len + 4);
+ try self.exitlude_jump_relocs.append(self.gpa, self.code.items.len - 4);
},
else => return self.fail(src, "TODO implement return for {}", .{self.target.cpu.arch}),
}
@@ -1932,6 +1958,13 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
mem.writeIntLittle(i32, self.code.addManyAsArrayAssumeCapacity(4), delta);
}
},
+ .arm => {
+ if (math.cast(i26, @intCast(i32, index) - @intCast(i32, self.code.items.len))) |delta| {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.b(.al, delta).toU32());
+ } else |err| {
+ return self.fail(src, "TODO: enable larger branch offset", .{});
+ }
+ },
else => return self.fail(src, "TODO implement jump for {}", .{self.target.cpu.arch}),
}
}
@@ -2167,6 +2200,58 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
fn genSetStack(self: *Self, src: usize, ty: Type, stack_offset: u32, mcv: MCValue) InnerError!void {
switch (arch) {
+ .arm => switch (mcv) {
+ .dead => unreachable,
+ .ptr_stack_offset => unreachable,
+ .ptr_embedded_in_code => unreachable,
+ .unreach, .none => return, // Nothing to do.
+ .undef => {
+ if (!self.wantSafety())
+ return; // The already existing value will do just fine.
+ // TODO Upgrade this to a memset call when we have that available.
+ switch (ty.abiSize(self.target.*)) {
+ 1 => return self.genSetStack(src, ty, stack_offset, .{ .immediate = 0xaa }),
+ 2 => return self.genSetStack(src, ty, stack_offset, .{ .immediate = 0xaaaa }),
+ 4 => return self.genSetStack(src, ty, stack_offset, .{ .immediate = 0xaaaaaaaa }),
+ 8 => return self.genSetStack(src, ty, stack_offset, .{ .immediate = 0xaaaaaaaaaaaaaaaa }),
+ else => return self.fail(src, "TODO implement memset", .{}),
+ }
+ },
+ .compare_flags_unsigned => |op| {
+ return self.fail(src, "TODO implement set stack variable with compare flags value (unsigned)", .{});
+ },
+ .compare_flags_signed => |op| {
+ return self.fail(src, "TODO implement set stack variable with compare flags value (signed)", .{});
+ },
+ .immediate => {
+ const reg = try self.copyToTmpRegister(src, mcv);
+ return self.genSetStack(src, ty, stack_offset, MCValue{ .register = reg });
+ },
+ .embedded_in_code => |code_offset| {
+ return self.fail(src, "TODO implement set stack variable from embedded_in_code", .{});
+ },
+ .register => |reg| {
+ // TODO: strb, strh
+ if (stack_offset <= math.maxInt(u12)) {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.str(.al, reg, .fp, .{
+ .offset = Instruction.Offset.imm(@intCast(u12, stack_offset)),
+ .positive = false,
+ }).toU32());
+ } else {
+ return self.fail(src, "TODO genSetStack with larger offsets", .{});
+ }
+ },
+ .memory => |vaddr| {
+ return self.fail(src, "TODO implement set stack variable from memory vaddr", .{});
+ },
+ .stack_offset => |off| {
+ if (stack_offset == off)
+ return; // Copy stack variable to itself; nothing to do.
+
+ const reg = try self.copyToTmpRegister(src, mcv);
+ return self.genSetStack(src, ty, stack_offset, MCValue{ .register = reg });
+ },
+ },
.x86_64 => switch (mcv) {
.dead => unreachable,
.ptr_stack_offset => unreachable,
@@ -2274,35 +2359,39 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
return self.genSetReg(src, reg, .{ .immediate = 0xaaaaaaaa });
},
.immediate => |x| {
- // TODO better analysis of x to determine the
- // least amount of necessary instructions (use
- // more intelligent rotating)
- if (x <= math.maxInt(u8)) {
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, reg, Instruction.Operand.imm(@truncate(u8, x), 0)).toU32());
- return;
- } else if (x <= math.maxInt(u16)) {
- // TODO Use movw Note: Not supported on
- // all ARM targets!
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, reg, Instruction.Operand.imm(@truncate(u8, x), 0)).toU32());
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 8), 12)).toU32());
- } else if (x <= math.maxInt(u32)) {
- // TODO Use movw and movt Note: Not
- // supported on all ARM targets! Also TODO
- // write constant to code and load
- // relative to pc
+ if (x > math.maxInt(u32)) return self.fail(src, "ARM registers are 32-bit wide", .{});
- // immediate: 0xaabbccdd
- // mov reg, #0xaa
- // orr reg, reg, #0xbb, 24
- // orr reg, reg, #0xcc, 16
- // orr reg, reg, #0xdd, 8
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, reg, Instruction.Operand.imm(@truncate(u8, x), 0)).toU32());
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 8), 12)).toU32());
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 16), 8)).toU32());
- mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 24), 4)).toU32());
- return;
+ if (Instruction.Operand.fromU32(@intCast(u32, x))) |op| {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, reg, op).toU32());
+ } else if (Instruction.Operand.fromU32(~@intCast(u32, x))) |op| {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mvn(.al, reg, op).toU32());
+ } else if (x <= math.maxInt(u16)) {
+ if (Target.arm.featureSetHas(self.target.cpu.features, .has_v7)) {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.movw(.al, reg, @intCast(u16, x)).toU32());
+ } else {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, reg, Instruction.Operand.imm(@truncate(u8, x), 0)).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 8), 12)).toU32());
+ }
} else {
- return self.fail(src, "ARM registers are 32-bit wide", .{});
+ // TODO write constant to code and load
+ // relative to pc
+ if (Target.arm.featureSetHas(self.target.cpu.features, .has_v7)) {
+ // immediate: 0xaaaabbbb
+ // movw reg, #0xbbbb
+ // movt reg, #0xaaaa
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.movw(.al, reg, @truncate(u16, x)).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.movt(.al, reg, @truncate(u16, x >> 16)).toU32());
+ } else {
+ // immediate: 0xaabbccdd
+ // mov reg, #0xaa
+ // orr reg, reg, #0xbb, 24
+ // orr reg, reg, #0xcc, 16
+ // orr reg, reg, #0xdd, 8
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.mov(.al, reg, Instruction.Operand.imm(@truncate(u8, x), 0)).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 8), 12)).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 16), 8)).toU32());
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.orr(.al, reg, reg, Instruction.Operand.imm(@truncate(u8, x >> 24), 4)).toU32());
+ }
}
},
.register => |src_reg| {
@@ -2319,6 +2408,18 @@ fn Function(comptime arch: std.Target.Cpu.Arch) type {
try self.genSetReg(src, reg, .{ .immediate = addr });
mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.ldr(.al, reg, reg, .{ .offset = Instruction.Offset.none }).toU32());
},
+ .stack_offset => |unadjusted_off| {
+ // TODO: ldrb, ldrh
+ // TODO: maybe addressing from sp instead of fp
+ if (unadjusted_off <= math.maxInt(u12)) {
+ mem.writeIntLittle(u32, try self.code.addManyAsArray(4), Instruction.ldr(.al, reg, .fp, .{
+ .offset = Instruction.Offset.imm(@intCast(u12, unadjusted_off)),
+ .positive = false,
+ }).toU32());
+ } else {
+ return self.fail(src, "TODO genSetReg with larger stack offset", .{});
+ }
+ },
else => return self.fail(src, "TODO implement getSetReg for arm {}", .{mcv}),
},
.riscv64 => switch (mcv) {
diff --git a/src/codegen/arm.zig b/src/codegen/arm.zig
index 36fb858022..33ff789648 100644
--- a/src/codegen/arm.zig
+++ b/src/codegen/arm.zig
@@ -138,6 +138,29 @@ pub const Instruction = union(enum) {
fixed: u2 = 0b00,
cond: u4,
},
+ Multiply: packed struct {
+ rn: u4,
+ fixed_1: u4 = 0b1001,
+ rm: u4,
+ ra: u4,
+ rd: u4,
+ set_cond: u1,
+ accumulate: u1,
+ fixed_2: u6 = 0b000000,
+ cond: u4,
+ },
+ MultiplyLong: packed struct {
+ rn: u4,
+ fixed_1: u4 = 0b1001,
+ rm: u4,
+ rdlo: u4,
+ rdhi: u4,
+ set_cond: u1,
+ accumulate: u1,
+ unsigned: u1,
+ fixed_2: u5 = 0b00001,
+ cond: u4,
+ },
SingleDataTransfer: packed struct {
offset: u12,
rd: u4,
@@ -317,6 +340,29 @@ pub const Instruction = union(enum) {
},
};
}
+
+ /// Tries to convert an unsigned 32 bit integer into an
+ /// immediate operand using rotation. Returns null when there
+ /// is no conversion
+ pub fn fromU32(x: u32) ?Operand {
+ const masks = comptime blk: {
+ const base_mask: u32 = std.math.maxInt(u8);
+ var result = [_]u32{0} ** 16;
+ for (result) |*mask, i| mask.* = std.math.rotr(u32, base_mask, 2 * i);
+ break :blk result;
+ };
+
+ return for (masks) |mask, i| {
+ if (x & mask == x) {
+ break Operand{
+ .Immediate = .{
+ .imm = @intCast(u8, std.math.rotl(u32, x, 2 * i)),
+ .rotate = @intCast(u4, i),
+ },
+ };
+ }
+ } else null;
+ }
};
/// Represents the offset operand of a load or store
@@ -349,7 +395,7 @@ pub const Instruction = union(enum) {
};
}
- pub fn imm(immediate: u8) Offset {
+ pub fn imm(immediate: u12) Offset {
return Offset{
.Immediate = immediate,
};
@@ -380,6 +426,8 @@ pub const Instruction = union(enum) {
pub fn toU32(self: Instruction) u32 {
return switch (self) {
.DataProcessing => |v| @bitCast(u32, v),
+ .Multiply => |v| @bitCast(u32, v),
+ .MultiplyLong => |v| @bitCast(u32, v),
.SingleDataTransfer => |v| @bitCast(u32, v),
.BlockDataTransfer => |v| @bitCast(u32, v),
.Branch => |v| @bitCast(u32, v),
@@ -412,6 +460,70 @@ pub const Instruction = union(enum) {
};
}
+ fn specialMov(
+ cond: Condition,
+ rd: Register,
+ imm: u16,
+ top: bool,
+ ) Instruction {
+ return Instruction{
+ .DataProcessing = .{
+ .cond = @enumToInt(cond),
+ .i = 1,
+ .opcode = if (top) 0b1010 else 0b1000,
+ .s = 0,
+ .rn = @truncate(u4, imm >> 12),
+ .rd = rd.id(),
+ .op2 = @truncate(u12, imm),
+ },
+ };
+ }
+
+ fn multiply(
+ cond: Condition,
+ set_cond: u1,
+ rd: Register,
+ rn: Register,
+ rm: Register,
+ ra: ?Register,
+ ) Instruction {
+ return Instruction{
+ .Multiply = .{
+ .cond = @enumToInt(cond),
+ .accumulate = @boolToInt(ra != null),
+ .set_cond = set_cond,
+ .rd = rd.id(),
+ .rn = rn.id(),
+ .ra = if (ra) |reg| reg.id() else 0b0000,
+ .rm = rm.id(),
+ },
+ };
+ }
+
+ fn multiplyLong(
+ cond: Condition,
+ signed: u1,
+ accumulate: u1,
+ set_cond: u1,
+ rdhi: Register,
+ rdlo: Register,
+ rm: Register,
+ rn: Register,
+ ) Instruction {
+ return Instruction{
+ .MultiplyLong = .{
+ .cond = @enumToInt(cond),
+ .unsigned = signed,
+ .accumulate = accumulate,
+ .set_cond = set_cond,
+ .rdlo = rdlo.id(),
+ .rdhi = rdhi.id(),
+ .rn = rn.id(),
+ .rm = rm.id(),
+ },
+ };
+ }
+
fn singleDataTransfer(
cond: Condition,
rd: Register,
@@ -463,12 +575,12 @@ pub const Instruction = union(enum) {
};
}
- fn branch(cond: Condition, offset: i24, link: u1) Instruction {
+ fn branch(cond: Condition, offset: i26, link: u1) Instruction {
return Instruction{
.Branch = .{
.cond = @enumToInt(cond),
.link = link,
- .offset = @bitCast(u24, offset),
+ .offset = @bitCast(u24, @intCast(i24, offset >> 2)),
},
};
}
@@ -618,10 +730,96 @@ pub const Instruction = union(enum) {
return dataProcessing(cond, .mvn, 1, rd, .r0, op2);
}
+ // movw and movt
+
+ pub fn movw(cond: Condition, rd: Register, imm: u16) Instruction {
+ return specialMov(cond, rd, imm, false);
+ }
+
+ pub fn movt(cond: Condition, rd: Register, imm: u16) Instruction {
+ return specialMov(cond, rd, imm, true);
+ }
+
// PSR transfer
pub fn mrs(cond: Condition, rd: Register, psr: Psr) Instruction {
- return dataProcessing(cond, if (psr == .cpsr) .tst else .cmp, 0, rd, .r15, Operand.reg(.r0, Operand.Shift.none));
+ return Instruction{
+ .DataProcessing = .{
+ .cond = @enumToInt(cond),
+ .i = 0,
+ .opcode = if (psr == .spsr) 0b1010 else 0b1000,
+ .s = 0,
+ .rn = 0b1111,
+ .rd = rd.id(),
+ .op2 = 0b0000_0000_0000,
+ },
+ };
+ }
+
+ pub fn msr(cond: Condition, psr: Psr, op: Operand) Instruction {
+ return Instruction{
+ .DataProcessing = .{
+ .cond = @enumToInt(cond),
+ .i = 0,
+ .opcode = if (psr == .spsr) 0b1011 else 0b1001,
+ .s = 0,
+ .rn = 0b1111,
+ .rd = 0b1111,
+ .op2 = op.toU12(),
+ },
+ };
+ }
+
+ // Multiply
+
+ pub fn mul(cond: Condition, rd: Register, rn: Register, rm: Register) Instruction {
+ return multiply(cond, 0, rd, rn, rm, null);
+ }
+
+ pub fn muls(cond: Condition, rd: Register, rn: Register, rm: Register) Instruction {
+ return multiply(cond, 1, rd, rn, rm, null);
+ }
+
+ pub fn mla(cond: Condition, rd: Register, rn: Register, rm: Register, ra: Register) Instruction {
+ return multiply(cond, 0, rd, rn, rm, ra);
+ }
+
+ pub fn mlas(cond: Condition, rd: Register, rn: Register, rm: Register, ra: Register) Instruction {
+ return multiply(cond, 1, rd, rn, rm, ra);
+ }
+
+ // Multiply long
+
+ pub fn umull(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 0, 0, 0, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn umulls(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 0, 0, 1, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn umlal(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 0, 1, 0, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn umlals(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 0, 1, 1, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn smull(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 1, 0, 0, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn smulls(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 1, 0, 1, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn smlal(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 1, 1, 0, rdhi, rdlo, rm, rn);
+ }
+
+ pub fn smlals(cond: Condition, rdlo: Register, rdhi: Register, rn: Register, rm: Register) Instruction {
+ return multiplyLong(cond, 1, 1, 1, rdhi, rdlo, rm, rn);
}
// Single data transfer
@@ -697,11 +895,11 @@ pub const Instruction = union(enum) {
// Branch
- pub fn b(cond: Condition, offset: i24) Instruction {
+ pub fn b(cond: Condition, offset: i26) Instruction {
return branch(cond, offset, 0);
}
- pub fn bl(cond: Condition, offset: i24) Instruction {
+ pub fn bl(cond: Condition, offset: i26) Instruction {
return branch(cond, offset, 1);
}
@@ -731,6 +929,10 @@ pub const Instruction = union(enum) {
// Aliases
+ pub fn nop() Instruction {
+ return mov(.al, .r0, Instruction.Operand.reg(.r0, Instruction.Operand.Shift.none));
+ }
+
pub fn pop(cond: Condition, args: anytype) Instruction {
if (@typeInfo(@TypeOf(args)) != .Struct) {
@compileError("Expected tuple or struct argument, found " ++ @typeName(@TypeOf(args)));
@@ -805,6 +1007,14 @@ test "serialize instructions" {
.inst = Instruction.mrs(.al, .r5, .cpsr),
.expected = 0b1110_00010_0_001111_0101_000000000000,
},
+ .{ // mul r0, r1, r2
+ .inst = Instruction.mul(.al, .r0, .r1, .r2),
+ .expected = 0b1110_000000_0_0_0000_0000_0010_1001_0001,
+ },
+ .{ // umlal r0, r1, r5, r6
+ .inst = Instruction.umlal(.al, .r0, .r1, .r5, .r6),
+ .expected = 0b1110_00001_0_1_0_0001_0000_0110_1001_0101,
+ },
.{ // ldr r0, [r2, #42]
.inst = Instruction.ldr(.al, .r0, .r2, .{
.offset = Instruction.Offset.imm(42),
@@ -819,11 +1029,11 @@ test "serialize instructions" {
},
.{ // b #12
.inst = Instruction.b(.al, 12),
- .expected = 0b1110_101_0_0000_0000_0000_0000_0000_1100,
+ .expected = 0b1110_101_0_0000_0000_0000_0000_0000_0011,
},
.{ // bl #-4
.inst = Instruction.bl(.al, -4),
- .expected = 0b1110_101_1_1111_1111_1111_1111_1111_1100,
+ .expected = 0b1110_101_1_1111_1111_1111_1111_1111_1111,
},
.{ // bx lr
.inst = Instruction.bx(.al, .lr),