aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--src/link.zig2
-rw-r--r--src/link/Coff.zig213
-rw-r--r--src/link/Coff/Atom.zig4
-rw-r--r--src/link/Coff/ImportTable.zig2
-rw-r--r--src/link/Coff/Relocation.zig72
-rw-r--r--src/link/MachO.zig2
-rw-r--r--src/link/MachO/Atom.zig4
-rw-r--r--src/link/MachO/Relocation.zig20
8 files changed, 134 insertions, 185 deletions
diff --git a/src/link.zig b/src/link.zig
index a7c3ac51bc..239dc646b2 100644
--- a/src/link.zig
+++ b/src/link.zig
@@ -395,6 +395,7 @@ pub const File = struct {
.macos => base.cast(MachO).?.ptraceAttach(pid) catch |err| {
log.warn("attaching failed with error: {s}", .{@errorName(err)});
},
+ .windows => {},
else => return error.HotSwapUnavailableOnHostOperatingSystem,
}
}
@@ -436,6 +437,7 @@ pub const File = struct {
.macos => base.cast(MachO).?.ptraceDetach(pid) catch |err| {
log.warn("detaching failed with error: {s}", .{@errorName(err)});
},
+ .windows => {},
else => return error.HotSwapUnavailableOnHostOperatingSystem,
}
}
diff --git a/src/link/Coff.zig b/src/link/Coff.zig
index df525bba90..7b5539287d 100644
--- a/src/link/Coff.zig
+++ b/src/link/Coff.zig
@@ -49,11 +49,8 @@ imports_count_dirty: bool = true,
/// Virtual address of the entry point procedure relative to image base.
entry_addr: ?u32 = null,
-/// Table of Decls that are currently alive.
-/// We store them here so that we can properly dispose of any allocated
-/// memory within the atom in the incremental linker.
-/// TODO consolidate this.
-decls: std.AutoHashMapUnmanaged(Module.Decl.Index, DeclMetadata) = .{},
+/// Table of tracked Decls.
+decls: std.AutoArrayHashMapUnmanaged(Module.Decl.Index, DeclMetadata) = .{},
/// List of atoms that are either synthetic or map directly to the Zig source program.
atoms: std.ArrayListUnmanaged(Atom) = .{},
@@ -98,9 +95,9 @@ const Entry = struct {
sym_index: u32,
};
-const RelocTable = std.AutoHashMapUnmanaged(Atom.Index, std.ArrayListUnmanaged(Relocation));
-const BaseRelocationTable = std.AutoHashMapUnmanaged(Atom.Index, std.ArrayListUnmanaged(u32));
-const UnnamedConstTable = std.AutoHashMapUnmanaged(Module.Decl.Index, std.ArrayListUnmanaged(Atom.Index));
+const RelocTable = std.AutoArrayHashMapUnmanaged(Atom.Index, std.ArrayListUnmanaged(Relocation));
+const BaseRelocationTable = std.AutoArrayHashMapUnmanaged(Atom.Index, std.ArrayListUnmanaged(u32));
+const UnnamedConstTable = std.AutoArrayHashMapUnmanaged(Module.Decl.Index, std.ArrayListUnmanaged(Atom.Index));
const default_file_alignment: u16 = 0x200;
const default_size_of_stack_reserve: u32 = 0x1000000;
@@ -137,6 +134,10 @@ const DeclMetadata = struct {
/// A list of all exports aliases of this Decl.
exports: std.ArrayListUnmanaged(u32) = .{},
+ fn deinit(m: *DeclMetadata, allocator: Allocator) void {
+ m.exports.deinit(allocator);
+ }
+
fn getExport(m: DeclMetadata, coff_file: *const Coff, name: []const u8) ?u32 {
for (m.exports.items) |exp| {
if (mem.eql(u8, name, coff_file.getSymbolName(.{
@@ -293,39 +294,27 @@ pub fn deinit(self: *Coff) void {
}
self.import_tables.deinit(gpa);
- {
- var it = self.decls.iterator();
- while (it.next()) |entry| {
- entry.value_ptr.exports.deinit(gpa);
- }
- self.decls.deinit(gpa);
+ for (self.decls.values()) |*metadata| {
+ metadata.deinit(gpa);
}
+ self.decls.deinit(gpa);
self.atom_by_index_table.deinit(gpa);
- {
- var it = self.unnamed_const_atoms.valueIterator();
- while (it.next()) |atoms| {
- atoms.deinit(gpa);
- }
- self.unnamed_const_atoms.deinit(gpa);
+ for (self.unnamed_const_atoms.values()) |*atoms| {
+ atoms.deinit(gpa);
}
+ self.unnamed_const_atoms.deinit(gpa);
- {
- var it = self.relocs.valueIterator();
- while (it.next()) |relocs| {
- relocs.deinit(gpa);
- }
- self.relocs.deinit(gpa);
+ for (self.relocs.values()) |*relocs| {
+ relocs.deinit(gpa);
}
+ self.relocs.deinit(gpa);
- {
- var it = self.base_relocs.valueIterator();
- while (it.next()) |relocs| {
- relocs.deinit(gpa);
- }
- self.base_relocs.deinit(gpa);
+ for (self.base_relocs.values()) |*relocs| {
+ relocs.deinit(gpa);
}
+ self.base_relocs.deinit(gpa);
}
fn populateMissingMetadata(self: *Coff) !void {
@@ -455,7 +444,44 @@ fn allocateSection(self: *Coff, name: []const u8, size: u32, flags: coff.Section
return index;
}
-fn growSectionVM(self: *Coff, sect_id: u32, needed_size: u32) !void {
+fn growSection(self: *Coff, sect_id: u32, needed_size: u32) !void {
+ const header = &self.sections.items(.header)[sect_id];
+ const maybe_last_atom_index = self.sections.items(.last_atom_index)[sect_id];
+ const sect_capacity = self.allocatedSize(header.pointer_to_raw_data);
+
+ if (needed_size > sect_capacity) {
+ const new_offset = self.findFreeSpace(needed_size, default_file_alignment);
+ const current_size = if (maybe_last_atom_index) |last_atom_index| blk: {
+ const last_atom = self.getAtom(last_atom_index);
+ const sym = last_atom.getSymbol(self);
+ break :blk (sym.value + last_atom.size) - header.virtual_address;
+ } else 0;
+ log.debug("moving {s} from 0x{x} to 0x{x}", .{
+ self.getSectionName(header),
+ header.pointer_to_raw_data,
+ new_offset,
+ });
+ const amt = try self.base.file.?.copyRangeAll(
+ header.pointer_to_raw_data,
+ self.base.file.?,
+ new_offset,
+ current_size,
+ );
+ if (amt != current_size) return error.InputOutput;
+ header.pointer_to_raw_data = new_offset;
+ }
+
+ const sect_vm_capacity = self.allocatedVirtualSize(header.virtual_address);
+ if (needed_size > sect_vm_capacity) {
+ try self.growSectionVirtualMemory(sect_id, needed_size);
+ self.markRelocsDirtyByAddress(header.virtual_address + needed_size);
+ }
+
+ header.virtual_size = @max(header.virtual_size, needed_size);
+ header.size_of_raw_data = needed_size;
+}
+
+fn growSectionVirtualMemory(self: *Coff, sect_id: u32, needed_size: u32) !void {
const header = &self.sections.items(.header)[sect_id];
const increased_size = padToIdeal(needed_size);
const old_aligned_end = header.virtual_address + mem.alignForwardGeneric(u32, header.virtual_size, self.page_size);
@@ -562,38 +588,8 @@ fn allocateAtom(self: *Coff, atom_index: Atom.Index, new_atom_size: u32, alignme
else
true;
if (expand_section) {
- const sect_capacity = self.allocatedSize(header.pointer_to_raw_data);
const needed_size: u32 = (vaddr + new_atom_size) - header.virtual_address;
- if (needed_size > sect_capacity) {
- const new_offset = self.findFreeSpace(needed_size, default_file_alignment);
- const current_size = if (maybe_last_atom_index.*) |last_atom_index| blk: {
- const last_atom = self.getAtom(last_atom_index);
- const sym = last_atom.getSymbol(self);
- break :blk (sym.value + last_atom.size) - header.virtual_address;
- } else 0;
- log.debug("moving {s} from 0x{x} to 0x{x}", .{
- self.getSectionName(header),
- header.pointer_to_raw_data,
- new_offset,
- });
- const amt = try self.base.file.?.copyRangeAll(
- header.pointer_to_raw_data,
- self.base.file.?,
- new_offset,
- current_size,
- );
- if (amt != current_size) return error.InputOutput;
- header.pointer_to_raw_data = new_offset;
- }
-
- const sect_vm_capacity = self.allocatedVirtualSize(header.virtual_address);
- if (needed_size > sect_vm_capacity) {
- try self.growSectionVM(sect_id, needed_size);
- self.markRelocsDirtyByAddress(header.virtual_address + needed_size);
- }
-
- header.virtual_size = @max(header.virtual_size, needed_size);
- header.size_of_raw_data = needed_size;
+ try self.growSection(sect_id, needed_size);
maybe_last_atom_index.* = atom_index;
}
@@ -771,7 +767,7 @@ fn shrinkAtom(self: *Coff, atom_index: Atom.Index, new_block_size: u32) void {
// capacity, insert a free list node for it.
}
-fn writeAtom(self: *Coff, atom_index: Atom.Index, code: []const u8) !void {
+fn writeAtom(self: *Coff, atom_index: Atom.Index, code: []u8) !void {
const atom = self.getAtom(atom_index);
const sym = atom.getSymbol(self);
const section = self.sections.get(@enumToInt(sym.section_number) - 1);
@@ -781,8 +777,8 @@ fn writeAtom(self: *Coff, atom_index: Atom.Index, code: []const u8) !void {
file_offset,
file_offset + code.len,
});
+ self.resolveRelocs(atom_index, code);
try self.base.file.?.pwriteAll(code, file_offset);
- try self.resolveRelocs(atom_index);
}
fn writePtrWidthAtom(self: *Coff, atom_index: Atom.Index) !void {
@@ -800,8 +796,7 @@ fn writePtrWidthAtom(self: *Coff, atom_index: Atom.Index) !void {
fn markRelocsDirtyByTarget(self: *Coff, target: SymbolWithLoc) void {
// TODO: reverse-lookup might come in handy here
- var it = self.relocs.valueIterator();
- while (it.next()) |relocs| {
+ for (self.relocs.values()) |*relocs| {
for (relocs.items) |*reloc| {
if (!reloc.target.eql(target)) continue;
reloc.dirty = true;
@@ -810,8 +805,7 @@ fn markRelocsDirtyByTarget(self: *Coff, target: SymbolWithLoc) void {
}
fn markRelocsDirtyByAddress(self: *Coff, addr: u32) void {
- var it = self.relocs.valueIterator();
- while (it.next()) |relocs| {
+ for (self.relocs.values()) |*relocs| {
for (relocs.items) |*reloc| {
const target_vaddr = reloc.getTargetAddress(self) orelse continue;
if (target_vaddr < addr) continue;
@@ -820,14 +814,16 @@ fn markRelocsDirtyByAddress(self: *Coff, addr: u32) void {
}
}
-fn resolveRelocs(self: *Coff, atom_index: Atom.Index) !void {
- const relocs = self.relocs.get(atom_index) orelse return;
+fn resolveRelocs(self: *Coff, atom_index: Atom.Index, code: []u8) void {
+ const relocs = self.relocs.getPtr(atom_index) orelse return;
log.debug("relocating '{s}'", .{self.getAtom(atom_index).getName(self)});
for (relocs.items) |*reloc| {
if (!reloc.dirty) continue;
- try reloc.resolve(atom_index, self);
+ if (reloc.resolve(atom_index, code, self)) {
+ reloc.dirty = false;
+ }
}
}
@@ -944,7 +940,7 @@ pub fn updateFunc(self: *Coff, module: *Module, func: *Module.Fn, air: Air, live
&code_buffer,
.none,
);
- const code = switch (res) {
+ var code = switch (res) {
.ok => code_buffer.items,
.fail => |em| {
decl.analysis = .codegen_failure;
@@ -994,7 +990,7 @@ pub fn lowerUnnamedConst(self: *Coff, tv: TypedValue, decl_index: Module.Decl.In
const res = try codegen.generateSymbol(&self.base, decl.srcLoc(), tv, &code_buffer, .none, .{
.parent_atom_index = self.getAtom(atom_index).getSymbolIndex().?,
});
- const code = switch (res) {
+ var code = switch (res) {
.ok => code_buffer.items,
.fail => |em| {
decl.analysis = .codegen_failure;
@@ -1057,7 +1053,7 @@ pub fn updateDecl(self: *Coff, module: *Module, decl_index: Module.Decl.Index) !
}, &code_buffer, .none, .{
.parent_atom_index = atom.getSymbolIndex().?,
});
- const code = switch (res) {
+ var code = switch (res) {
.ok => code_buffer.items,
.fail => |em| {
decl.analysis = .codegen_failure;
@@ -1110,7 +1106,7 @@ fn getDeclOutputSection(self: *Coff, decl_index: Module.Decl.Index) u16 {
return index;
}
-fn updateDeclCode(self: *Coff, decl_index: Module.Decl.Index, code: []const u8, complex_type: coff.ComplexType) !void {
+fn updateDeclCode(self: *Coff, decl_index: Module.Decl.Index, code: []u8, complex_type: coff.ComplexType) !void {
const gpa = self.base.allocator;
const mod = self.base.options.module.?;
const decl = mod.declPtr(decl_index);
@@ -1195,7 +1191,7 @@ pub fn freeDecl(self: *Coff, decl_index: Module.Decl.Index) void {
log.debug("freeDecl {*}", .{decl});
- if (self.decls.fetchRemove(decl_index)) |const_kv| {
+ if (self.decls.fetchOrderedRemove(decl_index)) |const_kv| {
var kv = const_kv;
self.freeAtom(kv.value.atom);
self.freeUnnamedConsts(decl_index);
@@ -1422,12 +1418,29 @@ pub fn flushModule(self: *Coff, comp: *Compilation, prog_node: *std.Progress.Nod
}
try self.writeImportTables();
- {
- var it = self.relocs.keyIterator();
- while (it.next()) |atom| {
- try self.resolveRelocs(atom.*);
- }
+
+ for (self.relocs.keys(), self.relocs.values()) |atom_index, relocs| {
+ const needs_update = for (relocs.items) |reloc| {
+ if (reloc.dirty) break true;
+ } else false;
+
+ if (!needs_update) continue;
+
+ const atom = self.getAtom(atom_index);
+ const sym = atom.getSymbol(self);
+ const section = self.sections.get(@enumToInt(sym.section_number) - 1).header;
+ const file_offset = section.pointer_to_raw_data + sym.value - section.virtual_address;
+
+ var code = std.ArrayList(u8).init(gpa);
+ defer code.deinit();
+ try code.resize(math.cast(usize, atom.size) orelse return error.Overflow);
+
+ const amt = try self.base.file.?.preadAll(code.items, file_offset);
+ if (amt != code.items.len) return error.InputOutput;
+
+ try self.writeAtom(atom_index, code.items);
}
+
try self.writeBaseRelocations();
if (self.getEntryPoint()) |entry_sym_loc| {
@@ -1576,25 +1589,8 @@ fn writeBaseRelocations(self: *Coff) !void {
}
const header = &self.sections.items(.header)[self.reloc_section_index.?];
- const sect_capacity = self.allocatedSize(header.pointer_to_raw_data);
const needed_size = @intCast(u32, buffer.items.len);
- if (needed_size > sect_capacity) {
- const new_offset = self.findFreeSpace(needed_size, default_file_alignment);
- log.debug("moving {s} from 0x{x} to 0x{x}", .{
- self.getSectionName(header),
- header.pointer_to_raw_data,
- new_offset,
- });
- header.pointer_to_raw_data = new_offset;
-
- const sect_vm_capacity = self.allocatedVirtualSize(header.virtual_address);
- if (needed_size > sect_vm_capacity) {
- // TODO: we want to enforce .reloc after every alloc section.
- try self.growSectionVM(self.reloc_section_index.?, needed_size);
- }
- }
- header.virtual_size = @max(header.virtual_size, needed_size);
- header.size_of_raw_data = needed_size;
+ try self.growSection(self.reloc_section_index.?, needed_size);
try self.base.file.?.pwriteAll(buffer.items, header.pointer_to_raw_data);
@@ -1633,20 +1629,7 @@ fn writeImportTables(self: *Coff) !void {
}
const needed_size = iat_size + dir_table_size + lookup_table_size + names_table_size + dll_names_size;
- const sect_capacity = self.allocatedSize(header.pointer_to_raw_data);
- if (needed_size > sect_capacity) {
- const new_offset = self.findFreeSpace(needed_size, default_file_alignment);
- log.debug("moving .idata from 0x{x} to 0x{x}", .{ header.pointer_to_raw_data, new_offset });
- header.pointer_to_raw_data = new_offset;
-
- const sect_vm_capacity = self.allocatedVirtualSize(header.virtual_address);
- if (needed_size > sect_vm_capacity) {
- try self.growSectionVM(self.idata_section_index.?, needed_size);
- }
-
- header.virtual_size = @max(header.virtual_size, needed_size);
- header.size_of_raw_data = needed_size;
- }
+ try self.growSection(self.idata_section_index.?, needed_size);
// Do the actual writes
var buffer = std.ArrayList(u8).init(gpa);
diff --git a/src/link/Coff/Atom.zig b/src/link/Coff/Atom.zig
index 80c04a8fa1..4afc1f7cb2 100644
--- a/src/link/Coff/Atom.zig
+++ b/src/link/Coff/Atom.zig
@@ -121,8 +121,8 @@ pub fn addBaseRelocation(coff_file: *Coff, atom_index: Index, offset: u32) !void
pub fn freeRelocations(coff_file: *Coff, atom_index: Index) void {
const gpa = coff_file.base.allocator;
- var removed_relocs = coff_file.relocs.fetchRemove(atom_index);
+ var removed_relocs = coff_file.relocs.fetchOrderedRemove(atom_index);
if (removed_relocs) |*relocs| relocs.value.deinit(gpa);
- var removed_base_relocs = coff_file.base_relocs.fetchRemove(atom_index);
+ var removed_base_relocs = coff_file.base_relocs.fetchOrderedRemove(atom_index);
if (removed_base_relocs) |*base_relocs| base_relocs.value.deinit(gpa);
}
diff --git a/src/link/Coff/ImportTable.zig b/src/link/Coff/ImportTable.zig
index ba58af2fe0..c3ba77e855 100644
--- a/src/link/Coff/ImportTable.zig
+++ b/src/link/Coff/ImportTable.zig
@@ -121,7 +121,7 @@ pub fn fmtDebug(itab: ImportTable, ctx: Context) std.fmt.Formatter(fmt) {
return .{ .data = .{ .itab = itab, .ctx = ctx } };
}
-const ImportIndex = u32;
+pub const ImportIndex = u32;
const ImportTable = @This();
const std = @import("std");
diff --git a/src/link/Coff/Relocation.zig b/src/link/Coff/Relocation.zig
index 45f3a97052..37bd3e292f 100644
--- a/src/link/Coff/Relocation.zig
+++ b/src/link/Coff/Relocation.zig
@@ -72,62 +72,50 @@ pub fn getTargetAddress(self: Relocation, coff_file: *const Coff) ?u32 {
}
}
-pub fn resolve(self: *Relocation, atom_index: Atom.Index, coff_file: *Coff) !void {
+/// Returns `false` if obtaining the target address has been deferred until `flushModule`.
+/// This can happen when trying to resolve address of an import table entry ahead of time.
+pub fn resolve(self: Relocation, atom_index: Atom.Index, code: []u8, coff_file: *Coff) bool {
const atom = coff_file.getAtom(atom_index);
const source_sym = atom.getSymbol(coff_file);
- const source_section = coff_file.sections.get(@enumToInt(source_sym.section_number) - 1).header;
const source_vaddr = source_sym.value + self.offset;
- const file_offset = source_section.pointer_to_raw_data + source_sym.value - source_section.virtual_address;
-
- const target_vaddr = self.getTargetAddress(coff_file) orelse return;
+ const target_vaddr = self.getTargetAddress(coff_file) orelse return false;
const target_vaddr_with_addend = target_vaddr + self.addend;
- log.debug(" ({x}: [() => 0x{x} ({s})) ({s}) (in file at 0x{x})", .{
+ log.debug(" ({x}: [() => 0x{x} ({s})) ({s}) ", .{
source_vaddr,
target_vaddr_with_addend,
coff_file.getSymbolName(self.target),
@tagName(self.type),
- file_offset + self.offset,
});
const ctx: Context = .{
.source_vaddr = source_vaddr,
.target_vaddr = target_vaddr_with_addend,
- .file_offset = file_offset,
.image_base = coff_file.getImageBase(),
+ .code = code,
+ .ptr_width = coff_file.ptr_width,
};
switch (coff_file.base.options.target.cpu.arch) {
- .aarch64 => try self.resolveAarch64(ctx, coff_file),
- .x86, .x86_64 => try self.resolveX86(ctx, coff_file),
+ .aarch64 => self.resolveAarch64(ctx),
+ .x86, .x86_64 => self.resolveX86(ctx),
else => unreachable, // unhandled target architecture
}
- self.dirty = false;
+ return true;
}
const Context = struct {
source_vaddr: u32,
target_vaddr: u32,
- file_offset: u32,
image_base: u64,
+ code: []u8,
+ ptr_width: Coff.PtrWidth,
};
-fn resolveAarch64(self: Relocation, ctx: Context, coff_file: *Coff) !void {
- var buffer: [@sizeOf(u64)]u8 = undefined;
- switch (self.length) {
- 2 => {
- const amt = try coff_file.base.file.?.preadAll(buffer[0..4], ctx.file_offset + self.offset);
- if (amt != 4) return error.InputOutput;
- },
- 3 => {
- const amt = try coff_file.base.file.?.preadAll(&buffer, ctx.file_offset + self.offset);
- if (amt != 8) return error.InputOutput;
- },
- else => unreachable,
- }
-
+fn resolveAarch64(self: Relocation, ctx: Context) void {
+ var buffer = ctx.code[self.offset..];
switch (self.type) {
.got_page, .import_page, .page => {
const source_page = @intCast(i32, ctx.source_vaddr >> 12);
@@ -188,7 +176,7 @@ fn resolveAarch64(self: Relocation, ctx: Context, coff_file: *Coff) !void {
buffer[0..4],
@truncate(u32, ctx.target_vaddr + ctx.image_base),
),
- 3 => mem.writeIntLittle(u64, &buffer, ctx.target_vaddr + ctx.image_base),
+ 3 => mem.writeIntLittle(u64, buffer[0..8], ctx.target_vaddr + ctx.image_base),
else => unreachable,
}
},
@@ -196,15 +184,10 @@ fn resolveAarch64(self: Relocation, ctx: Context, coff_file: *Coff) !void {
.got => unreachable,
.import => unreachable,
}
-
- switch (self.length) {
- 2 => try coff_file.base.file.?.pwriteAll(buffer[0..4], ctx.file_offset + self.offset),
- 3 => try coff_file.base.file.?.pwriteAll(&buffer, ctx.file_offset + self.offset),
- else => unreachable,
- }
}
-fn resolveX86(self: Relocation, ctx: Context, coff_file: *Coff) !void {
+fn resolveX86(self: Relocation, ctx: Context) void {
+ var buffer = ctx.code[self.offset..];
switch (self.type) {
.got_page => unreachable,
.got_pageoff => unreachable,
@@ -216,26 +199,17 @@ fn resolveX86(self: Relocation, ctx: Context, coff_file: *Coff) !void {
.got, .import => {
assert(self.pcrel);
const disp = @intCast(i32, ctx.target_vaddr) - @intCast(i32, ctx.source_vaddr) - 4;
- try coff_file.base.file.?.pwriteAll(mem.asBytes(&disp), ctx.file_offset + self.offset);
+ mem.writeIntLittle(i32, buffer[0..4], disp);
},
.direct => {
if (self.pcrel) {
const disp = @intCast(i32, ctx.target_vaddr) - @intCast(i32, ctx.source_vaddr) - 4;
- try coff_file.base.file.?.pwriteAll(mem.asBytes(&disp), ctx.file_offset + self.offset);
- } else switch (coff_file.ptr_width) {
- .p32 => try coff_file.base.file.?.pwriteAll(
- mem.asBytes(&@intCast(u32, ctx.target_vaddr + ctx.image_base)),
- ctx.file_offset + self.offset,
- ),
+ mem.writeIntLittle(i32, buffer[0..4], disp);
+ } else switch (ctx.ptr_width) {
+ .p32 => mem.writeIntLittle(u32, buffer[0..4], @intCast(u32, ctx.target_vaddr + ctx.image_base)),
.p64 => switch (self.length) {
- 2 => try coff_file.base.file.?.pwriteAll(
- mem.asBytes(&@truncate(u32, ctx.target_vaddr + ctx.image_base)),
- ctx.file_offset + self.offset,
- ),
- 3 => try coff_file.base.file.?.pwriteAll(
- mem.asBytes(&(ctx.target_vaddr + ctx.image_base)),
- ctx.file_offset + self.offset,
- ),
+ 2 => mem.writeIntLittle(u32, buffer[0..4], @truncate(u32, ctx.target_vaddr + ctx.image_base)),
+ 3 => mem.writeIntLittle(u64, buffer[0..8], ctx.target_vaddr + ctx.image_base),
else => unreachable,
},
}
diff --git a/src/link/MachO.zig b/src/link/MachO.zig
index 885afd37c3..3206b63fa3 100644
--- a/src/link/MachO.zig
+++ b/src/link/MachO.zig
@@ -1091,7 +1091,7 @@ pub fn writeAtom(self: *MachO, atom_index: Atom.Index, code: []u8) !void {
log.debug("writing atom for symbol {s} at file offset 0x{x}", .{ atom.getName(self), file_offset });
if (self.relocs.get(atom_index)) |relocs| {
- try Atom.resolveRelocations(self, atom_index, relocs.items, code);
+ Atom.resolveRelocations(self, atom_index, relocs.items, code);
}
if (is_hot_update_compatible) {
diff --git a/src/link/MachO/Atom.zig b/src/link/MachO/Atom.zig
index 36511dfd78..698e961db4 100644
--- a/src/link/MachO/Atom.zig
+++ b/src/link/MachO/Atom.zig
@@ -183,11 +183,11 @@ pub fn addLazyBinding(macho_file: *MachO, atom_index: Index, binding: Binding) !
try gop.value_ptr.append(gpa, binding);
}
-pub fn resolveRelocations(macho_file: *MachO, atom_index: Index, relocs: []Relocation, code: []u8) !void {
+pub fn resolveRelocations(macho_file: *MachO, atom_index: Index, relocs: []Relocation, code: []u8) void {
log.debug("relocating '{s}'", .{macho_file.getAtom(atom_index).getName(macho_file)});
for (relocs) |*reloc| {
if (!reloc.dirty) continue;
- try reloc.resolve(macho_file, atom_index, code);
+ reloc.resolve(macho_file, atom_index, code);
reloc.dirty = false;
}
}
diff --git a/src/link/MachO/Relocation.zig b/src/link/MachO/Relocation.zig
index 6beae8e8ea..cc565742d1 100644
--- a/src/link/MachO/Relocation.zig
+++ b/src/link/MachO/Relocation.zig
@@ -50,7 +50,7 @@ pub fn getTargetAtomIndex(self: Relocation, macho_file: *MachO) ?Atom.Index {
return macho_file.getAtomIndexForSymbol(self.target);
}
-pub fn resolve(self: Relocation, macho_file: *MachO, atom_index: Atom.Index, code: []u8) !void {
+pub fn resolve(self: Relocation, macho_file: *MachO, atom_index: Atom.Index, code: []u8) void {
const arch = macho_file.base.options.target.cpu.arch;
const atom = macho_file.getAtom(atom_index);
const source_sym = atom.getSymbol(macho_file);
@@ -68,18 +68,13 @@ pub fn resolve(self: Relocation, macho_file: *MachO, atom_index: Atom.Index, cod
});
switch (arch) {
- .aarch64 => return self.resolveAarch64(source_addr, target_addr, code),
- .x86_64 => return self.resolveX8664(source_addr, target_addr, code),
+ .aarch64 => self.resolveAarch64(source_addr, target_addr, code),
+ .x86_64 => self.resolveX8664(source_addr, target_addr, code),
else => unreachable,
}
}
-fn resolveAarch64(
- self: Relocation,
- source_addr: u64,
- target_addr: i64,
- code: []u8,
-) !void {
+fn resolveAarch64(self: Relocation, source_addr: u64, target_addr: i64, code: []u8) void {
const rel_type = @intToEnum(macho.reloc_type_arm64, self.type);
if (rel_type == .ARM64_RELOC_UNSIGNED) {
return switch (self.length) {
@@ -212,12 +207,7 @@ fn resolveAarch64(
}
}
-fn resolveX8664(
- self: Relocation,
- source_addr: u64,
- target_addr: i64,
- code: []u8,
-) !void {
+fn resolveX8664(self: Relocation, source_addr: u64, target_addr: i64, code: []u8) void {
const rel_type = @intToEnum(macho.reloc_type_x86_64, self.type);
switch (rel_type) {
.X86_64_RELOC_BRANCH,