Skip to content

Commit

Permalink
elf+aarch64: actually commit thunks.zig file
Browse files Browse the repository at this point in the history
  • Loading branch information
kubkon committed Feb 28, 2024
1 parent 5515dd7 commit 11b706a
Showing 1 changed file with 240 additions and 0 deletions.
240 changes: 240 additions & 0 deletions src/Elf/thunks.zig
Original file line number Diff line number Diff line change
@@ -0,0 +1,240 @@
pub fn createThunks(shndx: u32, elf_file: *Elf) !void {
const gpa = elf_file.base.allocator;
const cpu_arch = elf_file.options.cpu_arch.?;
const slice = elf_file.sections.slice();
const shdr = &slice.items(.shdr)[shndx];
const atoms = slice.items(.atoms)[shndx].items;
assert(atoms.len > 0);

for (atoms) |atom_index| {
elf_file.getAtom(atom_index).?.value = @bitCast(@as(i64, -1));
}

var i: usize = 0;
while (i < atoms.len) {
const start = i;
const start_atom = elf_file.getAtom(atoms[start]).?;
assert(start_atom.flags.alive);
start_atom.value = try advance(shdr, start_atom.size, start_atom.alignment);
i += 1;

while (i < atoms.len and
shdr.sh_size - start_atom.value < maxAllowedDistance(cpu_arch)) : (i += 1)
{
const atom_index = atoms[i];
const atom = elf_file.getAtom(atom_index).?;
assert(atom.flags.alive);
atom.value = try advance(shdr, atom.size, atom.alignment);
}

// Insert a thunk at the group end
const thunk_index = try elf_file.addThunk();
const thunk = elf_file.getThunk(thunk_index);
thunk.out_shndx = shndx;

// Scan relocs in the group and create trampolines for any unreachable callsite
for (atoms[start..i]) |atom_index| {
const atom = elf_file.getAtom(atom_index).?;
const object = atom.getObject(elf_file);
log.debug("atom({d}) {s}", .{ atom_index, atom.getName(elf_file) });
for (atom.getRelocs(elf_file)) |rel| {
const is_reachable = switch (cpu_arch) {
.aarch64 => aarch64.isReachable(atom, rel, elf_file),
.x86_64, .riscv64 => unreachable,
else => @panic("unsupported arch"),
};
if (is_reachable) continue;
const target = object.symbols.items[rel.r_sym()];
try thunk.symbols.put(gpa, target, {});
}
atom.thunk_index = thunk_index;
}

thunk.value = try advance(shdr, thunk.size(elf_file), 2);

log.debug("thunk({d}) : {}", .{ thunk_index, thunk.fmt(elf_file) });
}
}

fn advance(shdr: *elf.Elf64_Shdr, size: u64, pow2_align: u8) !u64 {
const alignment = try math.powi(u32, 2, pow2_align);
const offset = mem.alignForward(u64, shdr.sh_size, alignment);
const padding = offset - shdr.sh_size;
shdr.sh_size += padding + size;
shdr.sh_addralign = @max(shdr.sh_addralign, alignment);
return offset;
}

/// A branch will need an extender if its target is larger than
/// `2^(jump_bits - 1) - margin` where margin is some arbitrary number.
fn maxAllowedDistance(cpu_arch: std.Target.Cpu.Arch) u32 {
return switch (cpu_arch) {
.aarch64 => 0x500_000,
.x86_64, .riscv64 => unreachable,
else => @panic("unhandled arch"),
};
}

pub const Thunk = struct {
value: u64 = 0,
out_shndx: u32 = 0,
symbols: std.AutoArrayHashMapUnmanaged(Symbol.Index, void) = .{},
output_symtab_ctx: Elf.SymtabCtx = .{},

pub fn deinit(thunk: *Thunk, allocator: Allocator) void {
thunk.symbols.deinit(allocator);
}

pub fn size(thunk: Thunk, elf_file: *Elf) usize {
const cpu_arch = elf_file.options.cpu_arch.?;
return thunk.symbols.keys().len * trampolineSize(cpu_arch);
}

pub fn getAddress(thunk: Thunk, elf_file: *Elf) u64 {
const shdr = elf_file.sections.items(.shdr)[thunk.out_shndx];
return shdr.sh_addr + thunk.value;
}

pub fn getTargetAddress(thunk: Thunk, sym_index: Symbol.Index, elf_file: *Elf) u64 {
const cpu_arch = elf_file.options.cpu_arch.?;
return thunk.getAddress(elf_file) + thunk.symbols.getIndex(sym_index).? * trampolineSize(cpu_arch);
}

pub fn write(thunk: Thunk, elf_file: *Elf, writer: anytype) !void {
switch (elf_file.options.cpu_arch.?) {
.aarch64 => try aarch64.write(thunk, elf_file, writer),
.x86_64, .riscv64 => unreachable,
else => @panic("unhandled arch"),
}
}

pub fn calcSymtabSize(thunk: *Thunk, elf_file: *Elf) void {
if (elf_file.options.strip_all) return;

thunk.output_symtab_ctx.nlocals = @as(u32, @intCast(thunk.symbols.keys().len));
for (thunk.symbols.keys()) |sym_index| {
const sym = elf_file.getSymbol(sym_index);
thunk.output_symtab_ctx.strsize += @as(u32, @intCast(sym.getName(elf_file).len + "$thunk".len + 1));
}
}

pub fn writeSymtab(thunk: Thunk, elf_file: *Elf) void {
if (elf_file.options.strip_all) return;
const cpu_arch = elf_file.options.cpu_arch.?;

for (thunk.symbols.keys(), thunk.output_symtab_ctx.ilocal..) |sym_index, ilocal| {
const sym = elf_file.getSymbol(sym_index);
const st_name = @as(u32, @intCast(elf_file.strtab.items.len));
elf_file.strtab.appendSliceAssumeCapacity(sym.getName(elf_file));
elf_file.strtab.appendSliceAssumeCapacity("$thunk");
elf_file.strtab.appendAssumeCapacity(0);
elf_file.symtab.items[ilocal] = .{
.st_name = st_name,
.st_info = elf.STT_FUNC,
.st_other = 0,
.st_shndx = @intCast(thunk.out_shndx),
.st_value = thunk.getTargetAddress(sym_index, elf_file),
.st_size = trampolineSize(cpu_arch),
};
}
}

fn trampolineSize(cpu_arch: std.Target.Cpu.Arch) usize {
return switch (cpu_arch) {
.aarch64 => aarch64.trampoline_size,
.x86_64, .riscv64 => unreachable,
else => @panic("unhandled arch"),
};
}

pub fn format(
thunk: Thunk,
comptime unused_fmt_string: []const u8,
options: std.fmt.FormatOptions,
writer: anytype,
) !void {
_ = thunk;
_ = unused_fmt_string;
_ = options;
_ = writer;
@compileError("do not format Thunk directly");
}

pub fn fmt(thunk: Thunk, elf_file: *Elf) std.fmt.Formatter(format2) {
return .{ .data = .{
.thunk = thunk,
.elf_file = elf_file,
} };
}

const FormatContext = struct {
thunk: Thunk,
elf_file: *Elf,
};

fn format2(
ctx: FormatContext,
comptime unused_fmt_string: []const u8,
options: std.fmt.FormatOptions,
writer: anytype,
) !void {
_ = options;
_ = unused_fmt_string;
const thunk = ctx.thunk;
const elf_file = ctx.elf_file;
try writer.print("@{x} : size({x})\n", .{ thunk.value, thunk.size(elf_file) });
for (thunk.symbols.keys()) |index| {
const sym = elf_file.getSymbol(index);
try writer.print(" %{d} : {s} : @{x}\n", .{ index, sym.getName(elf_file), sym.value });
}
}

pub const Index = u32;
};

const aarch64 = struct {
fn isReachable(atom: *const Atom, rel: elf.Elf64_Rela, elf_file: *Elf) bool {
const r_type: elf.R_AARCH64 = @enumFromInt(rel.r_type());
if (r_type != .CALL26 and r_type != .JUMP26) return true;
const object = atom.getObject(elf_file);
const target = object.getSymbol(rel.r_sym(), elf_file);
if (target.flags.plt) return false;
if (atom.out_shndx != target.shndx) return false;
const target_atom = target.getAtom(elf_file).?;
if (target_atom.value == @as(u64, @bitCast(@as(i64, -1)))) return false;
const saddr = @as(i64, @intCast(atom.getAddress(elf_file) + rel.r_offset));
const taddr: i64 = @intCast(target.getAddress(.{}, elf_file));
_ = math.cast(i28, taddr + rel.r_addend - saddr) orelse return false;
return true;
}

fn write(thunk: Thunk, elf_file: *Elf, writer: anytype) !void {
for (thunk.symbols.keys(), 0..) |sym_index, i| {
const sym = elf_file.getSymbol(sym_index);
const saddr = thunk.getAddress(elf_file) + i * trampoline_size;
const taddr = sym.getAddress(.{}, elf_file);
const pages = try util.calcNumberOfPages(saddr, taddr);
try writer.writeInt(u32, Instruction.adrp(.x16, pages).toU32(), .little);
const off: u12 = @truncate(taddr);
try writer.writeInt(u32, Instruction.add(.x16, .x16, off, false).toU32(), .little);
try writer.writeInt(u32, Instruction.br(.x16).toU32(), .little);
}
}

const trampoline_size = 3 * @sizeOf(u32);

const util = @import("../aarch64.zig");
const Instruction = util.Instruction;
};

const assert = std.debug.assert;
const elf = std.elf;
const log = std.log.scoped(.elf);
const math = std.math;
const mem = std.mem;
const std = @import("std");

const Allocator = mem.Allocator;
const Atom = @import("Atom.zig");
const Elf = @import("../Elf.zig");
const Symbol = @import("Symbol.zig");

0 comments on commit 11b706a

Please sign in to comment.