aboutsummaryrefslogtreecommitdiff
path: root/src/AstGen.zig
diff options
context:
space:
mode:
authorAndrew Kelley <andrew@ziglang.org>2021-07-10 16:27:23 -0700
committerAndrew Kelley <andrew@ziglang.org>2021-07-20 12:19:14 -0700
commit9918a5fbe3dc910f90f2c60ad74edb51de53e0cf (patch)
treee46e46583051c8e9aa27347d91bafe193b2f0da2 /src/AstGen.zig
parent3c3abaf3907e344305620fb4565e7c1acb0a9c88 (diff)
downloadzig-9918a5fbe3dc910f90f2c60ad74edb51de53e0cf.tar.gz
zig-9918a5fbe3dc910f90f2c60ad74edb51de53e0cf.zip
AstGen: remove unneeded field ref_start_index
Previously, this field was used because the Zir.Inst.Ref encoding supported the concept of references to function parameters. However now thanks to whole-file-astgen, the implementations of indexToRef and refToIndex are trivial addition/subtraction of a comptime const integer.
Diffstat (limited to 'src/AstGen.zig')
-rw-r--r--src/AstGen.zig117
1 files changed, 56 insertions, 61 deletions
diff --git a/src/AstGen.zig b/src/AstGen.zig
index 3fdc097042..19906c94d3 100644
--- a/src/AstGen.zig
+++ b/src/AstGen.zig
@@ -989,7 +989,7 @@ fn suspendExpr(
}
try suspend_scope.setBlockBody(suspend_inst);
- return gz.indexToRef(suspend_inst);
+ return indexToRef(suspend_inst);
}
fn awaitExpr(
@@ -1300,7 +1300,7 @@ fn arrayInitExprRlPtr(
.lhs = result_ptr,
.rhs = index_inst,
});
- elem_ptr_list[i] = gz.refToIndex(elem_ptr).?;
+ elem_ptr_list[i] = refToIndex(elem_ptr).?;
_ = try expr(gz, scope, .{ .ptr = elem_ptr }, elem_init);
}
_ = try gz.addPlNode(.validate_array_init_ptr, node, Zir.Inst.Block{
@@ -1455,7 +1455,7 @@ fn structInitExprRlPtr(
.lhs = result_ptr,
.field_name_start = str_index,
});
- field_ptr_list[i] = gz.refToIndex(field_ptr).?;
+ field_ptr_list[i] = refToIndex(field_ptr).?;
_ = try expr(gz, scope, .{ .ptr = field_ptr }, field_init);
}
_ = try gz.addPlNode(.validate_struct_init_ptr, node, Zir.Inst.Block{
@@ -1489,7 +1489,7 @@ fn structInitExprRlTy(
.name_start = str_index,
});
fields_list[i] = .{
- .field_type = gz.refToIndex(field_ty_inst).?,
+ .field_type = refToIndex(field_ty_inst).?,
.init = try expr(gz, scope, .{ .ty = field_ty_inst }, field_init),
};
}
@@ -1786,7 +1786,7 @@ fn labeledBlockExpr(
}
try block_scope.setBlockBody(block_inst);
- return gz.indexToRef(block_inst);
+ return indexToRef(block_inst);
},
.break_operand => {
// All break operands are values that did not use the result location pointer.
@@ -1800,7 +1800,7 @@ fn labeledBlockExpr(
} else {
try block_scope.setBlockBody(block_inst);
}
- const block_ref = gz.indexToRef(block_inst);
+ const block_ref = indexToRef(block_inst);
switch (rl) {
.ref => return block_ref,
else => return rvalue(gz, rl, block_ref, block_node),
@@ -1878,7 +1878,7 @@ fn unusedResultExpr(gz: *GenZir, scope: *Scope, statement: ast.Node.Index) Inner
// we want to avoid adding the ZIR instruction if possible for performance.
const maybe_unused_result = try expr(gz, scope, .none, statement);
var noreturn_src_node: ast.Node.Index = 0;
- const elide_check = if (gz.refToIndex(maybe_unused_result)) |inst| b: {
+ const elide_check = if (refToIndex(maybe_unused_result)) |inst| b: {
// Note that this array becomes invalid after appending more items to it
// in the above while loop.
const zir_tags = gz.astgen.instructions.items(.tag);
@@ -2440,7 +2440,7 @@ fn varDecl(
// the alloc instruction and the store_to_block_ptr instruction.
try parent_zir.ensureUnusedCapacity(gpa, init_scope.instructions.items.len);
for (init_scope.instructions.items) |src_inst| {
- if (gz.indexToRef(src_inst) == init_scope.rl_ptr) continue;
+ if (indexToRef(src_inst) == init_scope.rl_ptr) continue;
if (zir_tags[src_inst] == .store_to_block_ptr) {
if (zir_datas[src_inst].bin.lhs == init_scope.rl_ptr) continue;
}
@@ -2743,7 +2743,7 @@ fn ptrType(
}
const new_index = @intCast(Zir.Inst.Index, gz.astgen.instructions.len);
- const result = gz.indexToRef(new_index);
+ const result = indexToRef(new_index);
gz.astgen.instructions.appendAssumeCapacity(.{ .tag = .ptr_type, .data = .{
.ptr_type = .{
.flags = .{
@@ -3473,7 +3473,7 @@ fn structDeclInner(
.body_len = 0,
.decls_len = 0,
});
- return gz.indexToRef(decl_inst);
+ return indexToRef(decl_inst);
}
const astgen = gz.astgen;
@@ -3492,7 +3492,6 @@ fn structDeclInner(
.astgen = astgen,
.force_comptime = true,
.in_defer = false,
- .ref_start_index = gz.ref_start_index,
};
defer block_scope.instructions.deinit(gpa);
@@ -3730,7 +3729,7 @@ fn structDeclInner(
}
astgen.extra.appendSliceAssumeCapacity(fields_data.items);
- return gz.indexToRef(decl_inst);
+ return indexToRef(decl_inst);
}
fn unionDeclInner(
@@ -3758,7 +3757,6 @@ fn unionDeclInner(
.astgen = astgen,
.force_comptime = true,
.in_defer = false,
- .ref_start_index = gz.ref_start_index,
};
defer block_scope.instructions.deinit(gpa);
@@ -4006,7 +4004,7 @@ fn unionDeclInner(
astgen.extra.appendAssumeCapacity(cur_bit_bag);
astgen.extra.appendSliceAssumeCapacity(fields_data.items);
- return gz.indexToRef(decl_inst);
+ return indexToRef(decl_inst);
}
fn containerDecl(
@@ -4170,7 +4168,6 @@ fn containerDecl(
.astgen = astgen,
.force_comptime = true,
.in_defer = false,
- .ref_start_index = gz.ref_start_index,
};
defer block_scope.instructions.deinit(gpa);
@@ -4398,7 +4395,7 @@ fn containerDecl(
astgen.extra.appendAssumeCapacity(cur_bit_bag);
astgen.extra.appendSliceAssumeCapacity(fields_data.items);
- return rvalue(gz, rl, gz.indexToRef(decl_inst), node);
+ return rvalue(gz, rl, indexToRef(decl_inst), node);
},
.keyword_opaque => {
var namespace: Scope.Namespace = .{ .parent = scope };
@@ -4559,7 +4556,7 @@ fn containerDecl(
}
astgen.extra.appendSliceAssumeCapacity(wip_decls.payload.items);
- return rvalue(gz, rl, gz.indexToRef(decl_inst), node);
+ return rvalue(gz, rl, indexToRef(decl_inst), node);
},
else => unreachable,
}
@@ -4797,7 +4794,7 @@ fn finishThenElseBlock(
}
assert(!strat.elide_store_to_block_ptr_instructions);
try setCondBrPayload(condbr, cond, then_scope, else_scope);
- return parent_gz.indexToRef(main_block);
+ return indexToRef(main_block);
},
.break_operand => {
if (!parent_gz.refIsNoReturn(then_result)) {
@@ -4815,7 +4812,7 @@ fn finishThenElseBlock(
} else {
try setCondBrPayload(condbr, cond, then_scope, else_scope);
}
- const block_ref = parent_gz.indexToRef(main_block);
+ const block_ref = indexToRef(main_block);
switch (rl) {
.ref => return block_ref,
else => return rvalue(parent_gz, rl, block_ref, node),
@@ -4937,7 +4934,7 @@ fn boolBinOp(
}
try rhs_scope.setBoolBrBody(bool_br);
- const block_ref = gz.indexToRef(bool_br);
+ const block_ref = indexToRef(bool_br);
return rvalue(gz, rl, block_ref, node);
}
@@ -5959,7 +5956,7 @@ fn switchExpr(
if (!strat.elide_store_to_block_ptr_instructions) {
astgen.extra.appendSliceAssumeCapacity(scalar_cases_payload.items);
astgen.extra.appendSliceAssumeCapacity(multi_cases_payload.items);
- return parent_gz.indexToRef(switch_block);
+ return indexToRef(switch_block);
}
// There will necessarily be a store_to_block_ptr for
@@ -6003,7 +6000,7 @@ fn switchExpr(
.lhs = block_scope.rl_ty_inst,
.rhs = zir_datas[break_inst].@"break".operand,
};
- zir_datas[break_inst].@"break".operand = parent_gz.indexToRef(store_inst);
+ zir_datas[break_inst].@"break".operand = indexToRef(store_inst);
} else {
scalar_cases_payload.items[body_len_index] -= 1;
astgen.extra.appendSliceAssumeCapacity(scalar_cases_payload.items[0..extra_index]);
@@ -6045,7 +6042,7 @@ fn switchExpr(
.lhs = block_scope.rl_ty_inst,
.rhs = zir_datas[break_inst].@"break".operand,
};
- zir_datas[break_inst].@"break".operand = parent_gz.indexToRef(store_inst);
+ zir_datas[break_inst].@"break".operand = indexToRef(store_inst);
} else {
scalar_cases_payload.items[body_len_index] -= 1;
astgen.extra.appendSliceAssumeCapacity(scalar_cases_payload.items[start_index..extra_index]);
@@ -6091,7 +6088,7 @@ fn switchExpr(
.lhs = block_scope.rl_ty_inst,
.rhs = zir_datas[break_inst].@"break".operand,
};
- zir_datas[break_inst].@"break".operand = parent_gz.indexToRef(store_inst);
+ zir_datas[break_inst].@"break".operand = indexToRef(store_inst);
} else {
assert(zir_datas[store_inst].bin.lhs == block_scope.rl_ptr);
multi_cases_payload.items[body_len_index] -= 1;
@@ -6102,7 +6099,7 @@ fn switchExpr(
}
}
- const block_ref = parent_gz.indexToRef(switch_block);
+ const block_ref = indexToRef(switch_block);
switch (rl) {
.ref => return block_ref,
else => return rvalue(parent_gz, rl, block_ref, switch_node),
@@ -6162,7 +6159,7 @@ fn switchExpr(
}
}
- return parent_gz.indexToRef(switch_block);
+ return indexToRef(switch_block);
},
}
}
@@ -6861,7 +6858,7 @@ fn asRlPtr(
const zir_datas = astgen.instructions.items(.data);
try parent_zir.ensureUnusedCapacity(astgen.gpa, as_scope.instructions.items.len);
for (as_scope.instructions.items) |src_inst| {
- if (parent_gz.indexToRef(src_inst) == as_scope.rl_ptr) continue;
+ if (indexToRef(src_inst) == as_scope.rl_ptr) continue;
if (zir_tags[src_inst] == .store_to_block_ptr) {
if (zir_datas[src_inst].bin.lhs == as_scope.rl_ptr) continue;
}
@@ -6992,10 +6989,10 @@ fn builtinCall(
const str_lit_token = main_tokens[operand_node];
const str = try astgen.strLitAsString(str_lit_token);
const result = try gz.addStrTok(.import, str.index, str_lit_token);
- const gop = try astgen.imports.getOrPut(astgen.gpa, str.index);
- if (!gop.found_existing) {
- gop.value_ptr.* = str_lit_token;
- }
+ const gop = try astgen.imports.getOrPut(astgen.gpa, str.index);
+ if (!gop.found_existing) {
+ gop.value_ptr.* = str_lit_token;
+ }
return rvalue(gz, rl, result, node);
},
.compile_log => {
@@ -8705,9 +8702,6 @@ const GenZir = struct {
in_defer: bool,
/// How decls created in this scope should be named.
anon_name_strategy: Zir.Inst.NameStrategy = .anon,
- /// The end of special indexes. `Zir.Inst.Ref` subtracts against this number to convert
- /// to `Zir.Inst.Index`. The default here is correct if there are 0 parameters.
- ref_start_index: u32 = Zir.Inst.Ref.typed_value_map.len,
/// The containing decl AST node.
decl_node_index: ast.Node.Index,
/// The containing decl line index, absolute.
@@ -8751,7 +8745,6 @@ const GenZir = struct {
return .{
.force_comptime = gz.force_comptime,
.in_defer = gz.in_defer,
- .ref_start_index = gz.ref_start_index,
.decl_node_index = gz.decl_node_index,
.decl_line = gz.decl_line,
.parent = scope,
@@ -8769,7 +8762,7 @@ const GenZir = struct {
fn refIsNoReturn(gz: GenZir, inst_ref: Zir.Inst.Ref) bool {
if (inst_ref == .unreachable_value) return true;
- if (gz.refToIndex(inst_ref)) |inst_index| {
+ if (refToIndex(inst_ref)) |inst_index| {
return gz.astgen.instructions.items(.tag)[inst_index].isNoReturn();
}
return false;
@@ -8807,19 +8800,6 @@ const GenZir = struct {
return gz.astgen.tree.firstToken(gz.decl_node_index);
}
- fn indexToRef(gz: GenZir, inst: Zir.Inst.Index) Zir.Inst.Ref {
- return @intToEnum(Zir.Inst.Ref, gz.ref_start_index + inst);
- }
-
- fn refToIndex(gz: GenZir, inst: Zir.Inst.Ref) ?Zir.Inst.Index {
- const ref_int = @enumToInt(inst);
- if (ref_int >= gz.ref_start_index) {
- return ref_int - gz.ref_start_index;
- } else {
- return null;
- }
- }
-
fn setBreakResultLoc(gz: *GenZir, parent_rl: AstGen.ResultLoc) void {
// Depending on whether the result location is a pointer or value, different
// ZIR needs to be generated. In the former case we rely on storing to the
@@ -8998,7 +8978,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
} else {
try gz.astgen.extra.ensureUnusedCapacity(
gpa,
@@ -9025,7 +9005,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
}
@@ -9079,7 +9059,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addCall(
@@ -9113,7 +9093,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
/// Note that this returns a `Zir.Inst.Index` not a ref.
@@ -9164,7 +9144,7 @@ const GenZir = struct {
});
gz.instructions.appendAssumeCapacity(new_index);
astgen.string_bytes.appendSliceAssumeCapacity(mem.sliceAsBytes(limbs));
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addFloat(gz: *GenZir, number: f32, src_node: ast.Node.Index) !Zir.Inst.Ref {
@@ -9215,7 +9195,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addExtendedPayload(
@@ -9239,7 +9219,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addExtendedMultiOp(
@@ -9272,7 +9252,7 @@ const GenZir = struct {
});
gz.instructions.appendAssumeCapacity(new_index);
astgen.appendRefsAssumeCapacity(operands);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addArrayTypeSentinel(
@@ -9298,7 +9278,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addUnTok(
@@ -9457,7 +9437,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
fn addAsm(
@@ -9515,7 +9495,7 @@ const GenZir = struct {
} },
});
gz.instructions.appendAssumeCapacity(new_index);
- return gz.indexToRef(new_index);
+ return indexToRef(new_index);
}
/// Note that this returns a `Zir.Inst.Index` not a ref.
@@ -9693,7 +9673,7 @@ const GenZir = struct {
}
fn add(gz: *GenZir, inst: Zir.Inst) !Zir.Inst.Ref {
- return gz.indexToRef(try gz.addAsIndex(inst));
+ return indexToRef(try gz.addAsIndex(inst));
}
fn addAsIndex(gz: *GenZir, inst: Zir.Inst) !Zir.Inst.Index {
@@ -9840,3 +9820,18 @@ fn advanceSourceCursor(astgen: *AstGen, source: []const u8, end: usize) void {
astgen.source_line = line;
astgen.source_column = column;
}
+
+const ref_start_index = Zir.Inst.Ref.typed_value_map.len;
+
+fn indexToRef(inst: Zir.Inst.Index) Zir.Inst.Ref {
+ return @intToEnum(Zir.Inst.Ref, ref_start_index + inst);
+}
+
+fn refToIndex(inst: Zir.Inst.Ref) ?Zir.Inst.Index {
+ const ref_int = @enumToInt(inst);
+ if (ref_int >= ref_start_index) {
+ return ref_int - ref_start_index;
+ } else {
+ return null;
+ }
+}