aboutsummaryrefslogtreecommitdiff
path: root/lib
diff options
context:
space:
mode:
Diffstat (limited to 'lib')
-rw-r--r--lib/c.zig18
-rw-r--r--lib/compiler/test_runner.zig4
-rw-r--r--lib/compiler_rt/aarch64_outline_atomics.zig200
-rw-r--r--lib/compiler_rt/absvdi2.zig2
-rw-r--r--lib/compiler_rt/absvsi2.zig2
-rw-r--r--lib/compiler_rt/absvti2.zig2
-rw-r--r--lib/compiler_rt/adddf3.zig4
-rw-r--r--lib/compiler_rt/addhf3.zig2
-rw-r--r--lib/compiler_rt/addo.zig6
-rw-r--r--lib/compiler_rt/addsf3.zig4
-rw-r--r--lib/compiler_rt/addtf3.zig4
-rw-r--r--lib/compiler_rt/addxf3.zig2
-rw-r--r--lib/compiler_rt/arm.zig44
-rw-r--r--lib/compiler_rt/atomics.zig128
-rw-r--r--lib/compiler_rt/aulldiv.zig4
-rw-r--r--lib/compiler_rt/aullrem.zig4
-rw-r--r--lib/compiler_rt/bcmp.zig2
-rw-r--r--lib/compiler_rt/bitreverse.zig6
-rw-r--r--lib/compiler_rt/bswap.zig6
-rw-r--r--lib/compiler_rt/ceil.zig12
-rw-r--r--lib/compiler_rt/clear_cache.zig2
-rw-r--r--lib/compiler_rt/clzsi2_test.zig2
-rw-r--r--lib/compiler_rt/cmp.zig12
-rw-r--r--lib/compiler_rt/cmpdf2.zig16
-rw-r--r--lib/compiler_rt/cmphf2.zig10
-rw-r--r--lib/compiler_rt/cmpsf2.zig16
-rw-r--r--lib/compiler_rt/cmptf2.zig24
-rw-r--r--lib/compiler_rt/cmpxf2.zig10
-rw-r--r--lib/compiler_rt/cos.zig12
-rw-r--r--lib/compiler_rt/count0bits.zig22
-rw-r--r--lib/compiler_rt/divc3_test.zig2
-rw-r--r--lib/compiler_rt/divdc3.zig2
-rw-r--r--lib/compiler_rt/divdf3.zig4
-rw-r--r--lib/compiler_rt/divhc3.zig2
-rw-r--r--lib/compiler_rt/divhf3.zig2
-rw-r--r--lib/compiler_rt/divmodei4.zig4
-rw-r--r--lib/compiler_rt/divsc3.zig2
-rw-r--r--lib/compiler_rt/divsf3.zig4
-rw-r--r--lib/compiler_rt/divtc3.zig2
-rw-r--r--lib/compiler_rt/divtf3.zig4
-rw-r--r--lib/compiler_rt/divti3.zig4
-rw-r--r--lib/compiler_rt/divxc3.zig2
-rw-r--r--lib/compiler_rt/divxf3.zig2
-rw-r--r--lib/compiler_rt/emutls.zig4
-rw-r--r--lib/compiler_rt/exp.zig12
-rw-r--r--lib/compiler_rt/exp2.zig12
-rw-r--r--lib/compiler_rt/extenddftf2.zig4
-rw-r--r--lib/compiler_rt/extenddfxf2.zig2
-rw-r--r--lib/compiler_rt/extendhfdf2.zig2
-rw-r--r--lib/compiler_rt/extendhfsf2.zig6
-rw-r--r--lib/compiler_rt/extendhftf2.zig2
-rw-r--r--lib/compiler_rt/extendhfxf2.zig2
-rw-r--r--lib/compiler_rt/extendsfdf2.zig4
-rw-r--r--lib/compiler_rt/extendsftf2.zig4
-rw-r--r--lib/compiler_rt/extendsfxf2.zig2
-rw-r--r--lib/compiler_rt/extendxftf2.zig2
-rw-r--r--lib/compiler_rt/fabs.zig12
-rw-r--r--lib/compiler_rt/fixdfdi.zig4
-rw-r--r--lib/compiler_rt/fixdfsi.zig4
-rw-r--r--lib/compiler_rt/fixdfti.zig4
-rw-r--r--lib/compiler_rt/fixhfdi.zig2
-rw-r--r--lib/compiler_rt/fixhfsi.zig2
-rw-r--r--lib/compiler_rt/fixhfti.zig4
-rw-r--r--lib/compiler_rt/fixsfdi.zig4
-rw-r--r--lib/compiler_rt/fixsfsi.zig4
-rw-r--r--lib/compiler_rt/fixsfti.zig4
-rw-r--r--lib/compiler_rt/fixtfdi.zig4
-rw-r--r--lib/compiler_rt/fixtfsi.zig4
-rw-r--r--lib/compiler_rt/fixtfti.zig4
-rw-r--r--lib/compiler_rt/fixunsdfdi.zig4
-rw-r--r--lib/compiler_rt/fixunsdfsi.zig4
-rw-r--r--lib/compiler_rt/fixunsdfti.zig4
-rw-r--r--lib/compiler_rt/fixunshfdi.zig2
-rw-r--r--lib/compiler_rt/fixunshfsi.zig2
-rw-r--r--lib/compiler_rt/fixunshfti.zig4
-rw-r--r--lib/compiler_rt/fixunssfdi.zig4
-rw-r--r--lib/compiler_rt/fixunssfsi.zig4
-rw-r--r--lib/compiler_rt/fixunssfti.zig4
-rw-r--r--lib/compiler_rt/fixunstfdi.zig4
-rw-r--r--lib/compiler_rt/fixunstfsi.zig4
-rw-r--r--lib/compiler_rt/fixunstfti.zig4
-rw-r--r--lib/compiler_rt/fixunsxfdi.zig2
-rw-r--r--lib/compiler_rt/fixunsxfsi.zig2
-rw-r--r--lib/compiler_rt/fixunsxfti.zig4
-rw-r--r--lib/compiler_rt/fixxfdi.zig2
-rw-r--r--lib/compiler_rt/fixxfsi.zig2
-rw-r--r--lib/compiler_rt/fixxfti.zig4
-rw-r--r--lib/compiler_rt/floatdidf.zig4
-rw-r--r--lib/compiler_rt/floatdihf.zig2
-rw-r--r--lib/compiler_rt/floatdisf.zig4
-rw-r--r--lib/compiler_rt/floatditf.zig4
-rw-r--r--lib/compiler_rt/floatdixf.zig2
-rw-r--r--lib/compiler_rt/floatsidf.zig4
-rw-r--r--lib/compiler_rt/floatsihf.zig2
-rw-r--r--lib/compiler_rt/floatsisf.zig4
-rw-r--r--lib/compiler_rt/floatsitf.zig4
-rw-r--r--lib/compiler_rt/floatsixf.zig2
-rw-r--r--lib/compiler_rt/floattidf.zig4
-rw-r--r--lib/compiler_rt/floattihf.zig4
-rw-r--r--lib/compiler_rt/floattisf.zig4
-rw-r--r--lib/compiler_rt/floattitf.zig4
-rw-r--r--lib/compiler_rt/floattixf.zig4
-rw-r--r--lib/compiler_rt/floatundidf.zig4
-rw-r--r--lib/compiler_rt/floatundihf.zig2
-rw-r--r--lib/compiler_rt/floatundisf.zig4
-rw-r--r--lib/compiler_rt/floatunditf.zig4
-rw-r--r--lib/compiler_rt/floatundixf.zig2
-rw-r--r--lib/compiler_rt/floatunsidf.zig4
-rw-r--r--lib/compiler_rt/floatunsihf.zig2
-rw-r--r--lib/compiler_rt/floatunsisf.zig4
-rw-r--r--lib/compiler_rt/floatunsitf.zig4
-rw-r--r--lib/compiler_rt/floatunsixf.zig2
-rw-r--r--lib/compiler_rt/floatuntidf.zig4
-rw-r--r--lib/compiler_rt/floatuntihf.zig4
-rw-r--r--lib/compiler_rt/floatuntisf.zig4
-rw-r--r--lib/compiler_rt/floatuntitf.zig4
-rw-r--r--lib/compiler_rt/floatuntixf.zig4
-rw-r--r--lib/compiler_rt/floor.zig12
-rw-r--r--lib/compiler_rt/fma.zig12
-rw-r--r--lib/compiler_rt/fmax.zig12
-rw-r--r--lib/compiler_rt/fmin.zig12
-rw-r--r--lib/compiler_rt/fmod.zig12
-rw-r--r--lib/compiler_rt/gedf2.zig8
-rw-r--r--lib/compiler_rt/gehf2.zig4
-rw-r--r--lib/compiler_rt/gesf2.zig8
-rw-r--r--lib/compiler_rt/getf2.zig4
-rw-r--r--lib/compiler_rt/gexf2.zig4
-rw-r--r--lib/compiler_rt/int.zig30
-rw-r--r--lib/compiler_rt/log.zig12
-rw-r--r--lib/compiler_rt/log10.zig12
-rw-r--r--lib/compiler_rt/log2.zig12
-rw-r--r--lib/compiler_rt/memcmp.zig2
-rw-r--r--lib/compiler_rt/memcpy.zig4
-rw-r--r--lib/compiler_rt/memmove.zig4
-rw-r--r--lib/compiler_rt/memset.zig4
-rw-r--r--lib/compiler_rt/modti3.zig4
-rw-r--r--lib/compiler_rt/mulXi3.zig10
-rw-r--r--lib/compiler_rt/mulc3_test.zig2
-rw-r--r--lib/compiler_rt/muldc3.zig2
-rw-r--r--lib/compiler_rt/muldf3.zig4
-rw-r--r--lib/compiler_rt/mulhc3.zig2
-rw-r--r--lib/compiler_rt/mulhf3.zig2
-rw-r--r--lib/compiler_rt/mulo.zig6
-rw-r--r--lib/compiler_rt/mulsc3.zig2
-rw-r--r--lib/compiler_rt/mulsf3.zig4
-rw-r--r--lib/compiler_rt/multc3.zig2
-rw-r--r--lib/compiler_rt/multf3.zig4
-rw-r--r--lib/compiler_rt/mulxc3.zig2
-rw-r--r--lib/compiler_rt/mulxf3.zig2
-rw-r--r--lib/compiler_rt/negXi2.zig6
-rw-r--r--lib/compiler_rt/negdf2.zig4
-rw-r--r--lib/compiler_rt/neghf2.zig2
-rw-r--r--lib/compiler_rt/negsf2.zig4
-rw-r--r--lib/compiler_rt/negtf2.zig2
-rw-r--r--lib/compiler_rt/negv.zig6
-rw-r--r--lib/compiler_rt/negxf2.zig2
-rw-r--r--lib/compiler_rt/os_version_check.zig2
-rw-r--r--lib/compiler_rt/parity.zig6
-rw-r--r--lib/compiler_rt/popcount.zig6
-rw-r--r--lib/compiler_rt/powiXf2.zig10
-rw-r--r--lib/compiler_rt/round.zig12
-rw-r--r--lib/compiler_rt/shift.zig24
-rw-r--r--lib/compiler_rt/sin.zig12
-rw-r--r--lib/compiler_rt/sincos.zig12
-rw-r--r--lib/compiler_rt/sqrt.zig12
-rw-r--r--lib/compiler_rt/ssp.zig24
-rw-r--r--lib/compiler_rt/stack_probe.zig12
-rw-r--r--lib/compiler_rt/subdf3.zig4
-rw-r--r--lib/compiler_rt/subhf3.zig2
-rw-r--r--lib/compiler_rt/subo.zig6
-rw-r--r--lib/compiler_rt/subsf3.zig4
-rw-r--r--lib/compiler_rt/subtf3.zig4
-rw-r--r--lib/compiler_rt/subxf3.zig2
-rw-r--r--lib/compiler_rt/tan.zig12
-rw-r--r--lib/compiler_rt/trunc.zig12
-rw-r--r--lib/compiler_rt/truncdfhf2.zig4
-rw-r--r--lib/compiler_rt/truncdfsf2.zig4
-rw-r--r--lib/compiler_rt/truncsfhf2.zig6
-rw-r--r--lib/compiler_rt/trunctfdf2.zig4
-rw-r--r--lib/compiler_rt/trunctfhf2.zig2
-rw-r--r--lib/compiler_rt/trunctfsf2.zig4
-rw-r--r--lib/compiler_rt/trunctfxf2.zig2
-rw-r--r--lib/compiler_rt/truncxfdf2.zig2
-rw-r--r--lib/compiler_rt/truncxfhf2.zig2
-rw-r--r--lib/compiler_rt/truncxfsf2.zig2
-rw-r--r--lib/compiler_rt/udivmodei4.zig4
-rw-r--r--lib/compiler_rt/udivmodti4.zig4
-rw-r--r--lib/compiler_rt/udivti3.zig4
-rw-r--r--lib/compiler_rt/umodti3.zig4
-rw-r--r--lib/compiler_rt/unorddf2.zig4
-rw-r--r--lib/compiler_rt/unordhf2.zig2
-rw-r--r--lib/compiler_rt/unordsf2.zig4
-rw-r--r--lib/compiler_rt/unordtf2.zig2
-rw-r--r--lib/compiler_rt/unordxf2.zig2
-rw-r--r--lib/fuzzer.zig2
-rw-r--r--lib/std/os/linux/aarch64.zig4
-rw-r--r--lib/std/os/linux/arm.zig6
-rw-r--r--lib/std/os/linux/hexagon.zig4
-rw-r--r--lib/std/os/linux/loongarch64.zig4
-rw-r--r--lib/std/os/linux/mips.zig6
-rw-r--r--lib/std/os/linux/mips64.zig6
-rw-r--r--lib/std/os/linux/powerpc.zig4
-rw-r--r--lib/std/os/linux/powerpc64.zig4
-rw-r--r--lib/std/os/linux/riscv32.zig4
-rw-r--r--lib/std/os/linux/riscv64.zig4
-rw-r--r--lib/std/os/linux/s390x.zig4
-rw-r--r--lib/std/os/linux/sparc64.zig2
-rw-r--r--lib/std/os/linux/thumb.zig4
-rw-r--r--lib/std/os/linux/x86.zig8
-rw-r--r--lib/std/os/linux/x86_64.zig6
-rw-r--r--lib/std/zig/parser_test.zig8
211 files changed, 729 insertions, 729 deletions
diff --git a/lib/c.zig b/lib/c.zig
index 734940ff05..bbb79eadab 100644
--- a/lib/c.zig
+++ b/lib/c.zig
@@ -54,11 +54,11 @@ pub fn panic(msg: []const u8, error_return_trace: ?*std.builtin.StackTrace, _: ?
}
extern fn main(argc: c_int, argv: [*:null]?[*:0]u8) c_int;
-fn wasm_start() callconv(.C) void {
+fn wasm_start() callconv(.c) void {
_ = main(0, undefined);
}
-fn strcpy(dest: [*:0]u8, src: [*:0]const u8) callconv(.C) [*:0]u8 {
+fn strcpy(dest: [*:0]u8, src: [*:0]const u8) callconv(.c) [*:0]u8 {
var i: usize = 0;
while (src[i] != 0) : (i += 1) {
dest[i] = src[i];
@@ -76,7 +76,7 @@ test "strcpy" {
try std.testing.expectEqualSlices(u8, "foobarbaz", std.mem.sliceTo(&s1, 0));
}
-fn strncpy(dest: [*:0]u8, src: [*:0]const u8, n: usize) callconv(.C) [*:0]u8 {
+fn strncpy(dest: [*:0]u8, src: [*:0]const u8, n: usize) callconv(.c) [*:0]u8 {
var i: usize = 0;
while (i < n and src[i] != 0) : (i += 1) {
dest[i] = src[i];
@@ -96,7 +96,7 @@ test "strncpy" {
try std.testing.expectEqualSlices(u8, "foobarbaz", std.mem.sliceTo(&s1, 0));
}
-fn strcat(dest: [*:0]u8, src: [*:0]const u8) callconv(.C) [*:0]u8 {
+fn strcat(dest: [*:0]u8, src: [*:0]const u8) callconv(.c) [*:0]u8 {
var dest_end: usize = 0;
while (dest[dest_end] != 0) : (dest_end += 1) {}
@@ -119,7 +119,7 @@ test "strcat" {
try std.testing.expectEqualSlices(u8, "foobarbaz", std.mem.sliceTo(&s1, 0));
}
-fn strncat(dest: [*:0]u8, src: [*:0]const u8, avail: usize) callconv(.C) [*:0]u8 {
+fn strncat(dest: [*:0]u8, src: [*:0]const u8, avail: usize) callconv(.c) [*:0]u8 {
var dest_end: usize = 0;
while (dest[dest_end] != 0) : (dest_end += 1) {}
@@ -142,7 +142,7 @@ test "strncat" {
try std.testing.expectEqualSlices(u8, "foobarbaz", std.mem.sliceTo(&s1, 0));
}
-fn strcmp(s1: [*:0]const u8, s2: [*:0]const u8) callconv(.C) c_int {
+fn strcmp(s1: [*:0]const u8, s2: [*:0]const u8) callconv(.c) c_int {
return switch (std.mem.orderZ(u8, s1, s2)) {
.lt => -1,
.eq => 0,
@@ -150,11 +150,11 @@ fn strcmp(s1: [*:0]const u8, s2: [*:0]const u8) callconv(.C) c_int {
};
}
-fn strlen(s: [*:0]const u8) callconv(.C) usize {
+fn strlen(s: [*:0]const u8) callconv(.c) usize {
return std.mem.len(s);
}
-fn strncmp(_l: [*:0]const u8, _r: [*:0]const u8, _n: usize) callconv(.C) c_int {
+fn strncmp(_l: [*:0]const u8, _r: [*:0]const u8, _n: usize) callconv(.c) c_int {
if (_n == 0) return 0;
var l = _l;
var r = _r;
@@ -167,7 +167,7 @@ fn strncmp(_l: [*:0]const u8, _r: [*:0]const u8, _n: usize) callconv(.C) c_int {
return @as(c_int, l[0]) - @as(c_int, r[0]);
}
-fn strerror(errnum: c_int) callconv(.C) [*:0]const u8 {
+fn strerror(errnum: c_int) callconv(.c) [*:0]const u8 {
_ = errnum;
return "TODO strerror implementation";
}
diff --git a/lib/compiler/test_runner.zig b/lib/compiler/test_runner.zig
index f0ff6ffe1c..ab1dae9a70 100644
--- a/lib/compiler/test_runner.zig
+++ b/lib/compiler/test_runner.zig
@@ -350,7 +350,7 @@ var is_fuzz_test: bool = undefined;
extern fn fuzzer_set_name(name_ptr: [*]const u8, name_len: usize) void;
extern fn fuzzer_init(cache_dir: FuzzerSlice) void;
extern fn fuzzer_init_corpus_elem(input_ptr: [*]const u8, input_len: usize) void;
-extern fn fuzzer_start(testOne: *const fn ([*]const u8, usize) callconv(.C) void) void;
+extern fn fuzzer_start(testOne: *const fn ([*]const u8, usize) callconv(.c) void) void;
extern fn fuzzer_coverage_id() u64;
pub fn fuzz(
@@ -382,7 +382,7 @@ pub fn fuzz(
const global = struct {
var ctx: @TypeOf(context) = undefined;
- fn fuzzer_one(input_ptr: [*]const u8, input_len: usize) callconv(.C) void {
+ fn fuzzer_one(input_ptr: [*]const u8, input_len: usize) callconv(.c) void {
@disableInstrumentation();
testing.allocator_instance = .{};
defer if (testing.allocator_instance.deinit() == .leak) std.process.exit(1);
diff --git a/lib/compiler_rt/aarch64_outline_atomics.zig b/lib/compiler_rt/aarch64_outline_atomics.zig
index e920dbb4b9..e78fda9b9b 100644
--- a/lib/compiler_rt/aarch64_outline_atomics.zig
+++ b/lib/compiler_rt/aarch64_outline_atomics.zig
@@ -10,7 +10,7 @@ const always_has_lse = std.Target.aarch64.featureSetHas(builtin.cpu.features, .l
/// which ARM is concerned would have too much overhead.
var __aarch64_have_lse_atomics: u8 = @intFromBool(always_has_lse);
-fn __aarch64_cas1_relax() align(16) callconv(.Naked) void {
+fn __aarch64_cas1_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -32,7 +32,7 @@ fn __aarch64_cas1_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp1_relax() align(16) callconv(.Naked) void {
+fn __aarch64_swp1_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -52,7 +52,7 @@ fn __aarch64_swp1_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd1_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd1_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -73,7 +73,7 @@ fn __aarch64_ldadd1_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr1_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr1_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -94,7 +94,7 @@ fn __aarch64_ldclr1_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor1_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor1_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -115,7 +115,7 @@ fn __aarch64_ldeor1_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset1_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldset1_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -136,7 +136,7 @@ fn __aarch64_ldset1_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas1_acq() align(16) callconv(.Naked) void {
+fn __aarch64_cas1_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -158,7 +158,7 @@ fn __aarch64_cas1_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp1_acq() align(16) callconv(.Naked) void {
+fn __aarch64_swp1_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -178,7 +178,7 @@ fn __aarch64_swp1_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd1_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd1_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -199,7 +199,7 @@ fn __aarch64_ldadd1_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr1_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr1_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -220,7 +220,7 @@ fn __aarch64_ldclr1_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor1_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor1_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -241,7 +241,7 @@ fn __aarch64_ldeor1_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset1_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldset1_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -262,7 +262,7 @@ fn __aarch64_ldset1_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas1_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas1_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -284,7 +284,7 @@ fn __aarch64_cas1_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp1_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp1_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -304,7 +304,7 @@ fn __aarch64_swp1_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd1_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd1_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -325,7 +325,7 @@ fn __aarch64_ldadd1_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr1_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr1_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -346,7 +346,7 @@ fn __aarch64_ldclr1_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor1_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor1_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -367,7 +367,7 @@ fn __aarch64_ldeor1_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset1_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset1_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -388,7 +388,7 @@ fn __aarch64_ldset1_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas1_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas1_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -410,7 +410,7 @@ fn __aarch64_cas1_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp1_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp1_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -430,7 +430,7 @@ fn __aarch64_swp1_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd1_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd1_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -451,7 +451,7 @@ fn __aarch64_ldadd1_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr1_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr1_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -472,7 +472,7 @@ fn __aarch64_ldclr1_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor1_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor1_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -493,7 +493,7 @@ fn __aarch64_ldeor1_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset1_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset1_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -514,7 +514,7 @@ fn __aarch64_ldset1_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas2_relax() align(16) callconv(.Naked) void {
+fn __aarch64_cas2_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -536,7 +536,7 @@ fn __aarch64_cas2_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp2_relax() align(16) callconv(.Naked) void {
+fn __aarch64_swp2_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -556,7 +556,7 @@ fn __aarch64_swp2_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd2_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd2_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -577,7 +577,7 @@ fn __aarch64_ldadd2_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr2_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr2_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -598,7 +598,7 @@ fn __aarch64_ldclr2_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor2_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor2_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -619,7 +619,7 @@ fn __aarch64_ldeor2_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset2_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldset2_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -640,7 +640,7 @@ fn __aarch64_ldset2_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas2_acq() align(16) callconv(.Naked) void {
+fn __aarch64_cas2_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -662,7 +662,7 @@ fn __aarch64_cas2_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp2_acq() align(16) callconv(.Naked) void {
+fn __aarch64_swp2_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -682,7 +682,7 @@ fn __aarch64_swp2_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd2_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd2_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -703,7 +703,7 @@ fn __aarch64_ldadd2_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr2_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr2_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -724,7 +724,7 @@ fn __aarch64_ldclr2_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor2_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor2_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -745,7 +745,7 @@ fn __aarch64_ldeor2_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset2_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldset2_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -766,7 +766,7 @@ fn __aarch64_ldset2_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas2_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas2_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -788,7 +788,7 @@ fn __aarch64_cas2_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp2_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp2_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -808,7 +808,7 @@ fn __aarch64_swp2_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd2_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd2_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -829,7 +829,7 @@ fn __aarch64_ldadd2_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr2_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr2_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -850,7 +850,7 @@ fn __aarch64_ldclr2_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor2_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor2_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -871,7 +871,7 @@ fn __aarch64_ldeor2_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset2_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset2_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -892,7 +892,7 @@ fn __aarch64_ldset2_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas2_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas2_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -914,7 +914,7 @@ fn __aarch64_cas2_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp2_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp2_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -934,7 +934,7 @@ fn __aarch64_swp2_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd2_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd2_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -955,7 +955,7 @@ fn __aarch64_ldadd2_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr2_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr2_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -976,7 +976,7 @@ fn __aarch64_ldclr2_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor2_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor2_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -997,7 +997,7 @@ fn __aarch64_ldeor2_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset2_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset2_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1018,7 +1018,7 @@ fn __aarch64_ldset2_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas4_relax() align(16) callconv(.Naked) void {
+fn __aarch64_cas4_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1040,7 +1040,7 @@ fn __aarch64_cas4_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp4_relax() align(16) callconv(.Naked) void {
+fn __aarch64_swp4_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1060,7 +1060,7 @@ fn __aarch64_swp4_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd4_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd4_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1081,7 +1081,7 @@ fn __aarch64_ldadd4_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr4_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr4_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1102,7 +1102,7 @@ fn __aarch64_ldclr4_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor4_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor4_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1123,7 +1123,7 @@ fn __aarch64_ldeor4_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset4_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldset4_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1144,7 +1144,7 @@ fn __aarch64_ldset4_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas4_acq() align(16) callconv(.Naked) void {
+fn __aarch64_cas4_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1166,7 +1166,7 @@ fn __aarch64_cas4_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp4_acq() align(16) callconv(.Naked) void {
+fn __aarch64_swp4_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1186,7 +1186,7 @@ fn __aarch64_swp4_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd4_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd4_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1207,7 +1207,7 @@ fn __aarch64_ldadd4_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr4_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr4_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1228,7 +1228,7 @@ fn __aarch64_ldclr4_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor4_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor4_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1249,7 +1249,7 @@ fn __aarch64_ldeor4_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset4_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldset4_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1270,7 +1270,7 @@ fn __aarch64_ldset4_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas4_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas4_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1292,7 +1292,7 @@ fn __aarch64_cas4_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp4_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp4_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1312,7 +1312,7 @@ fn __aarch64_swp4_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd4_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd4_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1333,7 +1333,7 @@ fn __aarch64_ldadd4_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr4_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr4_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1354,7 +1354,7 @@ fn __aarch64_ldclr4_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor4_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor4_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1375,7 +1375,7 @@ fn __aarch64_ldeor4_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset4_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset4_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1396,7 +1396,7 @@ fn __aarch64_ldset4_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas4_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas4_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1418,7 +1418,7 @@ fn __aarch64_cas4_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp4_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp4_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1438,7 +1438,7 @@ fn __aarch64_swp4_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd4_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd4_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1459,7 +1459,7 @@ fn __aarch64_ldadd4_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr4_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr4_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1480,7 +1480,7 @@ fn __aarch64_ldclr4_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor4_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor4_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1501,7 +1501,7 @@ fn __aarch64_ldeor4_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset4_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset4_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1522,7 +1522,7 @@ fn __aarch64_ldset4_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas8_relax() align(16) callconv(.Naked) void {
+fn __aarch64_cas8_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1544,7 +1544,7 @@ fn __aarch64_cas8_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp8_relax() align(16) callconv(.Naked) void {
+fn __aarch64_swp8_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1564,7 +1564,7 @@ fn __aarch64_swp8_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd8_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd8_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1585,7 +1585,7 @@ fn __aarch64_ldadd8_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr8_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr8_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1606,7 +1606,7 @@ fn __aarch64_ldclr8_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor8_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor8_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1627,7 +1627,7 @@ fn __aarch64_ldeor8_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset8_relax() align(16) callconv(.Naked) void {
+fn __aarch64_ldset8_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1648,7 +1648,7 @@ fn __aarch64_ldset8_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas8_acq() align(16) callconv(.Naked) void {
+fn __aarch64_cas8_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1670,7 +1670,7 @@ fn __aarch64_cas8_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp8_acq() align(16) callconv(.Naked) void {
+fn __aarch64_swp8_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1690,7 +1690,7 @@ fn __aarch64_swp8_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd8_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd8_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1711,7 +1711,7 @@ fn __aarch64_ldadd8_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr8_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr8_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1732,7 +1732,7 @@ fn __aarch64_ldclr8_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor8_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor8_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1753,7 +1753,7 @@ fn __aarch64_ldeor8_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset8_acq() align(16) callconv(.Naked) void {
+fn __aarch64_ldset8_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1774,7 +1774,7 @@ fn __aarch64_ldset8_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas8_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas8_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1796,7 +1796,7 @@ fn __aarch64_cas8_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp8_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp8_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1816,7 +1816,7 @@ fn __aarch64_swp8_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd8_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd8_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1837,7 +1837,7 @@ fn __aarch64_ldadd8_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr8_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr8_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1858,7 +1858,7 @@ fn __aarch64_ldclr8_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor8_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor8_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1879,7 +1879,7 @@ fn __aarch64_ldeor8_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset8_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset8_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1900,7 +1900,7 @@ fn __aarch64_ldset8_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas8_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas8_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1922,7 +1922,7 @@ fn __aarch64_cas8_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_swp8_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_swp8_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1942,7 +1942,7 @@ fn __aarch64_swp8_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldadd8_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldadd8_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1963,7 +1963,7 @@ fn __aarch64_ldadd8_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldclr8_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldclr8_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -1984,7 +1984,7 @@ fn __aarch64_ldclr8_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldeor8_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldeor8_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -2005,7 +2005,7 @@ fn __aarch64_ldeor8_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_ldset8_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_ldset8_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -2026,7 +2026,7 @@ fn __aarch64_ldset8_acq_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas16_relax() align(16) callconv(.Naked) void {
+fn __aarch64_cas16_relax() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -2050,7 +2050,7 @@ fn __aarch64_cas16_relax() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas16_acq() align(16) callconv(.Naked) void {
+fn __aarch64_cas16_acq() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -2074,7 +2074,7 @@ fn __aarch64_cas16_acq() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas16_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas16_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
@@ -2098,7 +2098,7 @@ fn __aarch64_cas16_rel() align(16) callconv(.Naked) void {
);
unreachable;
}
-fn __aarch64_cas16_acq_rel() align(16) callconv(.Naked) void {
+fn __aarch64_cas16_acq_rel() align(16) callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ cbz w16, 8f
diff --git a/lib/compiler_rt/absvdi2.zig b/lib/compiler_rt/absvdi2.zig
index 71f2467091..8b5e5d8a5d 100644
--- a/lib/compiler_rt/absvdi2.zig
+++ b/lib/compiler_rt/absvdi2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__absvdi2, .{ .name = "__absvdi2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __absvdi2(a: i64) callconv(.C) i64 {
+pub fn __absvdi2(a: i64) callconv(.c) i64 {
return absv(i64, a);
}
diff --git a/lib/compiler_rt/absvsi2.zig b/lib/compiler_rt/absvsi2.zig
index 83cc71bbaf..0030813db8 100644
--- a/lib/compiler_rt/absvsi2.zig
+++ b/lib/compiler_rt/absvsi2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__absvsi2, .{ .name = "__absvsi2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __absvsi2(a: i32) callconv(.C) i32 {
+pub fn __absvsi2(a: i32) callconv(.c) i32 {
return absv(i32, a);
}
diff --git a/lib/compiler_rt/absvti2.zig b/lib/compiler_rt/absvti2.zig
index 91e2d5fe7b..be232ece2c 100644
--- a/lib/compiler_rt/absvti2.zig
+++ b/lib/compiler_rt/absvti2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__absvti2, .{ .name = "__absvti2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __absvti2(a: i128) callconv(.C) i128 {
+pub fn __absvti2(a: i128) callconv(.c) i128 {
return absv(i128, a);
}
diff --git a/lib/compiler_rt/adddf3.zig b/lib/compiler_rt/adddf3.zig
index b63785a796..3a10f55520 100644
--- a/lib/compiler_rt/adddf3.zig
+++ b/lib/compiler_rt/adddf3.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-fn __adddf3(a: f64, b: f64) callconv(.C) f64 {
+fn __adddf3(a: f64, b: f64) callconv(.c) f64 {
return addf3(f64, a, b);
}
-fn __aeabi_dadd(a: f64, b: f64) callconv(.AAPCS) f64 {
+fn __aeabi_dadd(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) f64 {
return addf3(f64, a, b);
}
diff --git a/lib/compiler_rt/addhf3.zig b/lib/compiler_rt/addhf3.zig
index 5c337ea096..48a6328836 100644
--- a/lib/compiler_rt/addhf3.zig
+++ b/lib/compiler_rt/addhf3.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__addhf3, .{ .name = "__addhf3", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __addhf3(a: f16, b: f16) callconv(.C) f16 {
+fn __addhf3(a: f16, b: f16) callconv(.c) f16 {
return addf3(f16, a, b);
}
diff --git a/lib/compiler_rt/addo.zig b/lib/compiler_rt/addo.zig
index 00d0e95850..beb6249223 100644
--- a/lib/compiler_rt/addo.zig
+++ b/lib/compiler_rt/addo.zig
@@ -31,13 +31,13 @@ inline fn addoXi4_generic(comptime ST: type, a: ST, b: ST, overflow: *c_int) ST
return sum;
}
-pub fn __addosi4(a: i32, b: i32, overflow: *c_int) callconv(.C) i32 {
+pub fn __addosi4(a: i32, b: i32, overflow: *c_int) callconv(.c) i32 {
return addoXi4_generic(i32, a, b, overflow);
}
-pub fn __addodi4(a: i64, b: i64, overflow: *c_int) callconv(.C) i64 {
+pub fn __addodi4(a: i64, b: i64, overflow: *c_int) callconv(.c) i64 {
return addoXi4_generic(i64, a, b, overflow);
}
-pub fn __addoti4(a: i128, b: i128, overflow: *c_int) callconv(.C) i128 {
+pub fn __addoti4(a: i128, b: i128, overflow: *c_int) callconv(.c) i128 {
return addoXi4_generic(i128, a, b, overflow);
}
diff --git a/lib/compiler_rt/addsf3.zig b/lib/compiler_rt/addsf3.zig
index 244e3f1a25..c4b7773eb0 100644
--- a/lib/compiler_rt/addsf3.zig
+++ b/lib/compiler_rt/addsf3.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-fn __addsf3(a: f32, b: f32) callconv(.C) f32 {
+fn __addsf3(a: f32, b: f32) callconv(.c) f32 {
return addf3(f32, a, b);
}
-fn __aeabi_fadd(a: f32, b: f32) callconv(.AAPCS) f32 {
+fn __aeabi_fadd(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) f32 {
return addf3(f32, a, b);
}
diff --git a/lib/compiler_rt/addtf3.zig b/lib/compiler_rt/addtf3.zig
index 95bd593855..34e5bdf7af 100644
--- a/lib/compiler_rt/addtf3.zig
+++ b/lib/compiler_rt/addtf3.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__addtf3, .{ .name = "__addtf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __addtf3(a: f128, b: f128) callconv(.C) f128 {
+pub fn __addtf3(a: f128, b: f128) callconv(.c) f128 {
return addf3(f128, a, b);
}
-fn _Qp_add(c: *f128, a: *f128, b: *f128) callconv(.C) void {
+fn _Qp_add(c: *f128, a: *f128, b: *f128) callconv(.c) void {
c.* = addf3(f128, a.*, b.*);
}
diff --git a/lib/compiler_rt/addxf3.zig b/lib/compiler_rt/addxf3.zig
index 1aaf692ef3..1aac736750 100644
--- a/lib/compiler_rt/addxf3.zig
+++ b/lib/compiler_rt/addxf3.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__addxf3, .{ .name = "__addxf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __addxf3(a: f80, b: f80) callconv(.C) f80 {
+pub fn __addxf3(a: f80, b: f80) callconv(.c) f80 {
return addf3(f80, a, b);
}
diff --git a/lib/compiler_rt/arm.zig b/lib/compiler_rt/arm.zig
index 0cc268f189..8fd6f2938b 100644
--- a/lib/compiler_rt/arm.zig
+++ b/lib/compiler_rt/arm.zig
@@ -57,67 +57,67 @@ extern fn memset(dest: ?[*]u8, c: i32, n: usize) ?[*]u8;
extern fn memcpy(noalias dest: ?[*]u8, noalias src: ?[*]const u8, n: usize) ?[*]u8;
extern fn memmove(dest: ?[*]u8, src: ?[*]const u8, n: usize) ?[*]u8;
-pub fn __aeabi_memcpy(dest: [*]u8, src: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memcpy(dest: [*]u8, src: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memcpy(dest, src, n);
}
-pub fn __aeabi_memcpy4(dest: [*]u8, src: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memcpy4(dest: [*]u8, src: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memcpy(dest, src, n);
}
-pub fn __aeabi_memcpy8(dest: [*]u8, src: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memcpy8(dest: [*]u8, src: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memcpy(dest, src, n);
}
-pub fn __aeabi_memmove(dest: [*]u8, src: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memmove(dest: [*]u8, src: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memmove(dest, src, n);
}
-pub fn __aeabi_memmove4(dest: [*]u8, src: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memmove4(dest: [*]u8, src: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memmove(dest, src, n);
}
-pub fn __aeabi_memmove8(dest: [*]u8, src: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memmove8(dest: [*]u8, src: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memmove(dest, src, n);
}
-pub fn __aeabi_memset(dest: [*]u8, n: usize, c: i32) callconv(.AAPCS) void {
+pub fn __aeabi_memset(dest: [*]u8, n: usize, c: i32) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
// This is dentical to the standard `memset` definition but with the last
// two arguments swapped
_ = memset(dest, c, n);
}
-pub fn __aeabi_memset4(dest: [*]u8, n: usize, c: i32) callconv(.AAPCS) void {
+pub fn __aeabi_memset4(dest: [*]u8, n: usize, c: i32) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memset(dest, c, n);
}
-pub fn __aeabi_memset8(dest: [*]u8, n: usize, c: i32) callconv(.AAPCS) void {
+pub fn __aeabi_memset8(dest: [*]u8, n: usize, c: i32) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memset(dest, c, n);
}
-pub fn __aeabi_memclr(dest: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memclr(dest: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memset(dest, 0, n);
}
-pub fn __aeabi_memclr4(dest: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memclr4(dest: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memset(dest, 0, n);
}
-pub fn __aeabi_memclr8(dest: [*]u8, n: usize) callconv(.AAPCS) void {
+pub fn __aeabi_memclr8(dest: [*]u8, n: usize) callconv(.{ .arm_aapcs = .{} }) void {
@setRuntimeSafety(false);
_ = memset(dest, 0, n);
}
// Dummy functions to avoid errors during the linking phase
-pub fn __aeabi_unwind_cpp_pr0() callconv(.AAPCS) void {}
-pub fn __aeabi_unwind_cpp_pr1() callconv(.AAPCS) void {}
-pub fn __aeabi_unwind_cpp_pr2() callconv(.AAPCS) void {}
+pub fn __aeabi_unwind_cpp_pr0() callconv(.{ .arm_aapcs = .{} }) void {}
+pub fn __aeabi_unwind_cpp_pr1() callconv(.{ .arm_aapcs = .{} }) void {}
+pub fn __aeabi_unwind_cpp_pr2() callconv(.{ .arm_aapcs = .{} }) void {}
// This function can only clobber r0 according to the ABI
-pub fn __aeabi_read_tp() callconv(.Naked) void {
+pub fn __aeabi_read_tp() callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
\\ mrc p15, 0, r0, c13, c0, 3
@@ -129,7 +129,7 @@ pub fn __aeabi_read_tp() callconv(.Naked) void {
// The following functions are wrapped in an asm block to ensure the required
// calling convention is always respected
-pub fn __aeabi_uidivmod() callconv(.Naked) void {
+pub fn __aeabi_uidivmod() callconv(.naked) void {
@setRuntimeSafety(false);
// Divide r0 by r1; the quotient goes in r0, the remainder in r1
asm volatile (
@@ -147,7 +147,7 @@ pub fn __aeabi_uidivmod() callconv(.Naked) void {
unreachable;
}
-pub fn __aeabi_uldivmod() callconv(.Naked) void {
+pub fn __aeabi_uldivmod() callconv(.naked) void {
@setRuntimeSafety(false);
// Divide r1:r0 by r3:r2; the quotient goes in r1:r0, the remainder in r3:r2
asm volatile (
@@ -167,7 +167,7 @@ pub fn __aeabi_uldivmod() callconv(.Naked) void {
unreachable;
}
-pub fn __aeabi_idivmod() callconv(.Naked) void {
+pub fn __aeabi_idivmod() callconv(.naked) void {
@setRuntimeSafety(false);
// Divide r0 by r1; the quotient goes in r0, the remainder in r1
asm volatile (
@@ -185,7 +185,7 @@ pub fn __aeabi_idivmod() callconv(.Naked) void {
unreachable;
}
-pub fn __aeabi_ldivmod() callconv(.Naked) void {
+pub fn __aeabi_ldivmod() callconv(.naked) void {
@setRuntimeSafety(false);
// Divide r1:r0 by r3:r2; the quotient goes in r1:r0, the remainder in r3:r2
asm volatile (
@@ -207,12 +207,12 @@ pub fn __aeabi_ldivmod() callconv(.Naked) void {
// Float Arithmetic
-fn __aeabi_frsub(a: f32, b: f32) callconv(.AAPCS) f32 {
+fn __aeabi_frsub(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) f32 {
const neg_a: f32 = @bitCast(@as(u32, @bitCast(a)) ^ (@as(u32, 1) << 31));
return b + neg_a;
}
-fn __aeabi_drsub(a: f64, b: f64) callconv(.AAPCS) f64 {
+fn __aeabi_drsub(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) f64 {
const neg_a: f64 = @bitCast(@as(u64, @bitCast(a)) ^ (@as(u64, 1) << 63));
return b + neg_a;
}
diff --git a/lib/compiler_rt/atomics.zig b/lib/compiler_rt/atomics.zig
index 2cec398732..7005952cec 100644
--- a/lib/compiler_rt/atomics.zig
+++ b/lib/compiler_rt/atomics.zig
@@ -117,21 +117,21 @@ var spinlocks: SpinlockTable = SpinlockTable{};
// Generic version of GCC atomic builtin functions.
// Those work on any object no matter the pointer alignment nor its size.
-fn __atomic_load(size: u32, src: [*]u8, dest: [*]u8, model: i32) callconv(.C) void {
+fn __atomic_load(size: u32, src: [*]u8, dest: [*]u8, model: i32) callconv(.c) void {
_ = model;
var sl = spinlocks.get(@intFromPtr(src));
defer sl.release();
@memcpy(dest[0..size], src);
}
-fn __atomic_store(size: u32, dest: [*]u8, src: [*]u8, model: i32) callconv(.C) void {
+fn __atomic_store(size: u32, dest: [*]u8, src: [*]u8, model: i32) callconv(.c) void {
_ = model;
var sl = spinlocks.get(@intFromPtr(dest));
defer sl.release();
@memcpy(dest[0..size], src);
}
-fn __atomic_exchange(size: u32, ptr: [*]u8, val: [*]u8, old: [*]u8, model: i32) callconv(.C) void {
+fn __atomic_exchange(size: u32, ptr: [*]u8, val: [*]u8, old: [*]u8, model: i32) callconv(.c) void {
_ = model;
var sl = spinlocks.get(@intFromPtr(ptr));
defer sl.release();
@@ -146,7 +146,7 @@ fn __atomic_compare_exchange(
desired: [*]u8,
success: i32,
failure: i32,
-) callconv(.C) i32 {
+) callconv(.c) i32 {
_ = success;
_ = failure;
var sl = spinlocks.get(@intFromPtr(ptr));
@@ -176,23 +176,23 @@ inline fn atomic_load_N(comptime T: type, src: *T, model: i32) T {
}
}
-fn __atomic_load_1(src: *u8, model: i32) callconv(.C) u8 {
+fn __atomic_load_1(src: *u8, model: i32) callconv(.c) u8 {
return atomic_load_N(u8, src, model);
}
-fn __atomic_load_2(src: *u16, model: i32) callconv(.C) u16 {
+fn __atomic_load_2(src: *u16, model: i32) callconv(.c) u16 {
return atomic_load_N(u16, src, model);
}
-fn __atomic_load_4(src: *u32, model: i32) callconv(.C) u32 {
+fn __atomic_load_4(src: *u32, model: i32) callconv(.c) u32 {
return atomic_load_N(u32, src, model);
}
-fn __atomic_load_8(src: *u64, model: i32) callconv(.C) u64 {
+fn __atomic_load_8(src: *u64, model: i32) callconv(.c) u64 {
return atomic_load_N(u64, src, model);
}
-fn __atomic_load_16(src: *u128, model: i32) callconv(.C) u128 {
+fn __atomic_load_16(src: *u128, model: i32) callconv(.c) u128 {
return atomic_load_N(u128, src, model);
}
@@ -207,23 +207,23 @@ inline fn atomic_store_N(comptime T: type, dst: *T, value: T, model: i32) void {
}
}
-fn __atomic_store_1(dst: *u8, value: u8, model: i32) callconv(.C) void {
+fn __atomic_store_1(dst: *u8, value: u8, model: i32) callconv(.c) void {
return atomic_store_N(u8, dst, value, model);
}
-fn __atomic_store_2(dst: *u16, value: u16, model: i32) callconv(.C) void {
+fn __atomic_store_2(dst: *u16, value: u16, model: i32) callconv(.c) void {
return atomic_store_N(u16, dst, value, model);
}
-fn __atomic_store_4(dst: *u32, value: u32, model: i32) callconv(.C) void {
+fn __atomic_store_4(dst: *u32, value: u32, model: i32) callconv(.c) void {
return atomic_store_N(u32, dst, value, model);
}
-fn __atomic_store_8(dst: *u64, value: u64, model: i32) callconv(.C) void {
+fn __atomic_store_8(dst: *u64, value: u64, model: i32) callconv(.c) void {
return atomic_store_N(u64, dst, value, model);
}
-fn __atomic_store_16(dst: *u128, value: u128, model: i32) callconv(.C) void {
+fn __atomic_store_16(dst: *u128, value: u128, model: i32) callconv(.c) void {
return atomic_store_N(u128, dst, value, model);
}
@@ -274,23 +274,23 @@ inline fn atomic_exchange_N(comptime T: type, ptr: *T, val: T, model: i32) T {
}
}
-fn __atomic_exchange_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_exchange_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return atomic_exchange_N(u8, ptr, val, model);
}
-fn __atomic_exchange_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_exchange_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return atomic_exchange_N(u16, ptr, val, model);
}
-fn __atomic_exchange_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_exchange_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return atomic_exchange_N(u32, ptr, val, model);
}
-fn __atomic_exchange_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_exchange_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return atomic_exchange_N(u64, ptr, val, model);
}
-fn __atomic_exchange_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_exchange_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return atomic_exchange_N(u128, ptr, val, model);
}
@@ -323,23 +323,23 @@ inline fn atomic_compare_exchange_N(
}
}
-fn __atomic_compare_exchange_1(ptr: *u8, expected: *u8, desired: u8, success: i32, failure: i32) callconv(.C) i32 {
+fn __atomic_compare_exchange_1(ptr: *u8, expected: *u8, desired: u8, success: i32, failure: i32) callconv(.c) i32 {
return atomic_compare_exchange_N(u8, ptr, expected, desired, success, failure);
}
-fn __atomic_compare_exchange_2(ptr: *u16, expected: *u16, desired: u16, success: i32, failure: i32) callconv(.C) i32 {
+fn __atomic_compare_exchange_2(ptr: *u16, expected: *u16, desired: u16, success: i32, failure: i32) callconv(.c) i32 {
return atomic_compare_exchange_N(u16, ptr, expected, desired, success, failure);
}
-fn __atomic_compare_exchange_4(ptr: *u32, expected: *u32, desired: u32, success: i32, failure: i32) callconv(.C) i32 {
+fn __atomic_compare_exchange_4(ptr: *u32, expected: *u32, desired: u32, success: i32, failure: i32) callconv(.c) i32 {
return atomic_compare_exchange_N(u32, ptr, expected, desired, success, failure);
}
-fn __atomic_compare_exchange_8(ptr: *u64, expected: *u64, desired: u64, success: i32, failure: i32) callconv(.C) i32 {
+fn __atomic_compare_exchange_8(ptr: *u64, expected: *u64, desired: u64, success: i32, failure: i32) callconv(.c) i32 {
return atomic_compare_exchange_N(u64, ptr, expected, desired, success, failure);
}
-fn __atomic_compare_exchange_16(ptr: *u128, expected: *u128, desired: u128, success: i32, failure: i32) callconv(.C) i32 {
+fn __atomic_compare_exchange_16(ptr: *u128, expected: *u128, desired: u128, success: i32, failure: i32) callconv(.c) i32 {
return atomic_compare_exchange_N(u128, ptr, expected, desired, success, failure);
}
@@ -376,163 +376,163 @@ inline fn fetch_op_N(comptime T: type, comptime op: std.builtin.AtomicRmwOp, ptr
return @atomicRmw(T, ptr, op, val, .seq_cst);
}
-fn __atomic_fetch_add_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_add_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Add, ptr, val, model);
}
-fn __atomic_fetch_add_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_add_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Add, ptr, val, model);
}
-fn __atomic_fetch_add_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_add_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Add, ptr, val, model);
}
-fn __atomic_fetch_add_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_add_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Add, ptr, val, model);
}
-fn __atomic_fetch_add_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_add_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Add, ptr, val, model);
}
-fn __atomic_fetch_sub_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_sub_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Sub, ptr, val, model);
}
-fn __atomic_fetch_sub_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_sub_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Sub, ptr, val, model);
}
-fn __atomic_fetch_sub_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_sub_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Sub, ptr, val, model);
}
-fn __atomic_fetch_sub_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_sub_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Sub, ptr, val, model);
}
-fn __atomic_fetch_sub_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_sub_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Sub, ptr, val, model);
}
-fn __atomic_fetch_and_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_and_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .And, ptr, val, model);
}
-fn __atomic_fetch_and_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_and_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .And, ptr, val, model);
}
-fn __atomic_fetch_and_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_and_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .And, ptr, val, model);
}
-fn __atomic_fetch_and_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_and_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .And, ptr, val, model);
}
-fn __atomic_fetch_and_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_and_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .And, ptr, val, model);
}
-fn __atomic_fetch_or_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_or_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Or, ptr, val, model);
}
-fn __atomic_fetch_or_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_or_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Or, ptr, val, model);
}
-fn __atomic_fetch_or_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_or_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Or, ptr, val, model);
}
-fn __atomic_fetch_or_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_or_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Or, ptr, val, model);
}
-fn __atomic_fetch_or_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_or_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Or, ptr, val, model);
}
-fn __atomic_fetch_xor_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_xor_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Xor, ptr, val, model);
}
-fn __atomic_fetch_xor_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_xor_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Xor, ptr, val, model);
}
-fn __atomic_fetch_xor_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_xor_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Xor, ptr, val, model);
}
-fn __atomic_fetch_xor_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_xor_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Xor, ptr, val, model);
}
-fn __atomic_fetch_xor_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_xor_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Xor, ptr, val, model);
}
-fn __atomic_fetch_nand_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_nand_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Nand, ptr, val, model);
}
-fn __atomic_fetch_nand_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_nand_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Nand, ptr, val, model);
}
-fn __atomic_fetch_nand_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_nand_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Nand, ptr, val, model);
}
-fn __atomic_fetch_nand_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_nand_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Nand, ptr, val, model);
}
-fn __atomic_fetch_nand_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_nand_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Nand, ptr, val, model);
}
-fn __atomic_fetch_umax_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_umax_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Max, ptr, val, model);
}
-fn __atomic_fetch_umax_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_umax_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Max, ptr, val, model);
}
-fn __atomic_fetch_umax_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_umax_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Max, ptr, val, model);
}
-fn __atomic_fetch_umax_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_umax_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Max, ptr, val, model);
}
-fn __atomic_fetch_umax_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_umax_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Max, ptr, val, model);
}
-fn __atomic_fetch_umin_1(ptr: *u8, val: u8, model: i32) callconv(.C) u8 {
+fn __atomic_fetch_umin_1(ptr: *u8, val: u8, model: i32) callconv(.c) u8 {
return fetch_op_N(u8, .Min, ptr, val, model);
}
-fn __atomic_fetch_umin_2(ptr: *u16, val: u16, model: i32) callconv(.C) u16 {
+fn __atomic_fetch_umin_2(ptr: *u16, val: u16, model: i32) callconv(.c) u16 {
return fetch_op_N(u16, .Min, ptr, val, model);
}
-fn __atomic_fetch_umin_4(ptr: *u32, val: u32, model: i32) callconv(.C) u32 {
+fn __atomic_fetch_umin_4(ptr: *u32, val: u32, model: i32) callconv(.c) u32 {
return fetch_op_N(u32, .Min, ptr, val, model);
}
-fn __atomic_fetch_umin_8(ptr: *u64, val: u64, model: i32) callconv(.C) u64 {
+fn __atomic_fetch_umin_8(ptr: *u64, val: u64, model: i32) callconv(.c) u64 {
return fetch_op_N(u64, .Min, ptr, val, model);
}
-fn __atomic_fetch_umin_16(ptr: *u128, val: u128, model: i32) callconv(.C) u128 {
+fn __atomic_fetch_umin_16(ptr: *u128, val: u128, model: i32) callconv(.c) u128 {
return fetch_op_N(u128, .Min, ptr, val, model);
}
diff --git a/lib/compiler_rt/aulldiv.zig b/lib/compiler_rt/aulldiv.zig
index 8a7baaad3e..6832b66d01 100644
--- a/lib/compiler_rt/aulldiv.zig
+++ b/lib/compiler_rt/aulldiv.zig
@@ -15,7 +15,7 @@ comptime {
}
}
-pub fn _alldiv(a: i64, b: i64) callconv(.Stdcall) i64 {
+pub fn _alldiv(a: i64, b: i64) callconv(.{ .x86_stdcall = .{} }) i64 {
const s_a = a >> (64 - 1);
const s_b = b >> (64 - 1);
@@ -27,7 +27,7 @@ pub fn _alldiv(a: i64, b: i64) callconv(.Stdcall) i64 {
return (@as(i64, @bitCast(r)) ^ s) -% s;
}
-pub fn _aulldiv() callconv(.Naked) void {
+pub fn _aulldiv() callconv(.naked) void {
@setRuntimeSafety(false);
// The stack layout is:
diff --git a/lib/compiler_rt/aullrem.zig b/lib/compiler_rt/aullrem.zig
index 182ad4e1de..3934225c8c 100644
--- a/lib/compiler_rt/aullrem.zig
+++ b/lib/compiler_rt/aullrem.zig
@@ -15,7 +15,7 @@ comptime {
}
}
-pub fn _allrem(a: i64, b: i64) callconv(.Stdcall) i64 {
+pub fn _allrem(a: i64, b: i64) callconv(.{ .x86_stdcall = .{} }) i64 {
const s_a = a >> (64 - 1);
const s_b = b >> (64 - 1);
@@ -27,7 +27,7 @@ pub fn _allrem(a: i64, b: i64) callconv(.Stdcall) i64 {
return (@as(i64, @bitCast(r)) ^ s) -% s;
}
-pub fn _aullrem() callconv(.Naked) void {
+pub fn _aullrem() callconv(.naked) void {
@setRuntimeSafety(false);
// The stack layout is:
diff --git a/lib/compiler_rt/bcmp.zig b/lib/compiler_rt/bcmp.zig
index 829dc03715..eb1fca21f0 100644
--- a/lib/compiler_rt/bcmp.zig
+++ b/lib/compiler_rt/bcmp.zig
@@ -5,7 +5,7 @@ comptime {
@export(&bcmp, .{ .name = "bcmp", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn bcmp(vl: [*]allowzero const u8, vr: [*]allowzero const u8, n: usize) callconv(.C) c_int {
+pub fn bcmp(vl: [*]allowzero const u8, vr: [*]allowzero const u8, n: usize) callconv(.c) c_int {
@setRuntimeSafety(false);
var index: usize = 0;
diff --git a/lib/compiler_rt/bitreverse.zig b/lib/compiler_rt/bitreverse.zig
index 4022f8d72c..fe6cf8988d 100644
--- a/lib/compiler_rt/bitreverse.zig
+++ b/lib/compiler_rt/bitreverse.zig
@@ -46,15 +46,15 @@ inline fn bitreverseXi2(comptime T: type, a: T) T {
}
}
-pub fn __bitreversesi2(a: u32) callconv(.C) u32 {
+pub fn __bitreversesi2(a: u32) callconv(.c) u32 {
return bitreverseXi2(u32, a);
}
-pub fn __bitreversedi2(a: u64) callconv(.C) u64 {
+pub fn __bitreversedi2(a: u64) callconv(.c) u64 {
return bitreverseXi2(u64, a);
}
-pub fn __bitreverseti2(a: u128) callconv(.C) u128 {
+pub fn __bitreverseti2(a: u128) callconv(.c) u128 {
return bitreverseXi2(u128, a);
}
diff --git a/lib/compiler_rt/bswap.zig b/lib/compiler_rt/bswap.zig
index 4bd8ff688c..2d6df339df 100644
--- a/lib/compiler_rt/bswap.zig
+++ b/lib/compiler_rt/bswap.zig
@@ -66,15 +66,15 @@ inline fn bswapXi2(comptime T: type, a: T) T {
}
}
-pub fn __bswapsi2(a: u32) callconv(.C) u32 {
+pub fn __bswapsi2(a: u32) callconv(.c) u32 {
return bswapXi2(u32, a);
}
-pub fn __bswapdi2(a: u64) callconv(.C) u64 {
+pub fn __bswapdi2(a: u64) callconv(.c) u64 {
return bswapXi2(u64, a);
}
-pub fn __bswapti2(a: u128) callconv(.C) u128 {
+pub fn __bswapti2(a: u128) callconv(.c) u128 {
return bswapXi2(u128, a);
}
diff --git a/lib/compiler_rt/ceil.zig b/lib/compiler_rt/ceil.zig
index e1139ff9f4..5edca4a999 100644
--- a/lib/compiler_rt/ceil.zig
+++ b/lib/compiler_rt/ceil.zig
@@ -26,12 +26,12 @@ comptime {
@export(&ceill, .{ .name = "ceill", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __ceilh(x: f16) callconv(.C) f16 {
+pub fn __ceilh(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(ceilf(x));
}
-pub fn ceilf(x: f32) callconv(.C) f32 {
+pub fn ceilf(x: f32) callconv(.c) f32 {
var u: u32 = @bitCast(x);
const e = @as(i32, @intCast((u >> 23) & 0xFF)) - 0x7F;
var m: u32 = undefined;
@@ -64,7 +64,7 @@ pub fn ceilf(x: f32) callconv(.C) f32 {
}
}
-pub fn ceil(x: f64) callconv(.C) f64 {
+pub fn ceil(x: f64) callconv(.c) f64 {
const f64_toint = 1.0 / math.floatEps(f64);
const u: u64 = @bitCast(x);
@@ -95,12 +95,12 @@ pub fn ceil(x: f64) callconv(.C) f64 {
}
}
-pub fn __ceilx(x: f80) callconv(.C) f80 {
+pub fn __ceilx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(ceilq(x));
}
-pub fn ceilq(x: f128) callconv(.C) f128 {
+pub fn ceilq(x: f128) callconv(.c) f128 {
const f128_toint = 1.0 / math.floatEps(f128);
const u: u128 = @bitCast(x);
@@ -129,7 +129,7 @@ pub fn ceilq(x: f128) callconv(.C) f128 {
}
}
-pub fn ceill(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn ceill(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __ceilh(x),
32 => return ceilf(x),
diff --git a/lib/compiler_rt/clear_cache.zig b/lib/compiler_rt/clear_cache.zig
index c05a38f846..193535e874 100644
--- a/lib/compiler_rt/clear_cache.zig
+++ b/lib/compiler_rt/clear_cache.zig
@@ -15,7 +15,7 @@ comptime {
_ = &clear_cache;
}
-fn clear_cache(start: usize, end: usize) callconv(.C) void {
+fn clear_cache(start: usize, end: usize) callconv(.c) void {
const x86 = switch (arch) {
.x86, .x86_64 => true,
else => false,
diff --git a/lib/compiler_rt/clzsi2_test.zig b/lib/compiler_rt/clzsi2_test.zig
index 188290e90b..f0319d6795 100644
--- a/lib/compiler_rt/clzsi2_test.zig
+++ b/lib/compiler_rt/clzsi2_test.zig
@@ -4,7 +4,7 @@ const testing = @import("std").testing;
fn test__clzsi2(a: u32, expected: i32) !void {
const nakedClzsi2 = clz.__clzsi2;
- const actualClzsi2 = @as(*const fn (a: i32) callconv(.C) i32, @ptrCast(&nakedClzsi2));
+ const actualClzsi2 = @as(*const fn (a: i32) callconv(.c) i32, @ptrCast(&nakedClzsi2));
const x: i32 = @bitCast(a);
const result = actualClzsi2(x);
try testing.expectEqual(expected, result);
diff --git a/lib/compiler_rt/cmp.zig b/lib/compiler_rt/cmp.zig
index 6ae608fcb0..e1273aa622 100644
--- a/lib/compiler_rt/cmp.zig
+++ b/lib/compiler_rt/cmp.zig
@@ -34,27 +34,27 @@ inline fn XcmpXi2(comptime T: type, a: T, b: T) i32 {
return cmp1 - cmp2 + 1;
}
-pub fn __cmpsi2(a: i32, b: i32) callconv(.C) i32 {
+pub fn __cmpsi2(a: i32, b: i32) callconv(.c) i32 {
return XcmpXi2(i32, a, b);
}
-pub fn __cmpdi2(a: i64, b: i64) callconv(.C) i32 {
+pub fn __cmpdi2(a: i64, b: i64) callconv(.c) i32 {
return XcmpXi2(i64, a, b);
}
-pub fn __cmpti2(a: i128, b: i128) callconv(.C) i32 {
+pub fn __cmpti2(a: i128, b: i128) callconv(.c) i32 {
return XcmpXi2(i128, a, b);
}
-pub fn __ucmpsi2(a: u32, b: u32) callconv(.C) i32 {
+pub fn __ucmpsi2(a: u32, b: u32) callconv(.c) i32 {
return XcmpXi2(u32, a, b);
}
-pub fn __ucmpdi2(a: u64, b: u64) callconv(.C) i32 {
+pub fn __ucmpdi2(a: u64, b: u64) callconv(.c) i32 {
return XcmpXi2(u64, a, b);
}
-pub fn __ucmpti2(a: u128, b: u128) callconv(.C) i32 {
+pub fn __ucmpti2(a: u128, b: u128) callconv(.c) i32 {
return XcmpXi2(u128, a, b);
}
diff --git a/lib/compiler_rt/cmpdf2.zig b/lib/compiler_rt/cmpdf2.zig
index 47641145de..a0338c7a17 100644
--- a/lib/compiler_rt/cmpdf2.zig
+++ b/lib/compiler_rt/cmpdf2.zig
@@ -25,44 +25,44 @@ comptime {
///
/// Note that this matches the definition of `__ledf2`, `__eqdf2`, `__nedf2`, `__cmpdf2`,
/// and `__ltdf2`.
-fn __cmpdf2(a: f64, b: f64) callconv(.C) i32 {
+fn __cmpdf2(a: f64, b: f64) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f64, comparef.LE, a, b));
}
/// "These functions return a value less than or equal to zero if neither argument is NaN,
/// and a is less than or equal to b."
-pub fn __ledf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __ledf2(a: f64, b: f64) callconv(.c) i32 {
return __cmpdf2(a, b);
}
/// "These functions return zero if neither argument is NaN, and a and b are equal."
/// Note that due to some kind of historical accident, __eqdf2 and __nedf2 are defined
/// to have the same return value.
-pub fn __eqdf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __eqdf2(a: f64, b: f64) callconv(.c) i32 {
return __cmpdf2(a, b);
}
/// "These functions return a nonzero value if either argument is NaN, or if a and b are unequal."
/// Note that due to some kind of historical accident, __eqdf2 and __nedf2 are defined
/// to have the same return value.
-pub fn __nedf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __nedf2(a: f64, b: f64) callconv(.c) i32 {
return __cmpdf2(a, b);
}
/// "These functions return a value less than zero if neither argument is NaN, and a
/// is strictly less than b."
-pub fn __ltdf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __ltdf2(a: f64, b: f64) callconv(.c) i32 {
return __cmpdf2(a, b);
}
-fn __aeabi_dcmpeq(a: f64, b: f64) callconv(.AAPCS) i32 {
+fn __aeabi_dcmpeq(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f64, comparef.LE, a, b) == .Equal);
}
-fn __aeabi_dcmplt(a: f64, b: f64) callconv(.AAPCS) i32 {
+fn __aeabi_dcmplt(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f64, comparef.LE, a, b) == .Less);
}
-fn __aeabi_dcmple(a: f64, b: f64) callconv(.AAPCS) i32 {
+fn __aeabi_dcmple(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f64, comparef.LE, a, b) != .Greater);
}
diff --git a/lib/compiler_rt/cmphf2.zig b/lib/compiler_rt/cmphf2.zig
index 15b9ad610f..9fcc1e24d5 100644
--- a/lib/compiler_rt/cmphf2.zig
+++ b/lib/compiler_rt/cmphf2.zig
@@ -19,32 +19,32 @@ comptime {
///
/// Note that this matches the definition of `__lehf2`, `__eqhf2`, `__nehf2`, `__cmphf2`,
/// and `__lthf2`.
-fn __cmphf2(a: f16, b: f16) callconv(.C) i32 {
+fn __cmphf2(a: f16, b: f16) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f16, comparef.LE, a, b));
}
/// "These functions return a value less than or equal to zero if neither argument is NaN,
/// and a is less than or equal to b."
-pub fn __lehf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __lehf2(a: f16, b: f16) callconv(.c) i32 {
return __cmphf2(a, b);
}
/// "These functions return zero if neither argument is NaN, and a and b are equal."
/// Note that due to some kind of historical accident, __eqhf2 and __nehf2 are defined
/// to have the same return value.
-pub fn __eqhf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __eqhf2(a: f16, b: f16) callconv(.c) i32 {
return __cmphf2(a, b);
}
/// "These functions return a nonzero value if either argument is NaN, or if a and b are unequal."
/// Note that due to some kind of historical accident, __eqhf2 and __nehf2 are defined
/// to have the same return value.
-pub fn __nehf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __nehf2(a: f16, b: f16) callconv(.c) i32 {
return __cmphf2(a, b);
}
/// "These functions return a value less than zero if neither argument is NaN, and a
/// is strictly less than b."
-pub fn __lthf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __lthf2(a: f16, b: f16) callconv(.c) i32 {
return __cmphf2(a, b);
}
diff --git a/lib/compiler_rt/cmpsf2.zig b/lib/compiler_rt/cmpsf2.zig
index 3f65021b89..371319b072 100644
--- a/lib/compiler_rt/cmpsf2.zig
+++ b/lib/compiler_rt/cmpsf2.zig
@@ -25,44 +25,44 @@ comptime {
///
/// Note that this matches the definition of `__lesf2`, `__eqsf2`, `__nesf2`, `__cmpsf2`,
/// and `__ltsf2`.
-fn __cmpsf2(a: f32, b: f32) callconv(.C) i32 {
+fn __cmpsf2(a: f32, b: f32) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f32, comparef.LE, a, b));
}
/// "These functions return a value less than or equal to zero if neither argument is NaN,
/// and a is less than or equal to b."
-pub fn __lesf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __lesf2(a: f32, b: f32) callconv(.c) i32 {
return __cmpsf2(a, b);
}
/// "These functions return zero if neither argument is NaN, and a and b are equal."
/// Note that due to some kind of historical accident, __eqsf2 and __nesf2 are defined
/// to have the same return value.
-pub fn __eqsf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __eqsf2(a: f32, b: f32) callconv(.c) i32 {
return __cmpsf2(a, b);
}
/// "These functions return a nonzero value if either argument is NaN, or if a and b are unequal."
/// Note that due to some kind of historical accident, __eqsf2 and __nesf2 are defined
/// to have the same return value.
-pub fn __nesf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __nesf2(a: f32, b: f32) callconv(.c) i32 {
return __cmpsf2(a, b);
}
/// "These functions return a value less than zero if neither argument is NaN, and a
/// is strictly less than b."
-pub fn __ltsf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __ltsf2(a: f32, b: f32) callconv(.c) i32 {
return __cmpsf2(a, b);
}
-fn __aeabi_fcmpeq(a: f32, b: f32) callconv(.AAPCS) i32 {
+fn __aeabi_fcmpeq(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f32, comparef.LE, a, b) == .Equal);
}
-fn __aeabi_fcmplt(a: f32, b: f32) callconv(.AAPCS) i32 {
+fn __aeabi_fcmplt(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f32, comparef.LE, a, b) == .Less);
}
-fn __aeabi_fcmple(a: f32, b: f32) callconv(.AAPCS) i32 {
+fn __aeabi_fcmple(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f32, comparef.LE, a, b) != .Greater);
}
diff --git a/lib/compiler_rt/cmptf2.zig b/lib/compiler_rt/cmptf2.zig
index 41e7f4ba08..d5ee7970d4 100644
--- a/lib/compiler_rt/cmptf2.zig
+++ b/lib/compiler_rt/cmptf2.zig
@@ -33,33 +33,33 @@ comptime {
///
/// Note that this matches the definition of `__letf2`, `__eqtf2`, `__netf2`, `__cmptf2`,
/// and `__lttf2`.
-fn __cmptf2(a: f128, b: f128) callconv(.C) i32 {
+fn __cmptf2(a: f128, b: f128) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f128, comparef.LE, a, b));
}
/// "These functions return a value less than or equal to zero if neither argument is NaN,
/// and a is less than or equal to b."
-fn __letf2(a: f128, b: f128) callconv(.C) i32 {
+fn __letf2(a: f128, b: f128) callconv(.c) i32 {
return __cmptf2(a, b);
}
/// "These functions return zero if neither argument is NaN, and a and b are equal."
/// Note that due to some kind of historical accident, __eqtf2 and __netf2 are defined
/// to have the same return value.
-fn __eqtf2(a: f128, b: f128) callconv(.C) i32 {
+fn __eqtf2(a: f128, b: f128) callconv(.c) i32 {
return __cmptf2(a, b);
}
/// "These functions return a nonzero value if either argument is NaN, or if a and b are unequal."
/// Note that due to some kind of historical accident, __eqtf2 and __netf2 are defined
/// to have the same return value.
-fn __netf2(a: f128, b: f128) callconv(.C) i32 {
+fn __netf2(a: f128, b: f128) callconv(.c) i32 {
return __cmptf2(a, b);
}
/// "These functions return a value less than zero if neither argument is NaN, and a
/// is strictly less than b."
-fn __lttf2(a: f128, b: f128) callconv(.C) i32 {
+fn __lttf2(a: f128, b: f128) callconv(.c) i32 {
return __cmptf2(a, b);
}
@@ -70,34 +70,34 @@ const SparcFCMP = enum(i32) {
Unordered = 3,
};
-fn _Qp_cmp(a: *const f128, b: *const f128) callconv(.C) i32 {
+fn _Qp_cmp(a: *const f128, b: *const f128) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f128, SparcFCMP, a.*, b.*));
}
-fn _Qp_feq(a: *const f128, b: *const f128) callconv(.C) bool {
+fn _Qp_feq(a: *const f128, b: *const f128) callconv(.c) bool {
return @as(SparcFCMP, @enumFromInt(_Qp_cmp(a, b))) == .Equal;
}
-fn _Qp_fne(a: *const f128, b: *const f128) callconv(.C) bool {
+fn _Qp_fne(a: *const f128, b: *const f128) callconv(.c) bool {
return @as(SparcFCMP, @enumFromInt(_Qp_cmp(a, b))) != .Equal;
}
-fn _Qp_flt(a: *const f128, b: *const f128) callconv(.C) bool {
+fn _Qp_flt(a: *const f128, b: *const f128) callconv(.c) bool {
return @as(SparcFCMP, @enumFromInt(_Qp_cmp(a, b))) == .Less;
}
-fn _Qp_fgt(a: *const f128, b: *const f128) callconv(.C) bool {
+fn _Qp_fgt(a: *const f128, b: *const f128) callconv(.c) bool {
return @as(SparcFCMP, @enumFromInt(_Qp_cmp(a, b))) == .Greater;
}
-fn _Qp_fge(a: *const f128, b: *const f128) callconv(.C) bool {
+fn _Qp_fge(a: *const f128, b: *const f128) callconv(.c) bool {
return switch (@as(SparcFCMP, @enumFromInt(_Qp_cmp(a, b)))) {
.Equal, .Greater => true,
.Less, .Unordered => false,
};
}
-fn _Qp_fle(a: *const f128, b: *const f128) callconv(.C) bool {
+fn _Qp_fle(a: *const f128, b: *const f128) callconv(.c) bool {
return switch (@as(SparcFCMP, @enumFromInt(_Qp_cmp(a, b)))) {
.Equal, .Less => true,
.Greater, .Unordered => false,
diff --git a/lib/compiler_rt/cmpxf2.zig b/lib/compiler_rt/cmpxf2.zig
index a020f60824..dc90f87c34 100644
--- a/lib/compiler_rt/cmpxf2.zig
+++ b/lib/compiler_rt/cmpxf2.zig
@@ -19,32 +19,32 @@ comptime {
///
/// Note that this matches the definition of `__lexf2`, `__eqxf2`, `__nexf2`, `__cmpxf2`,
/// and `__ltxf2`.
-fn __cmpxf2(a: f80, b: f80) callconv(.C) i32 {
+fn __cmpxf2(a: f80, b: f80) callconv(.c) i32 {
return @intFromEnum(comparef.cmp_f80(comparef.LE, a, b));
}
/// "These functions return a value less than or equal to zero if neither argument is NaN,
/// and a is less than or equal to b."
-fn __lexf2(a: f80, b: f80) callconv(.C) i32 {
+fn __lexf2(a: f80, b: f80) callconv(.c) i32 {
return __cmpxf2(a, b);
}
/// "These functions return zero if neither argument is NaN, and a and b are equal."
/// Note that due to some kind of historical accident, __eqxf2 and __nexf2 are defined
/// to have the same return value.
-fn __eqxf2(a: f80, b: f80) callconv(.C) i32 {
+fn __eqxf2(a: f80, b: f80) callconv(.c) i32 {
return __cmpxf2(a, b);
}
/// "These functions return a nonzero value if either argument is NaN, or if a and b are unequal."
/// Note that due to some kind of historical accident, __eqxf2 and __nexf2 are defined
/// to have the same return value.
-fn __nexf2(a: f80, b: f80) callconv(.C) i32 {
+fn __nexf2(a: f80, b: f80) callconv(.c) i32 {
return __cmpxf2(a, b);
}
/// "These functions return a value less than zero if neither argument is NaN, and a
/// is strictly less than b."
-fn __ltxf2(a: f80, b: f80) callconv(.C) i32 {
+fn __ltxf2(a: f80, b: f80) callconv(.c) i32 {
return __cmpxf2(a, b);
}
diff --git a/lib/compiler_rt/cos.zig b/lib/compiler_rt/cos.zig
index 5cbe5af6e8..01a5196232 100644
--- a/lib/compiler_rt/cos.zig
+++ b/lib/compiler_rt/cos.zig
@@ -22,12 +22,12 @@ comptime {
@export(&cosl, .{ .name = "cosl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __cosh(a: f16) callconv(.C) f16 {
+pub fn __cosh(a: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(cosf(a));
}
-pub fn cosf(x: f32) callconv(.C) f32 {
+pub fn cosf(x: f32) callconv(.c) f32 {
// Small multiples of pi/2 rounded to double precision.
const c1pio2: f64 = 1.0 * math.pi / 2.0; // 0x3FF921FB, 0x54442D18
const c2pio2: f64 = 2.0 * math.pi / 2.0; // 0x400921FB, 0x54442D18
@@ -84,7 +84,7 @@ pub fn cosf(x: f32) callconv(.C) f32 {
};
}
-pub fn cos(x: f64) callconv(.C) f64 {
+pub fn cos(x: f64) callconv(.c) f64 {
var ix = @as(u64, @bitCast(x)) >> 32;
ix &= 0x7fffffff;
@@ -113,17 +113,17 @@ pub fn cos(x: f64) callconv(.C) f64 {
};
}
-pub fn __cosx(a: f80) callconv(.C) f80 {
+pub fn __cosx(a: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(cosq(a));
}
-pub fn cosq(a: f128) callconv(.C) f128 {
+pub fn cosq(a: f128) callconv(.c) f128 {
// TODO: more correct implementation
return cos(@floatCast(a));
}
-pub fn cosl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn cosl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __cosh(x),
32 => return cosf(x),
diff --git a/lib/compiler_rt/count0bits.zig b/lib/compiler_rt/count0bits.zig
index c2fba8ba1e..cf89d2f962 100644
--- a/lib/compiler_rt/count0bits.zig
+++ b/lib/compiler_rt/count0bits.zig
@@ -52,7 +52,7 @@ inline fn clzXi2(comptime T: type, a: T) i32 {
return @intCast(n - @as(T, @bitCast(x)));
}
-fn __clzsi2_thumb1() callconv(.Naked) void {
+fn __clzsi2_thumb1() callconv(.naked) void {
@setRuntimeSafety(false);
// Similar to the generic version with the last two rounds replaced by a LUT
@@ -86,7 +86,7 @@ fn __clzsi2_thumb1() callconv(.Naked) void {
unreachable;
}
-fn __clzsi2_arm32() callconv(.Naked) void {
+fn __clzsi2_arm32() callconv(.naked) void {
@setRuntimeSafety(false);
asm volatile (
@@ -135,7 +135,7 @@ fn __clzsi2_arm32() callconv(.Naked) void {
unreachable;
}
-fn clzsi2_generic(a: i32) callconv(.C) i32 {
+fn clzsi2_generic(a: i32) callconv(.c) i32 {
return clzXi2(i32, a);
}
@@ -159,11 +159,11 @@ pub const __clzsi2 = switch (builtin.cpu.arch) {
else => clzsi2_generic,
};
-pub fn __clzdi2(a: i64) callconv(.C) i32 {
+pub fn __clzdi2(a: i64) callconv(.c) i32 {
return clzXi2(i64, a);
}
-pub fn __clzti2(a: i128) callconv(.C) i32 {
+pub fn __clzti2(a: i128) callconv(.c) i32 {
return clzXi2(i128, a);
}
@@ -190,15 +190,15 @@ inline fn ctzXi2(comptime T: type, a: T) i32 {
return @intCast(n - @as(T, @bitCast((x & 1))));
}
-pub fn __ctzsi2(a: i32) callconv(.C) i32 {
+pub fn __ctzsi2(a: i32) callconv(.c) i32 {
return ctzXi2(i32, a);
}
-pub fn __ctzdi2(a: i64) callconv(.C) i32 {
+pub fn __ctzdi2(a: i64) callconv(.c) i32 {
return ctzXi2(i64, a);
}
-pub fn __ctzti2(a: i128) callconv(.C) i32 {
+pub fn __ctzti2(a: i128) callconv(.c) i32 {
return ctzXi2(i128, a);
}
@@ -222,15 +222,15 @@ inline fn ffsXi2(comptime T: type, a: T) i32 {
return @as(i32, @intCast(n - @as(T, @bitCast((x & 1))))) + 1;
}
-pub fn __ffssi2(a: i32) callconv(.C) i32 {
+pub fn __ffssi2(a: i32) callconv(.c) i32 {
return ffsXi2(i32, a);
}
-pub fn __ffsdi2(a: i64) callconv(.C) i32 {
+pub fn __ffsdi2(a: i64) callconv(.c) i32 {
return ffsXi2(i64, a);
}
-pub fn __ffsti2(a: i128) callconv(.C) i32 {
+pub fn __ffsti2(a: i128) callconv(.c) i32 {
return ffsXi2(i128, a);
}
diff --git a/lib/compiler_rt/divc3_test.zig b/lib/compiler_rt/divc3_test.zig
index f84ca9ba1f..d3c400d46c 100644
--- a/lib/compiler_rt/divc3_test.zig
+++ b/lib/compiler_rt/divc3_test.zig
@@ -17,7 +17,7 @@ test "divc3" {
try testDiv(f128, __divtc3);
}
-fn testDiv(comptime T: type, comptime f: fn (T, T, T, T) callconv(.C) Complex(T)) !void {
+fn testDiv(comptime T: type, comptime f: fn (T, T, T, T) callconv(.c) Complex(T)) !void {
{
const a: T = 1.0;
const b: T = 0.0;
diff --git a/lib/compiler_rt/divdc3.zig b/lib/compiler_rt/divdc3.zig
index 6a555b78db..ce2797ee39 100644
--- a/lib/compiler_rt/divdc3.zig
+++ b/lib/compiler_rt/divdc3.zig
@@ -8,6 +8,6 @@ comptime {
}
}
-pub fn __divdc3(a: f64, b: f64, c: f64, d: f64) callconv(.C) Complex(f64) {
+pub fn __divdc3(a: f64, b: f64, c: f64, d: f64) callconv(.c) Complex(f64) {
return divc3.divc3(f64, a, b, c, d);
}
diff --git a/lib/compiler_rt/divdf3.zig b/lib/compiler_rt/divdf3.zig
index 29b82d360d..0340404a69 100644
--- a/lib/compiler_rt/divdf3.zig
+++ b/lib/compiler_rt/divdf3.zig
@@ -21,11 +21,11 @@ comptime {
}
}
-pub fn __divdf3(a: f64, b: f64) callconv(.C) f64 {
+pub fn __divdf3(a: f64, b: f64) callconv(.c) f64 {
return div(a, b);
}
-fn __aeabi_ddiv(a: f64, b: f64) callconv(.AAPCS) f64 {
+fn __aeabi_ddiv(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) f64 {
return div(a, b);
}
diff --git a/lib/compiler_rt/divhc3.zig b/lib/compiler_rt/divhc3.zig
index 0a5053bfa8..8243c545b1 100644
--- a/lib/compiler_rt/divhc3.zig
+++ b/lib/compiler_rt/divhc3.zig
@@ -8,6 +8,6 @@ comptime {
}
}
-pub fn __divhc3(a: f16, b: f16, c: f16, d: f16) callconv(.C) Complex(f16) {
+pub fn __divhc3(a: f16, b: f16, c: f16, d: f16) callconv(.c) Complex(f16) {
return divc3.divc3(f16, a, b, c, d);
}
diff --git a/lib/compiler_rt/divhf3.zig b/lib/compiler_rt/divhf3.zig
index 23c6d60191..84bfe37871 100644
--- a/lib/compiler_rt/divhf3.zig
+++ b/lib/compiler_rt/divhf3.zig
@@ -5,7 +5,7 @@ comptime {
@export(&__divhf3, .{ .name = "__divhf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __divhf3(a: f16, b: f16) callconv(.C) f16 {
+pub fn __divhf3(a: f16, b: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(divsf3.__divsf3(a, b));
}
diff --git a/lib/compiler_rt/divmodei4.zig b/lib/compiler_rt/divmodei4.zig
index c26c7418fe..f5ed41d9d7 100644
--- a/lib/compiler_rt/divmodei4.zig
+++ b/lib/compiler_rt/divmodei4.zig
@@ -33,7 +33,7 @@ fn divmod(q: ?[]u32, r: ?[]u32, u: []u32, v: []u32) !void {
if (r) |x| if (u_sign < 0) neg(x);
}
-pub fn __divei4(r_q: [*]u32, u_p: [*]u32, v_p: [*]u32, bits: usize) callconv(.C) void {
+pub fn __divei4(r_q: [*]u32, u_p: [*]u32, v_p: [*]u32, bits: usize) callconv(.c) void {
@setRuntimeSafety(builtin.is_test);
const u = u_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
const v = v_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
@@ -41,7 +41,7 @@ pub fn __divei4(r_q: [*]u32, u_p: [*]u32, v_p: [*]u32, bits: usize) callconv(.C)
@call(.always_inline, divmod, .{ q, null, u, v }) catch unreachable;
}
-pub fn __modei4(r_p: [*]u32, u_p: [*]u32, v_p: [*]u32, bits: usize) callconv(.C) void {
+pub fn __modei4(r_p: [*]u32, u_p: [*]u32, v_p: [*]u32, bits: usize) callconv(.c) void {
@setRuntimeSafety(builtin.is_test);
const u = u_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
const v = v_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
diff --git a/lib/compiler_rt/divsc3.zig b/lib/compiler_rt/divsc3.zig
index d1af430508..504d8f7d67 100644
--- a/lib/compiler_rt/divsc3.zig
+++ b/lib/compiler_rt/divsc3.zig
@@ -8,6 +8,6 @@ comptime {
}
}
-pub fn __divsc3(a: f32, b: f32, c: f32, d: f32) callconv(.C) Complex(f32) {
+pub fn __divsc3(a: f32, b: f32, c: f32, d: f32) callconv(.c) Complex(f32) {
return divc3.divc3(f32, a, b, c, d);
}
diff --git a/lib/compiler_rt/divsf3.zig b/lib/compiler_rt/divsf3.zig
index 181f95cef6..a74d3f7abd 100644
--- a/lib/compiler_rt/divsf3.zig
+++ b/lib/compiler_rt/divsf3.zig
@@ -19,11 +19,11 @@ comptime {
}
}
-pub fn __divsf3(a: f32, b: f32) callconv(.C) f32 {
+pub fn __divsf3(a: f32, b: f32) callconv(.c) f32 {
return div(a, b);
}
-fn __aeabi_fdiv(a: f32, b: f32) callconv(.AAPCS) f32 {
+fn __aeabi_fdiv(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) f32 {
return div(a, b);
}
diff --git a/lib/compiler_rt/divtc3.zig b/lib/compiler_rt/divtc3.zig
index bfc8a4badf..3bdf2a9033 100644
--- a/lib/compiler_rt/divtc3.zig
+++ b/lib/compiler_rt/divtc3.zig
@@ -10,6 +10,6 @@ comptime {
}
}
-pub fn __divtc3(a: f128, b: f128, c: f128, d: f128) callconv(.C) Complex(f128) {
+pub fn __divtc3(a: f128, b: f128, c: f128, d: f128) callconv(.c) Complex(f128) {
return divc3.divc3(f128, a, b, c, d);
}
diff --git a/lib/compiler_rt/divtf3.zig b/lib/compiler_rt/divtf3.zig
index b6067e73d3..0bab96de39 100644
--- a/lib/compiler_rt/divtf3.zig
+++ b/lib/compiler_rt/divtf3.zig
@@ -16,11 +16,11 @@ comptime {
@export(&__divtf3, .{ .name = "__divtf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __divtf3(a: f128, b: f128) callconv(.C) f128 {
+pub fn __divtf3(a: f128, b: f128) callconv(.c) f128 {
return div(a, b);
}
-fn _Qp_div(c: *f128, a: *const f128, b: *const f128) callconv(.C) void {
+fn _Qp_div(c: *f128, a: *const f128, b: *const f128) callconv(.c) void {
c.* = div(a.*, b.*);
}
diff --git a/lib/compiler_rt/divti3.zig b/lib/compiler_rt/divti3.zig
index 0468175db1..5f24d196d7 100644
--- a/lib/compiler_rt/divti3.zig
+++ b/lib/compiler_rt/divti3.zig
@@ -14,13 +14,13 @@ comptime {
}
}
-pub fn __divti3(a: i128, b: i128) callconv(.C) i128 {
+pub fn __divti3(a: i128, b: i128) callconv(.c) i128 {
return div(a, b);
}
const v128 = @Vector(2, u64);
-fn __divti3_windows_x86_64(a: v128, b: v128) callconv(.C) v128 {
+fn __divti3_windows_x86_64(a: v128, b: v128) callconv(.c) v128 {
return @bitCast(div(@bitCast(a), @bitCast(b)));
}
diff --git a/lib/compiler_rt/divxc3.zig b/lib/compiler_rt/divxc3.zig
index 86ef5cbb73..6e5f7905dd 100644
--- a/lib/compiler_rt/divxc3.zig
+++ b/lib/compiler_rt/divxc3.zig
@@ -8,6 +8,6 @@ comptime {
}
}
-pub fn __divxc3(a: f80, b: f80, c: f80, d: f80) callconv(.C) Complex(f80) {
+pub fn __divxc3(a: f80, b: f80, c: f80, d: f80) callconv(.c) Complex(f80) {
return divc3.divc3(f80, a, b, c, d);
}
diff --git a/lib/compiler_rt/divxf3.zig b/lib/compiler_rt/divxf3.zig
index 0f50233546..1579db226c 100644
--- a/lib/compiler_rt/divxf3.zig
+++ b/lib/compiler_rt/divxf3.zig
@@ -12,7 +12,7 @@ comptime {
@export(&__divxf3, .{ .name = "__divxf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __divxf3(a: f80, b: f80) callconv(.C) f80 {
+pub fn __divxf3(a: f80, b: f80) callconv(.c) f80 {
const T = f80;
const Z = std.meta.Int(.unsigned, @bitSizeOf(T));
diff --git a/lib/compiler_rt/emutls.zig b/lib/compiler_rt/emutls.zig
index b0244a14f5..434d0c0913 100644
--- a/lib/compiler_rt/emutls.zig
+++ b/lib/compiler_rt/emutls.zig
@@ -24,7 +24,7 @@ comptime {
}
/// public entrypoint for generated code using EmulatedTLS
-pub fn __emutls_get_address(control: *emutls_control) callconv(.C) *anyopaque {
+pub fn __emutls_get_address(control: *emutls_control) callconv(.c) *anyopaque {
return control.getPointer();
}
@@ -191,7 +191,7 @@ const current_thread_storage = struct {
}
/// Invoked by pthread specific destructor. the passed argument is the ObjectArray pointer.
- fn deinit(arrayPtr: *anyopaque) callconv(.C) void {
+ fn deinit(arrayPtr: *anyopaque) callconv(.c) void {
var array: *ObjectArray = @ptrCast(@alignCast(arrayPtr));
array.deinit();
}
diff --git a/lib/compiler_rt/exp.zig b/lib/compiler_rt/exp.zig
index 7245abe191..178d46617e 100644
--- a/lib/compiler_rt/exp.zig
+++ b/lib/compiler_rt/exp.zig
@@ -26,12 +26,12 @@ comptime {
@export(&expl, .{ .name = "expl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __exph(a: f16) callconv(.C) f16 {
+pub fn __exph(a: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(expf(a));
}
-pub fn expf(x_: f32) callconv(.C) f32 {
+pub fn expf(x_: f32) callconv(.c) f32 {
const half = [_]f32{ 0.5, -0.5 };
const ln2hi = 6.9314575195e-1;
const ln2lo = 1.4286067653e-6;
@@ -106,7 +106,7 @@ pub fn expf(x_: f32) callconv(.C) f32 {
}
}
-pub fn exp(x_: f64) callconv(.C) f64 {
+pub fn exp(x_: f64) callconv(.c) f64 {
const half = [_]f64{ 0.5, -0.5 };
const ln2hi: f64 = 6.93147180369123816490e-01;
const ln2lo: f64 = 1.90821492927058770002e-10;
@@ -190,17 +190,17 @@ pub fn exp(x_: f64) callconv(.C) f64 {
}
}
-pub fn __expx(a: f80) callconv(.C) f80 {
+pub fn __expx(a: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(expq(a));
}
-pub fn expq(a: f128) callconv(.C) f128 {
+pub fn expq(a: f128) callconv(.c) f128 {
// TODO: more correct implementation
return exp(@floatCast(a));
}
-pub fn expl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn expl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __exph(x),
32 => return expf(x),
diff --git a/lib/compiler_rt/exp2.zig b/lib/compiler_rt/exp2.zig
index 5aa43a6543..66e58ee463 100644
--- a/lib/compiler_rt/exp2.zig
+++ b/lib/compiler_rt/exp2.zig
@@ -26,12 +26,12 @@ comptime {
@export(&exp2l, .{ .name = "exp2l", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __exp2h(x: f16) callconv(.C) f16 {
+pub fn __exp2h(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(exp2f(x));
}
-pub fn exp2f(x: f32) callconv(.C) f32 {
+pub fn exp2f(x: f32) callconv(.c) f32 {
const tblsiz: u32 = @intCast(exp2ft.len);
const redux: f32 = 0x1.8p23 / @as(f32, @floatFromInt(tblsiz));
const P1: f32 = 0x1.62e430p-1;
@@ -88,7 +88,7 @@ pub fn exp2f(x: f32) callconv(.C) f32 {
return @floatCast(r * uk);
}
-pub fn exp2(x: f64) callconv(.C) f64 {
+pub fn exp2(x: f64) callconv(.c) f64 {
const tblsiz: u32 = @intCast(exp2dt.len / 2);
const redux: f64 = 0x1.8p52 / @as(f64, @floatFromInt(tblsiz));
const P1: f64 = 0x1.62e42fefa39efp-1;
@@ -157,17 +157,17 @@ pub fn exp2(x: f64) callconv(.C) f64 {
return math.scalbn(r, ik);
}
-pub fn __exp2x(x: f80) callconv(.C) f80 {
+pub fn __exp2x(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(exp2q(x));
}
-pub fn exp2q(x: f128) callconv(.C) f128 {
+pub fn exp2q(x: f128) callconv(.c) f128 {
// TODO: more correct implementation
return exp2(@floatCast(x));
}
-pub fn exp2l(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn exp2l(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __exp2h(x),
32 => return exp2f(x),
diff --git a/lib/compiler_rt/extenddftf2.zig b/lib/compiler_rt/extenddftf2.zig
index 8c4ed97f76..457a64522a 100644
--- a/lib/compiler_rt/extenddftf2.zig
+++ b/lib/compiler_rt/extenddftf2.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__extenddftf2, .{ .name = "__extenddftf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __extenddftf2(a: f64) callconv(.C) f128 {
+pub fn __extenddftf2(a: f64) callconv(.c) f128 {
return extendf(f128, f64, @as(u64, @bitCast(a)));
}
-fn _Qp_dtoq(c: *f128, a: f64) callconv(.C) void {
+fn _Qp_dtoq(c: *f128, a: f64) callconv(.c) void {
c.* = extendf(f128, f64, @as(u64, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extenddfxf2.zig b/lib/compiler_rt/extenddfxf2.zig
index 3e36e86b70..55a1df5662 100644
--- a/lib/compiler_rt/extenddfxf2.zig
+++ b/lib/compiler_rt/extenddfxf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__extenddfxf2, .{ .name = "__extenddfxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __extenddfxf2(a: f64) callconv(.C) f80 {
+pub fn __extenddfxf2(a: f64) callconv(.c) f80 {
return extend_f80(f64, @as(u64, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendhfdf2.zig b/lib/compiler_rt/extendhfdf2.zig
index d4e693cc72..7b1a7929f7 100644
--- a/lib/compiler_rt/extendhfdf2.zig
+++ b/lib/compiler_rt/extendhfdf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__extendhfdf2, .{ .name = "__extendhfdf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __extendhfdf2(a: common.F16T(f64)) callconv(.C) f64 {
+pub fn __extendhfdf2(a: common.F16T(f64)) callconv(.c) f64 {
return extendf(f64, f16, @as(u16, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendhfsf2.zig b/lib/compiler_rt/extendhfsf2.zig
index 0047f878ba..37f3d79c7a 100644
--- a/lib/compiler_rt/extendhfsf2.zig
+++ b/lib/compiler_rt/extendhfsf2.zig
@@ -12,14 +12,14 @@ comptime {
@export(&__extendhfsf2, .{ .name = "__extendhfsf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __extendhfsf2(a: common.F16T(f32)) callconv(.C) f32 {
+pub fn __extendhfsf2(a: common.F16T(f32)) callconv(.c) f32 {
return extendf(f32, f16, @as(u16, @bitCast(a)));
}
-fn __gnu_h2f_ieee(a: common.F16T(f32)) callconv(.C) f32 {
+fn __gnu_h2f_ieee(a: common.F16T(f32)) callconv(.c) f32 {
return extendf(f32, f16, @as(u16, @bitCast(a)));
}
-fn __aeabi_h2f(a: u16) callconv(.AAPCS) f32 {
+fn __aeabi_h2f(a: u16) callconv(.{ .arm_aapcs = .{} }) f32 {
return extendf(f32, f16, @as(u16, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendhftf2.zig b/lib/compiler_rt/extendhftf2.zig
index ea82372a04..3b1f257fa9 100644
--- a/lib/compiler_rt/extendhftf2.zig
+++ b/lib/compiler_rt/extendhftf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__extendhftf2, .{ .name = "__extendhftf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __extendhftf2(a: common.F16T(f128)) callconv(.C) f128 {
+pub fn __extendhftf2(a: common.F16T(f128)) callconv(.c) f128 {
return extendf(f128, f16, @as(u16, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendhfxf2.zig b/lib/compiler_rt/extendhfxf2.zig
index c48fa9d90a..74e4fb64cc 100644
--- a/lib/compiler_rt/extendhfxf2.zig
+++ b/lib/compiler_rt/extendhfxf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__extendhfxf2, .{ .name = "__extendhfxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __extendhfxf2(a: common.F16T(f80)) callconv(.C) f80 {
+fn __extendhfxf2(a: common.F16T(f80)) callconv(.c) f80 {
return extend_f80(f16, @as(u16, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendsfdf2.zig b/lib/compiler_rt/extendsfdf2.zig
index 4fc0b83338..c87b721cdd 100644
--- a/lib/compiler_rt/extendsfdf2.zig
+++ b/lib/compiler_rt/extendsfdf2.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-fn __extendsfdf2(a: f32) callconv(.C) f64 {
+fn __extendsfdf2(a: f32) callconv(.c) f64 {
return extendf(f64, f32, @as(u32, @bitCast(a)));
}
-fn __aeabi_f2d(a: f32) callconv(.AAPCS) f64 {
+fn __aeabi_f2d(a: f32) callconv(.{ .arm_aapcs = .{} }) f64 {
return extendf(f64, f32, @as(u32, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendsftf2.zig b/lib/compiler_rt/extendsftf2.zig
index ce1af6d91e..4e45f85029 100644
--- a/lib/compiler_rt/extendsftf2.zig
+++ b/lib/compiler_rt/extendsftf2.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__extendsftf2, .{ .name = "__extendsftf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __extendsftf2(a: f32) callconv(.C) f128 {
+pub fn __extendsftf2(a: f32) callconv(.c) f128 {
return extendf(f128, f32, @as(u32, @bitCast(a)));
}
-fn _Qp_stoq(c: *f128, a: f32) callconv(.C) void {
+fn _Qp_stoq(c: *f128, a: f32) callconv(.c) void {
c.* = extendf(f128, f32, @as(u32, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendsfxf2.zig b/lib/compiler_rt/extendsfxf2.zig
index 4263b4663e..b883034d4e 100644
--- a/lib/compiler_rt/extendsfxf2.zig
+++ b/lib/compiler_rt/extendsfxf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__extendsfxf2, .{ .name = "__extendsfxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __extendsfxf2(a: f32) callconv(.C) f80 {
+fn __extendsfxf2(a: f32) callconv(.c) f80 {
return extend_f80(f32, @as(u32, @bitCast(a)));
}
diff --git a/lib/compiler_rt/extendxftf2.zig b/lib/compiler_rt/extendxftf2.zig
index 8809dbcc80..39d66d399c 100644
--- a/lib/compiler_rt/extendxftf2.zig
+++ b/lib/compiler_rt/extendxftf2.zig
@@ -7,7 +7,7 @@ comptime {
@export(&__extendxftf2, .{ .name = "__extendxftf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __extendxftf2(a: f80) callconv(.C) f128 {
+fn __extendxftf2(a: f80) callconv(.c) f128 {
const src_int_bit: u64 = 0x8000000000000000;
const src_sig_mask = ~src_int_bit;
const src_sig_bits = std.math.floatMantissaBits(f80) - 1; // -1 for the integer bit
diff --git a/lib/compiler_rt/fabs.zig b/lib/compiler_rt/fabs.zig
index 2ab6db5913..31c6cb9793 100644
--- a/lib/compiler_rt/fabs.zig
+++ b/lib/compiler_rt/fabs.zig
@@ -17,27 +17,27 @@ comptime {
@export(&fabsl, .{ .name = "fabsl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fabsh(a: f16) callconv(.C) f16 {
+pub fn __fabsh(a: f16) callconv(.c) f16 {
return generic_fabs(a);
}
-pub fn fabsf(a: f32) callconv(.C) f32 {
+pub fn fabsf(a: f32) callconv(.c) f32 {
return generic_fabs(a);
}
-pub fn fabs(a: f64) callconv(.C) f64 {
+pub fn fabs(a: f64) callconv(.c) f64 {
return generic_fabs(a);
}
-pub fn __fabsx(a: f80) callconv(.C) f80 {
+pub fn __fabsx(a: f80) callconv(.c) f80 {
return generic_fabs(a);
}
-pub fn fabsq(a: f128) callconv(.C) f128 {
+pub fn fabsq(a: f128) callconv(.c) f128 {
return generic_fabs(a);
}
-pub fn fabsl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn fabsl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __fabsh(x),
32 => return fabsf(x),
diff --git a/lib/compiler_rt/fixdfdi.zig b/lib/compiler_rt/fixdfdi.zig
index 044bc6abe1..50eb788827 100644
--- a/lib/compiler_rt/fixdfdi.zig
+++ b/lib/compiler_rt/fixdfdi.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __fixdfdi(a: f64) callconv(.C) i64 {
+pub fn __fixdfdi(a: f64) callconv(.c) i64 {
return intFromFloat(i64, a);
}
-fn __aeabi_d2lz(a: f64) callconv(.AAPCS) i64 {
+fn __aeabi_d2lz(a: f64) callconv(.{ .arm_aapcs = .{} }) i64 {
return intFromFloat(i64, a);
}
diff --git a/lib/compiler_rt/fixdfsi.zig b/lib/compiler_rt/fixdfsi.zig
index 611810dfc7..5076cdc4c0 100644
--- a/lib/compiler_rt/fixdfsi.zig
+++ b/lib/compiler_rt/fixdfsi.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __fixdfsi(a: f64) callconv(.C) i32 {
+pub fn __fixdfsi(a: f64) callconv(.c) i32 {
return intFromFloat(i32, a);
}
-fn __aeabi_d2iz(a: f64) callconv(.AAPCS) i32 {
+fn __aeabi_d2iz(a: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return intFromFloat(i32, a);
}
diff --git a/lib/compiler_rt/fixdfti.zig b/lib/compiler_rt/fixdfti.zig
index d46245d839..889e31db53 100644
--- a/lib/compiler_rt/fixdfti.zig
+++ b/lib/compiler_rt/fixdfti.zig
@@ -15,12 +15,12 @@ comptime {
}
}
-pub fn __fixdfti(a: f64) callconv(.C) i128 {
+pub fn __fixdfti(a: f64) callconv(.c) i128 {
return intFromFloat(i128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixdfti_windows_x86_64(a: f64) callconv(.C) v2u64 {
+fn __fixdfti_windows_x86_64(a: f64) callconv(.c) v2u64 {
return @bitCast(intFromFloat(i128, a));
}
diff --git a/lib/compiler_rt/fixhfdi.zig b/lib/compiler_rt/fixhfdi.zig
index 4582b4b7e1..154bb501ac 100644
--- a/lib/compiler_rt/fixhfdi.zig
+++ b/lib/compiler_rt/fixhfdi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixhfdi, .{ .name = "__fixhfdi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixhfdi(a: f16) callconv(.C) i64 {
+fn __fixhfdi(a: f16) callconv(.c) i64 {
return intFromFloat(i64, a);
}
diff --git a/lib/compiler_rt/fixhfsi.zig b/lib/compiler_rt/fixhfsi.zig
index 4a8ac2cfea..253e7b6450 100644
--- a/lib/compiler_rt/fixhfsi.zig
+++ b/lib/compiler_rt/fixhfsi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixhfsi, .{ .name = "__fixhfsi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixhfsi(a: f16) callconv(.C) i32 {
+fn __fixhfsi(a: f16) callconv(.c) i32 {
return intFromFloat(i32, a);
}
diff --git a/lib/compiler_rt/fixhfti.zig b/lib/compiler_rt/fixhfti.zig
index 1810ec6612..01586578fe 100644
--- a/lib/compiler_rt/fixhfti.zig
+++ b/lib/compiler_rt/fixhfti.zig
@@ -12,12 +12,12 @@ comptime {
}
}
-pub fn __fixhfti(a: f16) callconv(.C) i128 {
+pub fn __fixhfti(a: f16) callconv(.c) i128 {
return intFromFloat(i128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixhfti_windows_x86_64(a: f16) callconv(.C) v2u64 {
+fn __fixhfti_windows_x86_64(a: f16) callconv(.c) v2u64 {
return @bitCast(intFromFloat(i128, a));
}
diff --git a/lib/compiler_rt/fixsfdi.zig b/lib/compiler_rt/fixsfdi.zig
index bd6ba9276d..877e7a6d4c 100644
--- a/lib/compiler_rt/fixsfdi.zig
+++ b/lib/compiler_rt/fixsfdi.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __fixsfdi(a: f32) callconv(.C) i64 {
+pub fn __fixsfdi(a: f32) callconv(.c) i64 {
return intFromFloat(i64, a);
}
-fn __aeabi_f2lz(a: f32) callconv(.AAPCS) i64 {
+fn __aeabi_f2lz(a: f32) callconv(.{ .arm_aapcs = .{} }) i64 {
return intFromFloat(i64, a);
}
diff --git a/lib/compiler_rt/fixsfsi.zig b/lib/compiler_rt/fixsfsi.zig
index 8bc18e93b4..fb83d69227 100644
--- a/lib/compiler_rt/fixsfsi.zig
+++ b/lib/compiler_rt/fixsfsi.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __fixsfsi(a: f32) callconv(.C) i32 {
+pub fn __fixsfsi(a: f32) callconv(.c) i32 {
return intFromFloat(i32, a);
}
-fn __aeabi_f2iz(a: f32) callconv(.AAPCS) i32 {
+fn __aeabi_f2iz(a: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return intFromFloat(i32, a);
}
diff --git a/lib/compiler_rt/fixsfti.zig b/lib/compiler_rt/fixsfti.zig
index a0a3e3f3e5..ae37f0f0e6 100644
--- a/lib/compiler_rt/fixsfti.zig
+++ b/lib/compiler_rt/fixsfti.zig
@@ -15,12 +15,12 @@ comptime {
}
}
-pub fn __fixsfti(a: f32) callconv(.C) i128 {
+pub fn __fixsfti(a: f32) callconv(.c) i128 {
return intFromFloat(i128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixsfti_windows_x86_64(a: f32) callconv(.C) v2u64 {
+fn __fixsfti_windows_x86_64(a: f32) callconv(.c) v2u64 {
return @bitCast(intFromFloat(i128, a));
}
diff --git a/lib/compiler_rt/fixtfdi.zig b/lib/compiler_rt/fixtfdi.zig
index acf23a78dc..ebb7484c62 100644
--- a/lib/compiler_rt/fixtfdi.zig
+++ b/lib/compiler_rt/fixtfdi.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__fixtfdi, .{ .name = "__fixtfdi", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fixtfdi(a: f128) callconv(.C) i64 {
+pub fn __fixtfdi(a: f128) callconv(.c) i64 {
return intFromFloat(i64, a);
}
-fn _Qp_qtox(a: *const f128) callconv(.C) i64 {
+fn _Qp_qtox(a: *const f128) callconv(.c) i64 {
return intFromFloat(i64, a.*);
}
diff --git a/lib/compiler_rt/fixtfsi.zig b/lib/compiler_rt/fixtfsi.zig
index 89e0308d06..a6d2daf00c 100644
--- a/lib/compiler_rt/fixtfsi.zig
+++ b/lib/compiler_rt/fixtfsi.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__fixtfsi, .{ .name = "__fixtfsi", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fixtfsi(a: f128) callconv(.C) i32 {
+pub fn __fixtfsi(a: f128) callconv(.c) i32 {
return intFromFloat(i32, a);
}
-fn _Qp_qtoi(a: *const f128) callconv(.C) i32 {
+fn _Qp_qtoi(a: *const f128) callconv(.c) i32 {
return intFromFloat(i32, a.*);
}
diff --git a/lib/compiler_rt/fixtfti.zig b/lib/compiler_rt/fixtfti.zig
index 1c826d865c..44f88e8dd0 100644
--- a/lib/compiler_rt/fixtfti.zig
+++ b/lib/compiler_rt/fixtfti.zig
@@ -14,12 +14,12 @@ comptime {
}
}
-pub fn __fixtfti(a: f128) callconv(.C) i128 {
+pub fn __fixtfti(a: f128) callconv(.c) i128 {
return intFromFloat(i128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixtfti_windows_x86_64(a: f128) callconv(.C) v2u64 {
+fn __fixtfti_windows_x86_64(a: f128) callconv(.c) v2u64 {
return @bitCast(intFromFloat(i128, a));
}
diff --git a/lib/compiler_rt/fixunsdfdi.zig b/lib/compiler_rt/fixunsdfdi.zig
index b39b0538fb..e25a194bfc 100644
--- a/lib/compiler_rt/fixunsdfdi.zig
+++ b/lib/compiler_rt/fixunsdfdi.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __fixunsdfdi(a: f64) callconv(.C) u64 {
+pub fn __fixunsdfdi(a: f64) callconv(.c) u64 {
return intFromFloat(u64, a);
}
-fn __aeabi_d2ulz(a: f64) callconv(.AAPCS) u64 {
+fn __aeabi_d2ulz(a: f64) callconv(.{ .arm_aapcs = .{} }) u64 {
return intFromFloat(u64, a);
}
diff --git a/lib/compiler_rt/fixunsdfsi.zig b/lib/compiler_rt/fixunsdfsi.zig
index 908f60d615..246cbf97eb 100644
--- a/lib/compiler_rt/fixunsdfsi.zig
+++ b/lib/compiler_rt/fixunsdfsi.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __fixunsdfsi(a: f64) callconv(.C) u32 {
+pub fn __fixunsdfsi(a: f64) callconv(.c) u32 {
return intFromFloat(u32, a);
}
-fn __aeabi_d2uiz(a: f64) callconv(.AAPCS) u32 {
+fn __aeabi_d2uiz(a: f64) callconv(.{ .arm_aapcs = .{} }) u32 {
return intFromFloat(u32, a);
}
diff --git a/lib/compiler_rt/fixunsdfti.zig b/lib/compiler_rt/fixunsdfti.zig
index 8e9b9ef151..ddaa45bc66 100644
--- a/lib/compiler_rt/fixunsdfti.zig
+++ b/lib/compiler_rt/fixunsdfti.zig
@@ -15,12 +15,12 @@ comptime {
}
}
-pub fn __fixunsdfti(a: f64) callconv(.C) u128 {
+pub fn __fixunsdfti(a: f64) callconv(.c) u128 {
return intFromFloat(u128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixunsdfti_windows_x86_64(a: f64) callconv(.C) v2u64 {
+fn __fixunsdfti_windows_x86_64(a: f64) callconv(.c) v2u64 {
return @bitCast(intFromFloat(u128, a));
}
diff --git a/lib/compiler_rt/fixunshfdi.zig b/lib/compiler_rt/fixunshfdi.zig
index 262ba24bd7..dca13e17c4 100644
--- a/lib/compiler_rt/fixunshfdi.zig
+++ b/lib/compiler_rt/fixunshfdi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixunshfdi, .{ .name = "__fixunshfdi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixunshfdi(a: f16) callconv(.C) u64 {
+fn __fixunshfdi(a: f16) callconv(.c) u64 {
return intFromFloat(u64, a);
}
diff --git a/lib/compiler_rt/fixunshfsi.zig b/lib/compiler_rt/fixunshfsi.zig
index 06fae61d8a..27db84357f 100644
--- a/lib/compiler_rt/fixunshfsi.zig
+++ b/lib/compiler_rt/fixunshfsi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixunshfsi, .{ .name = "__fixunshfsi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixunshfsi(a: f16) callconv(.C) u32 {
+fn __fixunshfsi(a: f16) callconv(.c) u32 {
return intFromFloat(u32, a);
}
diff --git a/lib/compiler_rt/fixunshfti.zig b/lib/compiler_rt/fixunshfti.zig
index cd76694e5b..c92039212d 100644
--- a/lib/compiler_rt/fixunshfti.zig
+++ b/lib/compiler_rt/fixunshfti.zig
@@ -12,12 +12,12 @@ comptime {
}
}
-pub fn __fixunshfti(a: f16) callconv(.C) u128 {
+pub fn __fixunshfti(a: f16) callconv(.c) u128 {
return intFromFloat(u128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixunshfti_windows_x86_64(a: f16) callconv(.C) v2u64 {
+fn __fixunshfti_windows_x86_64(a: f16) callconv(.c) v2u64 {
return @bitCast(intFromFloat(u128, a));
}
diff --git a/lib/compiler_rt/fixunssfdi.zig b/lib/compiler_rt/fixunssfdi.zig
index 7f65c6c64d..e8a206d7b0 100644
--- a/lib/compiler_rt/fixunssfdi.zig
+++ b/lib/compiler_rt/fixunssfdi.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __fixunssfdi(a: f32) callconv(.C) u64 {
+pub fn __fixunssfdi(a: f32) callconv(.c) u64 {
return intFromFloat(u64, a);
}
-fn __aeabi_f2ulz(a: f32) callconv(.AAPCS) u64 {
+fn __aeabi_f2ulz(a: f32) callconv(.{ .arm_aapcs = .{} }) u64 {
return intFromFloat(u64, a);
}
diff --git a/lib/compiler_rt/fixunssfsi.zig b/lib/compiler_rt/fixunssfsi.zig
index 4ee183b3f2..d0038add1b 100644
--- a/lib/compiler_rt/fixunssfsi.zig
+++ b/lib/compiler_rt/fixunssfsi.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __fixunssfsi(a: f32) callconv(.C) u32 {
+pub fn __fixunssfsi(a: f32) callconv(.c) u32 {
return intFromFloat(u32, a);
}
-fn __aeabi_f2uiz(a: f32) callconv(.AAPCS) u32 {
+fn __aeabi_f2uiz(a: f32) callconv(.{ .arm_aapcs = .{} }) u32 {
return intFromFloat(u32, a);
}
diff --git a/lib/compiler_rt/fixunssfti.zig b/lib/compiler_rt/fixunssfti.zig
index 6aa3f17edb..0b3ac65fe9 100644
--- a/lib/compiler_rt/fixunssfti.zig
+++ b/lib/compiler_rt/fixunssfti.zig
@@ -15,12 +15,12 @@ comptime {
}
}
-pub fn __fixunssfti(a: f32) callconv(.C) u128 {
+pub fn __fixunssfti(a: f32) callconv(.c) u128 {
return intFromFloat(u128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixunssfti_windows_x86_64(a: f32) callconv(.C) v2u64 {
+fn __fixunssfti_windows_x86_64(a: f32) callconv(.c) v2u64 {
return @bitCast(intFromFloat(u128, a));
}
diff --git a/lib/compiler_rt/fixunstfdi.zig b/lib/compiler_rt/fixunstfdi.zig
index 5e71a8e8cb..87a06bc37b 100644
--- a/lib/compiler_rt/fixunstfdi.zig
+++ b/lib/compiler_rt/fixunstfdi.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__fixunstfdi, .{ .name = "__fixunstfdi", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fixunstfdi(a: f128) callconv(.C) u64 {
+pub fn __fixunstfdi(a: f128) callconv(.c) u64 {
return intFromFloat(u64, a);
}
-fn _Qp_qtoux(a: *const f128) callconv(.C) u64 {
+fn _Qp_qtoux(a: *const f128) callconv(.c) u64 {
return intFromFloat(u64, a.*);
}
diff --git a/lib/compiler_rt/fixunstfsi.zig b/lib/compiler_rt/fixunstfsi.zig
index cdaeaccd5d..8a57e2dbc1 100644
--- a/lib/compiler_rt/fixunstfsi.zig
+++ b/lib/compiler_rt/fixunstfsi.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__fixunstfsi, .{ .name = "__fixunstfsi", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fixunstfsi(a: f128) callconv(.C) u32 {
+pub fn __fixunstfsi(a: f128) callconv(.c) u32 {
return intFromFloat(u32, a);
}
-fn _Qp_qtoui(a: *const f128) callconv(.C) u32 {
+fn _Qp_qtoui(a: *const f128) callconv(.c) u32 {
return intFromFloat(u32, a.*);
}
diff --git a/lib/compiler_rt/fixunstfti.zig b/lib/compiler_rt/fixunstfti.zig
index 3aa9f41c2f..2c182e3a44 100644
--- a/lib/compiler_rt/fixunstfti.zig
+++ b/lib/compiler_rt/fixunstfti.zig
@@ -14,12 +14,12 @@ comptime {
}
}
-pub fn __fixunstfti(a: f128) callconv(.C) u128 {
+pub fn __fixunstfti(a: f128) callconv(.c) u128 {
return intFromFloat(u128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixunstfti_windows_x86_64(a: f128) callconv(.C) v2u64 {
+fn __fixunstfti_windows_x86_64(a: f128) callconv(.c) v2u64 {
return @bitCast(intFromFloat(u128, a));
}
diff --git a/lib/compiler_rt/fixunsxfdi.zig b/lib/compiler_rt/fixunsxfdi.zig
index 0d346365ae..1e10786fc8 100644
--- a/lib/compiler_rt/fixunsxfdi.zig
+++ b/lib/compiler_rt/fixunsxfdi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixunsxfdi, .{ .name = "__fixunsxfdi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixunsxfdi(a: f80) callconv(.C) u64 {
+fn __fixunsxfdi(a: f80) callconv(.c) u64 {
return intFromFloat(u64, a);
}
diff --git a/lib/compiler_rt/fixunsxfsi.zig b/lib/compiler_rt/fixunsxfsi.zig
index cfc91845f3..c876ea0d3e 100644
--- a/lib/compiler_rt/fixunsxfsi.zig
+++ b/lib/compiler_rt/fixunsxfsi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixunsxfsi, .{ .name = "__fixunsxfsi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixunsxfsi(a: f80) callconv(.C) u32 {
+fn __fixunsxfsi(a: f80) callconv(.c) u32 {
return intFromFloat(u32, a);
}
diff --git a/lib/compiler_rt/fixunsxfti.zig b/lib/compiler_rt/fixunsxfti.zig
index 4d8d46f939..bfe8a2c6e2 100644
--- a/lib/compiler_rt/fixunsxfti.zig
+++ b/lib/compiler_rt/fixunsxfti.zig
@@ -12,12 +12,12 @@ comptime {
}
}
-pub fn __fixunsxfti(a: f80) callconv(.C) u128 {
+pub fn __fixunsxfti(a: f80) callconv(.c) u128 {
return intFromFloat(u128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixunsxfti_windows_x86_64(a: f80) callconv(.C) v2u64 {
+fn __fixunsxfti_windows_x86_64(a: f80) callconv(.c) v2u64 {
return @bitCast(intFromFloat(u128, a));
}
diff --git a/lib/compiler_rt/fixxfdi.zig b/lib/compiler_rt/fixxfdi.zig
index c6dd6ece68..2477a4bb83 100644
--- a/lib/compiler_rt/fixxfdi.zig
+++ b/lib/compiler_rt/fixxfdi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixxfdi, .{ .name = "__fixxfdi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixxfdi(a: f80) callconv(.C) i64 {
+fn __fixxfdi(a: f80) callconv(.c) i64 {
return intFromFloat(i64, a);
}
diff --git a/lib/compiler_rt/fixxfsi.zig b/lib/compiler_rt/fixxfsi.zig
index 9d348d6f00..846df4d006 100644
--- a/lib/compiler_rt/fixxfsi.zig
+++ b/lib/compiler_rt/fixxfsi.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__fixxfsi, .{ .name = "__fixxfsi", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __fixxfsi(a: f80) callconv(.C) i32 {
+fn __fixxfsi(a: f80) callconv(.c) i32 {
return intFromFloat(i32, a);
}
diff --git a/lib/compiler_rt/fixxfti.zig b/lib/compiler_rt/fixxfti.zig
index 2f57f5c222..e1db47dc33 100644
--- a/lib/compiler_rt/fixxfti.zig
+++ b/lib/compiler_rt/fixxfti.zig
@@ -12,12 +12,12 @@ comptime {
}
}
-pub fn __fixxfti(a: f80) callconv(.C) i128 {
+pub fn __fixxfti(a: f80) callconv(.c) i128 {
return intFromFloat(i128, a);
}
const v2u64 = @Vector(2, u64);
-fn __fixxfti_windows_x86_64(a: f80) callconv(.C) v2u64 {
+fn __fixxfti_windows_x86_64(a: f80) callconv(.c) v2u64 {
return @bitCast(intFromFloat(i128, a));
}
diff --git a/lib/compiler_rt/floatdidf.zig b/lib/compiler_rt/floatdidf.zig
index 2396624dc3..50b6fc1aab 100644
--- a/lib/compiler_rt/floatdidf.zig
+++ b/lib/compiler_rt/floatdidf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatdidf(a: i64) callconv(.C) f64 {
+pub fn __floatdidf(a: i64) callconv(.c) f64 {
return floatFromInt(f64, a);
}
-fn __aeabi_l2d(a: i64) callconv(.AAPCS) f64 {
+fn __aeabi_l2d(a: i64) callconv(.{ .arm_aapcs = .{} }) f64 {
return floatFromInt(f64, a);
}
diff --git a/lib/compiler_rt/floatdihf.zig b/lib/compiler_rt/floatdihf.zig
index eb568c9ff2..37a03e4ed3 100644
--- a/lib/compiler_rt/floatdihf.zig
+++ b/lib/compiler_rt/floatdihf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatdihf, .{ .name = "__floatdihf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatdihf(a: i64) callconv(.C) f16 {
+fn __floatdihf(a: i64) callconv(.c) f16 {
return floatFromInt(f16, a);
}
diff --git a/lib/compiler_rt/floatdisf.zig b/lib/compiler_rt/floatdisf.zig
index 1cc159ccaa..54f57cc3ca 100644
--- a/lib/compiler_rt/floatdisf.zig
+++ b/lib/compiler_rt/floatdisf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatdisf(a: i64) callconv(.C) f32 {
+pub fn __floatdisf(a: i64) callconv(.c) f32 {
return floatFromInt(f32, a);
}
-fn __aeabi_l2f(a: i64) callconv(.AAPCS) f32 {
+fn __aeabi_l2f(a: i64) callconv(.{ .arm_aapcs = .{} }) f32 {
return floatFromInt(f32, a);
}
diff --git a/lib/compiler_rt/floatditf.zig b/lib/compiler_rt/floatditf.zig
index 7294c16802..b5cc262fa3 100644
--- a/lib/compiler_rt/floatditf.zig
+++ b/lib/compiler_rt/floatditf.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__floatditf, .{ .name = "__floatditf", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __floatditf(a: i64) callconv(.C) f128 {
+pub fn __floatditf(a: i64) callconv(.c) f128 {
return floatFromInt(f128, a);
}
-fn _Qp_xtoq(c: *f128, a: i64) callconv(.C) void {
+fn _Qp_xtoq(c: *f128, a: i64) callconv(.c) void {
c.* = floatFromInt(f128, a);
}
diff --git a/lib/compiler_rt/floatdixf.zig b/lib/compiler_rt/floatdixf.zig
index ccd5e5bd4d..eba666c053 100644
--- a/lib/compiler_rt/floatdixf.zig
+++ b/lib/compiler_rt/floatdixf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatdixf, .{ .name = "__floatdixf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatdixf(a: i64) callconv(.C) f80 {
+fn __floatdixf(a: i64) callconv(.c) f80 {
return floatFromInt(f80, a);
}
diff --git a/lib/compiler_rt/floatsidf.zig b/lib/compiler_rt/floatsidf.zig
index 183e886721..c3f249a455 100644
--- a/lib/compiler_rt/floatsidf.zig
+++ b/lib/compiler_rt/floatsidf.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __floatsidf(a: i32) callconv(.C) f64 {
+pub fn __floatsidf(a: i32) callconv(.c) f64 {
return floatFromInt(f64, a);
}
-fn __aeabi_i2d(a: i32) callconv(.AAPCS) f64 {
+fn __aeabi_i2d(a: i32) callconv(.{ .arm_aapcs = .{} }) f64 {
return floatFromInt(f64, a);
}
diff --git a/lib/compiler_rt/floatsihf.zig b/lib/compiler_rt/floatsihf.zig
index 4b27e9d00e..a5d5d2ba92 100644
--- a/lib/compiler_rt/floatsihf.zig
+++ b/lib/compiler_rt/floatsihf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatsihf, .{ .name = "__floatsihf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatsihf(a: i32) callconv(.C) f16 {
+fn __floatsihf(a: i32) callconv(.c) f16 {
return floatFromInt(f16, a);
}
diff --git a/lib/compiler_rt/floatsisf.zig b/lib/compiler_rt/floatsisf.zig
index 5839d20a0a..133bfaabbd 100644
--- a/lib/compiler_rt/floatsisf.zig
+++ b/lib/compiler_rt/floatsisf.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __floatsisf(a: i32) callconv(.C) f32 {
+pub fn __floatsisf(a: i32) callconv(.c) f32 {
return floatFromInt(f32, a);
}
-fn __aeabi_i2f(a: i32) callconv(.AAPCS) f32 {
+fn __aeabi_i2f(a: i32) callconv(.{ .arm_aapcs = .{} }) f32 {
return floatFromInt(f32, a);
}
diff --git a/lib/compiler_rt/floatsitf.zig b/lib/compiler_rt/floatsitf.zig
index 4619019ea0..6a5f81ae21 100644
--- a/lib/compiler_rt/floatsitf.zig
+++ b/lib/compiler_rt/floatsitf.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__floatsitf, .{ .name = "__floatsitf", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __floatsitf(a: i32) callconv(.C) f128 {
+pub fn __floatsitf(a: i32) callconv(.c) f128 {
return floatFromInt(f128, a);
}
-fn _Qp_itoq(c: *f128, a: i32) callconv(.C) void {
+fn _Qp_itoq(c: *f128, a: i32) callconv(.c) void {
c.* = floatFromInt(f128, a);
}
diff --git a/lib/compiler_rt/floatsixf.zig b/lib/compiler_rt/floatsixf.zig
index 0a22f5cd0c..31791eb9cc 100644
--- a/lib/compiler_rt/floatsixf.zig
+++ b/lib/compiler_rt/floatsixf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatsixf, .{ .name = "__floatsixf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatsixf(a: i32) callconv(.C) f80 {
+fn __floatsixf(a: i32) callconv(.c) f80 {
return floatFromInt(f80, a);
}
diff --git a/lib/compiler_rt/floattidf.zig b/lib/compiler_rt/floattidf.zig
index 9228f0adcd..78cef508a0 100644
--- a/lib/compiler_rt/floattidf.zig
+++ b/lib/compiler_rt/floattidf.zig
@@ -15,10 +15,10 @@ comptime {
}
}
-pub fn __floattidf(a: i128) callconv(.C) f64 {
+pub fn __floattidf(a: i128) callconv(.c) f64 {
return floatFromInt(f64, a);
}
-fn __floattidf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f64 {
+fn __floattidf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f64 {
return floatFromInt(f64, @as(i128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floattihf.zig b/lib/compiler_rt/floattihf.zig
index 4d0dfa1802..63f079b3a8 100644
--- a/lib/compiler_rt/floattihf.zig
+++ b/lib/compiler_rt/floattihf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floattihf(a: i128) callconv(.C) f16 {
+pub fn __floattihf(a: i128) callconv(.c) f16 {
return floatFromInt(f16, a);
}
-fn __floattihf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f16 {
+fn __floattihf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f16 {
return floatFromInt(f16, @as(i128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floattisf.zig b/lib/compiler_rt/floattisf.zig
index b2e70788da..4483643d45 100644
--- a/lib/compiler_rt/floattisf.zig
+++ b/lib/compiler_rt/floattisf.zig
@@ -15,10 +15,10 @@ comptime {
}
}
-pub fn __floattisf(a: i128) callconv(.C) f32 {
+pub fn __floattisf(a: i128) callconv(.c) f32 {
return floatFromInt(f32, a);
}
-fn __floattisf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f32 {
+fn __floattisf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f32 {
return floatFromInt(f32, @as(i128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floattitf.zig b/lib/compiler_rt/floattitf.zig
index d7aa35f0c3..c8d2090457 100644
--- a/lib/compiler_rt/floattitf.zig
+++ b/lib/compiler_rt/floattitf.zig
@@ -14,10 +14,10 @@ comptime {
}
}
-pub fn __floattitf(a: i128) callconv(.C) f128 {
+pub fn __floattitf(a: i128) callconv(.c) f128 {
return floatFromInt(f128, a);
}
-fn __floattitf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f128 {
+fn __floattitf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f128 {
return floatFromInt(f128, @as(i128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floattixf.zig b/lib/compiler_rt/floattixf.zig
index 3dcdac3a21..ebf83446d3 100644
--- a/lib/compiler_rt/floattixf.zig
+++ b/lib/compiler_rt/floattixf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floattixf(a: i128) callconv(.C) f80 {
+pub fn __floattixf(a: i128) callconv(.c) f80 {
return floatFromInt(f80, a);
}
-fn __floattixf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f80 {
+fn __floattixf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f80 {
return floatFromInt(f80, @as(i128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floatundidf.zig b/lib/compiler_rt/floatundidf.zig
index 159b6abd1a..2aa1c959d8 100644
--- a/lib/compiler_rt/floatundidf.zig
+++ b/lib/compiler_rt/floatundidf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatundidf(a: u64) callconv(.C) f64 {
+pub fn __floatundidf(a: u64) callconv(.c) f64 {
return floatFromInt(f64, a);
}
-fn __aeabi_ul2d(a: u64) callconv(.AAPCS) f64 {
+fn __aeabi_ul2d(a: u64) callconv(.{ .arm_aapcs = .{} }) f64 {
return floatFromInt(f64, a);
}
diff --git a/lib/compiler_rt/floatundihf.zig b/lib/compiler_rt/floatundihf.zig
index d05eb0a25e..5fefedec1d 100644
--- a/lib/compiler_rt/floatundihf.zig
+++ b/lib/compiler_rt/floatundihf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatundihf, .{ .name = "__floatundihf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatundihf(a: u64) callconv(.C) f16 {
+fn __floatundihf(a: u64) callconv(.c) f16 {
return floatFromInt(f16, a);
}
diff --git a/lib/compiler_rt/floatundisf.zig b/lib/compiler_rt/floatundisf.zig
index caecbd0cd8..3f5b01a4e0 100644
--- a/lib/compiler_rt/floatundisf.zig
+++ b/lib/compiler_rt/floatundisf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatundisf(a: u64) callconv(.C) f32 {
+pub fn __floatundisf(a: u64) callconv(.c) f32 {
return floatFromInt(f32, a);
}
-fn __aeabi_ul2f(a: u64) callconv(.AAPCS) f32 {
+fn __aeabi_ul2f(a: u64) callconv(.{ .arm_aapcs = .{} }) f32 {
return floatFromInt(f32, a);
}
diff --git a/lib/compiler_rt/floatunditf.zig b/lib/compiler_rt/floatunditf.zig
index 8f42107ba7..aadb251328 100644
--- a/lib/compiler_rt/floatunditf.zig
+++ b/lib/compiler_rt/floatunditf.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__floatunditf, .{ .name = "__floatunditf", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __floatunditf(a: u64) callconv(.C) f128 {
+pub fn __floatunditf(a: u64) callconv(.c) f128 {
return floatFromInt(f128, a);
}
-fn _Qp_uxtoq(c: *f128, a: u64) callconv(.C) void {
+fn _Qp_uxtoq(c: *f128, a: u64) callconv(.c) void {
c.* = floatFromInt(f128, a);
}
diff --git a/lib/compiler_rt/floatundixf.zig b/lib/compiler_rt/floatundixf.zig
index ed58f943d4..7f801fd6ba 100644
--- a/lib/compiler_rt/floatundixf.zig
+++ b/lib/compiler_rt/floatundixf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatundixf, .{ .name = "__floatundixf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatundixf(a: u64) callconv(.C) f80 {
+fn __floatundixf(a: u64) callconv(.c) f80 {
return floatFromInt(f80, a);
}
diff --git a/lib/compiler_rt/floatunsidf.zig b/lib/compiler_rt/floatunsidf.zig
index a988d80c07..c712428ce2 100644
--- a/lib/compiler_rt/floatunsidf.zig
+++ b/lib/compiler_rt/floatunsidf.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __floatunsidf(a: u32) callconv(.C) f64 {
+pub fn __floatunsidf(a: u32) callconv(.c) f64 {
return floatFromInt(f64, a);
}
-fn __aeabi_ui2d(a: u32) callconv(.AAPCS) f64 {
+fn __aeabi_ui2d(a: u32) callconv(.{ .arm_aapcs = .{} }) f64 {
return floatFromInt(f64, a);
}
diff --git a/lib/compiler_rt/floatunsihf.zig b/lib/compiler_rt/floatunsihf.zig
index 9f64e0e9a8..5c5778a460 100644
--- a/lib/compiler_rt/floatunsihf.zig
+++ b/lib/compiler_rt/floatunsihf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatunsihf, .{ .name = "__floatunsihf", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __floatunsihf(a: u32) callconv(.C) f16 {
+pub fn __floatunsihf(a: u32) callconv(.c) f16 {
return floatFromInt(f16, a);
}
diff --git a/lib/compiler_rt/floatunsisf.zig b/lib/compiler_rt/floatunsisf.zig
index 7519ab77c7..9fa213ed22 100644
--- a/lib/compiler_rt/floatunsisf.zig
+++ b/lib/compiler_rt/floatunsisf.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __floatunsisf(a: u32) callconv(.C) f32 {
+pub fn __floatunsisf(a: u32) callconv(.c) f32 {
return floatFromInt(f32, a);
}
-fn __aeabi_ui2f(a: u32) callconv(.AAPCS) f32 {
+fn __aeabi_ui2f(a: u32) callconv(.{ .arm_aapcs = .{} }) f32 {
return floatFromInt(f32, a);
}
diff --git a/lib/compiler_rt/floatunsitf.zig b/lib/compiler_rt/floatunsitf.zig
index c6f0bea610..d09d31d05d 100644
--- a/lib/compiler_rt/floatunsitf.zig
+++ b/lib/compiler_rt/floatunsitf.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__floatunsitf, .{ .name = "__floatunsitf", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __floatunsitf(a: u32) callconv(.C) f128 {
+pub fn __floatunsitf(a: u32) callconv(.c) f128 {
return floatFromInt(f128, a);
}
-fn _Qp_uitoq(c: *f128, a: u32) callconv(.C) void {
+fn _Qp_uitoq(c: *f128, a: u32) callconv(.c) void {
c.* = floatFromInt(f128, a);
}
diff --git a/lib/compiler_rt/floatunsixf.zig b/lib/compiler_rt/floatunsixf.zig
index cb4f77d73c..678085ad6f 100644
--- a/lib/compiler_rt/floatunsixf.zig
+++ b/lib/compiler_rt/floatunsixf.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__floatunsixf, .{ .name = "__floatunsixf", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __floatunsixf(a: u32) callconv(.C) f80 {
+fn __floatunsixf(a: u32) callconv(.c) f80 {
return floatFromInt(f80, a);
}
diff --git a/lib/compiler_rt/floatuntidf.zig b/lib/compiler_rt/floatuntidf.zig
index 52fdbfd930..cee154ed0b 100644
--- a/lib/compiler_rt/floatuntidf.zig
+++ b/lib/compiler_rt/floatuntidf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatuntidf(a: u128) callconv(.C) f64 {
+pub fn __floatuntidf(a: u128) callconv(.c) f64 {
return floatFromInt(f64, a);
}
-fn __floatuntidf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f64 {
+fn __floatuntidf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f64 {
return floatFromInt(f64, @as(u128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floatuntihf.zig b/lib/compiler_rt/floatuntihf.zig
index caa64d3a40..63cfdb604f 100644
--- a/lib/compiler_rt/floatuntihf.zig
+++ b/lib/compiler_rt/floatuntihf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatuntihf(a: u128) callconv(.C) f16 {
+pub fn __floatuntihf(a: u128) callconv(.c) f16 {
return floatFromInt(f16, a);
}
-fn __floatuntihf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f16 {
+fn __floatuntihf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f16 {
return floatFromInt(f16, @as(u128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floatuntisf.zig b/lib/compiler_rt/floatuntisf.zig
index 6db67b737d..1e65576f06 100644
--- a/lib/compiler_rt/floatuntisf.zig
+++ b/lib/compiler_rt/floatuntisf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatuntisf(a: u128) callconv(.C) f32 {
+pub fn __floatuntisf(a: u128) callconv(.c) f32 {
return floatFromInt(f32, a);
}
-fn __floatuntisf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f32 {
+fn __floatuntisf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f32 {
return floatFromInt(f32, @as(u128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floatuntitf.zig b/lib/compiler_rt/floatuntitf.zig
index c5152cee02..41c48bee99 100644
--- a/lib/compiler_rt/floatuntitf.zig
+++ b/lib/compiler_rt/floatuntitf.zig
@@ -14,10 +14,10 @@ comptime {
}
}
-pub fn __floatuntitf(a: u128) callconv(.C) f128 {
+pub fn __floatuntitf(a: u128) callconv(.c) f128 {
return floatFromInt(f128, a);
}
-fn __floatuntitf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f128 {
+fn __floatuntitf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f128 {
return floatFromInt(f128, @as(u128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floatuntixf.zig b/lib/compiler_rt/floatuntixf.zig
index d7592c1c75..353a3c3c0d 100644
--- a/lib/compiler_rt/floatuntixf.zig
+++ b/lib/compiler_rt/floatuntixf.zig
@@ -12,10 +12,10 @@ comptime {
}
}
-pub fn __floatuntixf(a: u128) callconv(.C) f80 {
+pub fn __floatuntixf(a: u128) callconv(.c) f80 {
return floatFromInt(f80, a);
}
-fn __floatuntixf_windows_x86_64(a: @Vector(2, u64)) callconv(.C) f80 {
+fn __floatuntixf_windows_x86_64(a: @Vector(2, u64)) callconv(.c) f80 {
return floatFromInt(f80, @as(u128, @bitCast(a)));
}
diff --git a/lib/compiler_rt/floor.zig b/lib/compiler_rt/floor.zig
index b9eb4e3251..fc8755e4ad 100644
--- a/lib/compiler_rt/floor.zig
+++ b/lib/compiler_rt/floor.zig
@@ -26,7 +26,7 @@ comptime {
@export(&floorl, .{ .name = "floorl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __floorh(x: f16) callconv(.C) f16 {
+pub fn __floorh(x: f16) callconv(.c) f16 {
var u: u16 = @bitCast(x);
const e = @as(i16, @intCast((u >> 10) & 31)) - 15;
var m: u16 = undefined;
@@ -60,7 +60,7 @@ pub fn __floorh(x: f16) callconv(.C) f16 {
}
}
-pub fn floorf(x: f32) callconv(.C) f32 {
+pub fn floorf(x: f32) callconv(.c) f32 {
var u: u32 = @bitCast(x);
const e = @as(i32, @intCast((u >> 23) & 0xFF)) - 0x7F;
var m: u32 = undefined;
@@ -94,7 +94,7 @@ pub fn floorf(x: f32) callconv(.C) f32 {
}
}
-pub fn floor(x: f64) callconv(.C) f64 {
+pub fn floor(x: f64) callconv(.c) f64 {
const f64_toint = 1.0 / math.floatEps(f64);
const u: u64 = @bitCast(x);
@@ -125,12 +125,12 @@ pub fn floor(x: f64) callconv(.C) f64 {
}
}
-pub fn __floorx(x: f80) callconv(.C) f80 {
+pub fn __floorx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(floorq(x));
}
-pub fn floorq(x: f128) callconv(.C) f128 {
+pub fn floorq(x: f128) callconv(.c) f128 {
const f128_toint = 1.0 / math.floatEps(f128);
const u: u128 = @bitCast(x);
@@ -159,7 +159,7 @@ pub fn floorq(x: f128) callconv(.C) f128 {
}
}
-pub fn floorl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn floorl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __floorh(x),
32 => return floorf(x),
diff --git a/lib/compiler_rt/fma.zig b/lib/compiler_rt/fma.zig
index 46377bdc2e..578139c5ac 100644
--- a/lib/compiler_rt/fma.zig
+++ b/lib/compiler_rt/fma.zig
@@ -24,12 +24,12 @@ comptime {
@export(&fmal, .{ .name = "fmal", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fmah(x: f16, y: f16, z: f16) callconv(.C) f16 {
+pub fn __fmah(x: f16, y: f16, z: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(fmaf(x, y, z));
}
-pub fn fmaf(x: f32, y: f32, z: f32) callconv(.C) f32 {
+pub fn fmaf(x: f32, y: f32, z: f32) callconv(.c) f32 {
const xy = @as(f64, x) * y;
const xy_z = xy + z;
const u = @as(u64, @bitCast(xy_z));
@@ -44,7 +44,7 @@ pub fn fmaf(x: f32, y: f32, z: f32) callconv(.C) f32 {
}
/// NOTE: Upstream fma.c has been rewritten completely to raise fp exceptions more accurately.
-pub fn fma(x: f64, y: f64, z: f64) callconv(.C) f64 {
+pub fn fma(x: f64, y: f64, z: f64) callconv(.c) f64 {
if (!math.isFinite(x) or !math.isFinite(y)) {
return x * y + z;
}
@@ -91,7 +91,7 @@ pub fn fma(x: f64, y: f64, z: f64) callconv(.C) f64 {
}
}
-pub fn __fmax(a: f80, b: f80, c: f80) callconv(.C) f80 {
+pub fn __fmax(a: f80, b: f80, c: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(fmaq(a, b, c));
}
@@ -103,7 +103,7 @@ pub fn __fmax(a: f80, b: f80, c: f80) callconv(.C) f80 {
///
/// Dekker, T. A Floating-Point Technique for Extending the
/// Available Precision. Numer. Math. 18, 224-242 (1971).
-pub fn fmaq(x: f128, y: f128, z: f128) callconv(.C) f128 {
+pub fn fmaq(x: f128, y: f128, z: f128) callconv(.c) f128 {
if (!math.isFinite(x) or !math.isFinite(y)) {
return x * y + z;
}
@@ -150,7 +150,7 @@ pub fn fmaq(x: f128, y: f128, z: f128) callconv(.C) f128 {
}
}
-pub fn fmal(x: c_longdouble, y: c_longdouble, z: c_longdouble) callconv(.C) c_longdouble {
+pub fn fmal(x: c_longdouble, y: c_longdouble, z: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __fmah(x, y, z),
32 => return fmaf(x, y, z),
diff --git a/lib/compiler_rt/fmax.zig b/lib/compiler_rt/fmax.zig
index a26d41a2b3..2384d37d87 100644
--- a/lib/compiler_rt/fmax.zig
+++ b/lib/compiler_rt/fmax.zig
@@ -18,27 +18,27 @@ comptime {
@export(&fmaxl, .{ .name = "fmaxl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fmaxh(x: f16, y: f16) callconv(.C) f16 {
+pub fn __fmaxh(x: f16, y: f16) callconv(.c) f16 {
return generic_fmax(f16, x, y);
}
-pub fn fmaxf(x: f32, y: f32) callconv(.C) f32 {
+pub fn fmaxf(x: f32, y: f32) callconv(.c) f32 {
return generic_fmax(f32, x, y);
}
-pub fn fmax(x: f64, y: f64) callconv(.C) f64 {
+pub fn fmax(x: f64, y: f64) callconv(.c) f64 {
return generic_fmax(f64, x, y);
}
-pub fn __fmaxx(x: f80, y: f80) callconv(.C) f80 {
+pub fn __fmaxx(x: f80, y: f80) callconv(.c) f80 {
return generic_fmax(f80, x, y);
}
-pub fn fmaxq(x: f128, y: f128) callconv(.C) f128 {
+pub fn fmaxq(x: f128, y: f128) callconv(.c) f128 {
return generic_fmax(f128, x, y);
}
-pub fn fmaxl(x: c_longdouble, y: c_longdouble) callconv(.C) c_longdouble {
+pub fn fmaxl(x: c_longdouble, y: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __fmaxh(x, y),
32 => return fmaxf(x, y),
diff --git a/lib/compiler_rt/fmin.zig b/lib/compiler_rt/fmin.zig
index 079e503218..0b47262ca2 100644
--- a/lib/compiler_rt/fmin.zig
+++ b/lib/compiler_rt/fmin.zig
@@ -18,27 +18,27 @@ comptime {
@export(&fminl, .{ .name = "fminl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fminh(x: f16, y: f16) callconv(.C) f16 {
+pub fn __fminh(x: f16, y: f16) callconv(.c) f16 {
return generic_fmin(f16, x, y);
}
-pub fn fminf(x: f32, y: f32) callconv(.C) f32 {
+pub fn fminf(x: f32, y: f32) callconv(.c) f32 {
return generic_fmin(f32, x, y);
}
-pub fn fmin(x: f64, y: f64) callconv(.C) f64 {
+pub fn fmin(x: f64, y: f64) callconv(.c) f64 {
return generic_fmin(f64, x, y);
}
-pub fn __fminx(x: f80, y: f80) callconv(.C) f80 {
+pub fn __fminx(x: f80, y: f80) callconv(.c) f80 {
return generic_fmin(f80, x, y);
}
-pub fn fminq(x: f128, y: f128) callconv(.C) f128 {
+pub fn fminq(x: f128, y: f128) callconv(.c) f128 {
return generic_fmin(f128, x, y);
}
-pub fn fminl(x: c_longdouble, y: c_longdouble) callconv(.C) c_longdouble {
+pub fn fminl(x: c_longdouble, y: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __fminh(x, y),
32 => return fminf(x, y),
diff --git a/lib/compiler_rt/fmod.zig b/lib/compiler_rt/fmod.zig
index 592a563e7e..5c055049e3 100644
--- a/lib/compiler_rt/fmod.zig
+++ b/lib/compiler_rt/fmod.zig
@@ -20,22 +20,22 @@ comptime {
@export(&fmodl, .{ .name = "fmodl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __fmodh(x: f16, y: f16) callconv(.C) f16 {
+pub fn __fmodh(x: f16, y: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(fmodf(x, y));
}
-pub fn fmodf(x: f32, y: f32) callconv(.C) f32 {
+pub fn fmodf(x: f32, y: f32) callconv(.c) f32 {
return generic_fmod(f32, x, y);
}
-pub fn fmod(x: f64, y: f64) callconv(.C) f64 {
+pub fn fmod(x: f64, y: f64) callconv(.c) f64 {
return generic_fmod(f64, x, y);
}
/// fmodx - floating modulo large, returns the remainder of division for f80 types
/// Logic and flow heavily inspired by MUSL fmodl for 113 mantissa digits
-pub fn __fmodx(a: f80, b: f80) callconv(.C) f80 {
+pub fn __fmodx(a: f80, b: f80) callconv(.c) f80 {
const T = f80;
const Z = std.meta.Int(.unsigned, @bitSizeOf(T));
@@ -133,7 +133,7 @@ pub fn __fmodx(a: f80, b: f80) callconv(.C) f80 {
/// fmodq - floating modulo large, returns the remainder of division for f128 types
/// Logic and flow heavily inspired by MUSL fmodl for 113 mantissa digits
-pub fn fmodq(a: f128, b: f128) callconv(.C) f128 {
+pub fn fmodq(a: f128, b: f128) callconv(.c) f128 {
var amod = a;
var bmod = b;
const aPtr_u64: [*]u64 = @ptrCast(&amod);
@@ -250,7 +250,7 @@ pub fn fmodq(a: f128, b: f128) callconv(.C) f128 {
return amod;
}
-pub fn fmodl(a: c_longdouble, b: c_longdouble) callconv(.C) c_longdouble {
+pub fn fmodl(a: c_longdouble, b: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __fmodh(a, b),
32 => return fmodf(a, b),
diff --git a/lib/compiler_rt/gedf2.zig b/lib/compiler_rt/gedf2.zig
index 967278f324..33b4fa9926 100644
--- a/lib/compiler_rt/gedf2.zig
+++ b/lib/compiler_rt/gedf2.zig
@@ -17,20 +17,20 @@ comptime {
/// "These functions return a value greater than or equal to zero if neither
/// argument is NaN, and a is greater than or equal to b."
-pub fn __gedf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __gedf2(a: f64, b: f64) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f64, comparef.GE, a, b));
}
/// "These functions return a value greater than zero if neither argument is NaN,
/// and a is strictly greater than b."
-pub fn __gtdf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __gtdf2(a: f64, b: f64) callconv(.c) i32 {
return __gedf2(a, b);
}
-fn __aeabi_dcmpge(a: f64, b: f64) callconv(.AAPCS) i32 {
+fn __aeabi_dcmpge(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f64, comparef.GE, a, b) != .Less);
}
-fn __aeabi_dcmpgt(a: f64, b: f64) callconv(.AAPCS) i32 {
+fn __aeabi_dcmpgt(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f64, comparef.GE, a, b) == .Greater);
}
diff --git a/lib/compiler_rt/gehf2.zig b/lib/compiler_rt/gehf2.zig
index edf45d9553..12eccf98d8 100644
--- a/lib/compiler_rt/gehf2.zig
+++ b/lib/compiler_rt/gehf2.zig
@@ -12,12 +12,12 @@ comptime {
/// "These functions return a value greater than or equal to zero if neither
/// argument is NaN, and a is greater than or equal to b."
-pub fn __gehf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __gehf2(a: f16, b: f16) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f16, comparef.GE, a, b));
}
/// "These functions return a value greater than zero if neither argument is NaN,
/// and a is strictly greater than b."
-pub fn __gthf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __gthf2(a: f16, b: f16) callconv(.c) i32 {
return __gehf2(a, b);
}
diff --git a/lib/compiler_rt/gesf2.zig b/lib/compiler_rt/gesf2.zig
index 569f66425d..5f9aefef47 100644
--- a/lib/compiler_rt/gesf2.zig
+++ b/lib/compiler_rt/gesf2.zig
@@ -17,20 +17,20 @@ comptime {
/// "These functions return a value greater than or equal to zero if neither
/// argument is NaN, and a is greater than or equal to b."
-pub fn __gesf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __gesf2(a: f32, b: f32) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f32, comparef.GE, a, b));
}
/// "These functions return a value greater than zero if neither argument is NaN,
/// and a is strictly greater than b."
-pub fn __gtsf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __gtsf2(a: f32, b: f32) callconv(.c) i32 {
return __gesf2(a, b);
}
-fn __aeabi_fcmpge(a: f32, b: f32) callconv(.AAPCS) i32 {
+fn __aeabi_fcmpge(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f32, comparef.GE, a, b) != .Less);
}
-fn __aeabi_fcmpgt(a: f32, b: f32) callconv(.AAPCS) i32 {
+fn __aeabi_fcmpgt(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return @intFromBool(comparef.cmpf2(f32, comparef.LE, a, b) == .Greater);
}
diff --git a/lib/compiler_rt/getf2.zig b/lib/compiler_rt/getf2.zig
index c5aebdd2b9..3a17843718 100644
--- a/lib/compiler_rt/getf2.zig
+++ b/lib/compiler_rt/getf2.zig
@@ -19,12 +19,12 @@ comptime {
/// "These functions return a value greater than or equal to zero if neither
/// argument is NaN, and a is greater than or equal to b."
-fn __getf2(a: f128, b: f128) callconv(.C) i32 {
+fn __getf2(a: f128, b: f128) callconv(.c) i32 {
return @intFromEnum(comparef.cmpf2(f128, comparef.GE, a, b));
}
/// "These functions return a value greater than zero if neither argument is NaN,
/// and a is strictly greater than b."
-fn __gttf2(a: f128, b: f128) callconv(.C) i32 {
+fn __gttf2(a: f128, b: f128) callconv(.c) i32 {
return __getf2(a, b);
}
diff --git a/lib/compiler_rt/gexf2.zig b/lib/compiler_rt/gexf2.zig
index 8aee88cbc2..7ea0d6eebe 100644
--- a/lib/compiler_rt/gexf2.zig
+++ b/lib/compiler_rt/gexf2.zig
@@ -8,10 +8,10 @@ comptime {
@export(&__gtxf2, .{ .name = "__gtxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __gexf2(a: f80, b: f80) callconv(.C) i32 {
+fn __gexf2(a: f80, b: f80) callconv(.c) i32 {
return @intFromEnum(comparef.cmp_f80(comparef.GE, a, b));
}
-fn __gtxf2(a: f80, b: f80) callconv(.C) i32 {
+fn __gtxf2(a: f80, b: f80) callconv(.c) i32 {
return __gexf2(a, b);
}
diff --git a/lib/compiler_rt/int.zig b/lib/compiler_rt/int.zig
index 758ca74781..4a89d0799d 100644
--- a/lib/compiler_rt/int.zig
+++ b/lib/compiler_rt/int.zig
@@ -34,7 +34,7 @@ comptime {
@export(&__udivmodsi4, .{ .name = "__udivmodsi4", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __divmodti4(a: i128, b: i128, rem: *i128) callconv(.C) i128 {
+pub fn __divmodti4(a: i128, b: i128, rem: *i128) callconv(.c) i128 {
const d = __divti3(a, b);
rem.* = a -% (d * b);
return d;
@@ -67,7 +67,7 @@ fn test_one_divmodti4(a: i128, b: i128, expected_q: i128, expected_r: i128) !voi
try testing.expect(q == expected_q and r == expected_r);
}
-pub fn __divmoddi4(a: i64, b: i64, rem: *i64) callconv(.C) i64 {
+pub fn __divmoddi4(a: i64, b: i64, rem: *i64) callconv(.c) i64 {
const d = __divdi3(a, b);
rem.* = a -% (d * b);
return d;
@@ -101,7 +101,7 @@ test "test_divmoddi4" {
}
}
-pub fn __udivmoddi4(a: u64, b: u64, maybe_rem: ?*u64) callconv(.C) u64 {
+pub fn __udivmoddi4(a: u64, b: u64, maybe_rem: ?*u64) callconv(.c) u64 {
return udivmod(u64, a, b, maybe_rem);
}
@@ -109,7 +109,7 @@ test "test_udivmoddi4" {
_ = @import("udivmoddi4_test.zig");
}
-pub fn __divdi3(a: i64, b: i64) callconv(.C) i64 {
+pub fn __divdi3(a: i64, b: i64) callconv(.c) i64 {
// Set aside the sign of the quotient.
const sign: u64 = @bitCast((a ^ b) >> 63);
// Take absolute value of a and b via abs(x) = (x^(x >> 63)) - (x >> 63).
@@ -146,7 +146,7 @@ fn test_one_divdi3(a: i64, b: i64, expected_q: i64) !void {
try testing.expect(q == expected_q);
}
-pub fn __moddi3(a: i64, b: i64) callconv(.C) i64 {
+pub fn __moddi3(a: i64, b: i64) callconv(.c) i64 {
// Take absolute value of a and b via abs(x) = (x^(x >> 63)) - (x >> 63).
const abs_a = (a ^ (a >> 63)) -% (a >> 63);
const abs_b = (b ^ (b >> 63)) -% (b >> 63);
@@ -184,11 +184,11 @@ fn test_one_moddi3(a: i64, b: i64, expected_r: i64) !void {
try testing.expect(r == expected_r);
}
-pub fn __udivdi3(a: u64, b: u64) callconv(.C) u64 {
+pub fn __udivdi3(a: u64, b: u64) callconv(.c) u64 {
return __udivmoddi4(a, b, null);
}
-pub fn __umoddi3(a: u64, b: u64) callconv(.C) u64 {
+pub fn __umoddi3(a: u64, b: u64) callconv(.c) u64 {
var r: u64 = undefined;
_ = __udivmoddi4(a, b, &r);
return r;
@@ -207,7 +207,7 @@ fn test_one_umoddi3(a: u64, b: u64, expected_r: u64) !void {
try testing.expect(r == expected_r);
}
-pub fn __divmodsi4(a: i32, b: i32, rem: *i32) callconv(.C) i32 {
+pub fn __divmodsi4(a: i32, b: i32, rem: *i32) callconv(.c) i32 {
const d = __divsi3(a, b);
rem.* = a -% (d * b);
return d;
@@ -241,17 +241,17 @@ test "test_divmodsi4" {
}
}
-pub fn __udivmodsi4(a: u32, b: u32, rem: *u32) callconv(.C) u32 {
+pub fn __udivmodsi4(a: u32, b: u32, rem: *u32) callconv(.c) u32 {
const d = __udivsi3(a, b);
rem.* = @bitCast(@as(i32, @bitCast(a)) -% (@as(i32, @bitCast(d)) * @as(i32, @bitCast(b))));
return d;
}
-pub fn __divsi3(n: i32, d: i32) callconv(.C) i32 {
+pub fn __divsi3(n: i32, d: i32) callconv(.c) i32 {
return div_i32(n, d);
}
-fn __aeabi_idiv(n: i32, d: i32) callconv(.AAPCS) i32 {
+fn __aeabi_idiv(n: i32, d: i32) callconv(.{ .arm_aapcs = .{} }) i32 {
return div_i32(n, d);
}
@@ -292,11 +292,11 @@ fn test_one_divsi3(a: i32, b: i32, expected_q: i32) !void {
try testing.expect(q == expected_q);
}
-pub fn __udivsi3(n: u32, d: u32) callconv(.C) u32 {
+pub fn __udivsi3(n: u32, d: u32) callconv(.c) u32 {
return div_u32(n, d);
}
-fn __aeabi_uidiv(n: u32, d: u32) callconv(.AAPCS) u32 {
+fn __aeabi_uidiv(n: u32, d: u32) callconv(.{ .arm_aapcs = .{} }) u32 {
return div_u32(n, d);
}
@@ -485,7 +485,7 @@ fn test_one_udivsi3(a: u32, b: u32, expected_q: u32) !void {
try testing.expect(q == expected_q);
}
-pub fn __modsi3(n: i32, d: i32) callconv(.C) i32 {
+pub fn __modsi3(n: i32, d: i32) callconv(.c) i32 {
return n -% __divsi3(n, d) * d;
}
@@ -514,7 +514,7 @@ fn test_one_modsi3(a: i32, b: i32, expected_r: i32) !void {
try testing.expect(r == expected_r);
}
-pub fn __umodsi3(n: u32, d: u32) callconv(.C) u32 {
+pub fn __umodsi3(n: u32, d: u32) callconv(.c) u32 {
return n -% __udivsi3(n, d) * d;
}
diff --git a/lib/compiler_rt/log.zig b/lib/compiler_rt/log.zig
index 6b179fd199..6b5c8fc683 100644
--- a/lib/compiler_rt/log.zig
+++ b/lib/compiler_rt/log.zig
@@ -25,12 +25,12 @@ comptime {
@export(&logl, .{ .name = "logl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __logh(a: f16) callconv(.C) f16 {
+pub fn __logh(a: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(logf(a));
}
-pub fn logf(x_: f32) callconv(.C) f32 {
+pub fn logf(x_: f32) callconv(.c) f32 {
const ln2_hi: f32 = 6.9313812256e-01;
const ln2_lo: f32 = 9.0580006145e-06;
const Lg1: f32 = 0xaaaaaa.0p-24;
@@ -82,7 +82,7 @@ pub fn logf(x_: f32) callconv(.C) f32 {
return s * (hfsq + R) + dk * ln2_lo - hfsq + f + dk * ln2_hi;
}
-pub fn log(x_: f64) callconv(.C) f64 {
+pub fn log(x_: f64) callconv(.c) f64 {
const ln2_hi: f64 = 6.93147180369123816490e-01;
const ln2_lo: f64 = 1.90821492927058770002e-10;
const Lg1: f64 = 6.666666666666735130e-01;
@@ -138,17 +138,17 @@ pub fn log(x_: f64) callconv(.C) f64 {
return s * (hfsq + R) + dk * ln2_lo - hfsq + f + dk * ln2_hi;
}
-pub fn __logx(a: f80) callconv(.C) f80 {
+pub fn __logx(a: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(logq(a));
}
-pub fn logq(a: f128) callconv(.C) f128 {
+pub fn logq(a: f128) callconv(.c) f128 {
// TODO: more correct implementation
return log(@floatCast(a));
}
-pub fn logl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn logl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __logh(x),
32 => return logf(x),
diff --git a/lib/compiler_rt/log10.zig b/lib/compiler_rt/log10.zig
index 7b32362534..f80c514ff6 100644
--- a/lib/compiler_rt/log10.zig
+++ b/lib/compiler_rt/log10.zig
@@ -26,12 +26,12 @@ comptime {
@export(&log10l, .{ .name = "log10l", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __log10h(a: f16) callconv(.C) f16 {
+pub fn __log10h(a: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(log10f(a));
}
-pub fn log10f(x_: f32) callconv(.C) f32 {
+pub fn log10f(x_: f32) callconv(.c) f32 {
const ivln10hi: f32 = 4.3432617188e-01;
const ivln10lo: f32 = -3.1689971365e-05;
const log10_2hi: f32 = 3.0102920532e-01;
@@ -91,7 +91,7 @@ pub fn log10f(x_: f32) callconv(.C) f32 {
return dk * log10_2lo + (lo + hi) * ivln10lo + lo * ivln10hi + hi * ivln10hi + dk * log10_2hi;
}
-pub fn log10(x_: f64) callconv(.C) f64 {
+pub fn log10(x_: f64) callconv(.c) f64 {
const ivln10hi: f64 = 4.34294481878168880939e-01;
const ivln10lo: f64 = 2.50829467116452752298e-11;
const log10_2hi: f64 = 3.01029995663611771306e-01;
@@ -166,17 +166,17 @@ pub fn log10(x_: f64) callconv(.C) f64 {
return val_lo + val_hi;
}
-pub fn __log10x(a: f80) callconv(.C) f80 {
+pub fn __log10x(a: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(log10q(a));
}
-pub fn log10q(a: f128) callconv(.C) f128 {
+pub fn log10q(a: f128) callconv(.c) f128 {
// TODO: more correct implementation
return log10(@floatCast(a));
}
-pub fn log10l(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn log10l(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __log10h(x),
32 => return log10f(x),
diff --git a/lib/compiler_rt/log2.zig b/lib/compiler_rt/log2.zig
index e152d5a6b4..7d52e437fd 100644
--- a/lib/compiler_rt/log2.zig
+++ b/lib/compiler_rt/log2.zig
@@ -26,12 +26,12 @@ comptime {
@export(&log2l, .{ .name = "log2l", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __log2h(a: f16) callconv(.C) f16 {
+pub fn __log2h(a: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(log2f(a));
}
-pub fn log2f(x_: f32) callconv(.C) f32 {
+pub fn log2f(x_: f32) callconv(.c) f32 {
const ivln2hi: f32 = 1.4428710938e+00;
const ivln2lo: f32 = -1.7605285393e-04;
const Lg1: f32 = 0xaaaaaa.0p-24;
@@ -87,7 +87,7 @@ pub fn log2f(x_: f32) callconv(.C) f32 {
return (lo + hi) * ivln2lo + lo * ivln2hi + hi * ivln2hi + @as(f32, @floatFromInt(k));
}
-pub fn log2(x_: f64) callconv(.C) f64 {
+pub fn log2(x_: f64) callconv(.c) f64 {
const ivln2hi: f64 = 1.44269504072144627571e+00;
const ivln2lo: f64 = 1.67517131648865118353e-10;
const Lg1: f64 = 6.666666666666735130e-01;
@@ -158,17 +158,17 @@ pub fn log2(x_: f64) callconv(.C) f64 {
return val_lo + val_hi;
}
-pub fn __log2x(a: f80) callconv(.C) f80 {
+pub fn __log2x(a: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(log2q(a));
}
-pub fn log2q(a: f128) callconv(.C) f128 {
+pub fn log2q(a: f128) callconv(.c) f128 {
// TODO: more correct implementation
return log2(@floatCast(a));
}
-pub fn log2l(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn log2l(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __log2h(x),
32 => return log2f(x),
diff --git a/lib/compiler_rt/memcmp.zig b/lib/compiler_rt/memcmp.zig
index b0e822e2dc..f332d82eac 100644
--- a/lib/compiler_rt/memcmp.zig
+++ b/lib/compiler_rt/memcmp.zig
@@ -5,7 +5,7 @@ comptime {
@export(&memcmp, .{ .name = "memcmp", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn memcmp(vl: [*]const u8, vr: [*]const u8, n: usize) callconv(.C) c_int {
+pub fn memcmp(vl: [*]const u8, vr: [*]const u8, n: usize) callconv(.c) c_int {
var i: usize = 0;
while (i < n) : (i += 1) {
const compared = @as(c_int, vl[i]) -% @as(c_int, vr[i]);
diff --git a/lib/compiler_rt/memcpy.zig b/lib/compiler_rt/memcpy.zig
index 0033df4d17..30971677ab 100644
--- a/lib/compiler_rt/memcpy.zig
+++ b/lib/compiler_rt/memcpy.zig
@@ -24,7 +24,7 @@ comptime {
assert(std.math.isPowerOfTwo(@sizeOf(Element)));
}
-fn memcpySmall(noalias dest: ?[*]u8, noalias src: ?[*]const u8, len: usize) callconv(.C) ?[*]u8 {
+fn memcpySmall(noalias dest: ?[*]u8, noalias src: ?[*]const u8, len: usize) callconv(.c) ?[*]u8 {
@setRuntimeSafety(false);
for (0..len) |i| {
@@ -34,7 +34,7 @@ fn memcpySmall(noalias dest: ?[*]u8, noalias src: ?[*]const u8, len: usize) call
return dest;
}
-fn memcpyFast(noalias dest: ?[*]u8, noalias src: ?[*]const u8, len: usize) callconv(.C) ?[*]u8 {
+fn memcpyFast(noalias dest: ?[*]u8, noalias src: ?[*]const u8, len: usize) callconv(.c) ?[*]u8 {
@setRuntimeSafety(false);
const small_limit = 2 * @sizeOf(Element);
diff --git a/lib/compiler_rt/memmove.zig b/lib/compiler_rt/memmove.zig
index c136e80c7c..71289a50ae 100644
--- a/lib/compiler_rt/memmove.zig
+++ b/lib/compiler_rt/memmove.zig
@@ -21,7 +21,7 @@ comptime {
}
}
-fn memmoveSmall(opt_dest: ?[*]u8, opt_src: ?[*]const u8, len: usize) callconv(.C) ?[*]u8 {
+fn memmoveSmall(opt_dest: ?[*]u8, opt_src: ?[*]const u8, len: usize) callconv(.c) ?[*]u8 {
const dest = opt_dest.?;
const src = opt_src.?;
@@ -38,7 +38,7 @@ fn memmoveSmall(opt_dest: ?[*]u8, opt_src: ?[*]const u8, len: usize) callconv(.C
return dest;
}
-fn memmoveFast(dest: ?[*]u8, src: ?[*]u8, len: usize) callconv(.C) ?[*]u8 {
+fn memmoveFast(dest: ?[*]u8, src: ?[*]u8, len: usize) callconv(.c) ?[*]u8 {
@setRuntimeSafety(builtin.is_test);
const small_limit = @max(2 * @sizeOf(Element), @sizeOf(Element));
diff --git a/lib/compiler_rt/memset.zig b/lib/compiler_rt/memset.zig
index 31c01f0948..6b51037b57 100644
--- a/lib/compiler_rt/memset.zig
+++ b/lib/compiler_rt/memset.zig
@@ -9,7 +9,7 @@ comptime {
}
}
-pub fn memset(dest: ?[*]u8, c: u8, len: usize) callconv(.C) ?[*]u8 {
+pub fn memset(dest: ?[*]u8, c: u8, len: usize) callconv(.c) ?[*]u8 {
@setRuntimeSafety(false);
if (len != 0) {
@@ -26,7 +26,7 @@ pub fn memset(dest: ?[*]u8, c: u8, len: usize) callconv(.C) ?[*]u8 {
return dest;
}
-pub fn __memset(dest: ?[*]u8, c: u8, n: usize, dest_n: usize) callconv(.C) ?[*]u8 {
+pub fn __memset(dest: ?[*]u8, c: u8, n: usize, dest_n: usize) callconv(.c) ?[*]u8 {
if (dest_n < n)
@panic("buffer overflow");
return memset(dest, c, n);
diff --git a/lib/compiler_rt/modti3.zig b/lib/compiler_rt/modti3.zig
index 0c13660d3a..b73f3d05fa 100644
--- a/lib/compiler_rt/modti3.zig
+++ b/lib/compiler_rt/modti3.zig
@@ -17,13 +17,13 @@ comptime {
}
}
-pub fn __modti3(a: i128, b: i128) callconv(.C) i128 {
+pub fn __modti3(a: i128, b: i128) callconv(.c) i128 {
return mod(a, b);
}
const v2u64 = @Vector(2, u64);
-fn __modti3_windows_x86_64(a: v2u64, b: v2u64) callconv(.C) v2u64 {
+fn __modti3_windows_x86_64(a: v2u64, b: v2u64) callconv(.c) v2u64 {
return @bitCast(mod(@as(i128, @bitCast(a)), @as(i128, @bitCast(b))));
}
diff --git a/lib/compiler_rt/mulXi3.zig b/lib/compiler_rt/mulXi3.zig
index d8b3c91914..a96ef135c4 100644
--- a/lib/compiler_rt/mulXi3.zig
+++ b/lib/compiler_rt/mulXi3.zig
@@ -20,7 +20,7 @@ comptime {
}
}
-pub fn __mulsi3(a: i32, b: i32) callconv(.C) i32 {
+pub fn __mulsi3(a: i32, b: i32) callconv(.c) i32 {
var ua: u32 = @bitCast(a);
var ub: u32 = @bitCast(b);
var r: u32 = 0;
@@ -34,11 +34,11 @@ pub fn __mulsi3(a: i32, b: i32) callconv(.C) i32 {
return @bitCast(r);
}
-pub fn __muldi3(a: i64, b: i64) callconv(.C) i64 {
+pub fn __muldi3(a: i64, b: i64) callconv(.c) i64 {
return mulX(i64, a, b);
}
-fn __aeabi_lmul(a: i64, b: i64) callconv(.AAPCS) i64 {
+fn __aeabi_lmul(a: i64, b: i64) callconv(.{ .arm_aapcs = .{} }) i64 {
return mulX(i64, a, b);
}
@@ -86,13 +86,13 @@ fn muldXi(comptime T: type, a: T, b: T) DoubleInt(T) {
return r.all;
}
-pub fn __multi3(a: i128, b: i128) callconv(.C) i128 {
+pub fn __multi3(a: i128, b: i128) callconv(.c) i128 {
return mulX(i128, a, b);
}
const v2u64 = @Vector(2, u64);
-fn __multi3_windows_x86_64(a: v2u64, b: v2u64) callconv(.C) v2u64 {
+fn __multi3_windows_x86_64(a: v2u64, b: v2u64) callconv(.c) v2u64 {
return @bitCast(mulX(i128, @as(i128, @bitCast(a)), @as(i128, @bitCast(b))));
}
diff --git a/lib/compiler_rt/mulc3_test.zig b/lib/compiler_rt/mulc3_test.zig
index 246a39cc90..3872cb34a6 100644
--- a/lib/compiler_rt/mulc3_test.zig
+++ b/lib/compiler_rt/mulc3_test.zig
@@ -17,7 +17,7 @@ test "mulc3" {
try testMul(f128, __multc3);
}
-fn testMul(comptime T: type, comptime f: fn (T, T, T, T) callconv(.C) Complex(T)) !void {
+fn testMul(comptime T: type, comptime f: fn (T, T, T, T) callconv(.c) Complex(T)) !void {
{
const a: T = 1.0;
const b: T = 0.0;
diff --git a/lib/compiler_rt/muldc3.zig b/lib/compiler_rt/muldc3.zig
index d23addeac5..1c53b4d073 100644
--- a/lib/compiler_rt/muldc3.zig
+++ b/lib/compiler_rt/muldc3.zig
@@ -9,6 +9,6 @@ comptime {
}
}
-pub fn __muldc3(a: f64, b: f64, c: f64, d: f64) callconv(.C) mulc3.Complex(f64) {
+pub fn __muldc3(a: f64, b: f64, c: f64, d: f64) callconv(.c) mulc3.Complex(f64) {
return mulc3.mulc3(f64, a, b, c, d);
}
diff --git a/lib/compiler_rt/muldf3.zig b/lib/compiler_rt/muldf3.zig
index d5af6d06c4..1fa4c9debb 100644
--- a/lib/compiler_rt/muldf3.zig
+++ b/lib/compiler_rt/muldf3.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __muldf3(a: f64, b: f64) callconv(.C) f64 {
+pub fn __muldf3(a: f64, b: f64) callconv(.c) f64 {
return mulf3(f64, a, b);
}
-fn __aeabi_dmul(a: f64, b: f64) callconv(.AAPCS) f64 {
+fn __aeabi_dmul(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) f64 {
return mulf3(f64, a, b);
}
diff --git a/lib/compiler_rt/mulhc3.zig b/lib/compiler_rt/mulhc3.zig
index 4d0fa1459d..4d97a0978d 100644
--- a/lib/compiler_rt/mulhc3.zig
+++ b/lib/compiler_rt/mulhc3.zig
@@ -9,6 +9,6 @@ comptime {
}
}
-pub fn __mulhc3(a: f16, b: f16, c: f16, d: f16) callconv(.C) mulc3.Complex(f16) {
+pub fn __mulhc3(a: f16, b: f16, c: f16, d: f16) callconv(.c) mulc3.Complex(f16) {
return mulc3.mulc3(f16, a, b, c, d);
}
diff --git a/lib/compiler_rt/mulhf3.zig b/lib/compiler_rt/mulhf3.zig
index 0eaf35e18f..2731625a16 100644
--- a/lib/compiler_rt/mulhf3.zig
+++ b/lib/compiler_rt/mulhf3.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__mulhf3, .{ .name = "__mulhf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __mulhf3(a: f16, b: f16) callconv(.C) f16 {
+pub fn __mulhf3(a: f16, b: f16) callconv(.c) f16 {
return mulf3(f16, a, b);
}
diff --git a/lib/compiler_rt/mulo.zig b/lib/compiler_rt/mulo.zig
index 44937cfa18..20fbff3b96 100644
--- a/lib/compiler_rt/mulo.zig
+++ b/lib/compiler_rt/mulo.zig
@@ -48,7 +48,7 @@ inline fn muloXi4_genericFast(comptime ST: type, a: ST, b: ST, overflow: *c_int)
return @as(ST, @truncate(res));
}
-pub fn __mulosi4(a: i32, b: i32, overflow: *c_int) callconv(.C) i32 {
+pub fn __mulosi4(a: i32, b: i32, overflow: *c_int) callconv(.c) i32 {
if (2 * @bitSizeOf(i32) <= @bitSizeOf(usize)) {
return muloXi4_genericFast(i32, a, b, overflow);
} else {
@@ -56,7 +56,7 @@ pub fn __mulosi4(a: i32, b: i32, overflow: *c_int) callconv(.C) i32 {
}
}
-pub fn __mulodi4(a: i64, b: i64, overflow: *c_int) callconv(.C) i64 {
+pub fn __mulodi4(a: i64, b: i64, overflow: *c_int) callconv(.c) i64 {
if (2 * @bitSizeOf(i64) <= @bitSizeOf(usize)) {
return muloXi4_genericFast(i64, a, b, overflow);
} else {
@@ -64,7 +64,7 @@ pub fn __mulodi4(a: i64, b: i64, overflow: *c_int) callconv(.C) i64 {
}
}
-pub fn __muloti4(a: i128, b: i128, overflow: *c_int) callconv(.C) i128 {
+pub fn __muloti4(a: i128, b: i128, overflow: *c_int) callconv(.c) i128 {
if (2 * @bitSizeOf(i128) <= @bitSizeOf(usize)) {
return muloXi4_genericFast(i128, a, b, overflow);
} else {
diff --git a/lib/compiler_rt/mulsc3.zig b/lib/compiler_rt/mulsc3.zig
index 9fb86d1bf4..f011d2fe23 100644
--- a/lib/compiler_rt/mulsc3.zig
+++ b/lib/compiler_rt/mulsc3.zig
@@ -9,6 +9,6 @@ comptime {
}
}
-pub fn __mulsc3(a: f32, b: f32, c: f32, d: f32) callconv(.C) mulc3.Complex(f32) {
+pub fn __mulsc3(a: f32, b: f32, c: f32, d: f32) callconv(.c) mulc3.Complex(f32) {
return mulc3.mulc3(f32, a, b, c, d);
}
diff --git a/lib/compiler_rt/mulsf3.zig b/lib/compiler_rt/mulsf3.zig
index 4c529cb2d8..8929b5deba 100644
--- a/lib/compiler_rt/mulsf3.zig
+++ b/lib/compiler_rt/mulsf3.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __mulsf3(a: f32, b: f32) callconv(.C) f32 {
+pub fn __mulsf3(a: f32, b: f32) callconv(.c) f32 {
return mulf3(f32, a, b);
}
-fn __aeabi_fmul(a: f32, b: f32) callconv(.AAPCS) f32 {
+fn __aeabi_fmul(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) f32 {
return mulf3(f32, a, b);
}
diff --git a/lib/compiler_rt/multc3.zig b/lib/compiler_rt/multc3.zig
index cc69e0b5fa..cf3621919d 100644
--- a/lib/compiler_rt/multc3.zig
+++ b/lib/compiler_rt/multc3.zig
@@ -11,6 +11,6 @@ comptime {
}
}
-pub fn __multc3(a: f128, b: f128, c: f128, d: f128) callconv(.C) mulc3.Complex(f128) {
+pub fn __multc3(a: f128, b: f128, c: f128, d: f128) callconv(.c) mulc3.Complex(f128) {
return mulc3.mulc3(f128, a, b, c, d);
}
diff --git a/lib/compiler_rt/multf3.zig b/lib/compiler_rt/multf3.zig
index 8ed84d0587..0da8c506e8 100644
--- a/lib/compiler_rt/multf3.zig
+++ b/lib/compiler_rt/multf3.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__multf3, .{ .name = "__multf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __multf3(a: f128, b: f128) callconv(.C) f128 {
+pub fn __multf3(a: f128, b: f128) callconv(.c) f128 {
return mulf3(f128, a, b);
}
-fn _Qp_mul(c: *f128, a: *const f128, b: *const f128) callconv(.C) void {
+fn _Qp_mul(c: *f128, a: *const f128, b: *const f128) callconv(.c) void {
c.* = mulf3(f128, a.*, b.*);
}
diff --git a/lib/compiler_rt/mulxc3.zig b/lib/compiler_rt/mulxc3.zig
index 27bcc0796d..ef6927ce08 100644
--- a/lib/compiler_rt/mulxc3.zig
+++ b/lib/compiler_rt/mulxc3.zig
@@ -9,6 +9,6 @@ comptime {
}
}
-pub fn __mulxc3(a: f80, b: f80, c: f80, d: f80) callconv(.C) mulc3.Complex(f80) {
+pub fn __mulxc3(a: f80, b: f80, c: f80, d: f80) callconv(.c) mulc3.Complex(f80) {
return mulc3.mulc3(f80, a, b, c, d);
}
diff --git a/lib/compiler_rt/mulxf3.zig b/lib/compiler_rt/mulxf3.zig
index f225ed1439..2db80b81b0 100644
--- a/lib/compiler_rt/mulxf3.zig
+++ b/lib/compiler_rt/mulxf3.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__mulxf3, .{ .name = "__mulxf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __mulxf3(a: f80, b: f80) callconv(.C) f80 {
+pub fn __mulxf3(a: f80, b: f80) callconv(.c) f80 {
return mulf3(f80, a, b);
}
diff --git a/lib/compiler_rt/negXi2.zig b/lib/compiler_rt/negXi2.zig
index c49df08428..eee5202e71 100644
--- a/lib/compiler_rt/negXi2.zig
+++ b/lib/compiler_rt/negXi2.zig
@@ -18,15 +18,15 @@ comptime {
@export(&__negti2, .{ .name = "__negti2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __negsi2(a: i32) callconv(.C) i32 {
+pub fn __negsi2(a: i32) callconv(.c) i32 {
return negXi2(i32, a);
}
-pub fn __negdi2(a: i64) callconv(.C) i64 {
+pub fn __negdi2(a: i64) callconv(.c) i64 {
return negXi2(i64, a);
}
-pub fn __negti2(a: i128) callconv(.C) i128 {
+pub fn __negti2(a: i128) callconv(.c) i128 {
return negXi2(i128, a);
}
diff --git a/lib/compiler_rt/negdf2.zig b/lib/compiler_rt/negdf2.zig
index afc9e31ffe..85e7f2bd82 100644
--- a/lib/compiler_rt/negdf2.zig
+++ b/lib/compiler_rt/negdf2.zig
@@ -10,10 +10,10 @@ comptime {
}
}
-fn __negdf2(a: f64) callconv(.C) f64 {
+fn __negdf2(a: f64) callconv(.c) f64 {
return common.fneg(a);
}
-fn __aeabi_dneg(a: f64) callconv(.AAPCS) f64 {
+fn __aeabi_dneg(a: f64) callconv(.{ .arm_aapcs = .{} }) f64 {
return common.fneg(a);
}
diff --git a/lib/compiler_rt/neghf2.zig b/lib/compiler_rt/neghf2.zig
index 455c3c7fc3..86b28d4802 100644
--- a/lib/compiler_rt/neghf2.zig
+++ b/lib/compiler_rt/neghf2.zig
@@ -6,6 +6,6 @@ comptime {
@export(&__neghf2, .{ .name = "__neghf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __neghf2(a: f16) callconv(.C) f16 {
+fn __neghf2(a: f16) callconv(.c) f16 {
return common.fneg(a);
}
diff --git a/lib/compiler_rt/negsf2.zig b/lib/compiler_rt/negsf2.zig
index 27c2eef5ce..08276e7f26 100644
--- a/lib/compiler_rt/negsf2.zig
+++ b/lib/compiler_rt/negsf2.zig
@@ -10,10 +10,10 @@ comptime {
}
}
-fn __negsf2(a: f32) callconv(.C) f32 {
+fn __negsf2(a: f32) callconv(.c) f32 {
return common.fneg(a);
}
-fn __aeabi_fneg(a: f32) callconv(.AAPCS) f32 {
+fn __aeabi_fneg(a: f32) callconv(.{ .arm_aapcs = .{} }) f32 {
return common.fneg(a);
}
diff --git a/lib/compiler_rt/negtf2.zig b/lib/compiler_rt/negtf2.zig
index 2c576c32f2..9fb9384446 100644
--- a/lib/compiler_rt/negtf2.zig
+++ b/lib/compiler_rt/negtf2.zig
@@ -8,6 +8,6 @@ comptime {
@export(&__negtf2, .{ .name = "__negtf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __negtf2(a: f128) callconv(.C) f128 {
+fn __negtf2(a: f128) callconv(.c) f128 {
return common.fneg(a);
}
diff --git a/lib/compiler_rt/negv.zig b/lib/compiler_rt/negv.zig
index 28122faca5..6a2e717db1 100644
--- a/lib/compiler_rt/negv.zig
+++ b/lib/compiler_rt/negv.zig
@@ -13,15 +13,15 @@ comptime {
@export(&__negvti2, .{ .name = "__negvti2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __negvsi2(a: i32) callconv(.C) i32 {
+pub fn __negvsi2(a: i32) callconv(.c) i32 {
return negvXi(i32, a);
}
-pub fn __negvdi2(a: i64) callconv(.C) i64 {
+pub fn __negvdi2(a: i64) callconv(.c) i64 {
return negvXi(i64, a);
}
-pub fn __negvti2(a: i128) callconv(.C) i128 {
+pub fn __negvti2(a: i128) callconv(.c) i128 {
return negvXi(i128, a);
}
diff --git a/lib/compiler_rt/negxf2.zig b/lib/compiler_rt/negxf2.zig
index 9bcd98463f..9a9c178654 100644
--- a/lib/compiler_rt/negxf2.zig
+++ b/lib/compiler_rt/negxf2.zig
@@ -6,6 +6,6 @@ comptime {
@export(&__negxf2, .{ .name = "__negxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __negxf2(a: f80) callconv(.C) f80 {
+fn __negxf2(a: f80) callconv(.c) f80 {
return common.fneg(a);
}
diff --git a/lib/compiler_rt/os_version_check.zig b/lib/compiler_rt/os_version_check.zig
index f61b9a71ba..9f3c9e74a7 100644
--- a/lib/compiler_rt/os_version_check.zig
+++ b/lib/compiler_rt/os_version_check.zig
@@ -34,7 +34,7 @@ const __isPlatformVersionAtLeast = if (have_availability_version_check) struct {
}
// Darwin-only
- fn __isPlatformVersionAtLeast(platform: u32, major: u32, minor: u32, subminor: u32) callconv(.C) i32 {
+ fn __isPlatformVersionAtLeast(platform: u32, major: u32, minor: u32, subminor: u32) callconv(.c) i32 {
const build_version = dyld_build_version_t{
.platform = platform,
.version = constructVersion(major, minor, subminor),
diff --git a/lib/compiler_rt/parity.zig b/lib/compiler_rt/parity.zig
index 37faa1a4b1..5a3e5a886d 100644
--- a/lib/compiler_rt/parity.zig
+++ b/lib/compiler_rt/parity.zig
@@ -13,15 +13,15 @@ comptime {
@export(&__parityti2, .{ .name = "__parityti2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __paritysi2(a: i32) callconv(.C) i32 {
+pub fn __paritysi2(a: i32) callconv(.c) i32 {
return parityXi2(i32, a);
}
-pub fn __paritydi2(a: i64) callconv(.C) i32 {
+pub fn __paritydi2(a: i64) callconv(.c) i32 {
return parityXi2(i64, a);
}
-pub fn __parityti2(a: i128) callconv(.C) i32 {
+pub fn __parityti2(a: i128) callconv(.c) i32 {
return parityXi2(i128, a);
}
diff --git a/lib/compiler_rt/popcount.zig b/lib/compiler_rt/popcount.zig
index 492423dda5..f937bd0ce0 100644
--- a/lib/compiler_rt/popcount.zig
+++ b/lib/compiler_rt/popcount.zig
@@ -18,15 +18,15 @@ comptime {
@export(&__popcountti2, .{ .name = "__popcountti2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __popcountsi2(a: i32) callconv(.C) i32 {
+pub fn __popcountsi2(a: i32) callconv(.c) i32 {
return popcountXi2(i32, a);
}
-pub fn __popcountdi2(a: i64) callconv(.C) i32 {
+pub fn __popcountdi2(a: i64) callconv(.c) i32 {
return popcountXi2(i64, a);
}
-pub fn __popcountti2(a: i128) callconv(.C) i32 {
+pub fn __popcountti2(a: i128) callconv(.c) i32 {
return popcountXi2(i128, a);
}
diff --git a/lib/compiler_rt/powiXf2.zig b/lib/compiler_rt/powiXf2.zig
index 4cf59950f0..21e7042a06 100644
--- a/lib/compiler_rt/powiXf2.zig
+++ b/lib/compiler_rt/powiXf2.zig
@@ -35,23 +35,23 @@ inline fn powiXf2(comptime FT: type, a: FT, b: i32) FT {
return if (is_recip) 1 / r else r;
}
-pub fn __powihf2(a: f16, b: i32) callconv(.C) f16 {
+pub fn __powihf2(a: f16, b: i32) callconv(.c) f16 {
return powiXf2(f16, a, b);
}
-pub fn __powisf2(a: f32, b: i32) callconv(.C) f32 {
+pub fn __powisf2(a: f32, b: i32) callconv(.c) f32 {
return powiXf2(f32, a, b);
}
-pub fn __powidf2(a: f64, b: i32) callconv(.C) f64 {
+pub fn __powidf2(a: f64, b: i32) callconv(.c) f64 {
return powiXf2(f64, a, b);
}
-pub fn __powitf2(a: f128, b: i32) callconv(.C) f128 {
+pub fn __powitf2(a: f128, b: i32) callconv(.c) f128 {
return powiXf2(f128, a, b);
}
-pub fn __powixf2(a: f80, b: i32) callconv(.C) f80 {
+pub fn __powixf2(a: f80, b: i32) callconv(.c) f80 {
return powiXf2(f80, a, b);
}
diff --git a/lib/compiler_rt/round.zig b/lib/compiler_rt/round.zig
index f0745feb49..9e8ea42d67 100644
--- a/lib/compiler_rt/round.zig
+++ b/lib/compiler_rt/round.zig
@@ -26,12 +26,12 @@ comptime {
@export(&roundl, .{ .name = "roundl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __roundh(x: f16) callconv(.C) f16 {
+pub fn __roundh(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(roundf(x));
}
-pub fn roundf(x_: f32) callconv(.C) f32 {
+pub fn roundf(x_: f32) callconv(.c) f32 {
const f32_toint = 1.0 / math.floatEps(f32);
var x = x_;
@@ -66,7 +66,7 @@ pub fn roundf(x_: f32) callconv(.C) f32 {
}
}
-pub fn round(x_: f64) callconv(.C) f64 {
+pub fn round(x_: f64) callconv(.c) f64 {
const f64_toint = 1.0 / math.floatEps(f64);
var x = x_;
@@ -101,12 +101,12 @@ pub fn round(x_: f64) callconv(.C) f64 {
}
}
-pub fn __roundx(x: f80) callconv(.C) f80 {
+pub fn __roundx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(roundq(x));
}
-pub fn roundq(x_: f128) callconv(.C) f128 {
+pub fn roundq(x_: f128) callconv(.c) f128 {
const f128_toint = 1.0 / math.floatEps(f128);
var x = x_;
@@ -141,7 +141,7 @@ pub fn roundq(x_: f128) callconv(.C) f128 {
}
}
-pub fn roundl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn roundl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __roundh(x),
32 => return roundf(x),
diff --git a/lib/compiler_rt/shift.zig b/lib/compiler_rt/shift.zig
index 5b46e8ec00..17860da943 100644
--- a/lib/compiler_rt/shift.zig
+++ b/lib/compiler_rt/shift.zig
@@ -93,48 +93,48 @@ inline fn lshrXi3(comptime T: type, a: T, b: i32) T {
return output.all;
}
-pub fn __ashlsi3(a: i32, b: i32) callconv(.C) i32 {
+pub fn __ashlsi3(a: i32, b: i32) callconv(.c) i32 {
return ashlXi3(i32, a, b);
}
-pub fn __ashrsi3(a: i32, b: i32) callconv(.C) i32 {
+pub fn __ashrsi3(a: i32, b: i32) callconv(.c) i32 {
return ashrXi3(i32, a, b);
}
-pub fn __lshrsi3(a: i32, b: i32) callconv(.C) i32 {
+pub fn __lshrsi3(a: i32, b: i32) callconv(.c) i32 {
return lshrXi3(i32, a, b);
}
-pub fn __ashldi3(a: i64, b: i32) callconv(.C) i64 {
+pub fn __ashldi3(a: i64, b: i32) callconv(.c) i64 {
return ashlXi3(i64, a, b);
}
-fn __aeabi_llsl(a: i64, b: i32) callconv(.AAPCS) i64 {
+fn __aeabi_llsl(a: i64, b: i32) callconv(.{ .arm_aapcs = .{} }) i64 {
return ashlXi3(i64, a, b);
}
-pub fn __ashlti3(a: i128, b: i32) callconv(.C) i128 {
+pub fn __ashlti3(a: i128, b: i32) callconv(.c) i128 {
return ashlXi3(i128, a, b);
}
-pub fn __ashrdi3(a: i64, b: i32) callconv(.C) i64 {
+pub fn __ashrdi3(a: i64, b: i32) callconv(.c) i64 {
return ashrXi3(i64, a, b);
}
-fn __aeabi_lasr(a: i64, b: i32) callconv(.AAPCS) i64 {
+fn __aeabi_lasr(a: i64, b: i32) callconv(.{ .arm_aapcs = .{} }) i64 {
return ashrXi3(i64, a, b);
}
-pub fn __ashrti3(a: i128, b: i32) callconv(.C) i128 {
+pub fn __ashrti3(a: i128, b: i32) callconv(.c) i128 {
return ashrXi3(i128, a, b);
}
-pub fn __lshrdi3(a: i64, b: i32) callconv(.C) i64 {
+pub fn __lshrdi3(a: i64, b: i32) callconv(.c) i64 {
return lshrXi3(i64, a, b);
}
-fn __aeabi_llsr(a: i64, b: i32) callconv(.AAPCS) i64 {
+fn __aeabi_llsr(a: i64, b: i32) callconv(.{ .arm_aapcs = .{} }) i64 {
return lshrXi3(i64, a, b);
}
-pub fn __lshrti3(a: i128, b: i32) callconv(.C) i128 {
+pub fn __lshrti3(a: i128, b: i32) callconv(.c) i128 {
return lshrXi3(i128, a, b);
}
diff --git a/lib/compiler_rt/sin.zig b/lib/compiler_rt/sin.zig
index 505bf59fd4..4ec2fa329e 100644
--- a/lib/compiler_rt/sin.zig
+++ b/lib/compiler_rt/sin.zig
@@ -30,12 +30,12 @@ comptime {
@export(&sinl, .{ .name = "sinl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __sinh(x: f16) callconv(.C) f16 {
+pub fn __sinh(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(sinf(x));
}
-pub fn sinf(x: f32) callconv(.C) f32 {
+pub fn sinf(x: f32) callconv(.c) f32 {
// Small multiples of pi/2 rounded to double precision.
const s1pio2: f64 = 1.0 * math.pi / 2.0; // 0x3FF921FB, 0x54442D18
const s2pio2: f64 = 2.0 * math.pi / 2.0; // 0x400921FB, 0x54442D18
@@ -90,7 +90,7 @@ pub fn sinf(x: f32) callconv(.C) f32 {
};
}
-pub fn sin(x: f64) callconv(.C) f64 {
+pub fn sin(x: f64) callconv(.c) f64 {
var ix = @as(u64, @bitCast(x)) >> 32;
ix &= 0x7fffffff;
@@ -119,17 +119,17 @@ pub fn sin(x: f64) callconv(.C) f64 {
};
}
-pub fn __sinx(x: f80) callconv(.C) f80 {
+pub fn __sinx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(sinq(x));
}
-pub fn sinq(x: f128) callconv(.C) f128 {
+pub fn sinq(x: f128) callconv(.c) f128 {
// TODO: more correct implementation
return sin(@floatCast(x));
}
-pub fn sinl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn sinl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __sinh(x),
32 => return sinf(x),
diff --git a/lib/compiler_rt/sincos.zig b/lib/compiler_rt/sincos.zig
index 7763be42db..ec082ff436 100644
--- a/lib/compiler_rt/sincos.zig
+++ b/lib/compiler_rt/sincos.zig
@@ -22,7 +22,7 @@ comptime {
@export(&sincosl, .{ .name = "sincosl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __sincosh(x: f16, r_sin: *f16, r_cos: *f16) callconv(.C) void {
+pub fn __sincosh(x: f16, r_sin: *f16, r_cos: *f16) callconv(.c) void {
// TODO: more efficient implementation
var big_sin: f32 = undefined;
var big_cos: f32 = undefined;
@@ -31,7 +31,7 @@ pub fn __sincosh(x: f16, r_sin: *f16, r_cos: *f16) callconv(.C) void {
r_cos.* = @as(f16, @floatCast(big_cos));
}
-pub fn sincosf(x: f32, r_sin: *f32, r_cos: *f32) callconv(.C) void {
+pub fn sincosf(x: f32, r_sin: *f32, r_cos: *f32) callconv(.c) void {
const sc1pio2: f64 = 1.0 * math.pi / 2.0; // 0x3FF921FB, 0x54442D18
const sc2pio2: f64 = 2.0 * math.pi / 2.0; // 0x400921FB, 0x54442D18
const sc3pio2: f64 = 3.0 * math.pi / 2.0; // 0x4012D97C, 0x7F3321D2
@@ -126,7 +126,7 @@ pub fn sincosf(x: f32, r_sin: *f32, r_cos: *f32) callconv(.C) void {
}
}
-pub fn sincos(x: f64, r_sin: *f64, r_cos: *f64) callconv(.C) void {
+pub fn sincos(x: f64, r_sin: *f64, r_cos: *f64) callconv(.c) void {
const ix = @as(u32, @truncate(@as(u64, @bitCast(x)) >> 32)) & 0x7fffffff;
// |x| ~< pi/4
@@ -177,7 +177,7 @@ pub fn sincos(x: f64, r_sin: *f64, r_cos: *f64) callconv(.C) void {
}
}
-pub fn __sincosx(x: f80, r_sin: *f80, r_cos: *f80) callconv(.C) void {
+pub fn __sincosx(x: f80, r_sin: *f80, r_cos: *f80) callconv(.c) void {
// TODO: more efficient implementation
//return sincos_generic(f80, x, r_sin, r_cos);
var big_sin: f128 = undefined;
@@ -187,7 +187,7 @@ pub fn __sincosx(x: f80, r_sin: *f80, r_cos: *f80) callconv(.C) void {
r_cos.* = @as(f80, @floatCast(big_cos));
}
-pub fn sincosq(x: f128, r_sin: *f128, r_cos: *f128) callconv(.C) void {
+pub fn sincosq(x: f128, r_sin: *f128, r_cos: *f128) callconv(.c) void {
// TODO: more correct implementation
//return sincos_generic(f128, x, r_sin, r_cos);
var small_sin: f64 = undefined;
@@ -197,7 +197,7 @@ pub fn sincosq(x: f128, r_sin: *f128, r_cos: *f128) callconv(.C) void {
r_cos.* = small_cos;
}
-pub fn sincosl(x: c_longdouble, r_sin: *c_longdouble, r_cos: *c_longdouble) callconv(.C) void {
+pub fn sincosl(x: c_longdouble, r_sin: *c_longdouble, r_cos: *c_longdouble) callconv(.c) void {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __sincosh(x, r_sin, r_cos),
32 => return sincosf(x, r_sin, r_cos),
diff --git a/lib/compiler_rt/sqrt.zig b/lib/compiler_rt/sqrt.zig
index b804ff9359..331606f0f4 100644
--- a/lib/compiler_rt/sqrt.zig
+++ b/lib/compiler_rt/sqrt.zig
@@ -18,12 +18,12 @@ comptime {
@export(&sqrtl, .{ .name = "sqrtl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __sqrth(x: f16) callconv(.C) f16 {
+pub fn __sqrth(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(sqrtf(x));
}
-pub fn sqrtf(x: f32) callconv(.C) f32 {
+pub fn sqrtf(x: f32) callconv(.c) f32 {
const tiny: f32 = 1.0e-30;
const sign: i32 = @bitCast(@as(u32, 0x80000000));
var ix: i32 = @bitCast(x);
@@ -102,7 +102,7 @@ pub fn sqrtf(x: f32) callconv(.C) f32 {
/// NOTE: The original code is full of implicit signed -> unsigned assumptions and u32 wraparound
/// behaviour. Most intermediate i32 values are changed to u32 where appropriate but there are
/// potentially some edge cases remaining that are not handled in the same way.
-pub fn sqrt(x: f64) callconv(.C) f64 {
+pub fn sqrt(x: f64) callconv(.c) f64 {
const tiny: f64 = 1.0e-300;
const sign: u32 = 0x80000000;
const u: u64 = @bitCast(x);
@@ -232,17 +232,17 @@ pub fn sqrt(x: f64) callconv(.C) f64 {
return @bitCast(uz);
}
-pub fn __sqrtx(x: f80) callconv(.C) f80 {
+pub fn __sqrtx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(sqrtq(x));
}
-pub fn sqrtq(x: f128) callconv(.C) f128 {
+pub fn sqrtq(x: f128) callconv(.c) f128 {
// TODO: more correct implementation
return sqrt(@floatCast(x));
}
-pub fn sqrtl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn sqrtl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __sqrth(x),
32 => return sqrtf(x),
diff --git a/lib/compiler_rt/ssp.zig b/lib/compiler_rt/ssp.zig
index d32895ea52..f0882cfff2 100644
--- a/lib/compiler_rt/ssp.zig
+++ b/lib/compiler_rt/ssp.zig
@@ -16,9 +16,9 @@ const std = @import("std");
const common = @import("./common.zig");
const builtin = @import("builtin");
-extern fn memset(dest: ?[*]u8, c: u8, n: usize) callconv(.C) ?[*]u8;
-extern fn memcpy(noalias dest: ?[*]u8, noalias src: ?[*]const u8, n: usize) callconv(.C) ?[*]u8;
-extern fn memmove(dest: ?[*]u8, src: ?[*]const u8, n: usize) callconv(.C) ?[*]u8;
+extern fn memset(dest: ?[*]u8, c: u8, n: usize) callconv(.c) ?[*]u8;
+extern fn memcpy(noalias dest: ?[*]u8, noalias src: ?[*]const u8, n: usize) callconv(.c) ?[*]u8;
+extern fn memmove(dest: ?[*]u8, src: ?[*]const u8, n: usize) callconv(.c) ?[*]u8;
comptime {
@export(&__stack_chk_fail, .{ .name = "__stack_chk_fail", .linkage = common.linkage, .visibility = common.visibility });
@@ -33,11 +33,11 @@ comptime {
@export(&__memset_chk, .{ .name = "__memset_chk", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __stack_chk_fail() callconv(.C) noreturn {
+fn __stack_chk_fail() callconv(.c) noreturn {
@panic("stack smashing detected");
}
-fn __chk_fail() callconv(.C) noreturn {
+fn __chk_fail() callconv(.c) noreturn {
@panic("buffer overflow detected");
}
@@ -49,7 +49,7 @@ var __stack_chk_guard: usize = blk: {
break :blk @as(usize, @bitCast(buf));
};
-fn __strcpy_chk(dest: [*:0]u8, src: [*:0]const u8, dest_n: usize) callconv(.C) [*:0]u8 {
+fn __strcpy_chk(dest: [*:0]u8, src: [*:0]const u8, dest_n: usize) callconv(.c) [*:0]u8 {
@setRuntimeSafety(false);
var i: usize = 0;
@@ -64,7 +64,7 @@ fn __strcpy_chk(dest: [*:0]u8, src: [*:0]const u8, dest_n: usize) callconv(.C) [
return dest;
}
-fn __strncpy_chk(dest: [*:0]u8, src: [*:0]const u8, n: usize, dest_n: usize) callconv(.C) [*:0]u8 {
+fn __strncpy_chk(dest: [*:0]u8, src: [*:0]const u8, n: usize, dest_n: usize) callconv(.c) [*:0]u8 {
@setRuntimeSafety(false);
if (dest_n < n) __chk_fail();
var i: usize = 0;
@@ -77,7 +77,7 @@ fn __strncpy_chk(dest: [*:0]u8, src: [*:0]const u8, n: usize, dest_n: usize) cal
return dest;
}
-fn __strcat_chk(dest: [*:0]u8, src: [*:0]const u8, dest_n: usize) callconv(.C) [*:0]u8 {
+fn __strcat_chk(dest: [*:0]u8, src: [*:0]const u8, dest_n: usize) callconv(.c) [*:0]u8 {
@setRuntimeSafety(false);
var avail = dest_n;
@@ -102,7 +102,7 @@ fn __strcat_chk(dest: [*:0]u8, src: [*:0]const u8, dest_n: usize) callconv(.C) [
return dest;
}
-fn __strncat_chk(dest: [*:0]u8, src: [*:0]const u8, n: usize, dest_n: usize) callconv(.C) [*:0]u8 {
+fn __strncat_chk(dest: [*:0]u8, src: [*:0]const u8, n: usize, dest_n: usize) callconv(.c) [*:0]u8 {
@setRuntimeSafety(false);
var avail = dest_n;
@@ -127,17 +127,17 @@ fn __strncat_chk(dest: [*:0]u8, src: [*:0]const u8, n: usize, dest_n: usize) cal
return dest;
}
-fn __memcpy_chk(noalias dest: ?[*]u8, noalias src: ?[*]const u8, n: usize, dest_n: usize) callconv(.C) ?[*]u8 {
+fn __memcpy_chk(noalias dest: ?[*]u8, noalias src: ?[*]const u8, n: usize, dest_n: usize) callconv(.c) ?[*]u8 {
if (dest_n < n) __chk_fail();
return memcpy(dest, src, n);
}
-fn __memmove_chk(dest: ?[*]u8, src: ?[*]const u8, n: usize, dest_n: usize) callconv(.C) ?[*]u8 {
+fn __memmove_chk(dest: ?[*]u8, src: ?[*]const u8, n: usize, dest_n: usize) callconv(.c) ?[*]u8 {
if (dest_n < n) __chk_fail();
return memmove(dest, src, n);
}
-fn __memset_chk(dest: ?[*]u8, c: u8, n: usize, dest_n: usize) callconv(.C) ?[*]u8 {
+fn __memset_chk(dest: ?[*]u8, c: u8, n: usize, dest_n: usize) callconv(.c) ?[*]u8 {
if (dest_n < n) __chk_fail();
return memset(dest, c, n);
}
diff --git a/lib/compiler_rt/stack_probe.zig b/lib/compiler_rt/stack_probe.zig
index 69de4c8575..31a0852832 100644
--- a/lib/compiler_rt/stack_probe.zig
+++ b/lib/compiler_rt/stack_probe.zig
@@ -37,7 +37,7 @@ comptime {
}
// Zig's own stack-probe routine (available only on x86 and x86_64)
-pub fn zig_probe_stack() callconv(.Naked) void {
+pub fn zig_probe_stack() callconv(.naked) void {
@setRuntimeSafety(false);
// Versions of the Linux kernel before 5.1 treat any access below SP as
@@ -245,11 +245,11 @@ fn win_probe_stack_adjust_sp() void {
// ___chkstk (__alloca) | yes | yes |
// ___chkstk_ms | no | no |
-pub fn _chkstk() callconv(.Naked) void {
+pub fn _chkstk() callconv(.naked) void {
@setRuntimeSafety(false);
@call(.always_inline, win_probe_stack_adjust_sp, .{});
}
-pub fn __chkstk() callconv(.Naked) void {
+pub fn __chkstk() callconv(.naked) void {
@setRuntimeSafety(false);
if (arch == .thumb or arch == .aarch64) {
@call(.always_inline, win_probe_stack_only, .{});
@@ -259,15 +259,15 @@ pub fn __chkstk() callconv(.Naked) void {
else => unreachable,
}
}
-pub fn ___chkstk() callconv(.Naked) void {
+pub fn ___chkstk() callconv(.naked) void {
@setRuntimeSafety(false);
@call(.always_inline, win_probe_stack_adjust_sp, .{});
}
-pub fn __chkstk_ms() callconv(.Naked) void {
+pub fn __chkstk_ms() callconv(.naked) void {
@setRuntimeSafety(false);
@call(.always_inline, win_probe_stack_only, .{});
}
-pub fn ___chkstk_ms() callconv(.Naked) void {
+pub fn ___chkstk_ms() callconv(.naked) void {
@setRuntimeSafety(false);
@call(.always_inline, win_probe_stack_only, .{});
}
diff --git a/lib/compiler_rt/subdf3.zig b/lib/compiler_rt/subdf3.zig
index 8970c63b23..24ded58e0c 100644
--- a/lib/compiler_rt/subdf3.zig
+++ b/lib/compiler_rt/subdf3.zig
@@ -11,11 +11,11 @@ comptime {
}
}
-fn __subdf3(a: f64, b: f64) callconv(.C) f64 {
+fn __subdf3(a: f64, b: f64) callconv(.c) f64 {
return sub(a, b);
}
-fn __aeabi_dsub(a: f64, b: f64) callconv(.AAPCS) f64 {
+fn __aeabi_dsub(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) f64 {
return sub(a, b);
}
diff --git a/lib/compiler_rt/subhf3.zig b/lib/compiler_rt/subhf3.zig
index f116a97d43..cc42eb9270 100644
--- a/lib/compiler_rt/subhf3.zig
+++ b/lib/compiler_rt/subhf3.zig
@@ -7,7 +7,7 @@ comptime {
@export(&__subhf3, .{ .name = "__subhf3", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __subhf3(a: f16, b: f16) callconv(.C) f16 {
+fn __subhf3(a: f16, b: f16) callconv(.c) f16 {
const neg_b = @as(f16, @bitCast(@as(u16, @bitCast(b)) ^ (@as(u16, 1) << 15)));
return addf3(f16, a, neg_b);
}
diff --git a/lib/compiler_rt/subo.zig b/lib/compiler_rt/subo.zig
index a03e2158f5..b4fb8f7710 100644
--- a/lib/compiler_rt/subo.zig
+++ b/lib/compiler_rt/subo.zig
@@ -15,13 +15,13 @@ comptime {
@export(&__suboti4, .{ .name = "__suboti4", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __subosi4(a: i32, b: i32, overflow: *c_int) callconv(.C) i32 {
+pub fn __subosi4(a: i32, b: i32, overflow: *c_int) callconv(.c) i32 {
return suboXi4_generic(i32, a, b, overflow);
}
-pub fn __subodi4(a: i64, b: i64, overflow: *c_int) callconv(.C) i64 {
+pub fn __subodi4(a: i64, b: i64, overflow: *c_int) callconv(.c) i64 {
return suboXi4_generic(i64, a, b, overflow);
}
-pub fn __suboti4(a: i128, b: i128, overflow: *c_int) callconv(.C) i128 {
+pub fn __suboti4(a: i128, b: i128, overflow: *c_int) callconv(.c) i128 {
return suboXi4_generic(i128, a, b, overflow);
}
diff --git a/lib/compiler_rt/subsf3.zig b/lib/compiler_rt/subsf3.zig
index 84b5e5341e..f3bb334bd8 100644
--- a/lib/compiler_rt/subsf3.zig
+++ b/lib/compiler_rt/subsf3.zig
@@ -11,11 +11,11 @@ comptime {
}
}
-fn __subsf3(a: f32, b: f32) callconv(.C) f32 {
+fn __subsf3(a: f32, b: f32) callconv(.c) f32 {
return sub(a, b);
}
-fn __aeabi_fsub(a: f32, b: f32) callconv(.AAPCS) f32 {
+fn __aeabi_fsub(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) f32 {
return sub(a, b);
}
diff --git a/lib/compiler_rt/subtf3.zig b/lib/compiler_rt/subtf3.zig
index 8c6626e882..aff4904adf 100644
--- a/lib/compiler_rt/subtf3.zig
+++ b/lib/compiler_rt/subtf3.zig
@@ -12,11 +12,11 @@ comptime {
@export(&__subtf3, .{ .name = "__subtf3", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __subtf3(a: f128, b: f128) callconv(.C) f128 {
+pub fn __subtf3(a: f128, b: f128) callconv(.c) f128 {
return sub(a, b);
}
-fn _Qp_sub(c: *f128, a: *const f128, b: *const f128) callconv(.C) void {
+fn _Qp_sub(c: *f128, a: *const f128, b: *const f128) callconv(.c) void {
c.* = sub(a.*, b.*);
}
diff --git a/lib/compiler_rt/subxf3.zig b/lib/compiler_rt/subxf3.zig
index 9dc7625b1e..e79dd230b5 100644
--- a/lib/compiler_rt/subxf3.zig
+++ b/lib/compiler_rt/subxf3.zig
@@ -7,7 +7,7 @@ comptime {
@export(&__subxf3, .{ .name = "__subxf3", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __subxf3(a: f80, b: f80) callconv(.C) f80 {
+fn __subxf3(a: f80, b: f80) callconv(.c) f80 {
var b_rep = std.math.F80.fromFloat(b);
b_rep.exp ^= 0x8000;
const neg_b = b_rep.toFloat();
diff --git a/lib/compiler_rt/tan.zig b/lib/compiler_rt/tan.zig
index 067164a3a7..7a87cfecad 100644
--- a/lib/compiler_rt/tan.zig
+++ b/lib/compiler_rt/tan.zig
@@ -32,12 +32,12 @@ comptime {
@export(&tanl, .{ .name = "tanl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __tanh(x: f16) callconv(.C) f16 {
+pub fn __tanh(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(tanf(x));
}
-pub fn tanf(x: f32) callconv(.C) f32 {
+pub fn tanf(x: f32) callconv(.c) f32 {
// Small multiples of pi/2 rounded to double precision.
const t1pio2: f64 = 1.0 * math.pi / 2.0; // 0x3FF921FB, 0x54442D18
const t2pio2: f64 = 2.0 * math.pi / 2.0; // 0x400921FB, 0x54442D18
@@ -81,7 +81,7 @@ pub fn tanf(x: f32) callconv(.C) f32 {
return kernel.__tandf(y, n & 1 != 0);
}
-pub fn tan(x: f64) callconv(.C) f64 {
+pub fn tan(x: f64) callconv(.c) f64 {
var ix = @as(u64, @bitCast(x)) >> 32;
ix &= 0x7fffffff;
@@ -105,17 +105,17 @@ pub fn tan(x: f64) callconv(.C) f64 {
return kernel.__tan(y[0], y[1], n & 1 != 0);
}
-pub fn __tanx(x: f80) callconv(.C) f80 {
+pub fn __tanx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(tanq(x));
}
-pub fn tanq(x: f128) callconv(.C) f128 {
+pub fn tanq(x: f128) callconv(.c) f128 {
// TODO: more correct implementation
return tan(@floatCast(x));
}
-pub fn tanl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn tanl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __tanh(x),
32 => return tanf(x),
diff --git a/lib/compiler_rt/trunc.zig b/lib/compiler_rt/trunc.zig
index ea655fbb18..ff829ea710 100644
--- a/lib/compiler_rt/trunc.zig
+++ b/lib/compiler_rt/trunc.zig
@@ -26,12 +26,12 @@ comptime {
@export(&truncl, .{ .name = "truncl", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __trunch(x: f16) callconv(.C) f16 {
+pub fn __trunch(x: f16) callconv(.c) f16 {
// TODO: more efficient implementation
return @floatCast(truncf(x));
}
-pub fn truncf(x: f32) callconv(.C) f32 {
+pub fn truncf(x: f32) callconv(.c) f32 {
const u: u32 = @bitCast(x);
var e = @as(i32, @intCast(((u >> 23) & 0xFF))) - 0x7F + 9;
var m: u32 = undefined;
@@ -52,7 +52,7 @@ pub fn truncf(x: f32) callconv(.C) f32 {
}
}
-pub fn trunc(x: f64) callconv(.C) f64 {
+pub fn trunc(x: f64) callconv(.c) f64 {
const u: u64 = @bitCast(x);
var e = @as(i32, @intCast(((u >> 52) & 0x7FF))) - 0x3FF + 12;
var m: u64 = undefined;
@@ -73,12 +73,12 @@ pub fn trunc(x: f64) callconv(.C) f64 {
}
}
-pub fn __truncx(x: f80) callconv(.C) f80 {
+pub fn __truncx(x: f80) callconv(.c) f80 {
// TODO: more efficient implementation
return @floatCast(truncq(x));
}
-pub fn truncq(x: f128) callconv(.C) f128 {
+pub fn truncq(x: f128) callconv(.c) f128 {
const u: u128 = @bitCast(x);
var e = @as(i32, @intCast(((u >> 112) & 0x7FFF))) - 0x3FFF + 16;
var m: u128 = undefined;
@@ -99,7 +99,7 @@ pub fn truncq(x: f128) callconv(.C) f128 {
}
}
-pub fn truncl(x: c_longdouble) callconv(.C) c_longdouble {
+pub fn truncl(x: c_longdouble) callconv(.c) c_longdouble {
switch (@typeInfo(c_longdouble).float.bits) {
16 => return __trunch(x),
32 => return truncf(x),
diff --git a/lib/compiler_rt/truncdfhf2.zig b/lib/compiler_rt/truncdfhf2.zig
index 10f13d5047..a24650ddb7 100644
--- a/lib/compiler_rt/truncdfhf2.zig
+++ b/lib/compiler_rt/truncdfhf2.zig
@@ -10,10 +10,10 @@ comptime {
@export(&__truncdfhf2, .{ .name = "__truncdfhf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __truncdfhf2(a: f64) callconv(.C) common.F16T(f64) {
+pub fn __truncdfhf2(a: f64) callconv(.c) common.F16T(f64) {
return @bitCast(truncf(f16, f64, a));
}
-fn __aeabi_d2h(a: f64) callconv(.AAPCS) u16 {
+fn __aeabi_d2h(a: f64) callconv(.{ .arm_aapcs = .{} }) u16 {
return @bitCast(truncf(f16, f64, a));
}
diff --git a/lib/compiler_rt/truncdfsf2.zig b/lib/compiler_rt/truncdfsf2.zig
index 160df412d6..fd42d5d344 100644
--- a/lib/compiler_rt/truncdfsf2.zig
+++ b/lib/compiler_rt/truncdfsf2.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __truncdfsf2(a: f64) callconv(.C) f32 {
+pub fn __truncdfsf2(a: f64) callconv(.c) f32 {
return truncf(f32, f64, a);
}
-fn __aeabi_d2f(a: f64) callconv(.AAPCS) f32 {
+fn __aeabi_d2f(a: f64) callconv(.{ .arm_aapcs = .{} }) f32 {
return truncf(f32, f64, a);
}
diff --git a/lib/compiler_rt/truncsfhf2.zig b/lib/compiler_rt/truncsfhf2.zig
index 9de9cecd4b..18d5854c91 100644
--- a/lib/compiler_rt/truncsfhf2.zig
+++ b/lib/compiler_rt/truncsfhf2.zig
@@ -12,14 +12,14 @@ comptime {
@export(&__truncsfhf2, .{ .name = "__truncsfhf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __truncsfhf2(a: f32) callconv(.C) common.F16T(f32) {
+pub fn __truncsfhf2(a: f32) callconv(.c) common.F16T(f32) {
return @bitCast(truncf(f16, f32, a));
}
-fn __gnu_f2h_ieee(a: f32) callconv(.C) common.F16T(f32) {
+fn __gnu_f2h_ieee(a: f32) callconv(.c) common.F16T(f32) {
return @bitCast(truncf(f16, f32, a));
}
-fn __aeabi_f2h(a: f32) callconv(.AAPCS) u16 {
+fn __aeabi_f2h(a: f32) callconv(.{ .arm_aapcs = .{} }) u16 {
return @bitCast(truncf(f16, f32, a));
}
diff --git a/lib/compiler_rt/trunctfdf2.zig b/lib/compiler_rt/trunctfdf2.zig
index 2dc3242b12..37b9227e0b 100644
--- a/lib/compiler_rt/trunctfdf2.zig
+++ b/lib/compiler_rt/trunctfdf2.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__trunctfdf2, .{ .name = "__trunctfdf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __trunctfdf2(a: f128) callconv(.C) f64 {
+pub fn __trunctfdf2(a: f128) callconv(.c) f64 {
return truncf(f64, f128, a);
}
-fn _Qp_qtod(a: *const f128) callconv(.C) f64 {
+fn _Qp_qtod(a: *const f128) callconv(.c) f64 {
return truncf(f64, f128, a.*);
}
diff --git a/lib/compiler_rt/trunctfhf2.zig b/lib/compiler_rt/trunctfhf2.zig
index d4dbc5b44b..daf4776106 100644
--- a/lib/compiler_rt/trunctfhf2.zig
+++ b/lib/compiler_rt/trunctfhf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__trunctfhf2, .{ .name = "__trunctfhf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __trunctfhf2(a: f128) callconv(.C) common.F16T(f128) {
+pub fn __trunctfhf2(a: f128) callconv(.c) common.F16T(f128) {
return @bitCast(truncf(f16, f128, a));
}
diff --git a/lib/compiler_rt/trunctfsf2.zig b/lib/compiler_rt/trunctfsf2.zig
index 42228d4690..e238e35fb3 100644
--- a/lib/compiler_rt/trunctfsf2.zig
+++ b/lib/compiler_rt/trunctfsf2.zig
@@ -12,10 +12,10 @@ comptime {
@export(&__trunctfsf2, .{ .name = "__trunctfsf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __trunctfsf2(a: f128) callconv(.C) f32 {
+pub fn __trunctfsf2(a: f128) callconv(.c) f32 {
return truncf(f32, f128, a);
}
-fn _Qp_qtos(a: *const f128) callconv(.C) f32 {
+fn _Qp_qtos(a: *const f128) callconv(.c) f32 {
return truncf(f32, f128, a.*);
}
diff --git a/lib/compiler_rt/trunctfxf2.zig b/lib/compiler_rt/trunctfxf2.zig
index be78200a16..ddc977943a 100644
--- a/lib/compiler_rt/trunctfxf2.zig
+++ b/lib/compiler_rt/trunctfxf2.zig
@@ -8,7 +8,7 @@ comptime {
@export(&__trunctfxf2, .{ .name = "__trunctfxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __trunctfxf2(a: f128) callconv(.C) f80 {
+pub fn __trunctfxf2(a: f128) callconv(.c) f80 {
const src_sig_bits = math.floatMantissaBits(f128);
const dst_sig_bits = math.floatMantissaBits(f80) - 1; // -1 for the integer bit
diff --git a/lib/compiler_rt/truncxfdf2.zig b/lib/compiler_rt/truncxfdf2.zig
index 90170db6a9..f45a07962e 100644
--- a/lib/compiler_rt/truncxfdf2.zig
+++ b/lib/compiler_rt/truncxfdf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__truncxfdf2, .{ .name = "__truncxfdf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __truncxfdf2(a: f80) callconv(.C) f64 {
+fn __truncxfdf2(a: f80) callconv(.c) f64 {
return trunc_f80(f64, a);
}
diff --git a/lib/compiler_rt/truncxfhf2.zig b/lib/compiler_rt/truncxfhf2.zig
index 276de030f8..34ee97c04b 100644
--- a/lib/compiler_rt/truncxfhf2.zig
+++ b/lib/compiler_rt/truncxfhf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__truncxfhf2, .{ .name = "__truncxfhf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __truncxfhf2(a: f80) callconv(.C) common.F16T(f80) {
+fn __truncxfhf2(a: f80) callconv(.c) common.F16T(f80) {
return @bitCast(trunc_f80(f16, a));
}
diff --git a/lib/compiler_rt/truncxfsf2.zig b/lib/compiler_rt/truncxfsf2.zig
index 0a11e06655..136e941af4 100644
--- a/lib/compiler_rt/truncxfsf2.zig
+++ b/lib/compiler_rt/truncxfsf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__truncxfsf2, .{ .name = "__truncxfsf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __truncxfsf2(a: f80) callconv(.C) f32 {
+fn __truncxfsf2(a: f80) callconv(.c) f32 {
return trunc_f80(f32, a);
}
diff --git a/lib/compiler_rt/udivmodei4.zig b/lib/compiler_rt/udivmodei4.zig
index 0938ac859b..fa7f9746b1 100644
--- a/lib/compiler_rt/udivmodei4.zig
+++ b/lib/compiler_rt/udivmodei4.zig
@@ -112,7 +112,7 @@ pub fn divmod(q: ?[]u32, r: ?[]u32, u: []const u32, v: []const u32) !void {
}
}
-pub fn __udivei4(r_q: [*]u32, u_p: [*]const u32, v_p: [*]const u32, bits: usize) callconv(.C) void {
+pub fn __udivei4(r_q: [*]u32, u_p: [*]const u32, v_p: [*]const u32, bits: usize) callconv(.c) void {
@setRuntimeSafety(builtin.is_test);
const u = u_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
const v = v_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
@@ -120,7 +120,7 @@ pub fn __udivei4(r_q: [*]u32, u_p: [*]const u32, v_p: [*]const u32, bits: usize)
@call(.always_inline, divmod, .{ q, null, u, v }) catch unreachable;
}
-pub fn __umodei4(r_p: [*]u32, u_p: [*]const u32, v_p: [*]const u32, bits: usize) callconv(.C) void {
+pub fn __umodei4(r_p: [*]u32, u_p: [*]const u32, v_p: [*]const u32, bits: usize) callconv(.c) void {
@setRuntimeSafety(builtin.is_test);
const u = u_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
const v = v_p[0 .. std.math.divCeil(usize, bits, 32) catch unreachable];
diff --git a/lib/compiler_rt/udivmodti4.zig b/lib/compiler_rt/udivmodti4.zig
index 20cd916693..9e9585c635 100644
--- a/lib/compiler_rt/udivmodti4.zig
+++ b/lib/compiler_rt/udivmodti4.zig
@@ -13,13 +13,13 @@ comptime {
}
}
-pub fn __udivmodti4(a: u128, b: u128, maybe_rem: ?*u128) callconv(.C) u128 {
+pub fn __udivmodti4(a: u128, b: u128, maybe_rem: ?*u128) callconv(.c) u128 {
return udivmod(u128, a, b, maybe_rem);
}
const v2u64 = @Vector(2, u64);
-fn __udivmodti4_windows_x86_64(a: v2u64, b: v2u64, maybe_rem: ?*u128) callconv(.C) v2u64 {
+fn __udivmodti4_windows_x86_64(a: v2u64, b: v2u64, maybe_rem: ?*u128) callconv(.c) v2u64 {
return @bitCast(udivmod(u128, @bitCast(a), @bitCast(b), maybe_rem));
}
diff --git a/lib/compiler_rt/udivti3.zig b/lib/compiler_rt/udivti3.zig
index 0fb193352c..8423c1df18 100644
--- a/lib/compiler_rt/udivti3.zig
+++ b/lib/compiler_rt/udivti3.zig
@@ -13,12 +13,12 @@ comptime {
}
}
-pub fn __udivti3(a: u128, b: u128) callconv(.C) u128 {
+pub fn __udivti3(a: u128, b: u128) callconv(.c) u128 {
return udivmod(u128, a, b, null);
}
const v2u64 = @Vector(2, u64);
-fn __udivti3_windows_x86_64(a: v2u64, b: v2u64) callconv(.C) v2u64 {
+fn __udivti3_windows_x86_64(a: v2u64, b: v2u64) callconv(.c) v2u64 {
return @bitCast(udivmod(u128, @bitCast(a), @bitCast(b), null));
}
diff --git a/lib/compiler_rt/umodti3.zig b/lib/compiler_rt/umodti3.zig
index 9b17f60c6b..592bc44390 100644
--- a/lib/compiler_rt/umodti3.zig
+++ b/lib/compiler_rt/umodti3.zig
@@ -13,7 +13,7 @@ comptime {
}
}
-pub fn __umodti3(a: u128, b: u128) callconv(.C) u128 {
+pub fn __umodti3(a: u128, b: u128) callconv(.c) u128 {
var r: u128 = undefined;
_ = udivmod(u128, a, b, &r);
return r;
@@ -21,7 +21,7 @@ pub fn __umodti3(a: u128, b: u128) callconv(.C) u128 {
const v2u64 = @Vector(2, u64);
-fn __umodti3_windows_x86_64(a: v2u64, b: v2u64) callconv(.C) v2u64 {
+fn __umodti3_windows_x86_64(a: v2u64, b: v2u64) callconv(.c) v2u64 {
var r: u128 = undefined;
_ = udivmod(u128, @bitCast(a), @bitCast(b), &r);
return @bitCast(r);
diff --git a/lib/compiler_rt/unorddf2.zig b/lib/compiler_rt/unorddf2.zig
index 6f0a1344dd..dcc8762c66 100644
--- a/lib/compiler_rt/unorddf2.zig
+++ b/lib/compiler_rt/unorddf2.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __unorddf2(a: f64, b: f64) callconv(.C) i32 {
+pub fn __unorddf2(a: f64, b: f64) callconv(.c) i32 {
return comparef.unordcmp(f64, a, b);
}
-fn __aeabi_dcmpun(a: f64, b: f64) callconv(.AAPCS) i32 {
+fn __aeabi_dcmpun(a: f64, b: f64) callconv(.{ .arm_aapcs = .{} }) i32 {
return comparef.unordcmp(f64, a, b);
}
diff --git a/lib/compiler_rt/unordhf2.zig b/lib/compiler_rt/unordhf2.zig
index 0c2ad16c8e..2c2edad816 100644
--- a/lib/compiler_rt/unordhf2.zig
+++ b/lib/compiler_rt/unordhf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__unordhf2, .{ .name = "__unordhf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __unordhf2(a: f16, b: f16) callconv(.C) i32 {
+pub fn __unordhf2(a: f16, b: f16) callconv(.c) i32 {
return comparef.unordcmp(f16, a, b);
}
diff --git a/lib/compiler_rt/unordsf2.zig b/lib/compiler_rt/unordsf2.zig
index c3209fe5fb..a403fa4fd3 100644
--- a/lib/compiler_rt/unordsf2.zig
+++ b/lib/compiler_rt/unordsf2.zig
@@ -11,10 +11,10 @@ comptime {
}
}
-pub fn __unordsf2(a: f32, b: f32) callconv(.C) i32 {
+pub fn __unordsf2(a: f32, b: f32) callconv(.c) i32 {
return comparef.unordcmp(f32, a, b);
}
-fn __aeabi_fcmpun(a: f32, b: f32) callconv(.AAPCS) i32 {
+fn __aeabi_fcmpun(a: f32, b: f32) callconv(.{ .arm_aapcs = .{} }) i32 {
return comparef.unordcmp(f32, a, b);
}
diff --git a/lib/compiler_rt/unordtf2.zig b/lib/compiler_rt/unordtf2.zig
index 6e80ba3d4b..618f148296 100644
--- a/lib/compiler_rt/unordtf2.zig
+++ b/lib/compiler_rt/unordtf2.zig
@@ -13,6 +13,6 @@ comptime {
@export(&__unordtf2, .{ .name = "__unordtf2", .linkage = common.linkage, .visibility = common.visibility });
}
-fn __unordtf2(a: f128, b: f128) callconv(.C) i32 {
+fn __unordtf2(a: f128, b: f128) callconv(.c) i32 {
return comparef.unordcmp(f128, a, b);
}
diff --git a/lib/compiler_rt/unordxf2.zig b/lib/compiler_rt/unordxf2.zig
index c985f9cd98..97a4f8d06f 100644
--- a/lib/compiler_rt/unordxf2.zig
+++ b/lib/compiler_rt/unordxf2.zig
@@ -7,6 +7,6 @@ comptime {
@export(&__unordxf2, .{ .name = "__unordxf2", .linkage = common.linkage, .visibility = common.visibility });
}
-pub fn __unordxf2(a: f80, b: f80) callconv(.C) i32 {
+pub fn __unordxf2(a: f80, b: f80) callconv(.c) i32 {
return comparef.unordcmp(f80, a, b);
}
diff --git a/lib/fuzzer.zig b/lib/fuzzer.zig
index 0c287c6afc..92b621d07f 100644
--- a/lib/fuzzer.zig
+++ b/lib/fuzzer.zig
@@ -453,7 +453,7 @@ export fn fuzzer_coverage_id() u64 {
return fuzzer.coverage_id;
}
-var fuzzer_one: *const fn (input_ptr: [*]const u8, input_len: usize) callconv(.C) void = undefined;
+var fuzzer_one: *const fn (input_ptr: [*]const u8, input_len: usize) callconv(.c) void = undefined;
export fn fuzzer_start(testOne: @TypeOf(fuzzer_one)) void {
fuzzer_one = testOne;
diff --git a/lib/std/os/linux/aarch64.zig b/lib/std/os/linux/aarch64.zig
index 302c58ac2b..649e474310 100644
--- a/lib/std/os/linux/aarch64.zig
+++ b/lib/std/os/linux/aarch64.zig
@@ -99,7 +99,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// x0, x1, w2, x3, x4, x5, x6
//
@@ -141,7 +141,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
switch (@import("builtin").zig_backend) {
.stage2_c => asm volatile (
\\ mov x8, %[number]
diff --git a/lib/std/os/linux/arm.zig b/lib/std/os/linux/arm.zig
index e9263b5cea..10561da270 100644
--- a/lib/std/os/linux/arm.zig
+++ b/lib/std/os/linux/arm.zig
@@ -98,7 +98,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// r0, r1, r2, r3, +0, +4, +8
//
@@ -134,7 +134,7 @@ pub fn clone() callconv(.Naked) usize {
);
}
-pub fn restore() callconv(.Naked) noreturn {
+pub fn restore() callconv(.naked) noreturn {
switch (@import("builtin").zig_backend) {
.stage2_c => asm volatile (
\\ mov r7, %[number]
@@ -152,7 +152,7 @@ pub fn restore() callconv(.Naked) noreturn {
}
}
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
switch (@import("builtin").zig_backend) {
.stage2_c => asm volatile (
\\ mov r7, %[number]
diff --git a/lib/std/os/linux/hexagon.zig b/lib/std/os/linux/hexagon.zig
index 0b5cbb208e..b6549fe4a9 100644
--- a/lib/std/os/linux/hexagon.zig
+++ b/lib/std/os/linux/hexagon.zig
@@ -96,7 +96,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// r0, r1, r2, r3, r4, r5, +0
//
@@ -137,7 +137,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ trap0(#0)
:
diff --git a/lib/std/os/linux/loongarch64.zig b/lib/std/os/linux/loongarch64.zig
index ce57a6713a..160615a966 100644
--- a/lib/std/os/linux/loongarch64.zig
+++ b/lib/std/os/linux/loongarch64.zig
@@ -100,7 +100,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// a0, a1, a2, a3, a4, a5, a6
// sys_clone(flags, stack, ptid, ctid, tls)
@@ -140,7 +140,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ or $a7, $zero, %[number]
\\ syscall 0
diff --git a/lib/std/os/linux/mips.zig b/lib/std/os/linux/mips.zig
index 5ebe6bd686..b31d5e4fed 100644
--- a/lib/std/os/linux/mips.zig
+++ b/lib/std/os/linux/mips.zig
@@ -199,7 +199,7 @@ pub fn syscall7(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// 3, 4, 5, 6, 7, 8, 9
//
@@ -250,7 +250,7 @@ pub fn clone() callconv(.Naked) usize {
);
}
-pub fn restore() callconv(.Naked) noreturn {
+pub fn restore() callconv(.naked) noreturn {
asm volatile (
\\ syscall
:
@@ -259,7 +259,7 @@ pub fn restore() callconv(.Naked) noreturn {
);
}
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ syscall
:
diff --git a/lib/std/os/linux/mips64.zig b/lib/std/os/linux/mips64.zig
index 0d0a1104c1..393d71dffe 100644
--- a/lib/std/os/linux/mips64.zig
+++ b/lib/std/os/linux/mips64.zig
@@ -182,7 +182,7 @@ pub fn syscall7(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// 3, 4, 5, 6, 7, 8, 9
//
@@ -229,7 +229,7 @@ pub fn clone() callconv(.Naked) usize {
);
}
-pub fn restore() callconv(.Naked) noreturn {
+pub fn restore() callconv(.naked) noreturn {
asm volatile (
\\ syscall
:
@@ -238,7 +238,7 @@ pub fn restore() callconv(.Naked) noreturn {
);
}
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ syscall
:
diff --git a/lib/std/os/linux/powerpc.zig b/lib/std/os/linux/powerpc.zig
index 591575cb02..200744ce2d 100644
--- a/lib/std/os/linux/powerpc.zig
+++ b/lib/std/os/linux/powerpc.zig
@@ -127,7 +127,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// 3, 4, 5, 6, 7, 8, 9
//
@@ -199,7 +199,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ sc
:
diff --git a/lib/std/os/linux/powerpc64.zig b/lib/std/os/linux/powerpc64.zig
index 419c9f0629..bb1032fb21 100644
--- a/lib/std/os/linux/powerpc64.zig
+++ b/lib/std/os/linux/powerpc64.zig
@@ -127,7 +127,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// 3, 4, 5, 6, 7, 8, 9
//
@@ -184,7 +184,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ sc
:
diff --git a/lib/std/os/linux/riscv32.zig b/lib/std/os/linux/riscv32.zig
index 4df6e1671f..f9240ceafc 100644
--- a/lib/std/os/linux/riscv32.zig
+++ b/lib/std/os/linux/riscv32.zig
@@ -96,7 +96,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// a0, a1, a2, a3, a4, a5, a6
//
@@ -142,7 +142,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ ecall
:
diff --git a/lib/std/os/linux/riscv64.zig b/lib/std/os/linux/riscv64.zig
index c402e9e894..b803e5b581 100644
--- a/lib/std/os/linux/riscv64.zig
+++ b/lib/std/os/linux/riscv64.zig
@@ -96,7 +96,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// a0, a1, a2, a3, a4, a5, a6
//
@@ -142,7 +142,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ ecall
:
diff --git a/lib/std/os/linux/s390x.zig b/lib/std/os/linux/s390x.zig
index 337e36a574..70cf53f6fd 100644
--- a/lib/std/os/linux/s390x.zig
+++ b/lib/std/os/linux/s390x.zig
@@ -90,7 +90,7 @@ pub fn syscall6(number: SYS, arg1: usize, arg2: usize, arg3: usize, arg4: usize,
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
asm volatile (
\\# int clone(
\\# fn, a = r2
@@ -156,7 +156,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\svc 0
:
diff --git a/lib/std/os/linux/sparc64.zig b/lib/std/os/linux/sparc64.zig
index 90cb799e02..9d550bb582 100644
--- a/lib/std/os/linux/sparc64.zig
+++ b/lib/std/os/linux/sparc64.zig
@@ -179,7 +179,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// i0, i1, i2, i3, i4, i5, sp
//
diff --git a/lib/std/os/linux/thumb.zig b/lib/std/os/linux/thumb.zig
index a464030858..6816b47c26 100644
--- a/lib/std/os/linux/thumb.zig
+++ b/lib/std/os/linux/thumb.zig
@@ -143,7 +143,7 @@ pub fn syscall6(
pub const clone = @import("arm.zig").clone;
-pub fn restore() callconv(.Naked) noreturn {
+pub fn restore() callconv(.naked) noreturn {
asm volatile (
\\ mov r7, %[number]
\\ svc #0
@@ -152,7 +152,7 @@ pub fn restore() callconv(.Naked) noreturn {
);
}
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
asm volatile (
\\ mov r7, %[number]
\\ svc #0
diff --git a/lib/std/os/linux/x86.zig b/lib/std/os/linux/x86.zig
index 0cde6d805e..4b4c76a4ca 100644
--- a/lib/std/os/linux/x86.zig
+++ b/lib/std/os/linux/x86.zig
@@ -122,7 +122,7 @@ pub fn socketcall(call: usize, args: [*]const usize) usize {
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
// __clone(func, stack, flags, arg, ptid, tls, ctid)
// +8, +12, +16, +20, +24, +28, +32
//
@@ -172,7 +172,7 @@ pub fn clone() callconv(.Naked) usize {
);
}
-pub fn restore() callconv(.Naked) noreturn {
+pub fn restore() callconv(.naked) noreturn {
switch (@import("builtin").zig_backend) {
.stage2_c => asm volatile (
\\ movl %[number], %%eax
@@ -190,7 +190,7 @@ pub fn restore() callconv(.Naked) noreturn {
}
}
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
switch (@import("builtin").zig_backend) {
.stage2_c => asm volatile (
\\ movl %[number], %%eax
@@ -405,7 +405,7 @@ noinline fn getContextReturnAddress() usize {
return @returnAddress();
}
-pub fn getContextInternal() callconv(.Naked) usize {
+pub fn getContextInternal() callconv(.naked) usize {
asm volatile (
\\ movl $0, %[flags_offset:c](%%edx)
\\ movl $0, %[link_offset:c](%%edx)
diff --git a/lib/std/os/linux/x86_64.zig b/lib/std/os/linux/x86_64.zig
index 1c5105d624..90147876c8 100644
--- a/lib/std/os/linux/x86_64.zig
+++ b/lib/std/os/linux/x86_64.zig
@@ -101,7 +101,7 @@ pub fn syscall6(
);
}
-pub fn clone() callconv(.Naked) usize {
+pub fn clone() callconv(.naked) usize {
asm volatile (
\\ movl $56,%%eax // SYS_clone
\\ movq %%rdi,%%r11
@@ -137,7 +137,7 @@ pub fn clone() callconv(.Naked) usize {
pub const restore = restore_rt;
-pub fn restore_rt() callconv(.Naked) noreturn {
+pub fn restore_rt() callconv(.naked) noreturn {
switch (@import("builtin").zig_backend) {
.stage2_c => asm volatile (
\\ movl %[number], %%eax
@@ -382,7 +382,7 @@ fn gpRegisterOffset(comptime reg_index: comptime_int) usize {
return @offsetOf(ucontext_t, "mcontext") + @offsetOf(mcontext_t, "gregs") + @sizeOf(usize) * reg_index;
}
-fn getContextInternal() callconv(.Naked) usize {
+fn getContextInternal() callconv(.naked) usize {
// TODO: Read GS/FS registers?
asm volatile (
\\ movq $0, %[flags_offset:c](%%rdi)
diff --git a/lib/std/zig/parser_test.zig b/lib/std/zig/parser_test.zig
index e0fc4e1785..e5961621bd 100644
--- a/lib/std/zig/parser_test.zig
+++ b/lib/std/zig/parser_test.zig
@@ -1096,10 +1096,10 @@ test "zig fmt: block in slice expression" {
test "zig fmt: async function" {
try testCanonical(
\\pub const Server = struct {
- \\ handleRequestFn: fn (*Server, *const std.net.Address, File) callconv(.Async) void,
+ \\ handleRequestFn: fn (*Server, *const std.net.Address, File) callconv(.@"async") void,
\\};
\\test "hi" {
- \\ var ptr: fn (i32) callconv(.Async) void = @ptrCast(other);
+ \\ var ptr: fn (i32) callconv(.@"async") void = @ptrCast(other);
\\}
\\
);
@@ -1259,7 +1259,7 @@ test "zig fmt: threadlocal" {
test "zig fmt: linksection" {
try testCanonical(
\\export var aoeu: u64 linksection(".text.derp") = 1234;
- \\export fn _start() linksection(".text.boot") callconv(.Naked) noreturn {}
+ \\export fn _start() linksection(".text.boot") callconv(.naked) noreturn {}
\\
);
}
@@ -3926,7 +3926,7 @@ test "zig fmt: fn type" {
\\}
\\
\\const a: fn (u8) u8 = undefined;
- \\const b: fn (u8) callconv(.Naked) u8 = undefined;
+ \\const b: fn (u8) callconv(.naked) u8 = undefined;
\\const ap: fn (u8) u8 = a;
\\
);