aboutsummaryrefslogtreecommitdiff
path: root/lib/std
diff options
context:
space:
mode:
Diffstat (limited to 'lib/std')
-rw-r--r--lib/std/cache_hash.zig709
-rw-r--r--lib/std/fs.zig12
-rw-r--r--lib/std/fs/file.zig6
-rw-r--r--lib/std/os/linux.zig2
-rw-r--r--lib/std/std.zig1
5 files changed, 721 insertions, 9 deletions
diff --git a/lib/std/cache_hash.zig b/lib/std/cache_hash.zig
new file mode 100644
index 0000000000..d160c4ebb2
--- /dev/null
+++ b/lib/std/cache_hash.zig
@@ -0,0 +1,709 @@
+const std = @import("std.zig");
+const Blake3 = std.crypto.Blake3;
+const fs = std.fs;
+const base64 = std.base64;
+const ArrayList = std.ArrayList;
+const assert = std.debug.assert;
+const testing = std.testing;
+const mem = std.mem;
+const fmt = std.fmt;
+const Allocator = std.mem.Allocator;
+
+const base64_encoder = fs.base64_encoder;
+const base64_decoder = fs.base64_decoder;
+/// This is 70 more bits than UUIDs. For an analysis of probability of collisions, see:
+/// https://en.wikipedia.org/wiki/Universally_unique_identifier#Collisions
+const BIN_DIGEST_LEN = 24;
+const BASE64_DIGEST_LEN = base64.Base64Encoder.calcSize(BIN_DIGEST_LEN);
+
+const MANIFEST_FILE_SIZE_MAX = 50 * 1024 * 1024;
+
+pub const File = struct {
+ path: ?[]const u8,
+ max_file_size: ?usize,
+ stat: fs.File.Stat,
+ bin_digest: [BIN_DIGEST_LEN]u8,
+ contents: ?[]const u8,
+
+ pub fn deinit(self: *File, allocator: *Allocator) void {
+ if (self.path) |owned_slice| {
+ allocator.free(owned_slice);
+ self.path = null;
+ }
+ if (self.contents) |contents| {
+ allocator.free(contents);
+ self.contents = null;
+ }
+ self.* = undefined;
+ }
+};
+
+pub const CacheHash = struct {
+ allocator: *Allocator,
+ blake3: Blake3,
+ manifest_dir: fs.Dir,
+ manifest_file: ?fs.File,
+ manifest_dirty: bool,
+ files: ArrayList(File),
+ b64_digest: [BASE64_DIGEST_LEN]u8,
+
+ /// Be sure to call release after successful initialization.
+ pub fn init(allocator: *Allocator, dir: fs.Dir, manifest_dir_path: []const u8) !CacheHash {
+ return CacheHash{
+ .allocator = allocator,
+ .blake3 = Blake3.init(),
+ .manifest_dir = try dir.makeOpenPath(manifest_dir_path, .{}),
+ .manifest_file = null,
+ .manifest_dirty = false,
+ .files = ArrayList(File).init(allocator),
+ .b64_digest = undefined,
+ };
+ }
+
+ /// Record a slice of bytes as an dependency of the process being cached
+ pub fn addSlice(self: *CacheHash, val: []const u8) void {
+ assert(self.manifest_file == null);
+
+ self.blake3.update(val);
+ self.blake3.update(&[_]u8{0});
+ }
+
+ /// Convert the input value into bytes and record it as a dependency of the
+ /// process being cached
+ pub fn add(self: *CacheHash, val: var) void {
+ assert(self.manifest_file == null);
+
+ const valPtr = switch (@typeInfo(@TypeOf(val))) {
+ .Int => &val,
+ .Pointer => val,
+ else => &val,
+ };
+
+ self.addSlice(mem.asBytes(valPtr));
+ }
+
+ /// Add a file as a dependency of process being cached. When `CacheHash.hit` is
+ /// called, the file's contents will be checked to ensure that it matches
+ /// the contents from previous times.
+ ///
+ /// Max file size will be used to determine the amount of space to the file contents
+ /// are allowed to take up in memory. If max_file_size is null, then the contents
+ /// will not be loaded into memory.
+ ///
+ /// Returns the index of the entry in the `CacheHash.files` ArrayList. You can use it
+ /// to access the contents of the file after calling `CacheHash.hit()` like so:
+ ///
+ /// ```
+ /// var file_contents = cache_hash.files.items[file_index].contents.?;
+ /// ```
+ pub fn addFile(self: *CacheHash, file_path: []const u8, max_file_size: ?usize) !usize {
+ assert(self.manifest_file == null);
+
+ try self.files.ensureCapacity(self.files.items.len + 1);
+ const resolved_path = try fs.path.resolve(self.allocator, &[_][]const u8{file_path});
+
+ const idx = self.files.items.len;
+ self.files.addOneAssumeCapacity().* = .{
+ .path = resolved_path,
+ .contents = null,
+ .max_file_size = max_file_size,
+ .stat = undefined,
+ .bin_digest = undefined,
+ };
+
+ self.addSlice(resolved_path);
+
+ return idx;
+ }
+
+ /// Check the cache to see if the input exists in it. If it exists, a base64 encoding
+ /// of it's hash will be returned; otherwise, null will be returned.
+ ///
+ /// This function will also acquire an exclusive lock to the manifest file. This means
+ /// that a process holding a CacheHash will block any other process attempting to
+ /// acquire the lock.
+ ///
+ /// The lock on the manifest file is released when `CacheHash.release` is called.
+ pub fn hit(self: *CacheHash) !?[BASE64_DIGEST_LEN]u8 {
+ assert(self.manifest_file == null);
+
+ var bin_digest: [BIN_DIGEST_LEN]u8 = undefined;
+ self.blake3.final(&bin_digest);
+
+ base64_encoder.encode(self.b64_digest[0..], &bin_digest);
+
+ self.blake3 = Blake3.init();
+ self.blake3.update(&bin_digest);
+
+ const manifest_file_path = try fmt.allocPrint(self.allocator, "{}.txt", .{self.b64_digest});
+ defer self.allocator.free(manifest_file_path);
+
+ if (self.files.items.len != 0) {
+ self.manifest_file = try self.manifest_dir.createFile(manifest_file_path, .{
+ .read = true,
+ .truncate = false,
+ .lock = .Exclusive,
+ });
+ } else {
+ // If there are no file inputs, we check if the manifest file exists instead of
+ // comparing the hashes on the files used for the cached item
+ self.manifest_file = self.manifest_dir.openFile(manifest_file_path, .{
+ .read = true,
+ .write = true,
+ .lock = .Exclusive,
+ }) catch |err| switch (err) {
+ error.FileNotFound => {
+ self.manifest_dirty = true;
+ self.manifest_file = try self.manifest_dir.createFile(manifest_file_path, .{
+ .read = true,
+ .truncate = false,
+ .lock = .Exclusive,
+ });
+ return null;
+ },
+ else => |e| return e,
+ };
+ }
+
+ const file_contents = try self.manifest_file.?.inStream().readAllAlloc(self.allocator, MANIFEST_FILE_SIZE_MAX);
+ defer self.allocator.free(file_contents);
+
+ const input_file_count = self.files.items.len;
+ var any_file_changed = false;
+ var line_iter = mem.tokenize(file_contents, "\n");
+ var idx: usize = 0;
+ while (line_iter.next()) |line| {
+ defer idx += 1;
+
+ const cache_hash_file = if (idx < input_file_count) &self.files.items[idx] else blk: {
+ const new = try self.files.addOne();
+ new.* = .{
+ .path = null,
+ .contents = null,
+ .max_file_size = null,
+ .stat = undefined,
+ .bin_digest = undefined,
+ };
+ break :blk new;
+ };
+
+ var iter = mem.tokenize(line, " ");
+ const inode = iter.next() orelse return error.InvalidFormat;
+ const mtime_nsec_str = iter.next() orelse return error.InvalidFormat;
+ const digest_str = iter.next() orelse return error.InvalidFormat;
+ const file_path = iter.rest();
+
+ cache_hash_file.stat.inode = fmt.parseInt(fs.File.INode, mtime_nsec_str, 10) catch return error.InvalidFormat;
+ cache_hash_file.stat.mtime = fmt.parseInt(i64, mtime_nsec_str, 10) catch return error.InvalidFormat;
+ base64_decoder.decode(&cache_hash_file.bin_digest, digest_str) catch return error.InvalidFormat;
+
+ if (file_path.len == 0) {
+ return error.InvalidFormat;
+ }
+ if (cache_hash_file.path) |p| {
+ if (!mem.eql(u8, file_path, p)) {
+ return error.InvalidFormat;
+ }
+ }
+
+ if (cache_hash_file.path == null) {
+ cache_hash_file.path = try mem.dupe(self.allocator, u8, file_path);
+ }
+
+ const this_file = fs.cwd().openFile(cache_hash_file.path.?, .{ .read = true }) catch {
+ return error.CacheUnavailable;
+ };
+ defer this_file.close();
+
+ const actual_stat = try this_file.stat();
+ const mtime_match = actual_stat.mtime == cache_hash_file.stat.mtime;
+ const inode_match = actual_stat.inode == cache_hash_file.stat.inode;
+
+ if (!mtime_match or !inode_match) {
+ self.manifest_dirty = true;
+
+ cache_hash_file.stat = actual_stat;
+
+ if (isProblematicTimestamp(cache_hash_file.stat.mtime)) {
+ cache_hash_file.stat.mtime = 0;
+ cache_hash_file.stat.inode = 0;
+ }
+
+ var actual_digest: [BIN_DIGEST_LEN]u8 = undefined;
+ try hashFile(this_file, &actual_digest);
+
+ if (!mem.eql(u8, &cache_hash_file.bin_digest, &actual_digest)) {
+ cache_hash_file.bin_digest = actual_digest;
+ // keep going until we have the input file digests
+ any_file_changed = true;
+ }
+ }
+
+ if (!any_file_changed) {
+ self.blake3.update(&cache_hash_file.bin_digest);
+ }
+ }
+
+ if (any_file_changed) {
+ // cache miss
+ // keep the manifest file open
+ // reset the hash
+ self.blake3 = Blake3.init();
+ self.blake3.update(&bin_digest);
+
+ // Remove files not in the initial hash
+ for (self.files.items[input_file_count..]) |*file| {
+ file.deinit(self.allocator);
+ }
+ self.files.shrink(input_file_count);
+
+ for (self.files.items) |file| {
+ self.blake3.update(&file.bin_digest);
+ }
+ return null;
+ }
+
+ if (idx < input_file_count) {
+ self.manifest_dirty = true;
+ while (idx < input_file_count) : (idx += 1) {
+ const ch_file = &self.files.items[idx];
+ try self.populateFileHash(ch_file);
+ }
+ return null;
+ }
+
+ return self.final();
+ }
+
+ fn populateFileHash(self: *CacheHash, ch_file: *File) !void {
+ const file = try fs.cwd().openFile(ch_file.path.?, .{});
+ defer file.close();
+
+ ch_file.stat = try file.stat();
+
+ if (isProblematicTimestamp(ch_file.stat.mtime)) {
+ ch_file.stat.mtime = 0;
+ ch_file.stat.inode = 0;
+ }
+
+ if (ch_file.max_file_size) |max_file_size| {
+ if (ch_file.stat.size > max_file_size) {
+ return error.FileTooBig;
+ }
+
+ const contents = try self.allocator.alloc(u8, @intCast(usize, ch_file.stat.size));
+ errdefer self.allocator.free(contents);
+
+ // Hash while reading from disk, to keep the contents in the cpu cache while
+ // doing hashing.
+ var blake3 = Blake3.init();
+ var off: usize = 0;
+ while (true) {
+ // give me everything you've got, captain
+ const bytes_read = try file.read(contents[off..]);
+ if (bytes_read == 0) break;
+ blake3.update(contents[off..][0..bytes_read]);
+ off += bytes_read;
+ }
+ blake3.final(&ch_file.bin_digest);
+
+ ch_file.contents = contents;
+ } else {
+ try hashFile(file, &ch_file.bin_digest);
+ }
+
+ self.blake3.update(&ch_file.bin_digest);
+ }
+
+ /// Add a file as a dependency of process being cached, after the initial hash has been
+ /// calculated. This is useful for processes that don't know the all the files that
+ /// are depended on ahead of time. For example, a source file that can import other files
+ /// will need to be recompiled if the imported file is changed.
+ pub fn addFilePostFetch(self: *CacheHash, file_path: []const u8, max_file_size: usize) ![]u8 {
+ assert(self.manifest_file != null);
+
+ const resolved_path = try fs.path.resolve(self.allocator, &[_][]const u8{file_path});
+ errdefer self.allocator.free(resolved_path);
+
+ const new_ch_file = try self.files.addOne();
+ new_ch_file.* = .{
+ .path = resolved_path,
+ .max_file_size = max_file_size,
+ .stat = undefined,
+ .bin_digest = undefined,
+ .contents = null,
+ };
+ errdefer self.files.shrink(self.files.items.len - 1);
+
+ try self.populateFileHash(new_ch_file);
+
+ return new_ch_file.contents.?;
+ }
+
+ /// Add a file as a dependency of process being cached, after the initial hash has been
+ /// calculated. This is useful for processes that don't know the all the files that
+ /// are depended on ahead of time. For example, a source file that can import other files
+ /// will need to be recompiled if the imported file is changed.
+ pub fn addFilePost(self: *CacheHash, file_path: []const u8) !void {
+ assert(self.manifest_file != null);
+
+ const resolved_path = try fs.path.resolve(self.allocator, &[_][]const u8{file_path});
+ errdefer self.allocator.free(resolved_path);
+
+ const new_ch_file = try self.files.addOne();
+ new_ch_file.* = .{
+ .path = resolved_path,
+ .max_file_size = null,
+ .stat = undefined,
+ .bin_digest = undefined,
+ .contents = null,
+ };
+ errdefer self.files.shrink(self.files.items.len - 1);
+
+ try self.populateFileHash(new_ch_file);
+ }
+
+ /// Returns a base64 encoded hash of the inputs.
+ pub fn final(self: *CacheHash) [BASE64_DIGEST_LEN]u8 {
+ assert(self.manifest_file != null);
+
+ // We don't close the manifest file yet, because we want to
+ // keep it locked until the API user is done using it.
+ // We also don't write out the manifest yet, because until
+ // cache_release is called we still might be working on creating
+ // the artifacts to cache.
+
+ var bin_digest: [BIN_DIGEST_LEN]u8 = undefined;
+ self.blake3.final(&bin_digest);
+
+ var out_digest: [BASE64_DIGEST_LEN]u8 = undefined;
+ base64_encoder.encode(&out_digest, &bin_digest);
+
+ return out_digest;
+ }
+
+ pub fn writeManifest(self: *CacheHash) !void {
+ assert(self.manifest_file != null);
+
+ var encoded_digest: [BASE64_DIGEST_LEN]u8 = undefined;
+ var contents = ArrayList(u8).init(self.allocator);
+ var outStream = contents.outStream();
+ defer contents.deinit();
+
+ for (self.files.items) |file| {
+ base64_encoder.encode(encoded_digest[0..], &file.bin_digest);
+ try outStream.print("{} {} {} {}\n", .{ file.stat.inode, file.stat.mtime, encoded_digest[0..], file.path });
+ }
+
+ try self.manifest_file.?.pwriteAll(contents.items, 0);
+ self.manifest_dirty = false;
+ }
+
+ /// Releases the manifest file and frees any memory the CacheHash was using.
+ /// `CacheHash.hit` must be called first.
+ ///
+ /// Will also attempt to write to the manifest file if the manifest is dirty.
+ /// Writing to the manifest file can fail, but this function ignores those errors.
+ /// To detect failures from writing the manifest, one may explicitly call
+ /// `writeManifest` before `release`.
+ pub fn release(self: *CacheHash) void {
+ if (self.manifest_file) |file| {
+ if (self.manifest_dirty) {
+ // To handle these errors, API users should call
+ // writeManifest before release().
+ self.writeManifest() catch {};
+ }
+
+ file.close();
+ }
+
+ for (self.files.items) |*file| {
+ file.deinit(self.allocator);
+ }
+ self.files.deinit();
+ self.manifest_dir.close();
+ }
+};
+
+fn hashFile(file: fs.File, bin_digest: []u8) !void {
+ var blake3 = Blake3.init();
+ var buf: [1024]u8 = undefined;
+
+ while (true) {
+ const bytes_read = try file.read(&buf);
+ if (bytes_read == 0) break;
+ blake3.update(buf[0..bytes_read]);
+ }
+
+ blake3.final(bin_digest);
+}
+
+/// If the wall clock time, rounded to the same precision as the
+/// mtime, is equal to the mtime, then we cannot rely on this mtime
+/// yet. We will instead save an mtime value that indicates the hash
+/// must be unconditionally computed.
+/// This function recognizes the precision of mtime by looking at trailing
+/// zero bits of the seconds and nanoseconds.
+fn isProblematicTimestamp(fs_clock: i128) bool {
+ const wall_clock = std.time.nanoTimestamp();
+
+ // We have to break the nanoseconds into seconds and remainder nanoseconds
+ // to detect precision of seconds, because looking at the zero bits in base
+ // 2 would not detect precision of the seconds value.
+ const fs_sec = @intCast(i64, @divFloor(fs_clock, std.time.ns_per_s));
+ const fs_nsec = @intCast(i64, @mod(fs_clock, std.time.ns_per_s));
+ var wall_sec = @intCast(i64, @divFloor(wall_clock, std.time.ns_per_s));
+ var wall_nsec = @intCast(i64, @mod(wall_clock, std.time.ns_per_s));
+
+ // First make all the least significant zero bits in the fs_clock, also zero bits in the wall clock.
+ if (fs_nsec == 0) {
+ wall_nsec = 0;
+ if (fs_sec == 0) {
+ wall_sec = 0;
+ } else {
+ wall_sec &= @as(i64, -1) << @intCast(u6, @ctz(i64, fs_sec));
+ }
+ } else {
+ wall_nsec &= @as(i64, -1) << @intCast(u6, @ctz(i64, fs_nsec));
+ }
+ return wall_nsec == fs_nsec and wall_sec == fs_sec;
+}
+
+test "cache file and then recall it" {
+ if (std.Target.current.os.tag == .wasi) {
+ // https://github.com/ziglang/zig/issues/5437
+ return error.SkipZigTest;
+ }
+ const cwd = fs.cwd();
+
+ const temp_file = "test.txt";
+ const temp_manifest_dir = "temp_manifest_dir";
+
+ try cwd.writeFile(temp_file, "Hello, world!\n");
+
+ while (isProblematicTimestamp(std.time.nanoTimestamp())) {
+ std.time.sleep(1);
+ }
+
+ var digest1: [BASE64_DIGEST_LEN]u8 = undefined;
+ var digest2: [BASE64_DIGEST_LEN]u8 = undefined;
+
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add(true);
+ ch.add(@as(u16, 1234));
+ ch.add("1234");
+ _ = try ch.addFile(temp_file, null);
+
+ // There should be nothing in the cache
+ testing.expectEqual(@as(?[32]u8, null), try ch.hit());
+
+ digest1 = ch.final();
+ }
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add(true);
+ ch.add(@as(u16, 1234));
+ ch.add("1234");
+ _ = try ch.addFile(temp_file, null);
+
+ // Cache hit! We just "built" the same file
+ digest2 = (try ch.hit()).?;
+ }
+
+ testing.expectEqual(digest1, digest2);
+
+ try cwd.deleteTree(temp_manifest_dir);
+ try cwd.deleteFile(temp_file);
+}
+
+test "give problematic timestamp" {
+ var fs_clock = std.time.nanoTimestamp();
+ // to make it problematic, we make it only accurate to the second
+ fs_clock = @divTrunc(fs_clock, std.time.ns_per_s);
+ fs_clock *= std.time.ns_per_s;
+ testing.expect(isProblematicTimestamp(fs_clock));
+}
+
+test "give nonproblematic timestamp" {
+ testing.expect(!isProblematicTimestamp(std.time.nanoTimestamp() - std.time.ns_per_s));
+}
+
+test "check that changing a file makes cache fail" {
+ if (std.Target.current.os.tag == .wasi) {
+ // https://github.com/ziglang/zig/issues/5437
+ return error.SkipZigTest;
+ }
+ const cwd = fs.cwd();
+
+ const temp_file = "cache_hash_change_file_test.txt";
+ const temp_manifest_dir = "cache_hash_change_file_manifest_dir";
+ const original_temp_file_contents = "Hello, world!\n";
+ const updated_temp_file_contents = "Hello, world; but updated!\n";
+
+ try cwd.writeFile(temp_file, original_temp_file_contents);
+
+ while (isProblematicTimestamp(std.time.nanoTimestamp())) {
+ std.time.sleep(1);
+ }
+
+ var digest1: [BASE64_DIGEST_LEN]u8 = undefined;
+ var digest2: [BASE64_DIGEST_LEN]u8 = undefined;
+
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+ const temp_file_idx = try ch.addFile(temp_file, 100);
+
+ // There should be nothing in the cache
+ testing.expectEqual(@as(?[32]u8, null), try ch.hit());
+
+ testing.expect(mem.eql(u8, original_temp_file_contents, ch.files.items[temp_file_idx].contents.?));
+
+ digest1 = ch.final();
+ }
+
+ try cwd.writeFile(temp_file, updated_temp_file_contents);
+
+ while (isProblematicTimestamp(std.time.nanoTimestamp())) {
+ std.time.sleep(1);
+ }
+
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+ const temp_file_idx = try ch.addFile(temp_file, 100);
+
+ // A file that we depend on has been updated, so the cache should not contain an entry for it
+ testing.expectEqual(@as(?[32]u8, null), try ch.hit());
+
+ // The cache system does not keep the contents of re-hashed input files.
+ testing.expect(ch.files.items[temp_file_idx].contents == null);
+
+ digest2 = ch.final();
+ }
+
+ testing.expect(!mem.eql(u8, digest1[0..], digest2[0..]));
+
+ try cwd.deleteTree(temp_manifest_dir);
+ try cwd.deleteFile(temp_file);
+}
+
+test "no file inputs" {
+ if (std.Target.current.os.tag == .wasi) {
+ // https://github.com/ziglang/zig/issues/5437
+ return error.SkipZigTest;
+ }
+ const cwd = fs.cwd();
+ const temp_manifest_dir = "no_file_inputs_manifest_dir";
+ defer cwd.deleteTree(temp_manifest_dir) catch unreachable;
+
+ var digest1: [BASE64_DIGEST_LEN]u8 = undefined;
+ var digest2: [BASE64_DIGEST_LEN]u8 = undefined;
+
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+
+ // There should be nothing in the cache
+ testing.expectEqual(@as(?[32]u8, null), try ch.hit());
+
+ digest1 = ch.final();
+ }
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+
+ digest2 = (try ch.hit()).?;
+ }
+
+ testing.expectEqual(digest1, digest2);
+}
+
+test "CacheHashes with files added after initial hash work" {
+ if (std.Target.current.os.tag == .wasi) {
+ // https://github.com/ziglang/zig/issues/5437
+ return error.SkipZigTest;
+ }
+ const cwd = fs.cwd();
+
+ const temp_file1 = "cache_hash_post_file_test1.txt";
+ const temp_file2 = "cache_hash_post_file_test2.txt";
+ const temp_manifest_dir = "cache_hash_post_file_manifest_dir";
+
+ try cwd.writeFile(temp_file1, "Hello, world!\n");
+ try cwd.writeFile(temp_file2, "Hello world the second!\n");
+
+ while (isProblematicTimestamp(std.time.nanoTimestamp())) {
+ std.time.sleep(1);
+ }
+
+ var digest1: [BASE64_DIGEST_LEN]u8 = undefined;
+ var digest2: [BASE64_DIGEST_LEN]u8 = undefined;
+ var digest3: [BASE64_DIGEST_LEN]u8 = undefined;
+
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+ _ = try ch.addFile(temp_file1, null);
+
+ // There should be nothing in the cache
+ testing.expectEqual(@as(?[32]u8, null), try ch.hit());
+
+ _ = try ch.addFilePost(temp_file2);
+
+ digest1 = ch.final();
+ }
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+ _ = try ch.addFile(temp_file1, null);
+
+ digest2 = (try ch.hit()).?;
+ }
+ testing.expect(mem.eql(u8, &digest1, &digest2));
+
+ // Modify the file added after initial hash
+ try cwd.writeFile(temp_file2, "Hello world the second, updated\n");
+
+ while (isProblematicTimestamp(std.time.nanoTimestamp())) {
+ std.time.sleep(1);
+ }
+
+ {
+ var ch = try CacheHash.init(testing.allocator, cwd, temp_manifest_dir);
+ defer ch.release();
+
+ ch.add("1234");
+ _ = try ch.addFile(temp_file1, null);
+
+ // A file that we depend on has been updated, so the cache should not contain an entry for it
+ testing.expectEqual(@as(?[32]u8, null), try ch.hit());
+
+ _ = try ch.addFilePost(temp_file2);
+
+ digest3 = ch.final();
+ }
+
+ testing.expect(!mem.eql(u8, &digest1, &digest3));
+
+ try cwd.deleteTree(temp_manifest_dir);
+ try cwd.deleteFile(temp_file1);
+ try cwd.deleteFile(temp_file2);
+}
diff --git a/lib/std/fs.zig b/lib/std/fs.zig
index a006d4ea05..99383277a7 100644
--- a/lib/std/fs.zig
+++ b/lib/std/fs.zig
@@ -49,11 +49,13 @@ pub const MAX_PATH_BYTES = switch (builtin.os.tag) {
else => @compileError("Unsupported OS"),
};
-/// Base64, replacing the standard `+/` with `-_` so that it can be used in a file name on any filesystem.
-pub const base64_encoder = base64.Base64Encoder.init(
- "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-_",
- base64.standard_pad_char,
-);
+pub const base64_alphabet = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-_";
+
+/// Base64 encoder, replacing the standard `+/` with `-_` so that it can be used in a file name on any filesystem.
+pub const base64_encoder = base64.Base64Encoder.init(base64_alphabet, base64.standard_pad_char);
+
+/// Base64 decoder, replacing the standard `+/` with `-_` so that it can be used in a file name on any filesystem.
+pub const base64_decoder = base64.Base64Decoder.init(base64_alphabet, base64.standard_pad_char);
/// Whether or not async file system syscalls need a dedicated thread because the operating
/// system does not support non-blocking I/O on the file system.
diff --git a/lib/std/fs/file.zig b/lib/std/fs/file.zig
index a9ee996e4b..ce9546a9d9 100644
--- a/lib/std/fs/file.zig
+++ b/lib/std/fs/file.zig
@@ -27,6 +27,7 @@ pub const File = struct {
intended_io_mode: io.ModeOverride = io.default_mode,
pub const Mode = os.mode_t;
+ pub const INode = os.ino_t;
pub const default_mode = switch (builtin.os.tag) {
.windows => 0,
@@ -215,15 +216,14 @@ pub const File = struct {
pub const Stat = struct {
/// A number that the system uses to point to the file metadata. This number is not guaranteed to be
- /// unique across time, as some file systems may reuse an inode after it's file has been deleted.
+ /// unique across time, as some file systems may reuse an inode after its file has been deleted.
/// Some systems may change the inode of a file over time.
///
/// On Linux, the inode _is_ structure that stores the metadata, and the inode _number_ is what
/// you see here: the index number of the inode.
///
/// The FileIndex on Windows is similar. It is a number for a file that is unique to each filesystem.
- inode: os.ino_t,
-
+ inode: INode,
size: u64,
mode: Mode,
diff --git a/lib/std/os/linux.zig b/lib/std/os/linux.zig
index 6653293e59..f00238141a 100644
--- a/lib/std/os/linux.zig
+++ b/lib/std/os/linux.zig
@@ -417,7 +417,7 @@ pub fn ftruncate(fd: i32, length: u64) usize {
}
}
-pub fn pwrite(fd: i32, buf: [*]const u8, count: usize, offset: usize) usize {
+pub fn pwrite(fd: i32, buf: [*]const u8, count: usize, offset: u64) usize {
if (@hasField(SYS, "pwrite64")) {
if (require_aligned_register_pair) {
return syscall6(
diff --git a/lib/std/std.zig b/lib/std/std.zig
index 376c200200..9920ca3378 100644
--- a/lib/std/std.zig
+++ b/lib/std/std.zig
@@ -31,6 +31,7 @@ pub const base64 = @import("base64.zig");
pub const build = @import("build.zig");
pub const builtin = @import("builtin.zig");
pub const c = @import("c.zig");
+pub const cache_hash = @import("cache_hash.zig");
pub const coff = @import("coff.zig");
pub const crypto = @import("crypto.zig");
pub const cstr = @import("cstr.zig");