Newer
Older
// SPDX-FileCopyrightText: 2021 Yoran Heling <projects@yorhel.nl>
// SPDX-License-Identifier: MIT
const std = @import("std");
const main = @import("main.zig");
usingnamespace @import("util.zig");
// While an arena allocator is optimimal for almost all scenarios in which ncdu
// is used, it doesn't allow for re-using deleted nodes after doing a delete or
// refresh operation, so a long-running ncdu session with regular refreshes
// will leak memory, but I'd say that's worth the efficiency gains.
// TODO: Can still implement a simple bucketed free list on top of this arena
// allocator to reuse nodes, if necessary.
var allocator = std.heap.ArenaAllocator.init(std.heap.page_allocator);
pub const EType = packed enum(u2) { dir, link, file };
// Type for the Entry.blocks field. Smaller than a u64 to make room for flags.
pub const Blocks = u60;
// (Ext +) Dir + name
// or: (Ext +) Link + name
// or: (Ext +) File + name
//
// Entry is always the first part of Dir, Link and File, so a pointer cast to
// *Entry is always safe and an *Entry can be casted to the full type. The Ext
// struct, if present, is placed before the *Entry pointer.
// These are all packed structs and hence do not have any alignment, which is
// great for saving memory but perhaps not very great for code size or
// performance.
// (TODO: What are the aliassing rules for Zig? There is a 'noalias' keyword,
// but does that mean all unmarked pointers are allowed to alias?)
pub const Entry = packed struct {
etype: EType,
isext: bool,
// Whether or not this entry's size has been counted in its parents.
// Counting of Link entries is deferred until the scan/delete operation has
// completed, so for those entries this flag indicates an intention to be
// counted.
counted: bool,
size: u64,
next: ?*Entry,
const Self = @This();
pub fn dir(self: *Self) ?*Dir {
return if (self.etype == .dir) @ptrCast(*Dir, self) else null;
}
pub fn link(self: *Self) ?*Link {
return if (self.etype == .link) @ptrCast(*Link, self) else null;
}
pub fn file(self: *Self) ?*File {
return if (self.etype == .file) @ptrCast(*File, self) else null;
}
// Whether this entry should be displayed as a "directory".
// Some dirs are actually represented in this data model as a File for efficiency.
pub fn isDirectory(self: *Self) bool {
return if (self.file()) |f| f.other_fs or f.kernfs else self.etype == .dir;
}
fn nameOffset(etype: EType) usize {
return switch (etype) {
.dir => @byteOffsetOf(Dir, "name"),
.link => @byteOffsetOf(Link, "name"),
.file => @byteOffsetOf(File, "name"),
};
}
pub fn name(self: *const Self) [:0]const u8 {
const ptr = @ptrCast([*:0]const u8, self) + nameOffset(self.etype);
return ptr[0..std.mem.lenZ(ptr) :0];
}
pub fn ext(self: *Self) ?*Ext {
if (!self.isext) return null;
pub fn create(etype: EType, isext: bool, ename: []const u8) *Entry {
const extsize = if (isext) @as(usize, @sizeOf(Ext)) else 0;
const size = nameOffset(etype) + ename.len + 1 + extsize;
if (allocator.allocator.allocWithOptions(u8, size, std.math.max(@alignOf(Ext), @alignOf(Entry)), null)) |p|
std.mem.set(u8, ptr, 0); // kind of ugly, but does the trick
e.etype = etype;
e.isext = isext;
std.mem.copy(u8, name_ptr[0..ename.len], ename);
return e;
}
// Set the 'err' flag on Dirs and Files, propagating 'suberr' to parents.
pub fn setErr(self: *Self, parent: *Dir) void {
if (self.dir()) |d| d.err = true
else if (self.file()) |f| f.err = true
else unreachable;
var it: ?*Dir = if (&parent.entry == self) parent.parent else parent;
while (it) |p| : (it = p.parent) {
if (p.suberr) break;
p.suberr = true;
}
}
pub fn addStats(self: *Entry, parent: *Dir, nlink: u31) void {
if (self.counted) return;
self.counted = true;
// Add link to the inode map, but don't count its size (yet).
if (self.link()) |l| {
l.parent = parent;
var d = inodes.map.getOrPut(l) catch unreachable;
if (!d.found_existing) {
d.value_ptr.* = .{ .counted = false, .nlink = nlink };
inodes.total_blocks = saturateAdd(inodes.total_blocks, self.blocks);
l.next = l;
} else {
inodes.setStats(.{ .key_ptr = d.key_ptr, .value_ptr = d.value_ptr }, false);
// If the nlink counts are not consistent, reset to 0 so we calculate with what we have instead.
if (d.value_ptr.nlink != nlink)
d.value_ptr.nlink = 0;
l.next = d.key_ptr.*.next;
d.key_ptr.*.next = l;
}
}
var it: ?*Dir = parent;
while(it) |p| : (it = p.parent) {
if (self.ext()) |e|
if (p.entry.ext()) |pe|
if (e.mtime > pe.mtime) { pe.mtime = e.mtime; };
p.items = saturateAdd(p.items, 1);
if (self.etype != .link) {
p.entry.size = saturateAdd(p.entry.size, self.size);
p.entry.blocks = saturateAdd(p.entry.blocks, self.blocks);
// Opposite of addStats(), but has some limitations:
// - If addStats() saturated adding sizes, then the sizes after delStats()
// will be incorrect.
// - mtime of parents is not adjusted (but that's a feature, possibly?)
//
// This function assumes that, for directories, all sub-entries have
// already been un-counted.
//
// When removing a Link, the entry's nlink counter is reset to zero, so
// that it will be recalculated based on our view of the tree. This means
// that links outside of the scanned directory will not be considered
// anymore, meaning that delStats() followed by addStats() with the same
// data may cause information to be lost.
pub fn delStats(self: *Entry, parent: *Dir) void {
defer self.counted = false; // defer, to make sure inodes.setStats() still sees it as counted.
if (self.link()) |l| {
var d = inodes.map.getEntry(l).?;
inodes.setStats(d, false);
d.value_ptr.nlink = 0;
if (l.next == l) {
_ = inodes.map.remove(l);
inodes.total_blocks = saturateSub(inodes.total_blocks, self.blocks);
} else {
if (d.key_ptr.* == l)
d.key_ptr.* = l.next;
// This is O(n), which in this context has the potential to
// slow ncdu down to a crawl. But this function is only called
// on refresh/delete operations and even then it's not common
// to have very long lists, so this blowing up should be very
// rare. This removal can also be deferred to setStats() to
// amortize the costs, if necessary.
var it = l.next;
while (it.next != l) it = it.next;
it.next = l.next;
}
}
var it: ?*Dir = parent;
while(it) |p| : (it = p.parent) {
if (self.etype != .link) {
p.entry.size = saturateSub(p.entry.size, self.size);
p.entry.blocks = saturateSub(p.entry.blocks, self.blocks);
}
}
}
pub fn delStatsRec(self: *Entry, parent: *Dir) void {
if (self.dir()) |d| {
var it = d.sub;
while (it) |e| : (it = e.next)
e.delStatsRec(d);
self.delStats(parent);
};
const DevId = u30; // Can be reduced to make room for more flags in Dir.
pub const Dir = packed struct {
entry: Entry,
sub: ?*Entry,
parent: ?*Dir,
// entry.{blocks,size}: Total size of all unique files + dirs. Non-shared hardlinks are counted only once.
// (i.e. the space you'll need if you created a filesystem with only this dir)
// shared_*: Unique hardlinks that still have references outside of this directory.
// (i.e. the space you won't reclaim by deleting this dir)
// (space reclaimed by deleting a dir =~ entry. - shared_)
shared_blocks: u64,
shared_size: u64,
// Indexes into the global 'devices.list' array
dev: DevId,
err: bool,
suberr: bool,
// Only used to find the @byteOffsetOff, the name is written at this point as a 0-terminated string.
// (Old C habits die hard)
name: u8,
pub fn fmtPath(self: *const @This(), withRoot: bool, out: *std.ArrayList(u8)) void {
if (!withRoot and self.parent == null) return;
var components = std.ArrayList([:0]const u8).init(main.allocator);
defer components.deinit();
var it: ?*const @This() = self;
while (it) |e| : (it = e.parent)
components.append(e.entry.name()) catch unreachable;
var i: usize = components.items.len-1;
while (true) {
if (i != components.items.len-1) out.append('/') catch unreachable;
out.appendSlice(components.items[i]) catch unreachable;
if (i == 0) break;
i -= 1;
}
}
};
// File that's been hardlinked (i.e. nlink > 1)
pub const Link = packed struct {
entry: Entry,
parent: *Dir,
next: *Link, // Singly circular linked list of all *Link nodes with the same dev,ino.
// dev is inherited from the parent Dir
// Return value should be freed with main.allocator.
pub fn path(self: @This(), withRoot: bool) [:0]const u8 {
var out = std.ArrayList(u8).init(main.allocator);
self.parent.fmtPath(withRoot, &out);
out.append('/') catch unreachable;
out.appendSlice(self.entry.name()) catch unreachable;
return out.toOwnedSliceSentinel(0) catch unreachable;
}
};
// Anything that's not an (indexed) directory or hardlink. Excluded directories are also "Files".
pub const File = packed struct {
entry: Entry,
err: bool,
excluded: bool,
other_fs: bool,
kernfs: bool,
notreg: bool,
_pad: u3,
name: u8,
pub fn resetFlags(f: *@This()) void {
f.err = false;
f.excluded = false;
f.other_fs = false;
f.kernfs = false;
f.notreg = false;
}
};
pub const Ext = packed struct {
mtime: u64 = 0,
uid: u32 = 0,
gid: u32 = 0,
mode: u16 = 0,
comptime {
std.debug.assert(@bitOffsetOf(Dir, "name") % 8 == 0);
std.debug.assert(@bitOffsetOf(Link, "name") % 8 == 0);
std.debug.assert(@bitOffsetOf(File, "name") % 8 == 0);
}
// List of st_dev entries. Those are typically 64bits, but that's quite a waste
// of space when a typical scan won't cover many unique devices.
// id -> dev
pub var list = std.ArrayList(u64).init(main.allocator);
var lookup = std.AutoHashMap(u64, DevId).init(main.allocator);
pub fn getId(dev: u64) DevId {
var d = lookup.getOrPut(dev) catch unreachable;
if (!d.found_existing) {
d.value_ptr.* = @intCast(DevId, list.items.len);
list.append(dev) catch unreachable;
// Lookup table for ino -> *Link entries, used for hard link counting.
pub const inodes = struct {
// Keys are hashed by their (dev,ino), the *Link points to an arbitrary
// node in the list. Link entries with the same dev/ino are part of a
// circular linked list, so you can iterate through all of them with this
// single pointer.
const Map = std.HashMap(*Link, Inode, HashContext, 80);
pub var map = Map.init(main.allocator);
// Cumulative size of all unique hard links in the map. This is a somewhat
// ugly workaround to provide accurate sizes during the initial scan, when
// the hard links are not counted as part of the parent directories yet.
pub var total_blocks: Blocks = 0;
const Inode = packed struct {
// Whether this Inode is counted towards the parent directories.
counted: bool,
// Number of links for this inode. When set to '0', we don't know the
// actual nlink count, either because it wasn't part of the imported
// JSON data or because we read inconsistent values from the
// filesystem. The count will then be updated by the actual number of
// links in our in-memory tree.
nlink: u31,
pub fn hash(_: @This(), l: *Link) u64 {
var h = std.hash.Wyhash.init(0);
h.update(std.mem.asBytes(&@as(u32, l.parent.dev)));
h.update(std.mem.asBytes(&l.ino));
return h.final();
}
pub fn eql(_: @This(), a: *Link, b: *Link) bool {
return a.ino == b.ino and a.parent.dev == b.parent.dev;
// Add/remove this inode from the parent Dir sizes. When removing stats,
// the list of *Links and their sizes and counts must be in the exact same
// state as when the stats were added. Hence, any modification to the Link
// state should be preceded by a setStats(.., false).
fn setStats(entry: Map.Entry, add: bool) void {
if (entry.value_ptr.counted == add) return;
entry.value_ptr.counted = add;
var nlink: u31 = 0;
var dirs = std.AutoHashMap(*Dir, u32).init(main.allocator);
defer dirs.deinit();
var it = entry.key_ptr.*;
while (true) {
if (it.entry.counted) {
nlink += 1;
var parent: ?*Dir = it.parent;
while (parent) |p| : (parent = p.parent) {
var de = dirs.getOrPut(p) catch unreachable;
if (de.found_existing) de.value_ptr.* += 1
else de.value_ptr.* = 1;
}
it = it.next;
if (it == entry.key_ptr.*)
break;
if (entry.value_ptr.nlink < nlink) entry.value_ptr.nlink = nlink
else nlink = entry.value_ptr.nlink;
var dir_iter = dirs.iterator();
if (add) {
while (dir_iter.next()) |de| {
de.key_ptr.*.entry.blocks = saturateAdd(de.key_ptr.*.entry.blocks, entry.key_ptr.*.entry.blocks);
de.key_ptr.*.entry.size = saturateAdd(de.key_ptr.*.entry.size, entry.key_ptr.*.entry.size);
if (de.value_ptr.* < nlink) {
de.key_ptr.*.shared_blocks = saturateAdd(de.key_ptr.*.shared_blocks, entry.key_ptr.*.entry.blocks);
de.key_ptr.*.shared_size = saturateAdd(de.key_ptr.*.shared_size, entry.key_ptr.*.entry.size);
}
}
} else {
while (dir_iter.next()) |de| {
de.key_ptr.*.entry.blocks = saturateSub(de.key_ptr.*.entry.blocks, entry.key_ptr.*.entry.blocks);
de.key_ptr.*.entry.size = saturateSub(de.key_ptr.*.entry.size, entry.key_ptr.*.entry.size);
if (de.value_ptr.* < nlink) {
de.key_ptr.*.shared_blocks = saturateSub(de.key_ptr.*.shared_blocks, entry.key_ptr.*.entry.blocks);
de.key_ptr.*.shared_size = saturateSub(de.key_ptr.*.shared_size, entry.key_ptr.*.entry.size);
}
// TODO: A version that doesn't have to iterate over the entire map, for
// smaller refresh/delete updates.
pub fn addAllStats() void {
var it = map.iterator();
while (it.next()) |e| setStats(e, true);
pub var root: *Dir = undefined;