mirror of
https://github.com/NixOS/nix
synced 2025-06-25 06:31:14 +02:00
Squashed commit of the following: commit 04fff3a637d455cbb1d75937a235950e43008db9 Author: Eelco Dolstra <edolstra@gmail.com> Date: Thu Jun 12 12:30:32 2025 +0200 Chown structured attr files safely commit 5417ad445e414c649d0cfc71a05661c7bf8f3ef5 Author: Eelco Dolstra <edolstra@gmail.com> Date: Thu Jun 12 12:14:04 2025 +0200 Replace 'bool sync' with an enum for clarity And drop writeFileAndSync(). commit 7ae0141f328d8e8e1094be24665789c05f974ba6 Author: Eelco Dolstra <edolstra@gmail.com> Date: Thu Jun 12 11:35:28 2025 +0200 Drop guessOrInventPathFromFD() No need to do hacky stuff like that when we already know the original path. commit 45b05098bd019da7c57cd4227a89bfd0fa65bb08 Author: Eelco Dolstra <edolstra@gmail.com> Date: Thu Jun 12 11:15:58 2025 +0200 Tweak comment commit 0af15b31209d1b7ec8addfae9a1a6b60d8f35848 Author: Raito Bezarius <raito@lix.systems> Date: Thu Mar 27 12:22:26 2025 +0100 libstore: ensure that temporary directory is always 0o000 before deletion In the case the deletion fails, we should ensure that the temporary directory cannot be used for nefarious purposes. Change-Id: I498a2dd0999a74195d13642f44a5de1e69d46120 Signed-off-by: Raito Bezarius <raito@lix.systems> commit 2c20fa37b15cfa03ac6a1a6a47cdb2ed66c0827e Author: Raito Bezarius <raito@lix.systems> Date: Wed Mar 26 12:42:55 2025 +0100 libutil: ensure that `_deletePath` does NOT use absolute paths with dirfds When calling `_deletePath` with a parent file descriptor, `openat` is made effective by using relative paths to the directory file descriptor. To avoid the problem, the signature is changed to resist misuse with an assert in the prologue of the function. Change-Id: I6b3fc766bad2afe54dc27d47d1df3873e188de96 Signed-off-by: Raito Bezarius <raito@lix.systems> commit d3c370bbcae48bb825ce19fd0f73bb4eefd2c9ea Author: Raito Bezarius <raito@lix.systems> Date: Wed Mar 26 01:07:47 2025 +0100 libstore: ensure that `passAsFile` is created in the original temp dir This ensures that `passAsFile` data is created inside the expected temporary build directory by `openat()` from the parent directory file descriptor. This avoids a TOCTOU which is part of the attack chain of CVE-????. Change-Id: Ie5273446c4a19403088d0389ae8e3f473af8879a Signed-off-by: Raito Bezarius <raito@lix.systems> commit 45d3598724f932d024ef6bc2ffb00c1bb90e6018 Author: Raito Bezarius <raito@lix.systems> Date: Wed Mar 26 01:06:03 2025 +0100 libutil: writeFile variant for file descriptors `writeFile` lose its `sync` boolean flag to make things simpler. A new `writeFileAndSync` function is created and all call sites are converted to it. Change-Id: Ib871a5283a9c047db1e4fe48a241506e4aab9192 Signed-off-by: Raito Bezarius <raito@lix.systems> commit 732bd9b98cabf4aaf95a01fd318923de303f9996 Author: Raito Bezarius <raito@lix.systems> Date: Wed Mar 26 01:05:34 2025 +0100 libstore: chown to builder variant for file descriptors We use it immediately for the build temporary directory. Change-Id: I180193c63a2b98721f5fb8e542c4e39c099bb947 Signed-off-by: Raito Bezarius <raito@lix.systems> commit 962c65f8dcd5570dd92c72370a862c7b38942e0d Author: Raito Bezarius <raito@lix.systems> Date: Wed Mar 26 01:04:59 2025 +0100 libstore: open build directory as a dirfd as well We now keep around a proper AutoCloseFD around the temporary directory which we plan to use for openat operations and avoiding the build directory being swapped out while we are doing something else. Change-Id: I18d387b0f123ebf2d20c6405cd47ebadc5505f2a Signed-off-by: Raito Bezarius <raito@lix.systems> commit c9b42462b75b5a37ee6564c2b53cff186c8323da Author: Raito Bezarius <raito@lix.systems> Date: Wed Mar 26 01:04:12 2025 +0100 libutil: guess or invent a path from file descriptors This is useful for certain error recovery paths (no pun intended) that does not thread through the original path name. Change-Id: I2d800740cb4f9912e64c923120d3f977c58ccb7e Signed-off-by: Raito Bezarius <raito@lix.systems>
1717 lines
57 KiB
C++
1717 lines
57 KiB
C++
#include "local-store.hh"
|
||
#include "globals.hh"
|
||
#include "git.hh"
|
||
#include "archive.hh"
|
||
#include "pathlocks.hh"
|
||
#include "worker-protocol.hh"
|
||
#include "derivations.hh"
|
||
#include "realisation.hh"
|
||
#include "nar-info.hh"
|
||
#include "references.hh"
|
||
#include "callback.hh"
|
||
#include "topo-sort.hh"
|
||
#include "finally.hh"
|
||
#include "compression.hh"
|
||
#include "signals.hh"
|
||
#include "posix-fs-canonicalise.hh"
|
||
#include "posix-source-accessor.hh"
|
||
#include "keys.hh"
|
||
#include "users.hh"
|
||
|
||
#include <iostream>
|
||
#include <algorithm>
|
||
#include <cstring>
|
||
|
||
#include <memory>
|
||
#include <new>
|
||
#include <sys/types.h>
|
||
#include <sys/stat.h>
|
||
#include <sys/time.h>
|
||
#include <unistd.h>
|
||
#include <utime.h>
|
||
#include <fcntl.h>
|
||
#include <errno.h>
|
||
#include <stdio.h>
|
||
#include <time.h>
|
||
|
||
#ifndef _WIN32
|
||
# include <grp.h>
|
||
#endif
|
||
|
||
#if __linux__
|
||
# include <sched.h>
|
||
# include <sys/statvfs.h>
|
||
# include <sys/mount.h>
|
||
#endif
|
||
|
||
#ifdef __CYGWIN__
|
||
# include <windows.h>
|
||
#endif
|
||
|
||
#include <sqlite3.h>
|
||
|
||
#include <nlohmann/json.hpp>
|
||
|
||
#include "strings.hh"
|
||
|
||
|
||
namespace nix {
|
||
|
||
LocalStoreConfig::LocalStoreConfig(
|
||
std::string_view scheme,
|
||
std::string_view authority,
|
||
const Params & params)
|
||
: StoreConfig(params)
|
||
, LocalFSStoreConfig(authority, params)
|
||
{
|
||
}
|
||
|
||
std::string LocalStoreConfig::doc()
|
||
{
|
||
return
|
||
#include "local-store.md"
|
||
;
|
||
}
|
||
|
||
struct LocalStore::State::Stmts {
|
||
/* Some precompiled SQLite statements. */
|
||
SQLiteStmt RegisterValidPath;
|
||
SQLiteStmt UpdatePathInfo;
|
||
SQLiteStmt AddReference;
|
||
SQLiteStmt QueryPathInfo;
|
||
SQLiteStmt QueryReferences;
|
||
SQLiteStmt QueryReferrers;
|
||
SQLiteStmt InvalidatePath;
|
||
SQLiteStmt AddDerivationOutput;
|
||
SQLiteStmt RegisterRealisedOutput;
|
||
SQLiteStmt UpdateRealisedOutput;
|
||
SQLiteStmt QueryValidDerivers;
|
||
SQLiteStmt QueryDerivationOutputs;
|
||
SQLiteStmt QueryRealisedOutput;
|
||
SQLiteStmt QueryAllRealisedOutputs;
|
||
SQLiteStmt QueryPathFromHashPart;
|
||
SQLiteStmt QueryValidPaths;
|
||
SQLiteStmt QueryRealisationReferences;
|
||
SQLiteStmt AddRealisationReference;
|
||
};
|
||
|
||
LocalStore::LocalStore(
|
||
std::string_view scheme,
|
||
PathView path,
|
||
const Params & params)
|
||
: StoreConfig(params)
|
||
, LocalFSStoreConfig(path, params)
|
||
, LocalStoreConfig(scheme, path, params)
|
||
, Store(params)
|
||
, LocalFSStore(params)
|
||
, dbDir(stateDir + "/db")
|
||
, linksDir(realStoreDir + "/.links")
|
||
, reservedPath(dbDir + "/reserved")
|
||
, schemaPath(dbDir + "/schema")
|
||
, tempRootsDir(stateDir + "/temproots")
|
||
, fnTempRoots(fmt("%s/%d", tempRootsDir, getpid()))
|
||
, locksHeld(tokenizeString<PathSet>(getEnv("NIX_HELD_LOCKS").value_or("")))
|
||
{
|
||
auto state(_state.lock());
|
||
state->stmts = std::make_unique<State::Stmts>();
|
||
|
||
/* Create missing state directories if they don't already exist. */
|
||
createDirs(realStoreDir.get());
|
||
if (readOnly) {
|
||
experimentalFeatureSettings.require(Xp::ReadOnlyLocalStore);
|
||
} else {
|
||
makeStoreWritable();
|
||
}
|
||
createDirs(linksDir);
|
||
Path profilesDir = stateDir + "/profiles";
|
||
createDirs(profilesDir);
|
||
createDirs(tempRootsDir);
|
||
createDirs(dbDir);
|
||
Path gcRootsDir = stateDir + "/gcroots";
|
||
if (!pathExists(gcRootsDir)) {
|
||
createDirs(gcRootsDir);
|
||
createSymlink(profilesDir, gcRootsDir + "/profiles");
|
||
}
|
||
|
||
for (auto & perUserDir : {profilesDir + "/per-user", gcRootsDir + "/per-user"}) {
|
||
createDirs(perUserDir);
|
||
if (!readOnly) {
|
||
auto st = lstat(perUserDir);
|
||
|
||
// Skip chmod call if the directory already has the correct permissions (0755).
|
||
// This is to avoid failing when the executing user lacks permissions to change the directory's permissions
|
||
// even if it would be no-op.
|
||
if ((st.st_mode & (S_IRWXU | S_IRWXG | S_IRWXO)) != 0755 && chmod(perUserDir.c_str(), 0755) == -1)
|
||
throw SysError("could not set permissions on '%s' to 755", perUserDir);
|
||
}
|
||
}
|
||
|
||
#ifndef _WIN32
|
||
/* Optionally, create directories and set permissions for a
|
||
multi-user install. */
|
||
if (isRootUser() && settings.buildUsersGroup != "") {
|
||
mode_t perm = 01775;
|
||
|
||
struct group * gr = getgrnam(settings.buildUsersGroup.get().c_str());
|
||
if (!gr)
|
||
printError("warning: the group '%1%' specified in 'build-users-group' does not exist", settings.buildUsersGroup);
|
||
else if (!readOnly) {
|
||
struct stat st;
|
||
if (stat(realStoreDir.get().c_str(), &st))
|
||
throw SysError("getting attributes of path '%1%'", realStoreDir);
|
||
|
||
if (st.st_uid != 0 || st.st_gid != gr->gr_gid || (st.st_mode & ~S_IFMT) != perm) {
|
||
if (chown(realStoreDir.get().c_str(), 0, gr->gr_gid) == -1)
|
||
throw SysError("changing ownership of path '%1%'", realStoreDir);
|
||
if (chmod(realStoreDir.get().c_str(), perm) == -1)
|
||
throw SysError("changing permissions on path '%1%'", realStoreDir);
|
||
}
|
||
}
|
||
}
|
||
#endif
|
||
|
||
/* Ensure that the store and its parents are not symlinks. */
|
||
if (!settings.allowSymlinkedStore) {
|
||
Path path = realStoreDir;
|
||
struct stat st;
|
||
while (path != "/") {
|
||
st = lstat(path);
|
||
if (S_ISLNK(st.st_mode))
|
||
throw Error(
|
||
"the path '%1%' is a symlink; "
|
||
"this is not allowed for the Nix store and its parent directories",
|
||
path);
|
||
path = dirOf(path);
|
||
}
|
||
}
|
||
|
||
/* We can't open a SQLite database if the disk is full. Since
|
||
this prevents the garbage collector from running when it's most
|
||
needed, we reserve some dummy space that we can free just
|
||
before doing a garbage collection. */
|
||
try {
|
||
struct stat st;
|
||
if (stat(reservedPath.c_str(), &st) == -1 ||
|
||
st.st_size != settings.reservedSize)
|
||
{
|
||
AutoCloseFD fd = toDescriptor(open(reservedPath.c_str(), O_WRONLY | O_CREAT
|
||
#ifndef _WIN32
|
||
| O_CLOEXEC
|
||
#endif
|
||
, 0600));
|
||
int res = -1;
|
||
#if HAVE_POSIX_FALLOCATE
|
||
res = posix_fallocate(fd.get(), 0, settings.reservedSize);
|
||
#endif
|
||
if (res == -1) {
|
||
writeFull(fd.get(), std::string(settings.reservedSize, 'X'));
|
||
[[gnu::unused]] auto res2 =
|
||
|
||
#ifdef _WIN32
|
||
SetEndOfFile(fd.get())
|
||
#else
|
||
ftruncate(fd.get(), settings.reservedSize)
|
||
#endif
|
||
;
|
||
}
|
||
}
|
||
} catch (SystemError & e) { /* don't care about errors */
|
||
}
|
||
|
||
/* Acquire the big fat lock in shared mode to make sure that no
|
||
schema upgrade is in progress. */
|
||
if (!readOnly) {
|
||
Path globalLockPath = dbDir + "/big-lock";
|
||
globalLock = openLockFile(globalLockPath.c_str(), true);
|
||
}
|
||
|
||
if (!readOnly && !lockFile(globalLock.get(), ltRead, false)) {
|
||
printInfo("waiting for the big Nix store lock...");
|
||
lockFile(globalLock.get(), ltRead, true);
|
||
}
|
||
|
||
/* Check the current database schema and if necessary do an
|
||
upgrade. */
|
||
int curSchema = getSchema();
|
||
if (readOnly && curSchema < nixSchemaVersion) {
|
||
debug("current schema version: %d", curSchema);
|
||
debug("supported schema version: %d", nixSchemaVersion);
|
||
throw Error(curSchema == 0 ?
|
||
"database does not exist, and cannot be created in read-only mode" :
|
||
"database schema needs migrating, but this cannot be done in read-only mode");
|
||
}
|
||
|
||
if (curSchema > nixSchemaVersion)
|
||
throw Error("current Nix store schema is version %1%, but I only support %2%",
|
||
curSchema, nixSchemaVersion);
|
||
|
||
else if (curSchema == 0) { /* new store */
|
||
curSchema = nixSchemaVersion;
|
||
openDB(*state, true);
|
||
writeFile(schemaPath, fmt("%1%", curSchema), 0666, FsSync::Yes);
|
||
}
|
||
|
||
else if (curSchema < nixSchemaVersion) {
|
||
if (curSchema < 5)
|
||
throw Error(
|
||
"Your Nix store has a database in Berkeley DB format,\n"
|
||
"which is no longer supported. To convert to the new format,\n"
|
||
"please upgrade Nix to version 0.12 first.");
|
||
|
||
if (curSchema < 6)
|
||
throw Error(
|
||
"Your Nix store has a database in flat file format,\n"
|
||
"which is no longer supported. To convert to the new format,\n"
|
||
"please upgrade Nix to version 1.11 first.");
|
||
|
||
if (!lockFile(globalLock.get(), ltWrite, false)) {
|
||
printInfo("waiting for exclusive access to the Nix store...");
|
||
lockFile(globalLock.get(), ltNone, false); // We have acquired a shared lock; release it to prevent deadlocks
|
||
lockFile(globalLock.get(), ltWrite, true);
|
||
}
|
||
|
||
/* Get the schema version again, because another process may
|
||
have performed the upgrade already. */
|
||
curSchema = getSchema();
|
||
|
||
openDB(*state, false);
|
||
|
||
/* Legacy database schema migrations. Don't bump 'schema' for
|
||
new migrations; instead, add a migration to
|
||
upgradeDBSchema(). */
|
||
|
||
if (curSchema < 8) {
|
||
SQLiteTxn txn(state->db);
|
||
state->db.exec("alter table ValidPaths add column ultimate integer");
|
||
state->db.exec("alter table ValidPaths add column sigs text");
|
||
txn.commit();
|
||
}
|
||
|
||
if (curSchema < 9) {
|
||
SQLiteTxn txn(state->db);
|
||
state->db.exec("drop table FailedPaths");
|
||
txn.commit();
|
||
}
|
||
|
||
if (curSchema < 10) {
|
||
SQLiteTxn txn(state->db);
|
||
state->db.exec("alter table ValidPaths add column ca text");
|
||
txn.commit();
|
||
}
|
||
|
||
writeFile(schemaPath, fmt("%1%", nixSchemaVersion), 0666, FsSync::Yes);
|
||
|
||
lockFile(globalLock.get(), ltRead, true);
|
||
}
|
||
|
||
else openDB(*state, false);
|
||
|
||
upgradeDBSchema(*state);
|
||
|
||
/* Prepare SQL statements. */
|
||
state->stmts->RegisterValidPath.create(state->db,
|
||
"insert into ValidPaths (path, hash, registrationTime, deriver, narSize, ultimate, sigs, ca) values (?, ?, ?, ?, ?, ?, ?, ?);");
|
||
state->stmts->UpdatePathInfo.create(state->db,
|
||
"update ValidPaths set narSize = ?, hash = ?, ultimate = ?, sigs = ?, ca = ? where path = ?;");
|
||
state->stmts->AddReference.create(state->db,
|
||
"insert or replace into Refs (referrer, reference) values (?, ?);");
|
||
state->stmts->QueryPathInfo.create(state->db,
|
||
"select id, hash, registrationTime, deriver, narSize, ultimate, sigs, ca from ValidPaths where path = ?;");
|
||
state->stmts->QueryReferences.create(state->db,
|
||
"select path from Refs join ValidPaths on reference = id where referrer = ?;");
|
||
state->stmts->QueryReferrers.create(state->db,
|
||
"select path from Refs join ValidPaths on referrer = id where reference = (select id from ValidPaths where path = ?);");
|
||
state->stmts->InvalidatePath.create(state->db,
|
||
"delete from ValidPaths where path = ?;");
|
||
state->stmts->AddDerivationOutput.create(state->db,
|
||
"insert or replace into DerivationOutputs (drv, id, path) values (?, ?, ?);");
|
||
state->stmts->QueryValidDerivers.create(state->db,
|
||
"select v.id, v.path from DerivationOutputs d join ValidPaths v on d.drv = v.id where d.path = ?;");
|
||
state->stmts->QueryDerivationOutputs.create(state->db,
|
||
"select id, path from DerivationOutputs where drv = ?;");
|
||
// Use "path >= ?" with limit 1 rather than "path like '?%'" to
|
||
// ensure efficient lookup.
|
||
state->stmts->QueryPathFromHashPart.create(state->db,
|
||
"select path from ValidPaths where path >= ? limit 1;");
|
||
state->stmts->QueryValidPaths.create(state->db, "select path from ValidPaths");
|
||
if (experimentalFeatureSettings.isEnabled(Xp::CaDerivations)) {
|
||
state->stmts->RegisterRealisedOutput.create(state->db,
|
||
R"(
|
||
insert into Realisations (drvPath, outputName, outputPath, signatures)
|
||
values (?, ?, (select id from ValidPaths where path = ?), ?)
|
||
;
|
||
)");
|
||
state->stmts->UpdateRealisedOutput.create(state->db,
|
||
R"(
|
||
update Realisations
|
||
set signatures = ?
|
||
where
|
||
drvPath = ? and
|
||
outputName = ?
|
||
;
|
||
)");
|
||
state->stmts->QueryRealisedOutput.create(state->db,
|
||
R"(
|
||
select Realisations.id, Output.path, Realisations.signatures from Realisations
|
||
inner join ValidPaths as Output on Output.id = Realisations.outputPath
|
||
where drvPath = ? and outputName = ?
|
||
;
|
||
)");
|
||
state->stmts->QueryAllRealisedOutputs.create(state->db,
|
||
R"(
|
||
select outputName, Output.path from Realisations
|
||
inner join ValidPaths as Output on Output.id = Realisations.outputPath
|
||
where drvPath = ?
|
||
;
|
||
)");
|
||
state->stmts->QueryRealisationReferences.create(state->db,
|
||
R"(
|
||
select drvPath, outputName from Realisations
|
||
join RealisationsRefs on realisationReference = Realisations.id
|
||
where referrer = ?;
|
||
)");
|
||
state->stmts->AddRealisationReference.create(state->db,
|
||
R"(
|
||
insert or replace into RealisationsRefs (referrer, realisationReference)
|
||
values (
|
||
(select id from Realisations where drvPath = ? and outputName = ?),
|
||
(select id from Realisations where drvPath = ? and outputName = ?));
|
||
)");
|
||
}
|
||
}
|
||
|
||
|
||
LocalStore::LocalStore(const Params & params)
|
||
: LocalStore("local", "", params)
|
||
{
|
||
}
|
||
|
||
|
||
AutoCloseFD LocalStore::openGCLock()
|
||
{
|
||
Path fnGCLock = stateDir + "/gc.lock";
|
||
auto fdGCLock = open(fnGCLock.c_str(), O_RDWR | O_CREAT
|
||
#ifndef _WIN32
|
||
| O_CLOEXEC
|
||
#endif
|
||
, 0600);
|
||
if (!fdGCLock)
|
||
throw SysError("opening global GC lock '%1%'", fnGCLock);
|
||
return toDescriptor(fdGCLock);
|
||
}
|
||
|
||
|
||
void LocalStore::deleteStorePath(const Path & path, uint64_t & bytesFreed)
|
||
{
|
||
deletePath(path, bytesFreed);
|
||
}
|
||
|
||
|
||
LocalStore::~LocalStore()
|
||
{
|
||
std::shared_future<void> future;
|
||
|
||
{
|
||
auto state(_state.lock());
|
||
if (state->gcRunning)
|
||
future = state->gcFuture;
|
||
}
|
||
|
||
if (future.valid()) {
|
||
printInfo("waiting for auto-GC to finish on exit...");
|
||
future.get();
|
||
}
|
||
|
||
try {
|
||
auto fdTempRoots(_fdTempRoots.lock());
|
||
if (*fdTempRoots) {
|
||
fdTempRoots->close();
|
||
unlink(fnTempRoots.c_str());
|
||
}
|
||
} catch (...) {
|
||
ignoreExceptionInDestructor();
|
||
}
|
||
}
|
||
|
||
|
||
std::string LocalStore::getUri()
|
||
{
|
||
return "local";
|
||
}
|
||
|
||
|
||
int LocalStore::getSchema()
|
||
{
|
||
int curSchema = 0;
|
||
if (pathExists(schemaPath)) {
|
||
auto s = readFile(schemaPath);
|
||
auto n = string2Int<int>(s);
|
||
if (!n)
|
||
throw Error("'%1%' is corrupt", schemaPath);
|
||
curSchema = *n;
|
||
}
|
||
return curSchema;
|
||
}
|
||
|
||
void LocalStore::openDB(State & state, bool create)
|
||
{
|
||
if (create && readOnly) {
|
||
throw Error("cannot create database while in read-only mode");
|
||
}
|
||
|
||
if (access(dbDir.c_str(), R_OK | (readOnly ? 0 : W_OK)))
|
||
throw SysError("Nix database directory '%1%' is not writable", dbDir);
|
||
|
||
/* Open the Nix database. */
|
||
std::string dbPath = dbDir + "/db.sqlite";
|
||
auto & db(state.db);
|
||
auto openMode = readOnly ? SQLiteOpenMode::Immutable
|
||
: create ? SQLiteOpenMode::Normal
|
||
: SQLiteOpenMode::NoCreate;
|
||
state.db = SQLite(dbPath, openMode);
|
||
|
||
#ifdef __CYGWIN__
|
||
/* The cygwin version of sqlite3 has a patch which calls
|
||
SetDllDirectory("/usr/bin") on init. It was intended to fix extension
|
||
loading, which we don't use, and the effect of SetDllDirectory is
|
||
inherited by child processes, and causes libraries to be loaded from
|
||
/usr/bin instead of $PATH. This breaks quite a few things (e.g.
|
||
checkPhase on openssh), so we set it back to default behaviour. */
|
||
SetDllDirectoryW(L"");
|
||
#endif
|
||
|
||
/* !!! check whether sqlite has been built with foreign key
|
||
support */
|
||
|
||
/* Whether SQLite should fsync(). "Normal" synchronous mode
|
||
should be safe enough. If the user asks for it, don't sync at
|
||
all. This can cause database corruption if the system
|
||
crashes. */
|
||
std::string syncMode = settings.fsyncMetadata ? "normal" : "off";
|
||
db.exec("pragma synchronous = " + syncMode);
|
||
|
||
/* Set the SQLite journal mode. WAL mode is fastest, so it's the
|
||
default. */
|
||
std::string mode = settings.useSQLiteWAL ? "wal" : "truncate";
|
||
std::string prevMode;
|
||
{
|
||
SQLiteStmt stmt;
|
||
stmt.create(db, "pragma main.journal_mode;");
|
||
if (sqlite3_step(stmt) != SQLITE_ROW)
|
||
SQLiteError::throw_(db, "querying journal mode");
|
||
prevMode = std::string((const char *) sqlite3_column_text(stmt, 0));
|
||
}
|
||
if (prevMode != mode &&
|
||
sqlite3_exec(db, ("pragma main.journal_mode = " + mode + ";").c_str(), 0, 0, 0) != SQLITE_OK)
|
||
SQLiteError::throw_(db, "setting journal mode");
|
||
|
||
if (mode == "wal") {
|
||
/* persist the WAL files when the db connection is closed. This allows
|
||
for read-only connections without write permissions on the
|
||
containing directory to succeed on a closed db. Setting the
|
||
journal_size_limit to 2^40 bytes results in the WAL files getting
|
||
truncated to 0 on exit and limits the on disk size of the WAL files
|
||
to 2^40 bytes following a checkpoint */
|
||
if (sqlite3_exec(db, "pragma main.journal_size_limit = 1099511627776;", 0, 0, 0) == SQLITE_OK) {
|
||
int enable = 1;
|
||
sqlite3_file_control(db, NULL, SQLITE_FCNTL_PERSIST_WAL, &enable);
|
||
}
|
||
}
|
||
|
||
/* Increase the auto-checkpoint interval to 40000 pages. This
|
||
seems enough to ensure that instantiating the NixOS system
|
||
derivation is done in a single fsync(). */
|
||
if (mode == "wal" && sqlite3_exec(db, "pragma wal_autocheckpoint = 40000;", 0, 0, 0) != SQLITE_OK)
|
||
SQLiteError::throw_(db, "setting autocheckpoint interval");
|
||
|
||
/* Initialise the database schema, if necessary. */
|
||
if (create) {
|
||
static const char schema[] =
|
||
#include "schema.sql.gen.hh"
|
||
;
|
||
db.exec(schema);
|
||
}
|
||
}
|
||
|
||
|
||
void LocalStore::upgradeDBSchema(State & state)
|
||
{
|
||
state.db.exec("create table if not exists SchemaMigrations (migration text primary key not null);");
|
||
|
||
std::set<std::string> schemaMigrations;
|
||
|
||
{
|
||
SQLiteStmt querySchemaMigrations;
|
||
querySchemaMigrations.create(state.db, "select migration from SchemaMigrations;");
|
||
auto useQuerySchemaMigrations(querySchemaMigrations.use());
|
||
while (useQuerySchemaMigrations.next())
|
||
schemaMigrations.insert(useQuerySchemaMigrations.getStr(0));
|
||
}
|
||
|
||
auto doUpgrade = [&](const std::string & migrationName, const std::string & stmt)
|
||
{
|
||
if (schemaMigrations.contains(migrationName))
|
||
return;
|
||
|
||
debug("executing Nix database schema migration '%s'...", migrationName);
|
||
|
||
SQLiteTxn txn(state.db);
|
||
state.db.exec(stmt + fmt(";\ninsert into SchemaMigrations values('%s')", migrationName));
|
||
txn.commit();
|
||
|
||
schemaMigrations.insert(migrationName);
|
||
};
|
||
|
||
if (experimentalFeatureSettings.isEnabled(Xp::CaDerivations))
|
||
doUpgrade(
|
||
"20220326-ca-derivations",
|
||
#include "ca-specific-schema.sql.gen.hh"
|
||
);
|
||
}
|
||
|
||
|
||
/* To improve purity, users may want to make the Nix store a read-only
|
||
bind mount. So make the Nix store writable for this process. */
|
||
void LocalStore::makeStoreWritable()
|
||
{
|
||
#if __linux__
|
||
if (!isRootUser()) return;
|
||
/* Check if /nix/store is on a read-only mount. */
|
||
struct statvfs stat;
|
||
if (statvfs(realStoreDir.get().c_str(), &stat) != 0)
|
||
throw SysError("getting info about the Nix store mount point");
|
||
|
||
if (stat.f_flag & ST_RDONLY) {
|
||
if (mount(0, realStoreDir.get().c_str(), "none", MS_REMOUNT | MS_BIND, 0) == -1)
|
||
throw SysError("remounting %1% writable", realStoreDir);
|
||
}
|
||
#endif
|
||
}
|
||
|
||
|
||
void LocalStore::registerDrvOutput(const Realisation & info, CheckSigsFlag checkSigs)
|
||
{
|
||
experimentalFeatureSettings.require(Xp::CaDerivations);
|
||
if (checkSigs == NoCheckSigs || !realisationIsUntrusted(info))
|
||
registerDrvOutput(info);
|
||
else
|
||
throw Error("cannot register realisation '%s' because it lacks a signature by a trusted key", info.outPath.to_string());
|
||
}
|
||
|
||
void LocalStore::registerDrvOutput(const Realisation & info)
|
||
{
|
||
experimentalFeatureSettings.require(Xp::CaDerivations);
|
||
retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
if (auto oldR = queryRealisation_(*state, info.id)) {
|
||
if (info.isCompatibleWith(*oldR)) {
|
||
auto combinedSignatures = oldR->signatures;
|
||
combinedSignatures.insert(info.signatures.begin(),
|
||
info.signatures.end());
|
||
state->stmts->UpdateRealisedOutput.use()
|
||
(concatStringsSep(" ", combinedSignatures))
|
||
(info.id.strHash())
|
||
(info.id.outputName)
|
||
.exec();
|
||
} else {
|
||
throw Error("Trying to register a realisation of '%s', but we already "
|
||
"have another one locally.\n"
|
||
"Local: %s\n"
|
||
"Remote: %s",
|
||
info.id.to_string(),
|
||
printStorePath(oldR->outPath),
|
||
printStorePath(info.outPath)
|
||
);
|
||
}
|
||
} else {
|
||
state->stmts->RegisterRealisedOutput.use()
|
||
(info.id.strHash())
|
||
(info.id.outputName)
|
||
(printStorePath(info.outPath))
|
||
(concatStringsSep(" ", info.signatures))
|
||
.exec();
|
||
}
|
||
for (auto & [outputId, depPath] : info.dependentRealisations) {
|
||
auto localRealisation = queryRealisationCore_(*state, outputId);
|
||
if (!localRealisation)
|
||
throw Error("unable to register the derivation '%s' as it "
|
||
"depends on the non existent '%s'",
|
||
info.id.to_string(), outputId.to_string());
|
||
if (localRealisation->second.outPath != depPath)
|
||
throw Error("unable to register the derivation '%s' as it "
|
||
"depends on a realisation of '%s' that doesn’t"
|
||
"match what we have locally",
|
||
info.id.to_string(), outputId.to_string());
|
||
state->stmts->AddRealisationReference.use()
|
||
(info.id.strHash())
|
||
(info.id.outputName)
|
||
(outputId.strHash())
|
||
(outputId.outputName)
|
||
.exec();
|
||
}
|
||
});
|
||
}
|
||
|
||
void LocalStore::cacheDrvOutputMapping(
|
||
State & state,
|
||
const uint64_t deriver,
|
||
const std::string & outputName,
|
||
const StorePath & output)
|
||
{
|
||
retrySQLite<void>([&]() {
|
||
state.stmts->AddDerivationOutput.use()
|
||
(deriver)
|
||
(outputName)
|
||
(printStorePath(output))
|
||
.exec();
|
||
});
|
||
}
|
||
|
||
|
||
uint64_t LocalStore::addValidPath(State & state,
|
||
const ValidPathInfo & info, bool checkOutputs)
|
||
{
|
||
if (info.ca.has_value() && !info.isContentAddressed(*this))
|
||
throw Error("cannot add path '%s' to the Nix store because it claims to be content-addressed but isn't",
|
||
printStorePath(info.path));
|
||
|
||
state.stmts->RegisterValidPath.use()
|
||
(printStorePath(info.path))
|
||
(info.narHash.to_string(HashFormat::Base16, true))
|
||
(info.registrationTime == 0 ? time(0) : info.registrationTime)
|
||
(info.deriver ? printStorePath(*info.deriver) : "", (bool) info.deriver)
|
||
(info.narSize, info.narSize != 0)
|
||
(info.ultimate ? 1 : 0, info.ultimate)
|
||
(concatStringsSep(" ", info.sigs), !info.sigs.empty())
|
||
(renderContentAddress(info.ca), (bool) info.ca)
|
||
.exec();
|
||
uint64_t id = state.db.getLastInsertedRowId();
|
||
|
||
/* If this is a derivation, then store the derivation outputs in
|
||
the database. This is useful for the garbage collector: it can
|
||
efficiently query whether a path is an output of some
|
||
derivation. */
|
||
if (info.path.isDerivation()) {
|
||
auto drv = readInvalidDerivation(info.path);
|
||
|
||
/* Verify that the output paths in the derivation are correct
|
||
(i.e., follow the scheme for computing output paths from
|
||
derivations). Note that if this throws an error, then the
|
||
DB transaction is rolled back, so the path validity
|
||
registration above is undone. */
|
||
if (checkOutputs) drv.checkInvariants(*this, info.path);
|
||
|
||
for (auto & i : drv.outputsAndOptPaths(*this)) {
|
||
/* Floating CA derivations have indeterminate output paths until
|
||
they are built, so don't register anything in that case */
|
||
if (i.second.second)
|
||
cacheDrvOutputMapping(state, id, i.first, *i.second.second);
|
||
}
|
||
}
|
||
|
||
{
|
||
auto state_(Store::state.lock());
|
||
state_->pathInfoCache.upsert(std::string(info.path.to_string()),
|
||
PathInfoCacheValue{ .value = std::make_shared<const ValidPathInfo>(info) });
|
||
}
|
||
|
||
return id;
|
||
}
|
||
|
||
|
||
void LocalStore::queryPathInfoUncached(const StorePath & path,
|
||
Callback<std::shared_ptr<const ValidPathInfo>> callback) noexcept
|
||
{
|
||
try {
|
||
callback(retrySQLite<std::shared_ptr<const ValidPathInfo>>([&]() {
|
||
auto state(_state.lock());
|
||
return queryPathInfoInternal(*state, path);
|
||
}));
|
||
|
||
} catch (...) { callback.rethrow(); }
|
||
}
|
||
|
||
|
||
std::shared_ptr<const ValidPathInfo> LocalStore::queryPathInfoInternal(State & state, const StorePath & path)
|
||
{
|
||
/* Get the path info. */
|
||
auto useQueryPathInfo(state.stmts->QueryPathInfo.use()(printStorePath(path)));
|
||
|
||
if (!useQueryPathInfo.next())
|
||
return std::shared_ptr<ValidPathInfo>();
|
||
|
||
auto id = useQueryPathInfo.getInt(0);
|
||
|
||
auto narHash = Hash::dummy;
|
||
try {
|
||
narHash = Hash::parseAnyPrefixed(useQueryPathInfo.getStr(1));
|
||
} catch (BadHash & e) {
|
||
throw Error("invalid-path entry for '%s': %s", printStorePath(path), e.what());
|
||
}
|
||
|
||
auto info = std::make_shared<ValidPathInfo>(path, narHash);
|
||
|
||
info->id = id;
|
||
|
||
info->registrationTime = useQueryPathInfo.getInt(2);
|
||
|
||
auto s = (const char *) sqlite3_column_text(state.stmts->QueryPathInfo, 3);
|
||
if (s) info->deriver = parseStorePath(s);
|
||
|
||
/* Note that narSize = NULL yields 0. */
|
||
info->narSize = useQueryPathInfo.getInt(4);
|
||
|
||
info->ultimate = useQueryPathInfo.getInt(5) == 1;
|
||
|
||
s = (const char *) sqlite3_column_text(state.stmts->QueryPathInfo, 6);
|
||
if (s) info->sigs = tokenizeString<StringSet>(s, " ");
|
||
|
||
s = (const char *) sqlite3_column_text(state.stmts->QueryPathInfo, 7);
|
||
if (s) info->ca = ContentAddress::parseOpt(s);
|
||
|
||
/* Get the references. */
|
||
auto useQueryReferences(state.stmts->QueryReferences.use()(info->id));
|
||
|
||
while (useQueryReferences.next())
|
||
info->references.insert(parseStorePath(useQueryReferences.getStr(0)));
|
||
|
||
return info;
|
||
}
|
||
|
||
|
||
/* Update path info in the database. */
|
||
void LocalStore::updatePathInfo(State & state, const ValidPathInfo & info)
|
||
{
|
||
state.stmts->UpdatePathInfo.use()
|
||
(info.narSize, info.narSize != 0)
|
||
(info.narHash.to_string(HashFormat::Base16, true))
|
||
(info.ultimate ? 1 : 0, info.ultimate)
|
||
(concatStringsSep(" ", info.sigs), !info.sigs.empty())
|
||
(renderContentAddress(info.ca), (bool) info.ca)
|
||
(printStorePath(info.path))
|
||
.exec();
|
||
}
|
||
|
||
|
||
uint64_t LocalStore::queryValidPathId(State & state, const StorePath & path)
|
||
{
|
||
auto use(state.stmts->QueryPathInfo.use()(printStorePath(path)));
|
||
if (!use.next())
|
||
throw InvalidPath("path '%s' is not valid", printStorePath(path));
|
||
return use.getInt(0);
|
||
}
|
||
|
||
|
||
bool LocalStore::isValidPath_(State & state, const StorePath & path)
|
||
{
|
||
return state.stmts->QueryPathInfo.use()(printStorePath(path)).next();
|
||
}
|
||
|
||
|
||
bool LocalStore::isValidPathUncached(const StorePath & path)
|
||
{
|
||
return retrySQLite<bool>([&]() {
|
||
auto state(_state.lock());
|
||
return isValidPath_(*state, path);
|
||
});
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::queryValidPaths(const StorePathSet & paths, SubstituteFlag maybeSubstitute)
|
||
{
|
||
StorePathSet res;
|
||
for (auto & i : paths)
|
||
if (isValidPath(i)) res.insert(i);
|
||
return res;
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::queryAllValidPaths()
|
||
{
|
||
return retrySQLite<StorePathSet>([&]() {
|
||
auto state(_state.lock());
|
||
auto use(state->stmts->QueryValidPaths.use());
|
||
StorePathSet res;
|
||
while (use.next()) res.insert(parseStorePath(use.getStr(0)));
|
||
return res;
|
||
});
|
||
}
|
||
|
||
|
||
void LocalStore::queryReferrers(State & state, const StorePath & path, StorePathSet & referrers)
|
||
{
|
||
auto useQueryReferrers(state.stmts->QueryReferrers.use()(printStorePath(path)));
|
||
|
||
while (useQueryReferrers.next())
|
||
referrers.insert(parseStorePath(useQueryReferrers.getStr(0)));
|
||
}
|
||
|
||
|
||
void LocalStore::queryReferrers(const StorePath & path, StorePathSet & referrers)
|
||
{
|
||
return retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
queryReferrers(*state, path, referrers);
|
||
});
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::queryValidDerivers(const StorePath & path)
|
||
{
|
||
return retrySQLite<StorePathSet>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
auto useQueryValidDerivers(state->stmts->QueryValidDerivers.use()(printStorePath(path)));
|
||
|
||
StorePathSet derivers;
|
||
while (useQueryValidDerivers.next())
|
||
derivers.insert(parseStorePath(useQueryValidDerivers.getStr(1)));
|
||
|
||
return derivers;
|
||
});
|
||
}
|
||
|
||
|
||
std::map<std::string, std::optional<StorePath>>
|
||
LocalStore::queryStaticPartialDerivationOutputMap(const StorePath & path)
|
||
{
|
||
return retrySQLite<std::map<std::string, std::optional<StorePath>>>([&]() {
|
||
auto state(_state.lock());
|
||
std::map<std::string, std::optional<StorePath>> outputs;
|
||
uint64_t drvId;
|
||
drvId = queryValidPathId(*state, path);
|
||
auto use(state->stmts->QueryDerivationOutputs.use()(drvId));
|
||
while (use.next())
|
||
outputs.insert_or_assign(
|
||
use.getStr(0), parseStorePath(use.getStr(1)));
|
||
|
||
return outputs;
|
||
});
|
||
}
|
||
|
||
std::optional<StorePath> LocalStore::queryPathFromHashPart(const std::string & hashPart)
|
||
{
|
||
if (hashPart.size() != StorePath::HashLen) throw Error("invalid hash part");
|
||
|
||
Path prefix = storeDir + "/" + hashPart;
|
||
|
||
return retrySQLite<std::optional<StorePath>>([&]() -> std::optional<StorePath> {
|
||
auto state(_state.lock());
|
||
|
||
auto useQueryPathFromHashPart(state->stmts->QueryPathFromHashPart.use()(prefix));
|
||
|
||
if (!useQueryPathFromHashPart.next()) return {};
|
||
|
||
const char * s = (const char *) sqlite3_column_text(state->stmts->QueryPathFromHashPart, 0);
|
||
if (s && prefix.compare(0, prefix.size(), s, prefix.size()) == 0)
|
||
return parseStorePath(s);
|
||
return {};
|
||
});
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::querySubstitutablePaths(const StorePathSet & paths)
|
||
{
|
||
if (!settings.useSubstitutes) return StorePathSet();
|
||
|
||
StorePathSet remaining;
|
||
for (auto & i : paths)
|
||
remaining.insert(i);
|
||
|
||
StorePathSet res;
|
||
|
||
for (auto & sub : getDefaultSubstituters()) {
|
||
if (remaining.empty()) break;
|
||
if (sub->storeDir != storeDir) continue;
|
||
if (!sub->wantMassQuery) continue;
|
||
|
||
auto valid = sub->queryValidPaths(remaining);
|
||
|
||
StorePathSet remaining2;
|
||
for (auto & path : remaining)
|
||
if (valid.count(path))
|
||
res.insert(path);
|
||
else
|
||
remaining2.insert(path);
|
||
|
||
std::swap(remaining, remaining2);
|
||
}
|
||
|
||
return res;
|
||
}
|
||
|
||
|
||
void LocalStore::registerValidPath(const ValidPathInfo & info)
|
||
{
|
||
registerValidPaths({{info.path, info}});
|
||
}
|
||
|
||
|
||
void LocalStore::registerValidPaths(const ValidPathInfos & infos)
|
||
{
|
||
#ifndef _WIN32
|
||
/* SQLite will fsync by default, but the new valid paths may not
|
||
be fsync-ed. So some may want to fsync them before registering
|
||
the validity, at the expense of some speed of the path
|
||
registering operation. */
|
||
if (settings.syncBeforeRegistering) sync();
|
||
#endif
|
||
|
||
return retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
SQLiteTxn txn(state->db);
|
||
StorePathSet paths;
|
||
|
||
for (auto & [_, i] : infos) {
|
||
assert(i.narHash.algo == HashAlgorithm::SHA256);
|
||
if (isValidPath_(*state, i.path))
|
||
updatePathInfo(*state, i);
|
||
else
|
||
addValidPath(*state, i, false);
|
||
paths.insert(i.path);
|
||
}
|
||
|
||
for (auto & [_, i] : infos) {
|
||
auto referrer = queryValidPathId(*state, i.path);
|
||
for (auto & j : i.references)
|
||
state->stmts->AddReference.use()(referrer)(queryValidPathId(*state, j)).exec();
|
||
}
|
||
|
||
/* Check that the derivation outputs are correct. We can't do
|
||
this in addValidPath() above, because the references might
|
||
not be valid yet. */
|
||
for (auto & [_, i] : infos)
|
||
if (i.path.isDerivation()) {
|
||
// FIXME: inefficient; we already loaded the derivation in addValidPath().
|
||
readInvalidDerivation(i.path).checkInvariants(*this, i.path);
|
||
}
|
||
|
||
/* Do a topological sort of the paths. This will throw an
|
||
error if a cycle is detected and roll back the
|
||
transaction. Cycles can only occur when a derivation
|
||
has multiple outputs. */
|
||
topoSort(paths,
|
||
{[&](const StorePath & path) {
|
||
auto i = infos.find(path);
|
||
return i == infos.end() ? StorePathSet() : i->second.references;
|
||
}},
|
||
{[&](const StorePath & path, const StorePath & parent) {
|
||
return BuildError(
|
||
"cycle detected in the references of '%s' from '%s'",
|
||
printStorePath(path),
|
||
printStorePath(parent));
|
||
}});
|
||
|
||
txn.commit();
|
||
});
|
||
}
|
||
|
||
|
||
/* Invalidate a path. The caller is responsible for checking that
|
||
there are no referrers. */
|
||
void LocalStore::invalidatePath(State & state, const StorePath & path)
|
||
{
|
||
debug("invalidating path '%s'", printStorePath(path));
|
||
|
||
state.stmts->InvalidatePath.use()(printStorePath(path)).exec();
|
||
|
||
/* Note that the foreign key constraints on the Refs table take
|
||
care of deleting the references entries for `path'. */
|
||
|
||
{
|
||
auto state_(Store::state.lock());
|
||
state_->pathInfoCache.erase(std::string(path.to_string()));
|
||
}
|
||
}
|
||
|
||
const PublicKeys & LocalStore::getPublicKeys()
|
||
{
|
||
auto state(_state.lock());
|
||
if (!state->publicKeys)
|
||
state->publicKeys = std::make_unique<PublicKeys>(getDefaultPublicKeys());
|
||
return *state->publicKeys;
|
||
}
|
||
|
||
bool LocalStore::pathInfoIsUntrusted(const ValidPathInfo & info)
|
||
{
|
||
return requireSigs && !info.checkSignatures(*this, getPublicKeys());
|
||
}
|
||
|
||
bool LocalStore::realisationIsUntrusted(const Realisation & realisation)
|
||
{
|
||
return requireSigs && !realisation.checkSignatures(getPublicKeys());
|
||
}
|
||
|
||
void LocalStore::addToStore(const ValidPathInfo & info, Source & source,
|
||
RepairFlag repair, CheckSigsFlag checkSigs)
|
||
{
|
||
if (checkSigs && pathInfoIsUntrusted(info))
|
||
throw Error("cannot add path '%s' because it lacks a signature by a trusted key", printStorePath(info.path));
|
||
|
||
{
|
||
/* In case we are not interested in reading the NAR: discard it. */
|
||
bool narRead = false;
|
||
Finally cleanup = [&]() {
|
||
if (!narRead) {
|
||
NullFileSystemObjectSink sink;
|
||
try {
|
||
parseDump(sink, source);
|
||
} catch (...) {
|
||
// TODO: should Interrupted be handled here?
|
||
ignoreExceptionInDestructor();
|
||
}
|
||
}
|
||
};
|
||
|
||
addTempRoot(info.path);
|
||
|
||
if (repair || !isValidPath(info.path)) {
|
||
|
||
PathLocks outputLock;
|
||
|
||
auto realPath = Store::toRealPath(info.path);
|
||
|
||
/* Lock the output path. But don't lock if we're being called
|
||
from a build hook (whose parent process already acquired a
|
||
lock on this path). */
|
||
if (!locksHeld.count(printStorePath(info.path)))
|
||
outputLock.lockPaths({realPath});
|
||
|
||
if (repair || !isValidPath(info.path)) {
|
||
|
||
deletePath(realPath);
|
||
|
||
/* While restoring the path from the NAR, compute the hash
|
||
of the NAR. */
|
||
HashSink hashSink(HashAlgorithm::SHA256);
|
||
|
||
TeeSource wrapperSource { source, hashSink };
|
||
|
||
narRead = true;
|
||
restorePath(realPath, wrapperSource, settings.fsyncStorePaths);
|
||
|
||
auto hashResult = hashSink.finish();
|
||
|
||
if (hashResult.first != info.narHash)
|
||
throw Error("hash mismatch importing path '%s';\n specified: %s\n got: %s",
|
||
printStorePath(info.path), info.narHash.to_string(HashFormat::Nix32, true), hashResult.first.to_string(HashFormat::Nix32, true));
|
||
|
||
if (hashResult.second != info.narSize)
|
||
throw Error("size mismatch importing path '%s';\n specified: %s\n got: %s",
|
||
printStorePath(info.path), info.narSize, hashResult.second);
|
||
|
||
if (info.ca) {
|
||
auto & specified = *info.ca;
|
||
auto actualHash = ({
|
||
auto accessor = getFSAccessor(false);
|
||
CanonPath path { printStorePath(info.path) };
|
||
Hash h { HashAlgorithm::SHA256 }; // throwaway def to appease C++
|
||
auto fim = specified.method.getFileIngestionMethod();
|
||
switch (fim) {
|
||
case FileIngestionMethod::Flat:
|
||
case FileIngestionMethod::NixArchive:
|
||
{
|
||
HashModuloSink caSink {
|
||
specified.hash.algo,
|
||
std::string { info.path.hashPart() },
|
||
};
|
||
dumpPath({accessor, path}, caSink, (FileSerialisationMethod) fim);
|
||
h = caSink.finish().first;
|
||
break;
|
||
}
|
||
case FileIngestionMethod::Git:
|
||
h = git::dumpHash(specified.hash.algo, {accessor, path}).hash;
|
||
break;
|
||
}
|
||
ContentAddress {
|
||
.method = specified.method,
|
||
.hash = std::move(h),
|
||
};
|
||
});
|
||
if (specified.hash != actualHash.hash) {
|
||
throw Error("ca hash mismatch importing path '%s';\n specified: %s\n got: %s",
|
||
printStorePath(info.path),
|
||
specified.hash.to_string(HashFormat::Nix32, true),
|
||
actualHash.hash.to_string(HashFormat::Nix32, true));
|
||
}
|
||
}
|
||
|
||
autoGC();
|
||
|
||
canonicalisePathMetaData(realPath);
|
||
|
||
optimisePath(realPath, repair); // FIXME: combine with hashPath()
|
||
|
||
if (settings.fsyncStorePaths) {
|
||
recursiveSync(realPath);
|
||
syncParent(realPath);
|
||
}
|
||
|
||
registerValidPath(info);
|
||
}
|
||
|
||
outputLock.setDeletion(true);
|
||
}
|
||
}
|
||
|
||
// In case `cleanup` ignored an `Interrupted` exception
|
||
checkInterrupt();
|
||
}
|
||
|
||
|
||
StorePath LocalStore::addToStoreFromDump(
|
||
Source & source0,
|
||
std::string_view name,
|
||
FileSerialisationMethod dumpMethod,
|
||
ContentAddressMethod hashMethod,
|
||
HashAlgorithm hashAlgo,
|
||
const StorePathSet & references,
|
||
RepairFlag repair)
|
||
{
|
||
/* For computing the store path. */
|
||
auto hashSink = std::make_unique<HashSink>(hashAlgo);
|
||
TeeSource source { source0, *hashSink };
|
||
|
||
/* Read the source path into memory, but only if it's up to
|
||
narBufferSize bytes. If it's larger, write it to a temporary
|
||
location in the Nix store. If the subsequently computed
|
||
destination store path is already valid, we just delete the
|
||
temporary path. Otherwise, we move it to the destination store
|
||
path. */
|
||
bool inMemory = false;
|
||
|
||
struct Free {
|
||
void operator()(void* v) { free(v); }
|
||
};
|
||
std::unique_ptr<char, Free> dumpBuffer(nullptr);
|
||
std::string_view dump;
|
||
|
||
/* Fill out buffer, and decide whether we are working strictly in
|
||
memory based on whether we break out because the buffer is full
|
||
or the original source is empty */
|
||
while (dump.size() < settings.narBufferSize) {
|
||
auto oldSize = dump.size();
|
||
constexpr size_t chunkSize = 65536;
|
||
auto want = std::min(chunkSize, settings.narBufferSize - oldSize);
|
||
if (auto tmp = realloc(dumpBuffer.get(), oldSize + want)) {
|
||
dumpBuffer.release();
|
||
dumpBuffer.reset((char*) tmp);
|
||
} else {
|
||
throw std::bad_alloc();
|
||
}
|
||
auto got = 0;
|
||
Finally cleanup([&]() {
|
||
dump = {dumpBuffer.get(), dump.size() + got};
|
||
});
|
||
try {
|
||
got = source.read(dumpBuffer.get() + oldSize, want);
|
||
} catch (EndOfFile &) {
|
||
inMemory = true;
|
||
break;
|
||
}
|
||
}
|
||
|
||
std::unique_ptr<AutoDelete> delTempDir;
|
||
std::filesystem::path tempPath;
|
||
std::filesystem::path tempDir;
|
||
AutoCloseFD tempDirFd;
|
||
|
||
bool methodsMatch = static_cast<FileIngestionMethod>(dumpMethod) == hashMethod.getFileIngestionMethod();
|
||
|
||
/* If the methods don't match, our streaming hash of the dump is the
|
||
wrong sort, and we need to rehash. */
|
||
bool inMemoryAndDontNeedRestore = inMemory && methodsMatch;
|
||
|
||
if (!inMemoryAndDontNeedRestore) {
|
||
/* Drain what we pulled so far, and then keep on pulling */
|
||
StringSource dumpSource { dump };
|
||
ChainSource bothSource { dumpSource, source };
|
||
|
||
std::tie(tempDir, tempDirFd) = createTempDirInStore();
|
||
delTempDir = std::make_unique<AutoDelete>(tempDir);
|
||
tempPath = tempDir / "x";
|
||
|
||
restorePath(tempPath.string(), bothSource, dumpMethod, settings.fsyncStorePaths);
|
||
|
||
dumpBuffer.reset();
|
||
dump = {};
|
||
}
|
||
|
||
auto [dumpHash, size] = hashSink->finish();
|
||
|
||
auto desc = ContentAddressWithReferences::fromParts(
|
||
hashMethod,
|
||
methodsMatch
|
||
? dumpHash
|
||
: hashPath(
|
||
PosixSourceAccessor::createAtRoot(tempPath),
|
||
hashMethod.getFileIngestionMethod(), hashAlgo).first,
|
||
{
|
||
.others = references,
|
||
// caller is not capable of creating a self-reference, because this is content-addressed without modulus
|
||
.self = false,
|
||
});
|
||
|
||
auto dstPath = makeFixedOutputPathFromCA(name, desc);
|
||
|
||
addTempRoot(dstPath);
|
||
|
||
if (repair || !isValidPath(dstPath)) {
|
||
|
||
/* The first check above is an optimisation to prevent
|
||
unnecessary lock acquisition. */
|
||
|
||
auto realPath = Store::toRealPath(dstPath);
|
||
|
||
PathLocks outputLock({realPath});
|
||
|
||
if (repair || !isValidPath(dstPath)) {
|
||
|
||
deletePath(realPath);
|
||
|
||
autoGC();
|
||
|
||
if (inMemoryAndDontNeedRestore) {
|
||
StringSource dumpSource { dump };
|
||
/* Restore from the buffer in memory. */
|
||
auto fim = hashMethod.getFileIngestionMethod();
|
||
switch (fim) {
|
||
case FileIngestionMethod::Flat:
|
||
case FileIngestionMethod::NixArchive:
|
||
restorePath(realPath, dumpSource, (FileSerialisationMethod) fim, settings.fsyncStorePaths);
|
||
break;
|
||
case FileIngestionMethod::Git:
|
||
// doesn't correspond to serialization method, so
|
||
// this should be unreachable
|
||
assert(false);
|
||
}
|
||
} else {
|
||
/* Move the temporary path we restored above. */
|
||
moveFile(tempPath.string(), realPath);
|
||
}
|
||
|
||
/* For computing the nar hash. In recursive SHA-256 mode, this
|
||
is the same as the store hash, so no need to do it again. */
|
||
auto narHash = std::pair { dumpHash, size };
|
||
if (dumpMethod != FileSerialisationMethod::NixArchive || hashAlgo != HashAlgorithm::SHA256) {
|
||
HashSink narSink { HashAlgorithm::SHA256 };
|
||
dumpPath(realPath, narSink);
|
||
narHash = narSink.finish();
|
||
}
|
||
|
||
canonicalisePathMetaData(realPath); // FIXME: merge into restorePath
|
||
|
||
optimisePath(realPath, repair);
|
||
|
||
if (settings.fsyncStorePaths) {
|
||
recursiveSync(realPath);
|
||
syncParent(realPath);
|
||
}
|
||
|
||
ValidPathInfo info {
|
||
*this,
|
||
name,
|
||
std::move(desc),
|
||
narHash.first
|
||
};
|
||
info.narSize = narHash.second;
|
||
registerValidPath(info);
|
||
}
|
||
|
||
outputLock.setDeletion(true);
|
||
}
|
||
|
||
return dstPath;
|
||
}
|
||
|
||
|
||
/* Create a temporary directory in the store that won't be
|
||
garbage-collected until the returned FD is closed. */
|
||
std::pair<std::filesystem::path, AutoCloseFD> LocalStore::createTempDirInStore()
|
||
{
|
||
std::filesystem::path tmpDirFn;
|
||
AutoCloseFD tmpDirFd;
|
||
bool lockedByUs = false;
|
||
do {
|
||
/* There is a slight possibility that `tmpDir' gets deleted by
|
||
the GC between createTempDir() and when we acquire a lock on it.
|
||
We'll repeat until 'tmpDir' exists and we've locked it. */
|
||
tmpDirFn = createTempDir(realStoreDir, "tmp");
|
||
tmpDirFd = openDirectory(tmpDirFn);
|
||
if (!tmpDirFd) {
|
||
continue;
|
||
}
|
||
lockedByUs = lockFile(tmpDirFd.get(), ltWrite, true);
|
||
} while (!pathExists(tmpDirFn.string()) || !lockedByUs);
|
||
return {tmpDirFn, std::move(tmpDirFd)};
|
||
}
|
||
|
||
|
||
void LocalStore::invalidatePathChecked(const StorePath & path)
|
||
{
|
||
retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
SQLiteTxn txn(state->db);
|
||
|
||
if (isValidPath_(*state, path)) {
|
||
StorePathSet referrers; queryReferrers(*state, path, referrers);
|
||
referrers.erase(path); /* ignore self-references */
|
||
if (!referrers.empty())
|
||
throw PathInUse("cannot delete path '%s' because it is in use by %s",
|
||
printStorePath(path), showPaths(referrers));
|
||
invalidatePath(*state, path);
|
||
}
|
||
|
||
txn.commit();
|
||
});
|
||
}
|
||
|
||
|
||
bool LocalStore::verifyStore(bool checkContents, RepairFlag repair)
|
||
{
|
||
printInfo("reading the Nix store...");
|
||
|
||
/* Acquire the global GC lock to get a consistent snapshot of
|
||
existing and valid paths. */
|
||
auto fdGCLock = openGCLock();
|
||
FdLock gcLock(fdGCLock.get(), ltRead, true, "waiting for the big garbage collector lock...");
|
||
|
||
auto [errors, validPaths] = verifyAllValidPaths(repair);
|
||
|
||
/* Optionally, check the content hashes (slow). */
|
||
if (checkContents) {
|
||
|
||
printInfo("checking link hashes...");
|
||
|
||
for (auto & link : std::filesystem::directory_iterator{linksDir}) {
|
||
checkInterrupt();
|
||
auto name = link.path().filename();
|
||
printMsg(lvlTalkative, "checking contents of '%s'", name);
|
||
PosixSourceAccessor accessor;
|
||
std::string hash = hashPath(
|
||
PosixSourceAccessor::createAtRoot(link.path()),
|
||
FileIngestionMethod::NixArchive, HashAlgorithm::SHA256).first.to_string(HashFormat::Nix32, false);
|
||
if (hash != name.string()) {
|
||
printError("link '%s' was modified! expected hash '%s', got '%s'",
|
||
link.path(), name, hash);
|
||
if (repair) {
|
||
std::filesystem::remove(link.path());
|
||
printInfo("removed link '%s'", link.path());
|
||
} else {
|
||
errors = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
printInfo("checking store hashes...");
|
||
|
||
Hash nullHash(HashAlgorithm::SHA256);
|
||
|
||
for (auto & i : validPaths) {
|
||
try {
|
||
auto info = std::const_pointer_cast<ValidPathInfo>(std::shared_ptr<const ValidPathInfo>(queryPathInfo(i)));
|
||
|
||
/* Check the content hash (optionally - slow). */
|
||
printMsg(lvlTalkative, "checking contents of '%s'", printStorePath(i));
|
||
|
||
auto hashSink = HashSink(info->narHash.algo);
|
||
|
||
dumpPath(Store::toRealPath(i), hashSink);
|
||
auto current = hashSink.finish();
|
||
|
||
if (info->narHash != nullHash && info->narHash != current.first) {
|
||
printError("path '%s' was modified! expected hash '%s', got '%s'",
|
||
printStorePath(i), info->narHash.to_string(HashFormat::Nix32, true), current.first.to_string(HashFormat::Nix32, true));
|
||
if (repair) repairPath(i); else errors = true;
|
||
} else {
|
||
|
||
bool update = false;
|
||
|
||
/* Fill in missing hashes. */
|
||
if (info->narHash == nullHash) {
|
||
printInfo("fixing missing hash on '%s'", printStorePath(i));
|
||
info->narHash = current.first;
|
||
update = true;
|
||
}
|
||
|
||
/* Fill in missing narSize fields (from old stores). */
|
||
if (info->narSize == 0) {
|
||
printInfo("updating size field on '%s' to %s", printStorePath(i), current.second);
|
||
info->narSize = current.second;
|
||
update = true;
|
||
}
|
||
|
||
if (update) {
|
||
auto state(_state.lock());
|
||
updatePathInfo(*state, *info);
|
||
}
|
||
|
||
}
|
||
|
||
} catch (Error & e) {
|
||
/* It's possible that the path got GC'ed, so ignore
|
||
errors on invalid paths. */
|
||
if (isValidPath(i))
|
||
logError(e.info());
|
||
else
|
||
warn(e.msg());
|
||
errors = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
return errors;
|
||
}
|
||
|
||
|
||
LocalStore::VerificationResult LocalStore::verifyAllValidPaths(RepairFlag repair)
|
||
{
|
||
StorePathSet storePathsInStoreDir;
|
||
/* Why aren't we using `queryAllValidPaths`? Because that would
|
||
tell us about all the paths than the database knows about. Here we
|
||
want to know about all the store paths in the store directory,
|
||
regardless of what the database thinks.
|
||
|
||
We will end up cross-referencing these two sources of truth (the
|
||
database and the filesystem) in the loop below, in order to catch
|
||
invalid states.
|
||
*/
|
||
for (auto & i : std::filesystem::directory_iterator{realStoreDir.to_string()}) {
|
||
checkInterrupt();
|
||
try {
|
||
storePathsInStoreDir.insert({i.path().filename().string()});
|
||
} catch (BadStorePath &) { }
|
||
}
|
||
|
||
/* Check whether all valid paths actually exist. */
|
||
printInfo("checking path existence...");
|
||
|
||
StorePathSet done;
|
||
|
||
auto existsInStoreDir = [&](const StorePath & storePath) {
|
||
return storePathsInStoreDir.count(storePath);
|
||
};
|
||
|
||
bool errors = false;
|
||
StorePathSet validPaths;
|
||
|
||
for (auto & i : queryAllValidPaths())
|
||
verifyPath(i, existsInStoreDir, done, validPaths, repair, errors);
|
||
|
||
return {
|
||
.errors = errors,
|
||
.validPaths = validPaths,
|
||
};
|
||
}
|
||
|
||
|
||
void LocalStore::verifyPath(const StorePath & path, std::function<bool(const StorePath &)> existsInStoreDir,
|
||
StorePathSet & done, StorePathSet & validPaths, RepairFlag repair, bool & errors)
|
||
{
|
||
checkInterrupt();
|
||
|
||
if (!done.insert(path).second) return;
|
||
|
||
if (!existsInStoreDir(path)) {
|
||
/* Check any referrers first. If we can invalidate them
|
||
first, then we can invalidate this path as well. */
|
||
bool canInvalidate = true;
|
||
StorePathSet referrers; queryReferrers(path, referrers);
|
||
for (auto & i : referrers)
|
||
if (i != path) {
|
||
verifyPath(i, existsInStoreDir, done, validPaths, repair, errors);
|
||
if (validPaths.count(i))
|
||
canInvalidate = false;
|
||
}
|
||
|
||
auto pathS = printStorePath(path);
|
||
|
||
if (canInvalidate) {
|
||
printInfo("path '%s' disappeared, removing from database...", pathS);
|
||
auto state(_state.lock());
|
||
invalidatePath(*state, path);
|
||
} else {
|
||
printError("path '%s' disappeared, but it still has valid referrers!", pathS);
|
||
if (repair)
|
||
try {
|
||
repairPath(path);
|
||
} catch (Error & e) {
|
||
logWarning(e.info());
|
||
errors = true;
|
||
}
|
||
else errors = true;
|
||
}
|
||
|
||
return;
|
||
}
|
||
|
||
validPaths.insert(std::move(path));
|
||
}
|
||
|
||
|
||
unsigned int LocalStore::getProtocol()
|
||
{
|
||
return PROTOCOL_VERSION;
|
||
}
|
||
|
||
std::optional<TrustedFlag> LocalStore::isTrustedClient()
|
||
{
|
||
return Trusted;
|
||
}
|
||
|
||
|
||
void LocalStore::vacuumDB()
|
||
{
|
||
auto state(_state.lock());
|
||
state->db.exec("vacuum");
|
||
}
|
||
|
||
|
||
void LocalStore::addSignatures(const StorePath & storePath, const StringSet & sigs)
|
||
{
|
||
retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
SQLiteTxn txn(state->db);
|
||
|
||
auto info = std::const_pointer_cast<ValidPathInfo>(queryPathInfoInternal(*state, storePath));
|
||
|
||
info->sigs.insert(sigs.begin(), sigs.end());
|
||
|
||
updatePathInfo(*state, *info);
|
||
|
||
txn.commit();
|
||
});
|
||
}
|
||
|
||
|
||
void LocalStore::signRealisation(Realisation & realisation)
|
||
{
|
||
// FIXME: keep secret keys in memory.
|
||
|
||
auto secretKeyFiles = settings.secretKeyFiles;
|
||
|
||
for (auto & secretKeyFile : secretKeyFiles.get()) {
|
||
SecretKey secretKey(readFile(secretKeyFile));
|
||
LocalSigner signer(std::move(secretKey));
|
||
realisation.sign(signer);
|
||
}
|
||
}
|
||
|
||
void LocalStore::signPathInfo(ValidPathInfo & info)
|
||
{
|
||
// FIXME: keep secret keys in memory.
|
||
|
||
auto secretKeyFiles = settings.secretKeyFiles;
|
||
|
||
for (auto & secretKeyFile : secretKeyFiles.get()) {
|
||
SecretKey secretKey(readFile(secretKeyFile));
|
||
LocalSigner signer(std::move(secretKey));
|
||
info.sign(*this, signer);
|
||
}
|
||
}
|
||
|
||
|
||
std::optional<std::pair<int64_t, Realisation>> LocalStore::queryRealisationCore_(
|
||
LocalStore::State & state,
|
||
const DrvOutput & id)
|
||
{
|
||
auto useQueryRealisedOutput(
|
||
state.stmts->QueryRealisedOutput.use()
|
||
(id.strHash())
|
||
(id.outputName));
|
||
if (!useQueryRealisedOutput.next())
|
||
return std::nullopt;
|
||
auto realisationDbId = useQueryRealisedOutput.getInt(0);
|
||
auto outputPath = parseStorePath(useQueryRealisedOutput.getStr(1));
|
||
auto signatures =
|
||
tokenizeString<StringSet>(useQueryRealisedOutput.getStr(2));
|
||
|
||
return {{
|
||
realisationDbId,
|
||
Realisation{
|
||
.id = id,
|
||
.outPath = outputPath,
|
||
.signatures = signatures,
|
||
}
|
||
}};
|
||
}
|
||
|
||
std::optional<const Realisation> LocalStore::queryRealisation_(
|
||
LocalStore::State & state,
|
||
const DrvOutput & id)
|
||
{
|
||
auto maybeCore = queryRealisationCore_(state, id);
|
||
if (!maybeCore)
|
||
return std::nullopt;
|
||
auto [realisationDbId, res] = *maybeCore;
|
||
|
||
std::map<DrvOutput, StorePath> dependentRealisations;
|
||
auto useRealisationRefs(
|
||
state.stmts->QueryRealisationReferences.use()
|
||
(realisationDbId));
|
||
while (useRealisationRefs.next()) {
|
||
auto depId = DrvOutput {
|
||
Hash::parseAnyPrefixed(useRealisationRefs.getStr(0)),
|
||
useRealisationRefs.getStr(1),
|
||
};
|
||
auto dependentRealisation = queryRealisationCore_(state, depId);
|
||
assert(dependentRealisation); // Enforced by the db schema
|
||
auto outputPath = dependentRealisation->second.outPath;
|
||
dependentRealisations.insert({depId, outputPath});
|
||
}
|
||
|
||
res.dependentRealisations = dependentRealisations;
|
||
|
||
return { res };
|
||
}
|
||
|
||
void LocalStore::queryRealisationUncached(const DrvOutput & id,
|
||
Callback<std::shared_ptr<const Realisation>> callback) noexcept
|
||
{
|
||
try {
|
||
auto maybeRealisation
|
||
= retrySQLite<std::optional<const Realisation>>([&]() {
|
||
auto state(_state.lock());
|
||
return queryRealisation_(*state, id);
|
||
});
|
||
if (maybeRealisation)
|
||
callback(
|
||
std::make_shared<const Realisation>(maybeRealisation.value()));
|
||
else
|
||
callback(nullptr);
|
||
|
||
} catch (...) {
|
||
callback.rethrow();
|
||
}
|
||
}
|
||
|
||
void LocalStore::addBuildLog(const StorePath & drvPath, std::string_view log)
|
||
{
|
||
assert(drvPath.isDerivation());
|
||
|
||
auto baseName = drvPath.to_string();
|
||
|
||
auto logPath = fmt("%s/%s/%s/%s.bz2", logDir, drvsLogDir, baseName.substr(0, 2), baseName.substr(2));
|
||
|
||
if (pathExists(logPath)) return;
|
||
|
||
createDirs(dirOf(logPath));
|
||
|
||
auto tmpFile = fmt("%s.tmp.%d", logPath, getpid());
|
||
|
||
writeFile(tmpFile, compress("bzip2", log));
|
||
|
||
std::filesystem::rename(tmpFile, logPath);
|
||
}
|
||
|
||
std::optional<std::string> LocalStore::getVersion()
|
||
{
|
||
return nixVersion;
|
||
}
|
||
|
||
static RegisterStoreImplementation<LocalStore, LocalStoreConfig> regLocalStore;
|
||
|
||
} // namespace nix
|