mirror of
https://github.com/privatevoid-net/nix-super.git
synced 2024-11-28 08:36:15 +02:00
15e5684884
This caused nlohmann/json.hpp to leak into a lot of compilation units, which is slow (when not using precompiled headers). Cuts build time from 46m24s to 42m5s (real time with -j24: 2m42s to 2m24s).
1713 lines
57 KiB
C++
1713 lines
57 KiB
C++
#include "local-store.hh"
|
||
#include "globals.hh"
|
||
#include "git.hh"
|
||
#include "archive.hh"
|
||
#include "pathlocks.hh"
|
||
#include "worker-protocol.hh"
|
||
#include "derivations.hh"
|
||
#include "realisation.hh"
|
||
#include "nar-info.hh"
|
||
#include "references.hh"
|
||
#include "callback.hh"
|
||
#include "topo-sort.hh"
|
||
#include "finally.hh"
|
||
#include "compression.hh"
|
||
#include "signals.hh"
|
||
#include "posix-fs-canonicalise.hh"
|
||
#include "posix-source-accessor.hh"
|
||
#include "keys.hh"
|
||
#include "users.hh"
|
||
|
||
#include <iostream>
|
||
#include <algorithm>
|
||
#include <cstring>
|
||
|
||
#include <memory>
|
||
#include <new>
|
||
#include <sys/types.h>
|
||
#include <sys/stat.h>
|
||
#include <sys/time.h>
|
||
#include <unistd.h>
|
||
#include <utime.h>
|
||
#include <fcntl.h>
|
||
#include <errno.h>
|
||
#include <stdio.h>
|
||
#include <time.h>
|
||
|
||
#ifndef _WIN32
|
||
# include <grp.h>
|
||
#endif
|
||
|
||
#if __linux__
|
||
# include <sched.h>
|
||
# include <sys/statvfs.h>
|
||
# include <sys/mount.h>
|
||
#endif
|
||
|
||
#ifdef __CYGWIN__
|
||
# include <windows.h>
|
||
#endif
|
||
|
||
#include <sqlite3.h>
|
||
|
||
#include <nlohmann/json.hpp>
|
||
|
||
#include "strings.hh"
|
||
|
||
|
||
namespace nix {
|
||
|
||
LocalStoreConfig::LocalStoreConfig(
|
||
std::string_view scheme,
|
||
std::string_view authority,
|
||
const Params & params)
|
||
: StoreConfig(params)
|
||
, LocalFSStoreConfig(authority, params)
|
||
{
|
||
}
|
||
|
||
std::string LocalStoreConfig::doc()
|
||
{
|
||
return
|
||
#include "local-store.md"
|
||
;
|
||
}
|
||
|
||
struct LocalStore::State::Stmts {
|
||
/* Some precompiled SQLite statements. */
|
||
SQLiteStmt RegisterValidPath;
|
||
SQLiteStmt UpdatePathInfo;
|
||
SQLiteStmt AddReference;
|
||
SQLiteStmt QueryPathInfo;
|
||
SQLiteStmt QueryReferences;
|
||
SQLiteStmt QueryReferrers;
|
||
SQLiteStmt InvalidatePath;
|
||
SQLiteStmt AddDerivationOutput;
|
||
SQLiteStmt RegisterRealisedOutput;
|
||
SQLiteStmt UpdateRealisedOutput;
|
||
SQLiteStmt QueryValidDerivers;
|
||
SQLiteStmt QueryDerivationOutputs;
|
||
SQLiteStmt QueryRealisedOutput;
|
||
SQLiteStmt QueryAllRealisedOutputs;
|
||
SQLiteStmt QueryPathFromHashPart;
|
||
SQLiteStmt QueryValidPaths;
|
||
SQLiteStmt QueryRealisationReferences;
|
||
SQLiteStmt AddRealisationReference;
|
||
};
|
||
|
||
static int getSchema(Path schemaPath)
|
||
{
|
||
int curSchema = 0;
|
||
if (pathExists(schemaPath)) {
|
||
auto s = readFile(schemaPath);
|
||
auto n = string2Int<int>(s);
|
||
if (!n)
|
||
throw Error("'%1%' is corrupt", schemaPath);
|
||
curSchema = *n;
|
||
}
|
||
return curSchema;
|
||
}
|
||
|
||
void migrateCASchema(SQLite& db, Path schemaPath, AutoCloseFD& lockFd)
|
||
{
|
||
const int nixCASchemaVersion = 4;
|
||
int curCASchema = getSchema(schemaPath);
|
||
if (curCASchema != nixCASchemaVersion) {
|
||
if (curCASchema > nixCASchemaVersion) {
|
||
throw Error("current Nix store ca-schema is version %1%, but I only support %2%",
|
||
curCASchema, nixCASchemaVersion);
|
||
}
|
||
|
||
if (!lockFile(lockFd.get(), ltWrite, false)) {
|
||
printInfo("waiting for exclusive access to the Nix store for ca drvs...");
|
||
lockFile(lockFd.get(), ltNone, false); // We have acquired a shared lock; release it to prevent deadlocks
|
||
lockFile(lockFd.get(), ltWrite, true);
|
||
}
|
||
|
||
if (curCASchema == 0) {
|
||
static const char schema[] =
|
||
#include "ca-specific-schema.sql.gen.hh"
|
||
;
|
||
db.exec(schema);
|
||
curCASchema = nixCASchemaVersion;
|
||
}
|
||
|
||
if (curCASchema < 4)
|
||
throw Error("experimental CA schema version %d is no longer supported", curCASchema);
|
||
|
||
writeFile(schemaPath, fmt("%d", nixCASchemaVersion), 0666, true);
|
||
lockFile(lockFd.get(), ltRead, true);
|
||
}
|
||
}
|
||
|
||
LocalStore::LocalStore(
|
||
std::string_view scheme,
|
||
PathView path,
|
||
const Params & params)
|
||
: StoreConfig(params)
|
||
, LocalFSStoreConfig(path, params)
|
||
, LocalStoreConfig(scheme, path, params)
|
||
, Store(params)
|
||
, LocalFSStore(params)
|
||
, dbDir(stateDir + "/db")
|
||
, linksDir(realStoreDir + "/.links")
|
||
, reservedPath(dbDir + "/reserved")
|
||
, schemaPath(dbDir + "/schema")
|
||
, tempRootsDir(stateDir + "/temproots")
|
||
, fnTempRoots(fmt("%s/%d", tempRootsDir, getpid()))
|
||
, locksHeld(tokenizeString<PathSet>(getEnv("NIX_HELD_LOCKS").value_or("")))
|
||
{
|
||
auto state(_state.lock());
|
||
state->stmts = std::make_unique<State::Stmts>();
|
||
|
||
/* Create missing state directories if they don't already exist. */
|
||
createDirs(realStoreDir);
|
||
if (readOnly) {
|
||
experimentalFeatureSettings.require(Xp::ReadOnlyLocalStore);
|
||
} else {
|
||
makeStoreWritable();
|
||
}
|
||
createDirs(linksDir);
|
||
Path profilesDir = stateDir + "/profiles";
|
||
createDirs(profilesDir);
|
||
createDirs(tempRootsDir);
|
||
createDirs(dbDir);
|
||
Path gcRootsDir = stateDir + "/gcroots";
|
||
if (!pathExists(gcRootsDir)) {
|
||
createDirs(gcRootsDir);
|
||
createSymlink(profilesDir, gcRootsDir + "/profiles");
|
||
}
|
||
|
||
for (auto & perUserDir : {profilesDir + "/per-user", gcRootsDir + "/per-user"}) {
|
||
createDirs(perUserDir);
|
||
if (!readOnly) {
|
||
if (chmod(perUserDir.c_str(), 0755) == -1)
|
||
throw SysError("could not set permissions on '%s' to 755", perUserDir);
|
||
}
|
||
}
|
||
|
||
#ifndef _WIN32
|
||
/* Optionally, create directories and set permissions for a
|
||
multi-user install. */
|
||
if (isRootUser() && settings.buildUsersGroup != "") {
|
||
mode_t perm = 01775;
|
||
|
||
struct group * gr = getgrnam(settings.buildUsersGroup.get().c_str());
|
||
if (!gr)
|
||
printError("warning: the group '%1%' specified in 'build-users-group' does not exist", settings.buildUsersGroup);
|
||
else if (!readOnly) {
|
||
struct stat st;
|
||
if (stat(realStoreDir.get().c_str(), &st))
|
||
throw SysError("getting attributes of path '%1%'", realStoreDir);
|
||
|
||
if (st.st_uid != 0 || st.st_gid != gr->gr_gid || (st.st_mode & ~S_IFMT) != perm) {
|
||
if (chown(realStoreDir.get().c_str(), 0, gr->gr_gid) == -1)
|
||
throw SysError("changing ownership of path '%1%'", realStoreDir);
|
||
if (chmod(realStoreDir.get().c_str(), perm) == -1)
|
||
throw SysError("changing permissions on path '%1%'", realStoreDir);
|
||
}
|
||
}
|
||
}
|
||
#endif
|
||
|
||
/* Ensure that the store and its parents are not symlinks. */
|
||
if (!settings.allowSymlinkedStore) {
|
||
Path path = realStoreDir;
|
||
struct stat st;
|
||
while (path != "/") {
|
||
st = lstat(path);
|
||
if (S_ISLNK(st.st_mode))
|
||
throw Error(
|
||
"the path '%1%' is a symlink; "
|
||
"this is not allowed for the Nix store and its parent directories",
|
||
path);
|
||
path = dirOf(path);
|
||
}
|
||
}
|
||
|
||
/* We can't open a SQLite database if the disk is full. Since
|
||
this prevents the garbage collector from running when it's most
|
||
needed, we reserve some dummy space that we can free just
|
||
before doing a garbage collection. */
|
||
try {
|
||
struct stat st;
|
||
if (stat(reservedPath.c_str(), &st) == -1 ||
|
||
st.st_size != settings.reservedSize)
|
||
{
|
||
AutoCloseFD fd = toDescriptor(open(reservedPath.c_str(), O_WRONLY | O_CREAT
|
||
#ifndef _WIN32
|
||
| O_CLOEXEC
|
||
#endif
|
||
, 0600));
|
||
int res = -1;
|
||
#if HAVE_POSIX_FALLOCATE
|
||
res = posix_fallocate(fd.get(), 0, settings.reservedSize);
|
||
#endif
|
||
if (res == -1) {
|
||
writeFull(fd.get(), std::string(settings.reservedSize, 'X'));
|
||
[[gnu::unused]] auto res2 =
|
||
|
||
#ifdef _WIN32
|
||
SetEndOfFile(fd.get())
|
||
#else
|
||
ftruncate(fd.get(), settings.reservedSize)
|
||
#endif
|
||
;
|
||
}
|
||
}
|
||
} catch (SystemError & e) { /* don't care about errors */
|
||
}
|
||
|
||
/* Acquire the big fat lock in shared mode to make sure that no
|
||
schema upgrade is in progress. */
|
||
if (!readOnly) {
|
||
Path globalLockPath = dbDir + "/big-lock";
|
||
globalLock = openLockFile(globalLockPath.c_str(), true);
|
||
}
|
||
|
||
if (!readOnly && !lockFile(globalLock.get(), ltRead, false)) {
|
||
printInfo("waiting for the big Nix store lock...");
|
||
lockFile(globalLock.get(), ltRead, true);
|
||
}
|
||
|
||
/* Check the current database schema and if necessary do an
|
||
upgrade. */
|
||
int curSchema = getSchema();
|
||
if (readOnly && curSchema < nixSchemaVersion) {
|
||
debug("current schema version: %d", curSchema);
|
||
debug("supported schema version: %d", nixSchemaVersion);
|
||
throw Error(curSchema == 0 ?
|
||
"database does not exist, and cannot be created in read-only mode" :
|
||
"database schema needs migrating, but this cannot be done in read-only mode");
|
||
}
|
||
|
||
if (curSchema > nixSchemaVersion)
|
||
throw Error("current Nix store schema is version %1%, but I only support %2%",
|
||
curSchema, nixSchemaVersion);
|
||
|
||
else if (curSchema == 0) { /* new store */
|
||
curSchema = nixSchemaVersion;
|
||
openDB(*state, true);
|
||
writeFile(schemaPath, fmt("%1%", nixSchemaVersion), 0666, true);
|
||
}
|
||
|
||
else if (curSchema < nixSchemaVersion) {
|
||
if (curSchema < 5)
|
||
throw Error(
|
||
"Your Nix store has a database in Berkeley DB format,\n"
|
||
"which is no longer supported. To convert to the new format,\n"
|
||
"please upgrade Nix to version 0.12 first.");
|
||
|
||
if (curSchema < 6)
|
||
throw Error(
|
||
"Your Nix store has a database in flat file format,\n"
|
||
"which is no longer supported. To convert to the new format,\n"
|
||
"please upgrade Nix to version 1.11 first.");
|
||
|
||
if (!lockFile(globalLock.get(), ltWrite, false)) {
|
||
printInfo("waiting for exclusive access to the Nix store...");
|
||
lockFile(globalLock.get(), ltNone, false); // We have acquired a shared lock; release it to prevent deadlocks
|
||
lockFile(globalLock.get(), ltWrite, true);
|
||
}
|
||
|
||
/* Get the schema version again, because another process may
|
||
have performed the upgrade already. */
|
||
curSchema = getSchema();
|
||
|
||
openDB(*state, false);
|
||
|
||
if (curSchema < 8) {
|
||
SQLiteTxn txn(state->db);
|
||
state->db.exec("alter table ValidPaths add column ultimate integer");
|
||
state->db.exec("alter table ValidPaths add column sigs text");
|
||
txn.commit();
|
||
}
|
||
|
||
if (curSchema < 9) {
|
||
SQLiteTxn txn(state->db);
|
||
state->db.exec("drop table FailedPaths");
|
||
txn.commit();
|
||
}
|
||
|
||
if (curSchema < 10) {
|
||
SQLiteTxn txn(state->db);
|
||
state->db.exec("alter table ValidPaths add column ca text");
|
||
txn.commit();
|
||
}
|
||
|
||
writeFile(schemaPath, fmt("%1%", nixSchemaVersion), 0666, true);
|
||
|
||
lockFile(globalLock.get(), ltRead, true);
|
||
}
|
||
|
||
else openDB(*state, false);
|
||
|
||
if (experimentalFeatureSettings.isEnabled(Xp::CaDerivations)) {
|
||
if (!readOnly) {
|
||
migrateCASchema(state->db, dbDir + "/ca-schema", globalLock);
|
||
} else {
|
||
throw Error("need to migrate to content-addressed schema, but this cannot be done in read-only mode");
|
||
}
|
||
}
|
||
|
||
/* Prepare SQL statements. */
|
||
state->stmts->RegisterValidPath.create(state->db,
|
||
"insert into ValidPaths (path, hash, registrationTime, deriver, narSize, ultimate, sigs, ca) values (?, ?, ?, ?, ?, ?, ?, ?);");
|
||
state->stmts->UpdatePathInfo.create(state->db,
|
||
"update ValidPaths set narSize = ?, hash = ?, ultimate = ?, sigs = ?, ca = ? where path = ?;");
|
||
state->stmts->AddReference.create(state->db,
|
||
"insert or replace into Refs (referrer, reference) values (?, ?);");
|
||
state->stmts->QueryPathInfo.create(state->db,
|
||
"select id, hash, registrationTime, deriver, narSize, ultimate, sigs, ca from ValidPaths where path = ?;");
|
||
state->stmts->QueryReferences.create(state->db,
|
||
"select path from Refs join ValidPaths on reference = id where referrer = ?;");
|
||
state->stmts->QueryReferrers.create(state->db,
|
||
"select path from Refs join ValidPaths on referrer = id where reference = (select id from ValidPaths where path = ?);");
|
||
state->stmts->InvalidatePath.create(state->db,
|
||
"delete from ValidPaths where path = ?;");
|
||
state->stmts->AddDerivationOutput.create(state->db,
|
||
"insert or replace into DerivationOutputs (drv, id, path) values (?, ?, ?);");
|
||
state->stmts->QueryValidDerivers.create(state->db,
|
||
"select v.id, v.path from DerivationOutputs d join ValidPaths v on d.drv = v.id where d.path = ?;");
|
||
state->stmts->QueryDerivationOutputs.create(state->db,
|
||
"select id, path from DerivationOutputs where drv = ?;");
|
||
// Use "path >= ?" with limit 1 rather than "path like '?%'" to
|
||
// ensure efficient lookup.
|
||
state->stmts->QueryPathFromHashPart.create(state->db,
|
||
"select path from ValidPaths where path >= ? limit 1;");
|
||
state->stmts->QueryValidPaths.create(state->db, "select path from ValidPaths");
|
||
if (experimentalFeatureSettings.isEnabled(Xp::CaDerivations)) {
|
||
state->stmts->RegisterRealisedOutput.create(state->db,
|
||
R"(
|
||
insert into Realisations (drvPath, outputName, outputPath, signatures)
|
||
values (?, ?, (select id from ValidPaths where path = ?), ?)
|
||
;
|
||
)");
|
||
state->stmts->UpdateRealisedOutput.create(state->db,
|
||
R"(
|
||
update Realisations
|
||
set signatures = ?
|
||
where
|
||
drvPath = ? and
|
||
outputName = ?
|
||
;
|
||
)");
|
||
state->stmts->QueryRealisedOutput.create(state->db,
|
||
R"(
|
||
select Realisations.id, Output.path, Realisations.signatures from Realisations
|
||
inner join ValidPaths as Output on Output.id = Realisations.outputPath
|
||
where drvPath = ? and outputName = ?
|
||
;
|
||
)");
|
||
state->stmts->QueryAllRealisedOutputs.create(state->db,
|
||
R"(
|
||
select outputName, Output.path from Realisations
|
||
inner join ValidPaths as Output on Output.id = Realisations.outputPath
|
||
where drvPath = ?
|
||
;
|
||
)");
|
||
state->stmts->QueryRealisationReferences.create(state->db,
|
||
R"(
|
||
select drvPath, outputName from Realisations
|
||
join RealisationsRefs on realisationReference = Realisations.id
|
||
where referrer = ?;
|
||
)");
|
||
state->stmts->AddRealisationReference.create(state->db,
|
||
R"(
|
||
insert or replace into RealisationsRefs (referrer, realisationReference)
|
||
values (
|
||
(select id from Realisations where drvPath = ? and outputName = ?),
|
||
(select id from Realisations where drvPath = ? and outputName = ?));
|
||
)");
|
||
}
|
||
}
|
||
|
||
|
||
LocalStore::LocalStore(const Params & params)
|
||
: LocalStore("local", "", params)
|
||
{
|
||
}
|
||
|
||
|
||
AutoCloseFD LocalStore::openGCLock()
|
||
{
|
||
Path fnGCLock = stateDir + "/gc.lock";
|
||
auto fdGCLock = open(fnGCLock.c_str(), O_RDWR | O_CREAT
|
||
#ifndef _WIN32
|
||
| O_CLOEXEC
|
||
#endif
|
||
, 0600);
|
||
if (!fdGCLock)
|
||
throw SysError("opening global GC lock '%1%'", fnGCLock);
|
||
return toDescriptor(fdGCLock);
|
||
}
|
||
|
||
|
||
void LocalStore::deleteStorePath(const Path & path, uint64_t & bytesFreed)
|
||
{
|
||
deletePath(path, bytesFreed);
|
||
}
|
||
|
||
|
||
LocalStore::~LocalStore()
|
||
{
|
||
std::shared_future<void> future;
|
||
|
||
{
|
||
auto state(_state.lock());
|
||
if (state->gcRunning)
|
||
future = state->gcFuture;
|
||
}
|
||
|
||
if (future.valid()) {
|
||
printInfo("waiting for auto-GC to finish on exit...");
|
||
future.get();
|
||
}
|
||
|
||
try {
|
||
auto fdTempRoots(_fdTempRoots.lock());
|
||
if (*fdTempRoots) {
|
||
fdTempRoots->close();
|
||
unlink(fnTempRoots.c_str());
|
||
}
|
||
} catch (...) {
|
||
ignoreExceptionInDestructor();
|
||
}
|
||
}
|
||
|
||
|
||
std::string LocalStore::getUri()
|
||
{
|
||
return "local";
|
||
}
|
||
|
||
|
||
int LocalStore::getSchema()
|
||
{ return nix::getSchema(schemaPath); }
|
||
|
||
void LocalStore::openDB(State & state, bool create)
|
||
{
|
||
if (create && readOnly) {
|
||
throw Error("cannot create database while in read-only mode");
|
||
}
|
||
|
||
if (access(dbDir.c_str(), R_OK | (readOnly ? 0 : W_OK)))
|
||
throw SysError("Nix database directory '%1%' is not writable", dbDir);
|
||
|
||
/* Open the Nix database. */
|
||
std::string dbPath = dbDir + "/db.sqlite";
|
||
auto & db(state.db);
|
||
auto openMode = readOnly ? SQLiteOpenMode::Immutable
|
||
: create ? SQLiteOpenMode::Normal
|
||
: SQLiteOpenMode::NoCreate;
|
||
state.db = SQLite(dbPath, openMode);
|
||
|
||
#ifdef __CYGWIN__
|
||
/* The cygwin version of sqlite3 has a patch which calls
|
||
SetDllDirectory("/usr/bin") on init. It was intended to fix extension
|
||
loading, which we don't use, and the effect of SetDllDirectory is
|
||
inherited by child processes, and causes libraries to be loaded from
|
||
/usr/bin instead of $PATH. This breaks quite a few things (e.g.
|
||
checkPhase on openssh), so we set it back to default behaviour. */
|
||
SetDllDirectoryW(L"");
|
||
#endif
|
||
|
||
/* !!! check whether sqlite has been built with foreign key
|
||
support */
|
||
|
||
/* Whether SQLite should fsync(). "Normal" synchronous mode
|
||
should be safe enough. If the user asks for it, don't sync at
|
||
all. This can cause database corruption if the system
|
||
crashes. */
|
||
std::string syncMode = settings.fsyncMetadata ? "normal" : "off";
|
||
db.exec("pragma synchronous = " + syncMode);
|
||
|
||
/* Set the SQLite journal mode. WAL mode is fastest, so it's the
|
||
default. */
|
||
std::string mode = settings.useSQLiteWAL ? "wal" : "truncate";
|
||
std::string prevMode;
|
||
{
|
||
SQLiteStmt stmt;
|
||
stmt.create(db, "pragma main.journal_mode;");
|
||
if (sqlite3_step(stmt) != SQLITE_ROW)
|
||
SQLiteError::throw_(db, "querying journal mode");
|
||
prevMode = std::string((const char *) sqlite3_column_text(stmt, 0));
|
||
}
|
||
if (prevMode != mode &&
|
||
sqlite3_exec(db, ("pragma main.journal_mode = " + mode + ";").c_str(), 0, 0, 0) != SQLITE_OK)
|
||
SQLiteError::throw_(db, "setting journal mode");
|
||
|
||
if (mode == "wal") {
|
||
/* persist the WAL files when the db connection is closed. This allows
|
||
for read-only connections without write permissions on the
|
||
containing directory to succeed on a closed db. Setting the
|
||
journal_size_limit to 2^40 bytes results in the WAL files getting
|
||
truncated to 0 on exit and limits the on disk size of the WAL files
|
||
to 2^40 bytes following a checkpoint */
|
||
if (sqlite3_exec(db, "pragma main.journal_size_limit = 1099511627776;", 0, 0, 0) == SQLITE_OK) {
|
||
int enable = 1;
|
||
sqlite3_file_control(db, NULL, SQLITE_FCNTL_PERSIST_WAL, &enable);
|
||
}
|
||
}
|
||
|
||
/* Increase the auto-checkpoint interval to 40000 pages. This
|
||
seems enough to ensure that instantiating the NixOS system
|
||
derivation is done in a single fsync(). */
|
||
if (mode == "wal" && sqlite3_exec(db, "pragma wal_autocheckpoint = 40000;", 0, 0, 0) != SQLITE_OK)
|
||
SQLiteError::throw_(db, "setting autocheckpoint interval");
|
||
|
||
/* Initialise the database schema, if necessary. */
|
||
if (create) {
|
||
static const char schema[] =
|
||
#include "schema.sql.gen.hh"
|
||
;
|
||
db.exec(schema);
|
||
}
|
||
}
|
||
|
||
|
||
/* To improve purity, users may want to make the Nix store a read-only
|
||
bind mount. So make the Nix store writable for this process. */
|
||
void LocalStore::makeStoreWritable()
|
||
{
|
||
#if __linux__
|
||
if (!isRootUser()) return;
|
||
/* Check if /nix/store is on a read-only mount. */
|
||
struct statvfs stat;
|
||
if (statvfs(realStoreDir.get().c_str(), &stat) != 0)
|
||
throw SysError("getting info about the Nix store mount point");
|
||
|
||
if (stat.f_flag & ST_RDONLY) {
|
||
if (mount(0, realStoreDir.get().c_str(), "none", MS_REMOUNT | MS_BIND, 0) == -1)
|
||
throw SysError("remounting %1% writable", realStoreDir);
|
||
}
|
||
#endif
|
||
}
|
||
|
||
|
||
void LocalStore::registerDrvOutput(const Realisation & info, CheckSigsFlag checkSigs)
|
||
{
|
||
experimentalFeatureSettings.require(Xp::CaDerivations);
|
||
if (checkSigs == NoCheckSigs || !realisationIsUntrusted(info))
|
||
registerDrvOutput(info);
|
||
else
|
||
throw Error("cannot register realisation '%s' because it lacks a signature by a trusted key", info.outPath.to_string());
|
||
}
|
||
|
||
void LocalStore::registerDrvOutput(const Realisation & info)
|
||
{
|
||
experimentalFeatureSettings.require(Xp::CaDerivations);
|
||
retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
if (auto oldR = queryRealisation_(*state, info.id)) {
|
||
if (info.isCompatibleWith(*oldR)) {
|
||
auto combinedSignatures = oldR->signatures;
|
||
combinedSignatures.insert(info.signatures.begin(),
|
||
info.signatures.end());
|
||
state->stmts->UpdateRealisedOutput.use()
|
||
(concatStringsSep(" ", combinedSignatures))
|
||
(info.id.strHash())
|
||
(info.id.outputName)
|
||
.exec();
|
||
} else {
|
||
throw Error("Trying to register a realisation of '%s', but we already "
|
||
"have another one locally.\n"
|
||
"Local: %s\n"
|
||
"Remote: %s",
|
||
info.id.to_string(),
|
||
printStorePath(oldR->outPath),
|
||
printStorePath(info.outPath)
|
||
);
|
||
}
|
||
} else {
|
||
state->stmts->RegisterRealisedOutput.use()
|
||
(info.id.strHash())
|
||
(info.id.outputName)
|
||
(printStorePath(info.outPath))
|
||
(concatStringsSep(" ", info.signatures))
|
||
.exec();
|
||
}
|
||
for (auto & [outputId, depPath] : info.dependentRealisations) {
|
||
auto localRealisation = queryRealisationCore_(*state, outputId);
|
||
if (!localRealisation)
|
||
throw Error("unable to register the derivation '%s' as it "
|
||
"depends on the non existent '%s'",
|
||
info.id.to_string(), outputId.to_string());
|
||
if (localRealisation->second.outPath != depPath)
|
||
throw Error("unable to register the derivation '%s' as it "
|
||
"depends on a realisation of '%s' that doesn’t"
|
||
"match what we have locally",
|
||
info.id.to_string(), outputId.to_string());
|
||
state->stmts->AddRealisationReference.use()
|
||
(info.id.strHash())
|
||
(info.id.outputName)
|
||
(outputId.strHash())
|
||
(outputId.outputName)
|
||
.exec();
|
||
}
|
||
});
|
||
}
|
||
|
||
void LocalStore::cacheDrvOutputMapping(
|
||
State & state,
|
||
const uint64_t deriver,
|
||
const std::string & outputName,
|
||
const StorePath & output)
|
||
{
|
||
retrySQLite<void>([&]() {
|
||
state.stmts->AddDerivationOutput.use()
|
||
(deriver)
|
||
(outputName)
|
||
(printStorePath(output))
|
||
.exec();
|
||
});
|
||
}
|
||
|
||
|
||
uint64_t LocalStore::addValidPath(State & state,
|
||
const ValidPathInfo & info, bool checkOutputs)
|
||
{
|
||
if (info.ca.has_value() && !info.isContentAddressed(*this))
|
||
throw Error("cannot add path '%s' to the Nix store because it claims to be content-addressed but isn't",
|
||
printStorePath(info.path));
|
||
|
||
state.stmts->RegisterValidPath.use()
|
||
(printStorePath(info.path))
|
||
(info.narHash.to_string(HashFormat::Base16, true))
|
||
(info.registrationTime == 0 ? time(0) : info.registrationTime)
|
||
(info.deriver ? printStorePath(*info.deriver) : "", (bool) info.deriver)
|
||
(info.narSize, info.narSize != 0)
|
||
(info.ultimate ? 1 : 0, info.ultimate)
|
||
(concatStringsSep(" ", info.sigs), !info.sigs.empty())
|
||
(renderContentAddress(info.ca), (bool) info.ca)
|
||
.exec();
|
||
uint64_t id = state.db.getLastInsertedRowId();
|
||
|
||
/* If this is a derivation, then store the derivation outputs in
|
||
the database. This is useful for the garbage collector: it can
|
||
efficiently query whether a path is an output of some
|
||
derivation. */
|
||
if (info.path.isDerivation()) {
|
||
auto drv = readInvalidDerivation(info.path);
|
||
|
||
/* Verify that the output paths in the derivation are correct
|
||
(i.e., follow the scheme for computing output paths from
|
||
derivations). Note that if this throws an error, then the
|
||
DB transaction is rolled back, so the path validity
|
||
registration above is undone. */
|
||
if (checkOutputs) drv.checkInvariants(*this, info.path);
|
||
|
||
for (auto & i : drv.outputsAndOptPaths(*this)) {
|
||
/* Floating CA derivations have indeterminate output paths until
|
||
they are built, so don't register anything in that case */
|
||
if (i.second.second)
|
||
cacheDrvOutputMapping(state, id, i.first, *i.second.second);
|
||
}
|
||
}
|
||
|
||
{
|
||
auto state_(Store::state.lock());
|
||
state_->pathInfoCache.upsert(std::string(info.path.to_string()),
|
||
PathInfoCacheValue{ .value = std::make_shared<const ValidPathInfo>(info) });
|
||
}
|
||
|
||
return id;
|
||
}
|
||
|
||
|
||
void LocalStore::queryPathInfoUncached(const StorePath & path,
|
||
Callback<std::shared_ptr<const ValidPathInfo>> callback) noexcept
|
||
{
|
||
try {
|
||
callback(retrySQLite<std::shared_ptr<const ValidPathInfo>>([&]() {
|
||
auto state(_state.lock());
|
||
return queryPathInfoInternal(*state, path);
|
||
}));
|
||
|
||
} catch (...) { callback.rethrow(); }
|
||
}
|
||
|
||
|
||
std::shared_ptr<const ValidPathInfo> LocalStore::queryPathInfoInternal(State & state, const StorePath & path)
|
||
{
|
||
/* Get the path info. */
|
||
auto useQueryPathInfo(state.stmts->QueryPathInfo.use()(printStorePath(path)));
|
||
|
||
if (!useQueryPathInfo.next())
|
||
return std::shared_ptr<ValidPathInfo>();
|
||
|
||
auto id = useQueryPathInfo.getInt(0);
|
||
|
||
auto narHash = Hash::dummy;
|
||
try {
|
||
narHash = Hash::parseAnyPrefixed(useQueryPathInfo.getStr(1));
|
||
} catch (BadHash & e) {
|
||
throw Error("invalid-path entry for '%s': %s", printStorePath(path), e.what());
|
||
}
|
||
|
||
auto info = std::make_shared<ValidPathInfo>(path, narHash);
|
||
|
||
info->id = id;
|
||
|
||
info->registrationTime = useQueryPathInfo.getInt(2);
|
||
|
||
auto s = (const char *) sqlite3_column_text(state.stmts->QueryPathInfo, 3);
|
||
if (s) info->deriver = parseStorePath(s);
|
||
|
||
/* Note that narSize = NULL yields 0. */
|
||
info->narSize = useQueryPathInfo.getInt(4);
|
||
|
||
info->ultimate = useQueryPathInfo.getInt(5) == 1;
|
||
|
||
s = (const char *) sqlite3_column_text(state.stmts->QueryPathInfo, 6);
|
||
if (s) info->sigs = tokenizeString<StringSet>(s, " ");
|
||
|
||
s = (const char *) sqlite3_column_text(state.stmts->QueryPathInfo, 7);
|
||
if (s) info->ca = ContentAddress::parseOpt(s);
|
||
|
||
/* Get the references. */
|
||
auto useQueryReferences(state.stmts->QueryReferences.use()(info->id));
|
||
|
||
while (useQueryReferences.next())
|
||
info->references.insert(parseStorePath(useQueryReferences.getStr(0)));
|
||
|
||
return info;
|
||
}
|
||
|
||
|
||
/* Update path info in the database. */
|
||
void LocalStore::updatePathInfo(State & state, const ValidPathInfo & info)
|
||
{
|
||
state.stmts->UpdatePathInfo.use()
|
||
(info.narSize, info.narSize != 0)
|
||
(info.narHash.to_string(HashFormat::Base16, true))
|
||
(info.ultimate ? 1 : 0, info.ultimate)
|
||
(concatStringsSep(" ", info.sigs), !info.sigs.empty())
|
||
(renderContentAddress(info.ca), (bool) info.ca)
|
||
(printStorePath(info.path))
|
||
.exec();
|
||
}
|
||
|
||
|
||
uint64_t LocalStore::queryValidPathId(State & state, const StorePath & path)
|
||
{
|
||
auto use(state.stmts->QueryPathInfo.use()(printStorePath(path)));
|
||
if (!use.next())
|
||
throw InvalidPath("path '%s' is not valid", printStorePath(path));
|
||
return use.getInt(0);
|
||
}
|
||
|
||
|
||
bool LocalStore::isValidPath_(State & state, const StorePath & path)
|
||
{
|
||
return state.stmts->QueryPathInfo.use()(printStorePath(path)).next();
|
||
}
|
||
|
||
|
||
bool LocalStore::isValidPathUncached(const StorePath & path)
|
||
{
|
||
return retrySQLite<bool>([&]() {
|
||
auto state(_state.lock());
|
||
return isValidPath_(*state, path);
|
||
});
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::queryValidPaths(const StorePathSet & paths, SubstituteFlag maybeSubstitute)
|
||
{
|
||
StorePathSet res;
|
||
for (auto & i : paths)
|
||
if (isValidPath(i)) res.insert(i);
|
||
return res;
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::queryAllValidPaths()
|
||
{
|
||
return retrySQLite<StorePathSet>([&]() {
|
||
auto state(_state.lock());
|
||
auto use(state->stmts->QueryValidPaths.use());
|
||
StorePathSet res;
|
||
while (use.next()) res.insert(parseStorePath(use.getStr(0)));
|
||
return res;
|
||
});
|
||
}
|
||
|
||
|
||
void LocalStore::queryReferrers(State & state, const StorePath & path, StorePathSet & referrers)
|
||
{
|
||
auto useQueryReferrers(state.stmts->QueryReferrers.use()(printStorePath(path)));
|
||
|
||
while (useQueryReferrers.next())
|
||
referrers.insert(parseStorePath(useQueryReferrers.getStr(0)));
|
||
}
|
||
|
||
|
||
void LocalStore::queryReferrers(const StorePath & path, StorePathSet & referrers)
|
||
{
|
||
return retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
queryReferrers(*state, path, referrers);
|
||
});
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::queryValidDerivers(const StorePath & path)
|
||
{
|
||
return retrySQLite<StorePathSet>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
auto useQueryValidDerivers(state->stmts->QueryValidDerivers.use()(printStorePath(path)));
|
||
|
||
StorePathSet derivers;
|
||
while (useQueryValidDerivers.next())
|
||
derivers.insert(parseStorePath(useQueryValidDerivers.getStr(1)));
|
||
|
||
return derivers;
|
||
});
|
||
}
|
||
|
||
|
||
std::map<std::string, std::optional<StorePath>>
|
||
LocalStore::queryStaticPartialDerivationOutputMap(const StorePath & path)
|
||
{
|
||
return retrySQLite<std::map<std::string, std::optional<StorePath>>>([&]() {
|
||
auto state(_state.lock());
|
||
std::map<std::string, std::optional<StorePath>> outputs;
|
||
uint64_t drvId;
|
||
drvId = queryValidPathId(*state, path);
|
||
auto use(state->stmts->QueryDerivationOutputs.use()(drvId));
|
||
while (use.next())
|
||
outputs.insert_or_assign(
|
||
use.getStr(0), parseStorePath(use.getStr(1)));
|
||
|
||
return outputs;
|
||
});
|
||
}
|
||
|
||
std::optional<StorePath> LocalStore::queryPathFromHashPart(const std::string & hashPart)
|
||
{
|
||
if (hashPart.size() != StorePath::HashLen) throw Error("invalid hash part");
|
||
|
||
Path prefix = storeDir + "/" + hashPart;
|
||
|
||
return retrySQLite<std::optional<StorePath>>([&]() -> std::optional<StorePath> {
|
||
auto state(_state.lock());
|
||
|
||
auto useQueryPathFromHashPart(state->stmts->QueryPathFromHashPart.use()(prefix));
|
||
|
||
if (!useQueryPathFromHashPart.next()) return {};
|
||
|
||
const char * s = (const char *) sqlite3_column_text(state->stmts->QueryPathFromHashPart, 0);
|
||
if (s && prefix.compare(0, prefix.size(), s, prefix.size()) == 0)
|
||
return parseStorePath(s);
|
||
return {};
|
||
});
|
||
}
|
||
|
||
|
||
StorePathSet LocalStore::querySubstitutablePaths(const StorePathSet & paths)
|
||
{
|
||
if (!settings.useSubstitutes) return StorePathSet();
|
||
|
||
StorePathSet remaining;
|
||
for (auto & i : paths)
|
||
remaining.insert(i);
|
||
|
||
StorePathSet res;
|
||
|
||
for (auto & sub : getDefaultSubstituters()) {
|
||
if (remaining.empty()) break;
|
||
if (sub->storeDir != storeDir) continue;
|
||
if (!sub->wantMassQuery) continue;
|
||
|
||
auto valid = sub->queryValidPaths(remaining);
|
||
|
||
StorePathSet remaining2;
|
||
for (auto & path : remaining)
|
||
if (valid.count(path))
|
||
res.insert(path);
|
||
else
|
||
remaining2.insert(path);
|
||
|
||
std::swap(remaining, remaining2);
|
||
}
|
||
|
||
return res;
|
||
}
|
||
|
||
|
||
void LocalStore::registerValidPath(const ValidPathInfo & info)
|
||
{
|
||
registerValidPaths({{info.path, info}});
|
||
}
|
||
|
||
|
||
void LocalStore::registerValidPaths(const ValidPathInfos & infos)
|
||
{
|
||
#ifndef _WIN32
|
||
/* SQLite will fsync by default, but the new valid paths may not
|
||
be fsync-ed. So some may want to fsync them before registering
|
||
the validity, at the expense of some speed of the path
|
||
registering operation. */
|
||
if (settings.syncBeforeRegistering) sync();
|
||
#endif
|
||
|
||
return retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
SQLiteTxn txn(state->db);
|
||
StorePathSet paths;
|
||
|
||
for (auto & [_, i] : infos) {
|
||
assert(i.narHash.algo == HashAlgorithm::SHA256);
|
||
if (isValidPath_(*state, i.path))
|
||
updatePathInfo(*state, i);
|
||
else
|
||
addValidPath(*state, i, false);
|
||
paths.insert(i.path);
|
||
}
|
||
|
||
for (auto & [_, i] : infos) {
|
||
auto referrer = queryValidPathId(*state, i.path);
|
||
for (auto & j : i.references)
|
||
state->stmts->AddReference.use()(referrer)(queryValidPathId(*state, j)).exec();
|
||
}
|
||
|
||
/* Check that the derivation outputs are correct. We can't do
|
||
this in addValidPath() above, because the references might
|
||
not be valid yet. */
|
||
for (auto & [_, i] : infos)
|
||
if (i.path.isDerivation()) {
|
||
// FIXME: inefficient; we already loaded the derivation in addValidPath().
|
||
readInvalidDerivation(i.path).checkInvariants(*this, i.path);
|
||
}
|
||
|
||
/* Do a topological sort of the paths. This will throw an
|
||
error if a cycle is detected and roll back the
|
||
transaction. Cycles can only occur when a derivation
|
||
has multiple outputs. */
|
||
topoSort(paths,
|
||
{[&](const StorePath & path) {
|
||
auto i = infos.find(path);
|
||
return i == infos.end() ? StorePathSet() : i->second.references;
|
||
}},
|
||
{[&](const StorePath & path, const StorePath & parent) {
|
||
return BuildError(
|
||
"cycle detected in the references of '%s' from '%s'",
|
||
printStorePath(path),
|
||
printStorePath(parent));
|
||
}});
|
||
|
||
txn.commit();
|
||
});
|
||
}
|
||
|
||
|
||
/* Invalidate a path. The caller is responsible for checking that
|
||
there are no referrers. */
|
||
void LocalStore::invalidatePath(State & state, const StorePath & path)
|
||
{
|
||
debug("invalidating path '%s'", printStorePath(path));
|
||
|
||
state.stmts->InvalidatePath.use()(printStorePath(path)).exec();
|
||
|
||
/* Note that the foreign key constraints on the Refs table take
|
||
care of deleting the references entries for `path'. */
|
||
|
||
{
|
||
auto state_(Store::state.lock());
|
||
state_->pathInfoCache.erase(std::string(path.to_string()));
|
||
}
|
||
}
|
||
|
||
const PublicKeys & LocalStore::getPublicKeys()
|
||
{
|
||
auto state(_state.lock());
|
||
if (!state->publicKeys)
|
||
state->publicKeys = std::make_unique<PublicKeys>(getDefaultPublicKeys());
|
||
return *state->publicKeys;
|
||
}
|
||
|
||
bool LocalStore::pathInfoIsUntrusted(const ValidPathInfo & info)
|
||
{
|
||
return requireSigs && !info.checkSignatures(*this, getPublicKeys());
|
||
}
|
||
|
||
bool LocalStore::realisationIsUntrusted(const Realisation & realisation)
|
||
{
|
||
return requireSigs && !realisation.checkSignatures(getPublicKeys());
|
||
}
|
||
|
||
void LocalStore::addToStore(const ValidPathInfo & info, Source & source,
|
||
RepairFlag repair, CheckSigsFlag checkSigs)
|
||
{
|
||
if (checkSigs && pathInfoIsUntrusted(info))
|
||
throw Error("cannot add path '%s' because it lacks a signature by a trusted key", printStorePath(info.path));
|
||
|
||
{
|
||
/* In case we are not interested in reading the NAR: discard it. */
|
||
bool narRead = false;
|
||
Finally cleanup = [&]() {
|
||
if (!narRead) {
|
||
NullFileSystemObjectSink sink;
|
||
try {
|
||
parseDump(sink, source);
|
||
} catch (...) {
|
||
// TODO: should Interrupted be handled here?
|
||
ignoreExceptionInDestructor();
|
||
}
|
||
}
|
||
};
|
||
|
||
addTempRoot(info.path);
|
||
|
||
if (repair || !isValidPath(info.path)) {
|
||
|
||
PathLocks outputLock;
|
||
|
||
auto realPath = Store::toRealPath(info.path);
|
||
|
||
/* Lock the output path. But don't lock if we're being called
|
||
from a build hook (whose parent process already acquired a
|
||
lock on this path). */
|
||
if (!locksHeld.count(printStorePath(info.path)))
|
||
outputLock.lockPaths({realPath});
|
||
|
||
if (repair || !isValidPath(info.path)) {
|
||
|
||
deletePath(realPath);
|
||
|
||
/* While restoring the path from the NAR, compute the hash
|
||
of the NAR. */
|
||
HashSink hashSink(HashAlgorithm::SHA256);
|
||
|
||
TeeSource wrapperSource { source, hashSink };
|
||
|
||
narRead = true;
|
||
restorePath(realPath, wrapperSource, settings.fsyncStorePaths);
|
||
|
||
auto hashResult = hashSink.finish();
|
||
|
||
if (hashResult.first != info.narHash)
|
||
throw Error("hash mismatch importing path '%s';\n specified: %s\n got: %s",
|
||
printStorePath(info.path), info.narHash.to_string(HashFormat::Nix32, true), hashResult.first.to_string(HashFormat::Nix32, true));
|
||
|
||
if (hashResult.second != info.narSize)
|
||
throw Error("size mismatch importing path '%s';\n specified: %s\n got: %s",
|
||
printStorePath(info.path), info.narSize, hashResult.second);
|
||
|
||
if (info.ca) {
|
||
auto & specified = *info.ca;
|
||
auto actualHash = ({
|
||
auto accessor = getFSAccessor(false);
|
||
CanonPath path { printStorePath(info.path) };
|
||
Hash h { HashAlgorithm::SHA256 }; // throwaway def to appease C++
|
||
auto fim = specified.method.getFileIngestionMethod();
|
||
switch (fim) {
|
||
case FileIngestionMethod::Flat:
|
||
case FileIngestionMethod::NixArchive:
|
||
{
|
||
HashModuloSink caSink {
|
||
specified.hash.algo,
|
||
std::string { info.path.hashPart() },
|
||
};
|
||
dumpPath({accessor, path}, caSink, (FileSerialisationMethod) fim);
|
||
h = caSink.finish().first;
|
||
break;
|
||
}
|
||
case FileIngestionMethod::Git:
|
||
h = git::dumpHash(specified.hash.algo, {accessor, path}).hash;
|
||
break;
|
||
}
|
||
ContentAddress {
|
||
.method = specified.method,
|
||
.hash = std::move(h),
|
||
};
|
||
});
|
||
if (specified.hash != actualHash.hash) {
|
||
throw Error("ca hash mismatch importing path '%s';\n specified: %s\n got: %s",
|
||
printStorePath(info.path),
|
||
specified.hash.to_string(HashFormat::Nix32, true),
|
||
actualHash.hash.to_string(HashFormat::Nix32, true));
|
||
}
|
||
}
|
||
|
||
autoGC();
|
||
|
||
canonicalisePathMetaData(realPath);
|
||
|
||
optimisePath(realPath, repair); // FIXME: combine with hashPath()
|
||
|
||
if (settings.fsyncStorePaths) {
|
||
recursiveSync(realPath);
|
||
syncParent(realPath);
|
||
}
|
||
|
||
registerValidPath(info);
|
||
}
|
||
|
||
outputLock.setDeletion(true);
|
||
}
|
||
}
|
||
|
||
// In case `cleanup` ignored an `Interrupted` exception
|
||
checkInterrupt();
|
||
}
|
||
|
||
|
||
StorePath LocalStore::addToStoreFromDump(
|
||
Source & source0,
|
||
std::string_view name,
|
||
FileSerialisationMethod dumpMethod,
|
||
ContentAddressMethod hashMethod,
|
||
HashAlgorithm hashAlgo,
|
||
const StorePathSet & references,
|
||
RepairFlag repair)
|
||
{
|
||
/* For computing the store path. */
|
||
auto hashSink = std::make_unique<HashSink>(hashAlgo);
|
||
TeeSource source { source0, *hashSink };
|
||
|
||
/* Read the source path into memory, but only if it's up to
|
||
narBufferSize bytes. If it's larger, write it to a temporary
|
||
location in the Nix store. If the subsequently computed
|
||
destination store path is already valid, we just delete the
|
||
temporary path. Otherwise, we move it to the destination store
|
||
path. */
|
||
bool inMemory = false;
|
||
|
||
struct Free {
|
||
void operator()(void* v) { free(v); }
|
||
};
|
||
std::unique_ptr<char, Free> dumpBuffer(nullptr);
|
||
std::string_view dump;
|
||
|
||
/* Fill out buffer, and decide whether we are working strictly in
|
||
memory based on whether we break out because the buffer is full
|
||
or the original source is empty */
|
||
while (dump.size() < settings.narBufferSize) {
|
||
auto oldSize = dump.size();
|
||
constexpr size_t chunkSize = 65536;
|
||
auto want = std::min(chunkSize, settings.narBufferSize - oldSize);
|
||
if (auto tmp = realloc(dumpBuffer.get(), oldSize + want)) {
|
||
dumpBuffer.release();
|
||
dumpBuffer.reset((char*) tmp);
|
||
} else {
|
||
throw std::bad_alloc();
|
||
}
|
||
auto got = 0;
|
||
Finally cleanup([&]() {
|
||
dump = {dumpBuffer.get(), dump.size() + got};
|
||
});
|
||
try {
|
||
got = source.read(dumpBuffer.get() + oldSize, want);
|
||
} catch (EndOfFile &) {
|
||
inMemory = true;
|
||
break;
|
||
}
|
||
}
|
||
|
||
std::unique_ptr<AutoDelete> delTempDir;
|
||
std::filesystem::path tempPath;
|
||
std::filesystem::path tempDir;
|
||
AutoCloseFD tempDirFd;
|
||
|
||
bool methodsMatch = static_cast<FileIngestionMethod>(dumpMethod) == hashMethod.getFileIngestionMethod();
|
||
|
||
/* If the methods don't match, our streaming hash of the dump is the
|
||
wrong sort, and we need to rehash. */
|
||
bool inMemoryAndDontNeedRestore = inMemory && methodsMatch;
|
||
|
||
if (!inMemoryAndDontNeedRestore) {
|
||
/* Drain what we pulled so far, and then keep on pulling */
|
||
StringSource dumpSource { dump };
|
||
ChainSource bothSource { dumpSource, source };
|
||
|
||
std::tie(tempDir, tempDirFd) = createTempDirInStore();
|
||
delTempDir = std::make_unique<AutoDelete>(tempDir);
|
||
tempPath = tempDir / "x";
|
||
|
||
restorePath(tempPath.string(), bothSource, dumpMethod, settings.fsyncStorePaths);
|
||
|
||
dumpBuffer.reset();
|
||
dump = {};
|
||
}
|
||
|
||
auto [dumpHash, size] = hashSink->finish();
|
||
|
||
auto desc = ContentAddressWithReferences::fromParts(
|
||
hashMethod,
|
||
methodsMatch
|
||
? dumpHash
|
||
: hashPath(
|
||
PosixSourceAccessor::createAtRoot(tempPath),
|
||
hashMethod.getFileIngestionMethod(), hashAlgo).first,
|
||
{
|
||
.others = references,
|
||
// caller is not capable of creating a self-reference, because this is content-addressed without modulus
|
||
.self = false,
|
||
});
|
||
|
||
auto dstPath = makeFixedOutputPathFromCA(name, desc);
|
||
|
||
addTempRoot(dstPath);
|
||
|
||
if (repair || !isValidPath(dstPath)) {
|
||
|
||
/* The first check above is an optimisation to prevent
|
||
unnecessary lock acquisition. */
|
||
|
||
auto realPath = Store::toRealPath(dstPath);
|
||
|
||
PathLocks outputLock({realPath});
|
||
|
||
if (repair || !isValidPath(dstPath)) {
|
||
|
||
deletePath(realPath);
|
||
|
||
autoGC();
|
||
|
||
if (inMemoryAndDontNeedRestore) {
|
||
StringSource dumpSource { dump };
|
||
/* Restore from the buffer in memory. */
|
||
auto fim = hashMethod.getFileIngestionMethod();
|
||
switch (fim) {
|
||
case FileIngestionMethod::Flat:
|
||
case FileIngestionMethod::NixArchive:
|
||
restorePath(realPath, dumpSource, (FileSerialisationMethod) fim, settings.fsyncStorePaths);
|
||
break;
|
||
case FileIngestionMethod::Git:
|
||
// doesn't correspond to serialization method, so
|
||
// this should be unreachable
|
||
assert(false);
|
||
}
|
||
} else {
|
||
/* Move the temporary path we restored above. */
|
||
moveFile(tempPath.string(), realPath);
|
||
}
|
||
|
||
/* For computing the nar hash. In recursive SHA-256 mode, this
|
||
is the same as the store hash, so no need to do it again. */
|
||
auto narHash = std::pair { dumpHash, size };
|
||
if (dumpMethod != FileSerialisationMethod::NixArchive || hashAlgo != HashAlgorithm::SHA256) {
|
||
HashSink narSink { HashAlgorithm::SHA256 };
|
||
dumpPath(realPath, narSink);
|
||
narHash = narSink.finish();
|
||
}
|
||
|
||
canonicalisePathMetaData(realPath); // FIXME: merge into restorePath
|
||
|
||
optimisePath(realPath, repair);
|
||
|
||
if (settings.fsyncStorePaths) {
|
||
recursiveSync(realPath);
|
||
syncParent(realPath);
|
||
}
|
||
|
||
ValidPathInfo info {
|
||
*this,
|
||
name,
|
||
std::move(desc),
|
||
narHash.first
|
||
};
|
||
info.narSize = narHash.second;
|
||
registerValidPath(info);
|
||
}
|
||
|
||
outputLock.setDeletion(true);
|
||
}
|
||
|
||
return dstPath;
|
||
}
|
||
|
||
|
||
/* Create a temporary directory in the store that won't be
|
||
garbage-collected until the returned FD is closed. */
|
||
std::pair<std::filesystem::path, AutoCloseFD> LocalStore::createTempDirInStore()
|
||
{
|
||
std::filesystem::path tmpDirFn;
|
||
AutoCloseFD tmpDirFd;
|
||
bool lockedByUs = false;
|
||
do {
|
||
/* There is a slight possibility that `tmpDir' gets deleted by
|
||
the GC between createTempDir() and when we acquire a lock on it.
|
||
We'll repeat until 'tmpDir' exists and we've locked it. */
|
||
tmpDirFn = createTempDir(realStoreDir, "tmp");
|
||
tmpDirFd = openDirectory(tmpDirFn);
|
||
if (!tmpDirFd) {
|
||
continue;
|
||
}
|
||
lockedByUs = lockFile(tmpDirFd.get(), ltWrite, true);
|
||
} while (!pathExists(tmpDirFn.string()) || !lockedByUs);
|
||
return {tmpDirFn, std::move(tmpDirFd)};
|
||
}
|
||
|
||
|
||
void LocalStore::invalidatePathChecked(const StorePath & path)
|
||
{
|
||
retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
SQLiteTxn txn(state->db);
|
||
|
||
if (isValidPath_(*state, path)) {
|
||
StorePathSet referrers; queryReferrers(*state, path, referrers);
|
||
referrers.erase(path); /* ignore self-references */
|
||
if (!referrers.empty())
|
||
throw PathInUse("cannot delete path '%s' because it is in use by %s",
|
||
printStorePath(path), showPaths(referrers));
|
||
invalidatePath(*state, path);
|
||
}
|
||
|
||
txn.commit();
|
||
});
|
||
}
|
||
|
||
|
||
bool LocalStore::verifyStore(bool checkContents, RepairFlag repair)
|
||
{
|
||
printInfo("reading the Nix store...");
|
||
|
||
/* Acquire the global GC lock to get a consistent snapshot of
|
||
existing and valid paths. */
|
||
auto fdGCLock = openGCLock();
|
||
FdLock gcLock(fdGCLock.get(), ltRead, true, "waiting for the big garbage collector lock...");
|
||
|
||
auto [errors, validPaths] = verifyAllValidPaths(repair);
|
||
|
||
/* Optionally, check the content hashes (slow). */
|
||
if (checkContents) {
|
||
|
||
printInfo("checking link hashes...");
|
||
|
||
for (auto & link : std::filesystem::directory_iterator{linksDir}) {
|
||
checkInterrupt();
|
||
auto name = link.path().filename();
|
||
printMsg(lvlTalkative, "checking contents of '%s'", name);
|
||
PosixSourceAccessor accessor;
|
||
std::string hash = hashPath(
|
||
PosixSourceAccessor::createAtRoot(link.path()),
|
||
FileIngestionMethod::NixArchive, HashAlgorithm::SHA256).first.to_string(HashFormat::Nix32, false);
|
||
if (hash != name.string()) {
|
||
printError("link '%s' was modified! expected hash '%s', got '%s'",
|
||
link.path(), name, hash);
|
||
if (repair) {
|
||
std::filesystem::remove(link.path());
|
||
printInfo("removed link '%s'", link.path());
|
||
} else {
|
||
errors = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
printInfo("checking store hashes...");
|
||
|
||
Hash nullHash(HashAlgorithm::SHA256);
|
||
|
||
for (auto & i : validPaths) {
|
||
try {
|
||
auto info = std::const_pointer_cast<ValidPathInfo>(std::shared_ptr<const ValidPathInfo>(queryPathInfo(i)));
|
||
|
||
/* Check the content hash (optionally - slow). */
|
||
printMsg(lvlTalkative, "checking contents of '%s'", printStorePath(i));
|
||
|
||
auto hashSink = HashSink(info->narHash.algo);
|
||
|
||
dumpPath(Store::toRealPath(i), hashSink);
|
||
auto current = hashSink.finish();
|
||
|
||
if (info->narHash != nullHash && info->narHash != current.first) {
|
||
printError("path '%s' was modified! expected hash '%s', got '%s'",
|
||
printStorePath(i), info->narHash.to_string(HashFormat::Nix32, true), current.first.to_string(HashFormat::Nix32, true));
|
||
if (repair) repairPath(i); else errors = true;
|
||
} else {
|
||
|
||
bool update = false;
|
||
|
||
/* Fill in missing hashes. */
|
||
if (info->narHash == nullHash) {
|
||
printInfo("fixing missing hash on '%s'", printStorePath(i));
|
||
info->narHash = current.first;
|
||
update = true;
|
||
}
|
||
|
||
/* Fill in missing narSize fields (from old stores). */
|
||
if (info->narSize == 0) {
|
||
printInfo("updating size field on '%s' to %s", printStorePath(i), current.second);
|
||
info->narSize = current.second;
|
||
update = true;
|
||
}
|
||
|
||
if (update) {
|
||
auto state(_state.lock());
|
||
updatePathInfo(*state, *info);
|
||
}
|
||
|
||
}
|
||
|
||
} catch (Error & e) {
|
||
/* It's possible that the path got GC'ed, so ignore
|
||
errors on invalid paths. */
|
||
if (isValidPath(i))
|
||
logError(e.info());
|
||
else
|
||
warn(e.msg());
|
||
errors = true;
|
||
}
|
||
}
|
||
}
|
||
|
||
return errors;
|
||
}
|
||
|
||
|
||
LocalStore::VerificationResult LocalStore::verifyAllValidPaths(RepairFlag repair)
|
||
{
|
||
StorePathSet storePathsInStoreDir;
|
||
/* Why aren't we using `queryAllValidPaths`? Because that would
|
||
tell us about all the paths than the database knows about. Here we
|
||
want to know about all the store paths in the store directory,
|
||
regardless of what the database thinks.
|
||
|
||
We will end up cross-referencing these two sources of truth (the
|
||
database and the filesystem) in the loop below, in order to catch
|
||
invalid states.
|
||
*/
|
||
for (auto & i : std::filesystem::directory_iterator{realStoreDir.to_string()}) {
|
||
checkInterrupt();
|
||
try {
|
||
storePathsInStoreDir.insert({i.path().filename().string()});
|
||
} catch (BadStorePath &) { }
|
||
}
|
||
|
||
/* Check whether all valid paths actually exist. */
|
||
printInfo("checking path existence...");
|
||
|
||
StorePathSet done;
|
||
|
||
auto existsInStoreDir = [&](const StorePath & storePath) {
|
||
return storePathsInStoreDir.count(storePath);
|
||
};
|
||
|
||
bool errors = false;
|
||
StorePathSet validPaths;
|
||
|
||
for (auto & i : queryAllValidPaths())
|
||
verifyPath(i, existsInStoreDir, done, validPaths, repair, errors);
|
||
|
||
return {
|
||
.errors = errors,
|
||
.validPaths = validPaths,
|
||
};
|
||
}
|
||
|
||
|
||
void LocalStore::verifyPath(const StorePath & path, std::function<bool(const StorePath &)> existsInStoreDir,
|
||
StorePathSet & done, StorePathSet & validPaths, RepairFlag repair, bool & errors)
|
||
{
|
||
checkInterrupt();
|
||
|
||
if (!done.insert(path).second) return;
|
||
|
||
if (!existsInStoreDir(path)) {
|
||
/* Check any referrers first. If we can invalidate them
|
||
first, then we can invalidate this path as well. */
|
||
bool canInvalidate = true;
|
||
StorePathSet referrers; queryReferrers(path, referrers);
|
||
for (auto & i : referrers)
|
||
if (i != path) {
|
||
verifyPath(i, existsInStoreDir, done, validPaths, repair, errors);
|
||
if (validPaths.count(i))
|
||
canInvalidate = false;
|
||
}
|
||
|
||
auto pathS = printStorePath(path);
|
||
|
||
if (canInvalidate) {
|
||
printInfo("path '%s' disappeared, removing from database...", pathS);
|
||
auto state(_state.lock());
|
||
invalidatePath(*state, path);
|
||
} else {
|
||
printError("path '%s' disappeared, but it still has valid referrers!", pathS);
|
||
if (repair)
|
||
try {
|
||
repairPath(path);
|
||
} catch (Error & e) {
|
||
logWarning(e.info());
|
||
errors = true;
|
||
}
|
||
else errors = true;
|
||
}
|
||
|
||
return;
|
||
}
|
||
|
||
validPaths.insert(std::move(path));
|
||
}
|
||
|
||
|
||
unsigned int LocalStore::getProtocol()
|
||
{
|
||
return PROTOCOL_VERSION;
|
||
}
|
||
|
||
std::optional<TrustedFlag> LocalStore::isTrustedClient()
|
||
{
|
||
return Trusted;
|
||
}
|
||
|
||
|
||
void LocalStore::vacuumDB()
|
||
{
|
||
auto state(_state.lock());
|
||
state->db.exec("vacuum");
|
||
}
|
||
|
||
|
||
void LocalStore::addSignatures(const StorePath & storePath, const StringSet & sigs)
|
||
{
|
||
retrySQLite<void>([&]() {
|
||
auto state(_state.lock());
|
||
|
||
SQLiteTxn txn(state->db);
|
||
|
||
auto info = std::const_pointer_cast<ValidPathInfo>(queryPathInfoInternal(*state, storePath));
|
||
|
||
info->sigs.insert(sigs.begin(), sigs.end());
|
||
|
||
updatePathInfo(*state, *info);
|
||
|
||
txn.commit();
|
||
});
|
||
}
|
||
|
||
|
||
void LocalStore::signRealisation(Realisation & realisation)
|
||
{
|
||
// FIXME: keep secret keys in memory.
|
||
|
||
auto secretKeyFiles = settings.secretKeyFiles;
|
||
|
||
for (auto & secretKeyFile : secretKeyFiles.get()) {
|
||
SecretKey secretKey(readFile(secretKeyFile));
|
||
LocalSigner signer(std::move(secretKey));
|
||
realisation.sign(signer);
|
||
}
|
||
}
|
||
|
||
void LocalStore::signPathInfo(ValidPathInfo & info)
|
||
{
|
||
// FIXME: keep secret keys in memory.
|
||
|
||
auto secretKeyFiles = settings.secretKeyFiles;
|
||
|
||
for (auto & secretKeyFile : secretKeyFiles.get()) {
|
||
SecretKey secretKey(readFile(secretKeyFile));
|
||
LocalSigner signer(std::move(secretKey));
|
||
info.sign(*this, signer);
|
||
}
|
||
}
|
||
|
||
|
||
std::optional<std::pair<int64_t, Realisation>> LocalStore::queryRealisationCore_(
|
||
LocalStore::State & state,
|
||
const DrvOutput & id)
|
||
{
|
||
auto useQueryRealisedOutput(
|
||
state.stmts->QueryRealisedOutput.use()
|
||
(id.strHash())
|
||
(id.outputName));
|
||
if (!useQueryRealisedOutput.next())
|
||
return std::nullopt;
|
||
auto realisationDbId = useQueryRealisedOutput.getInt(0);
|
||
auto outputPath = parseStorePath(useQueryRealisedOutput.getStr(1));
|
||
auto signatures =
|
||
tokenizeString<StringSet>(useQueryRealisedOutput.getStr(2));
|
||
|
||
return {{
|
||
realisationDbId,
|
||
Realisation{
|
||
.id = id,
|
||
.outPath = outputPath,
|
||
.signatures = signatures,
|
||
}
|
||
}};
|
||
}
|
||
|
||
std::optional<const Realisation> LocalStore::queryRealisation_(
|
||
LocalStore::State & state,
|
||
const DrvOutput & id)
|
||
{
|
||
auto maybeCore = queryRealisationCore_(state, id);
|
||
if (!maybeCore)
|
||
return std::nullopt;
|
||
auto [realisationDbId, res] = *maybeCore;
|
||
|
||
std::map<DrvOutput, StorePath> dependentRealisations;
|
||
auto useRealisationRefs(
|
||
state.stmts->QueryRealisationReferences.use()
|
||
(realisationDbId));
|
||
while (useRealisationRefs.next()) {
|
||
auto depId = DrvOutput {
|
||
Hash::parseAnyPrefixed(useRealisationRefs.getStr(0)),
|
||
useRealisationRefs.getStr(1),
|
||
};
|
||
auto dependentRealisation = queryRealisationCore_(state, depId);
|
||
assert(dependentRealisation); // Enforced by the db schema
|
||
auto outputPath = dependentRealisation->second.outPath;
|
||
dependentRealisations.insert({depId, outputPath});
|
||
}
|
||
|
||
res.dependentRealisations = dependentRealisations;
|
||
|
||
return { res };
|
||
}
|
||
|
||
void LocalStore::queryRealisationUncached(const DrvOutput & id,
|
||
Callback<std::shared_ptr<const Realisation>> callback) noexcept
|
||
{
|
||
try {
|
||
auto maybeRealisation
|
||
= retrySQLite<std::optional<const Realisation>>([&]() {
|
||
auto state(_state.lock());
|
||
return queryRealisation_(*state, id);
|
||
});
|
||
if (maybeRealisation)
|
||
callback(
|
||
std::make_shared<const Realisation>(maybeRealisation.value()));
|
||
else
|
||
callback(nullptr);
|
||
|
||
} catch (...) {
|
||
callback.rethrow();
|
||
}
|
||
}
|
||
|
||
void LocalStore::addBuildLog(const StorePath & drvPath, std::string_view log)
|
||
{
|
||
assert(drvPath.isDerivation());
|
||
|
||
auto baseName = drvPath.to_string();
|
||
|
||
auto logPath = fmt("%s/%s/%s/%s.bz2", logDir, drvsLogDir, baseName.substr(0, 2), baseName.substr(2));
|
||
|
||
if (pathExists(logPath)) return;
|
||
|
||
createDirs(dirOf(logPath));
|
||
|
||
auto tmpFile = fmt("%s.tmp.%d", logPath, getpid());
|
||
|
||
writeFile(tmpFile, compress("bzip2", log));
|
||
|
||
std::filesystem::rename(tmpFile, logPath);
|
||
}
|
||
|
||
std::optional<std::string> LocalStore::getVersion()
|
||
{
|
||
return nixVersion;
|
||
}
|
||
|
||
static RegisterStoreImplementation<LocalStore, LocalStoreConfig> regLocalStore;
|
||
|
||
} // namespace nix
|