2021-02-26 17:20:33 +02:00
|
|
|
|
#include "local-derivation-goal.hh"
|
2022-03-01 20:31:36 +02:00
|
|
|
|
#include "gc-store.hh"
|
2021-02-23 08:26:35 +02:00
|
|
|
|
#include "hook-instance.hh"
|
|
|
|
|
#include "worker.hh"
|
|
|
|
|
#include "builtins.hh"
|
|
|
|
|
#include "builtins/buildenv.hh"
|
|
|
|
|
#include "references.hh"
|
|
|
|
|
#include "finally.hh"
|
|
|
|
|
#include "util.hh"
|
|
|
|
|
#include "archive.hh"
|
|
|
|
|
#include "compression.hh"
|
|
|
|
|
#include "daemon.hh"
|
|
|
|
|
#include "worker-protocol.hh"
|
|
|
|
|
#include "topo-sort.hh"
|
|
|
|
|
#include "callback.hh"
|
2022-05-04 12:22:06 +03:00
|
|
|
|
#include "json-utils.hh"
|
2022-11-18 11:39:28 +02:00
|
|
|
|
#include "cgroup.hh"
|
2022-12-23 17:28:26 +02:00
|
|
|
|
#include "personality.hh"
|
2023-01-25 18:31:27 +02:00
|
|
|
|
#include "namespaces.hh"
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
#include <regex>
|
|
|
|
|
#include <queue>
|
|
|
|
|
|
|
|
|
|
#include <sys/un.h>
|
|
|
|
|
#include <fcntl.h>
|
|
|
|
|
#include <termios.h>
|
|
|
|
|
#include <unistd.h>
|
|
|
|
|
#include <sys/mman.h>
|
|
|
|
|
#include <sys/resource.h>
|
2021-09-22 13:14:50 +03:00
|
|
|
|
#include <sys/socket.h>
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
#if HAVE_STATVFS
|
|
|
|
|
#include <sys/statvfs.h>
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Includes required for chroot support. */
|
|
|
|
|
#if __linux__
|
|
|
|
|
#include <sys/ioctl.h>
|
|
|
|
|
#include <net/if.h>
|
|
|
|
|
#include <netinet/ip.h>
|
|
|
|
|
#include <sys/mman.h>
|
|
|
|
|
#include <sched.h>
|
|
|
|
|
#include <sys/param.h>
|
|
|
|
|
#include <sys/mount.h>
|
|
|
|
|
#include <sys/syscall.h>
|
|
|
|
|
#if HAVE_SECCOMP
|
|
|
|
|
#include <seccomp.h>
|
|
|
|
|
#endif
|
|
|
|
|
#define pivot_root(new_root, put_old) (syscall(SYS_pivot_root, new_root, put_old))
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
#include <spawn.h>
|
|
|
|
|
#include <sys/sysctl.h>
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
#include <pwd.h>
|
|
|
|
|
#include <grp.h>
|
2022-11-16 17:49:49 +02:00
|
|
|
|
#include <iostream>
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
namespace nix {
|
|
|
|
|
|
|
|
|
|
void handleDiffHook(
|
|
|
|
|
uid_t uid, uid_t gid,
|
|
|
|
|
const Path & tryA, const Path & tryB,
|
|
|
|
|
const Path & drvPath, const Path & tmpDir)
|
|
|
|
|
{
|
|
|
|
|
auto diffHook = settings.diffHook;
|
|
|
|
|
if (diffHook != "" && settings.runDiffHook) {
|
|
|
|
|
try {
|
2021-09-14 09:19:41 +03:00
|
|
|
|
auto diffRes = runProgram(RunOptions {
|
2021-09-14 00:22:09 +03:00
|
|
|
|
.program = diffHook,
|
|
|
|
|
.searchPath = true,
|
|
|
|
|
.args = {tryA, tryB, drvPath, tmpDir},
|
|
|
|
|
.uid = uid,
|
|
|
|
|
.gid = gid,
|
|
|
|
|
.chdir = "/"
|
|
|
|
|
});
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (!statusOk(diffRes.first))
|
|
|
|
|
throw ExecError(diffRes.first,
|
|
|
|
|
"diff-hook program '%1%' %2%",
|
|
|
|
|
diffHook,
|
|
|
|
|
statusToString(diffRes.first));
|
|
|
|
|
|
|
|
|
|
if (diffRes.second != "")
|
|
|
|
|
printError(chomp(diffRes.second));
|
|
|
|
|
} catch (Error & error) {
|
|
|
|
|
ErrorInfo ei = error.info();
|
|
|
|
|
// FIXME: wrap errors.
|
|
|
|
|
ei.msg = hintfmt("diff hook execution failed: %s", ei.msg.str());
|
|
|
|
|
logError(ei);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
const Path LocalDerivationGoal::homeDir = "/homeless-shelter";
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
LocalDerivationGoal::~LocalDerivationGoal()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
/* Careful: we should never ever throw an exception from a
|
|
|
|
|
destructor. */
|
2021-02-26 17:20:33 +02:00
|
|
|
|
try { deleteTmpDir(false); } catch (...) { ignoreException(); }
|
2021-02-23 08:26:35 +02:00
|
|
|
|
try { killChild(); } catch (...) { ignoreException(); }
|
|
|
|
|
try { stopDaemon(); } catch (...) { ignoreException(); }
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
inline bool LocalDerivationGoal::needsHashRewrite()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
#if __linux__
|
|
|
|
|
return !useChroot;
|
|
|
|
|
#else
|
|
|
|
|
/* Darwin requires hash rewriting even when sandboxing is enabled. */
|
|
|
|
|
return true;
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
LocalStore & LocalDerivationGoal::getLocalStore()
|
|
|
|
|
{
|
|
|
|
|
auto p = dynamic_cast<LocalStore *>(&worker.store);
|
|
|
|
|
assert(p);
|
|
|
|
|
return *p;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void LocalDerivationGoal::killChild()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
if (pid != -1) {
|
|
|
|
|
worker.childTerminated(this);
|
|
|
|
|
|
2022-11-18 11:39:28 +02:00
|
|
|
|
/* If we're using a build user, then there is a tricky race
|
|
|
|
|
condition: if we kill the build user before the child has
|
|
|
|
|
done its setuid() to the build user uid, then it won't be
|
|
|
|
|
killed, and we'll potentially lock up in pid.wait(). So
|
|
|
|
|
also send a conventional kill to the child. */
|
|
|
|
|
::kill(-pid, SIGKILL); /* ignore the result */
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-11-18 14:40:59 +02:00
|
|
|
|
killSandbox(true);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-11-18 11:39:28 +02:00
|
|
|
|
pid.wait();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
DerivationGoal::killChild();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2022-11-18 14:40:59 +02:00
|
|
|
|
void LocalDerivationGoal::killSandbox(bool getStats)
|
2022-11-18 11:39:28 +02:00
|
|
|
|
{
|
|
|
|
|
if (cgroup) {
|
2022-11-21 12:45:41 +02:00
|
|
|
|
#if __linux__
|
2022-11-18 14:40:59 +02:00
|
|
|
|
auto stats = destroyCgroup(*cgroup);
|
|
|
|
|
if (getStats) {
|
|
|
|
|
buildResult.cpuUser = stats.cpuUser;
|
|
|
|
|
buildResult.cpuSystem = stats.cpuSystem;
|
|
|
|
|
}
|
2022-11-21 12:45:41 +02:00
|
|
|
|
#else
|
|
|
|
|
abort();
|
|
|
|
|
#endif
|
2022-11-18 11:39:28 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
else if (buildUser) {
|
|
|
|
|
auto uid = buildUser->getUID();
|
|
|
|
|
assert(uid != 0);
|
|
|
|
|
killUser(uid);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2023-01-25 18:31:27 +02:00
|
|
|
|
void LocalDerivationGoal::tryLocalBuild()
|
|
|
|
|
{
|
2021-02-23 08:26:35 +02:00
|
|
|
|
unsigned int curBuilds = worker.getNrLocalBuilds();
|
|
|
|
|
if (curBuilds >= settings.maxBuildJobs) {
|
2021-05-12 20:15:32 +03:00
|
|
|
|
state = &DerivationGoal::tryToBuild;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
worker.waitForBuildSlot(shared_from_this());
|
|
|
|
|
outputLocks.unlock();
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
|
2022-11-21 13:55:49 +02:00
|
|
|
|
/* Are we doing a chroot build? */
|
|
|
|
|
{
|
|
|
|
|
auto noChroot = parsedDrv->getBoolAttr("__noChroot");
|
|
|
|
|
if (settings.sandboxMode == smEnabled) {
|
|
|
|
|
if (noChroot)
|
|
|
|
|
throw Error("derivation '%s' has '__noChroot' set, "
|
|
|
|
|
"but that's not allowed when 'sandbox' is 'true'", worker.store.printStorePath(drvPath));
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
if (additionalSandboxProfile != "")
|
|
|
|
|
throw Error("derivation '%s' specifies a sandbox profile, "
|
|
|
|
|
"but this is only allowed when 'sandbox' is 'relaxed'", worker.store.printStorePath(drvPath));
|
|
|
|
|
#endif
|
|
|
|
|
useChroot = true;
|
|
|
|
|
}
|
|
|
|
|
else if (settings.sandboxMode == smDisabled)
|
|
|
|
|
useChroot = false;
|
|
|
|
|
else if (settings.sandboxMode == smRelaxed)
|
|
|
|
|
useChroot = derivationType.isSandboxed() && !noChroot;
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-11-21 13:55:49 +02:00
|
|
|
|
auto & localStore = getLocalStore();
|
|
|
|
|
if (localStore.storeDir != localStore.realStoreDir.get()) {
|
|
|
|
|
#if __linux__
|
|
|
|
|
useChroot = true;
|
|
|
|
|
#else
|
|
|
|
|
throw Error("building using a diverted store is not supported on this platform");
|
|
|
|
|
#endif
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2023-01-25 18:31:27 +02:00
|
|
|
|
#if __linux__
|
|
|
|
|
if (useChroot) {
|
2023-02-10 15:38:14 +02:00
|
|
|
|
if (!mountAndPidNamespacesSupported()) {
|
2023-01-25 18:31:27 +02:00
|
|
|
|
if (!settings.sandboxFallback)
|
2023-02-07 23:59:46 +02:00
|
|
|
|
throw Error("this system does not support the kernel namespaces that are required for sandboxing; use '--no-sandbox' to disable sandboxing");
|
2023-01-27 16:25:56 +02:00
|
|
|
|
debug("auto-disabling sandboxing because the prerequisite namespaces are not available");
|
2023-01-25 18:31:27 +02:00
|
|
|
|
useChroot = false;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
2021-04-27 22:06:58 +03:00
|
|
|
|
if (useBuildUsers()) {
|
|
|
|
|
if (!buildUser)
|
2022-11-21 13:55:49 +02:00
|
|
|
|
buildUser = acquireUserLock(parsedDrv->useUidRange() ? 65536 : 1, useChroot);
|
2021-04-27 22:06:58 +03:00
|
|
|
|
|
|
|
|
|
if (!buildUser) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (!actLock)
|
|
|
|
|
actLock = std::make_unique<Activity>(*logger, lvlWarn, actBuildWaiting,
|
|
|
|
|
fmt("waiting for UID to build '%s'", yellowtxt(worker.store.printStorePath(drvPath))));
|
|
|
|
|
worker.waitForAWhile(shared_from_this());
|
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
actLock.reset();
|
|
|
|
|
|
|
|
|
|
try {
|
|
|
|
|
|
|
|
|
|
/* Okay, we have to build. */
|
|
|
|
|
startBuilder();
|
|
|
|
|
|
|
|
|
|
} catch (BuildError & e) {
|
|
|
|
|
outputLocks.unlock();
|
|
|
|
|
buildUser.reset();
|
|
|
|
|
worker.permanentFailure = true;
|
2022-12-07 13:58:58 +02:00
|
|
|
|
done(BuildResult::InputRejected, {}, std::move(e));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
return;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* This state will be reached when we get EOF on the child's
|
|
|
|
|
log pipe. */
|
|
|
|
|
state = &DerivationGoal::buildDone;
|
|
|
|
|
|
|
|
|
|
started();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void chmod_(const Path & path, mode_t mode)
|
|
|
|
|
{
|
|
|
|
|
if (chmod(path.c_str(), mode) == -1)
|
|
|
|
|
throw SysError("setting permissions on '%s'", path);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/* Move/rename path 'src' to 'dst'. Temporarily make 'src' writable if
|
|
|
|
|
it's a directory and we're not root (to be able to update the
|
|
|
|
|
directory's parent link ".."). */
|
|
|
|
|
static void movePath(const Path & src, const Path & dst)
|
|
|
|
|
{
|
|
|
|
|
auto st = lstat(src);
|
|
|
|
|
|
|
|
|
|
bool changePerm = (geteuid() && S_ISDIR(st.st_mode) && !(st.st_mode & S_IWUSR));
|
|
|
|
|
|
|
|
|
|
if (changePerm)
|
|
|
|
|
chmod_(src, st.st_mode | S_IWUSR);
|
|
|
|
|
|
2022-04-13 15:10:36 +03:00
|
|
|
|
renameFile(src, dst);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (changePerm)
|
|
|
|
|
chmod_(dst, st.st_mode);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
extern void replaceValidPath(const Path & storePath, const Path & tmpPath);
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
int LocalDerivationGoal::getChildStatus()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
2021-02-26 17:20:33 +02:00
|
|
|
|
return hook ? DerivationGoal::getChildStatus() : pid.kill();
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::closeReadPipes()
|
|
|
|
|
{
|
|
|
|
|
if (hook) {
|
|
|
|
|
DerivationGoal::closeReadPipes();
|
|
|
|
|
} else
|
2023-03-15 11:37:39 +02:00
|
|
|
|
builderOut.close();
|
2021-02-26 17:20:33 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void LocalDerivationGoal::cleanupHookFinally()
|
|
|
|
|
{
|
|
|
|
|
/* Release the build user at the end of this function. We don't do
|
|
|
|
|
it right away because we don't want another build grabbing this
|
|
|
|
|
uid and then messing around with our output. */
|
|
|
|
|
buildUser.reset();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void LocalDerivationGoal::cleanupPreChildKill()
|
|
|
|
|
{
|
|
|
|
|
sandboxMountNamespace = -1;
|
2021-10-09 01:55:08 +03:00
|
|
|
|
sandboxUserNamespace = -1;
|
2021-02-26 17:20:33 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void LocalDerivationGoal::cleanupPostChildKill()
|
|
|
|
|
{
|
|
|
|
|
/* When running under a build user, make sure that all processes
|
|
|
|
|
running under that uid are gone. This is to prevent a
|
|
|
|
|
malicious user from leaving behind a process that keeps files
|
|
|
|
|
open and modifies them after they have been chown'ed to
|
|
|
|
|
root. */
|
2022-11-18 14:40:59 +02:00
|
|
|
|
killSandbox(true);
|
2021-02-26 17:20:33 +02:00
|
|
|
|
|
|
|
|
|
/* Terminate the recursive Nix daemon. */
|
|
|
|
|
stopDaemon();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
bool LocalDerivationGoal::cleanupDecideWhetherDiskFull()
|
|
|
|
|
{
|
|
|
|
|
bool diskFull = false;
|
|
|
|
|
|
|
|
|
|
/* Heuristically check whether the build failure may have
|
|
|
|
|
been caused by a disk full condition. We have no way
|
|
|
|
|
of knowing whether the build actually got an ENOSPC.
|
|
|
|
|
So instead, check if the disk is (nearly) full now. If
|
|
|
|
|
so, we don't mark this build as a permanent failure. */
|
|
|
|
|
#if HAVE_STATVFS
|
2021-02-27 07:23:14 +02:00
|
|
|
|
{
|
2021-02-26 17:20:33 +02:00
|
|
|
|
auto & localStore = getLocalStore();
|
|
|
|
|
uint64_t required = 8ULL * 1024 * 1024; // FIXME: make configurable
|
|
|
|
|
struct statvfs st;
|
2021-06-18 18:04:11 +03:00
|
|
|
|
if (statvfs(localStore.realStoreDir.get().c_str(), &st) == 0 &&
|
2021-02-26 17:20:33 +02:00
|
|
|
|
(uint64_t) st.f_bavail * st.f_bsize < required)
|
|
|
|
|
diskFull = true;
|
|
|
|
|
if (statvfs(tmpDir.c_str(), &st) == 0 &&
|
|
|
|
|
(uint64_t) st.f_bavail * st.f_bsize < required)
|
|
|
|
|
diskFull = true;
|
2021-02-27 07:23:14 +02:00
|
|
|
|
}
|
2021-02-26 17:20:33 +02:00
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
deleteTmpDir(false);
|
|
|
|
|
|
|
|
|
|
/* Move paths out of the chroot for easier debugging of
|
|
|
|
|
build failures. */
|
|
|
|
|
if (useChroot && buildMode == bmNormal)
|
|
|
|
|
for (auto & [_, status] : initialOutputs) {
|
|
|
|
|
if (!status.known) continue;
|
|
|
|
|
if (buildMode != bmCheck && status.known->isValid()) continue;
|
|
|
|
|
auto p = worker.store.printStorePath(status.known->path);
|
|
|
|
|
if (pathExists(chrootRootDir + p))
|
2022-04-13 15:10:36 +03:00
|
|
|
|
renameFile((chrootRootDir + p), p);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
return diskFull;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void LocalDerivationGoal::cleanupPostOutputsRegisteredModeCheck()
|
|
|
|
|
{
|
|
|
|
|
deleteTmpDir(true);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::cleanupPostOutputsRegisteredModeNonCheck()
|
|
|
|
|
{
|
|
|
|
|
/* Delete unused redirected outputs (when doing hash rewriting). */
|
|
|
|
|
for (auto & i : redirectedOutputs)
|
|
|
|
|
deletePath(worker.store.Store::toRealPath(i.second));
|
|
|
|
|
|
|
|
|
|
/* Delete the chroot (if we were using one). */
|
|
|
|
|
autoDelChroot.reset(); /* this runs the destructor */
|
|
|
|
|
|
|
|
|
|
cleanupPostOutputsRegisteredModeCheck();
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
|
2021-11-19 16:22:31 +02:00
|
|
|
|
#if __linux__
|
2021-02-26 17:20:33 +02:00
|
|
|
|
static void linkOrCopy(const Path & from, const Path & to)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
if (link(from.c_str(), to.c_str()) == -1) {
|
|
|
|
|
/* Hard-linking fails if we exceed the maximum link count on a
|
|
|
|
|
file (e.g. 32000 of ext3), which is quite possible after a
|
|
|
|
|
'nix-store --optimise'. FIXME: actually, why don't we just
|
|
|
|
|
bind-mount in this case?
|
|
|
|
|
|
|
|
|
|
It can also fail with EPERM in BeegFS v7 and earlier versions
|
|
|
|
|
which don't allow hard-links to other directories */
|
|
|
|
|
if (errno != EMLINK && errno != EPERM)
|
|
|
|
|
throw SysError("linking '%s' to '%s'", to, from);
|
|
|
|
|
copyPath(from, to);
|
|
|
|
|
}
|
|
|
|
|
}
|
2021-11-19 16:22:31 +02:00
|
|
|
|
#endif
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::startBuilder()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
2022-11-18 11:39:28 +02:00
|
|
|
|
if ((buildUser && buildUser->getUIDCount() != 1)
|
2022-11-28 22:54:02 +02:00
|
|
|
|
#if __linux__
|
|
|
|
|
|| settings.useCgroups
|
|
|
|
|
#endif
|
|
|
|
|
)
|
2022-11-18 11:39:28 +02:00
|
|
|
|
{
|
|
|
|
|
#if __linux__
|
2023-03-17 16:33:48 +02:00
|
|
|
|
experimentalFeatureSettings.require(Xp::Cgroups);
|
2022-11-28 22:54:02 +02:00
|
|
|
|
|
2022-12-02 13:57:41 +02:00
|
|
|
|
auto cgroupFS = getCgroupFS();
|
|
|
|
|
if (!cgroupFS)
|
|
|
|
|
throw Error("cannot determine the cgroups file system");
|
|
|
|
|
|
2022-11-18 11:39:28 +02:00
|
|
|
|
auto ourCgroups = getCgroups("/proc/self/cgroup");
|
|
|
|
|
auto ourCgroup = ourCgroups[""];
|
|
|
|
|
if (ourCgroup == "")
|
|
|
|
|
throw Error("cannot determine cgroup name from /proc/self/cgroup");
|
|
|
|
|
|
2022-12-02 13:57:41 +02:00
|
|
|
|
auto ourCgroupPath = canonPath(*cgroupFS + "/" + ourCgroup);
|
2022-11-18 11:39:28 +02:00
|
|
|
|
|
|
|
|
|
if (!pathExists(ourCgroupPath))
|
|
|
|
|
throw Error("expected cgroup directory '%s'", ourCgroupPath);
|
|
|
|
|
|
|
|
|
|
static std::atomic<unsigned int> counter{0};
|
|
|
|
|
|
|
|
|
|
cgroup = buildUser
|
|
|
|
|
? fmt("%s/nix-build-uid-%d", ourCgroupPath, buildUser->getUID())
|
|
|
|
|
: fmt("%s/nix-build-pid-%d-%d", ourCgroupPath, getpid(), counter++);
|
|
|
|
|
|
|
|
|
|
debug("using cgroup '%s'", *cgroup);
|
|
|
|
|
|
|
|
|
|
/* When using a build user, record the cgroup we used for that
|
|
|
|
|
user so that if we got interrupted previously, we can kill
|
|
|
|
|
any left-over cgroup first. */
|
|
|
|
|
if (buildUser) {
|
|
|
|
|
auto cgroupsDir = settings.nixStateDir + "/cgroups";
|
|
|
|
|
createDirs(cgroupsDir);
|
|
|
|
|
|
|
|
|
|
auto cgroupFile = fmt("%s/%d", cgroupsDir, buildUser->getUID());
|
|
|
|
|
|
|
|
|
|
if (pathExists(cgroupFile)) {
|
|
|
|
|
auto prevCgroup = readFile(cgroupFile);
|
|
|
|
|
destroyCgroup(prevCgroup);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
writeFile(cgroupFile, *cgroup);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
throw Error("cgroups are not supported on this platform");
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Make sure that no other processes are executing under the
|
|
|
|
|
sandbox uids. This must be done before any chownToBuilder()
|
|
|
|
|
calls. */
|
2022-11-18 14:40:59 +02:00
|
|
|
|
killSandbox(false);
|
2022-11-18 11:39:28 +02:00
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Right platform? */
|
|
|
|
|
if (!parsedDrv->canBuildLocally(worker.store))
|
|
|
|
|
throw Error("a '%s' with features {%s} is required to build '%s', but I am a '%s' with features {%s}",
|
|
|
|
|
drv->platform,
|
|
|
|
|
concatStringsSep(", ", parsedDrv->getRequiredSystemFeatures()),
|
|
|
|
|
worker.store.printStorePath(drvPath),
|
|
|
|
|
settings.thisSystem,
|
|
|
|
|
concatStringsSep<StringSet>(", ", worker.store.systemFeatures));
|
|
|
|
|
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
additionalSandboxProfile = parsedDrv->getStringAttr("__sandboxProfile").value_or("");
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Create a temporary directory where the build will take
|
|
|
|
|
place. */
|
|
|
|
|
tmpDir = createTempDir("", "nix-build-" + std::string(drvPath.name()), false, false, 0700);
|
|
|
|
|
|
|
|
|
|
chownToBuilder(tmpDir);
|
|
|
|
|
|
|
|
|
|
for (auto & [outputName, status] : initialOutputs) {
|
|
|
|
|
/* Set scratch path we'll actually use during the build.
|
|
|
|
|
|
|
|
|
|
If we're not doing a chroot build, but we have some valid
|
|
|
|
|
output paths. Since we can't just overwrite or delete
|
|
|
|
|
them, we have to do hash rewriting: i.e. in the
|
|
|
|
|
environment/arguments passed to the build, we replace the
|
|
|
|
|
hashes of the valid outputs with unique dummy strings;
|
|
|
|
|
after the build, we discard the redirected outputs
|
|
|
|
|
corresponding to the valid outputs, and rewrite the
|
|
|
|
|
contents of the new outputs to replace the dummy strings
|
|
|
|
|
with the actual hashes. */
|
|
|
|
|
auto scratchPath =
|
|
|
|
|
!status.known
|
|
|
|
|
? makeFallbackPath(outputName)
|
|
|
|
|
: !needsHashRewrite()
|
|
|
|
|
/* Can always use original path in sandbox */
|
|
|
|
|
? status.known->path
|
|
|
|
|
: !status.known->isPresent()
|
|
|
|
|
/* If path doesn't yet exist can just use it */
|
|
|
|
|
? status.known->path
|
|
|
|
|
: buildMode != bmRepair && !status.known->isValid()
|
|
|
|
|
/* If we aren't repairing we'll delete a corrupted path, so we
|
|
|
|
|
can use original path */
|
|
|
|
|
? status.known->path
|
|
|
|
|
: /* If we are repairing or the path is totally valid, we'll need
|
|
|
|
|
to use a temporary path */
|
|
|
|
|
makeFallbackPath(status.known->path);
|
|
|
|
|
scratchOutputs.insert_or_assign(outputName, scratchPath);
|
|
|
|
|
|
|
|
|
|
/* Substitute output placeholders with the scratch output paths.
|
|
|
|
|
We'll use during the build. */
|
|
|
|
|
inputRewrites[hashPlaceholder(outputName)] = worker.store.printStorePath(scratchPath);
|
|
|
|
|
|
|
|
|
|
/* Additional tasks if we know the final path a priori. */
|
|
|
|
|
if (!status.known) continue;
|
|
|
|
|
auto fixedFinalPath = status.known->path;
|
|
|
|
|
|
|
|
|
|
/* Additional tasks if the final and scratch are both known and
|
|
|
|
|
differ. */
|
|
|
|
|
if (fixedFinalPath == scratchPath) continue;
|
|
|
|
|
|
|
|
|
|
/* Ensure scratch path is ours to use. */
|
|
|
|
|
deletePath(worker.store.printStorePath(scratchPath));
|
|
|
|
|
|
|
|
|
|
/* Rewrite and unrewrite paths */
|
|
|
|
|
{
|
|
|
|
|
std::string h1 { fixedFinalPath.hashPart() };
|
|
|
|
|
std::string h2 { scratchPath.hashPart() };
|
|
|
|
|
inputRewrites[h1] = h2;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
redirectedOutputs.insert_or_assign(std::move(fixedFinalPath), std::move(scratchPath));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Construct the environment passed to the builder. */
|
|
|
|
|
initEnv();
|
|
|
|
|
|
|
|
|
|
writeStructuredAttrs();
|
|
|
|
|
|
|
|
|
|
/* Handle exportReferencesGraph(), if set. */
|
|
|
|
|
if (!parsedDrv->getStructuredAttrs()) {
|
|
|
|
|
/* The `exportReferencesGraph' feature allows the references graph
|
|
|
|
|
to be passed to a builder. This attribute should be a list of
|
|
|
|
|
pairs [name1 path1 name2 path2 ...]. The references graph of
|
|
|
|
|
each `pathN' will be stored in a text file `nameN' in the
|
|
|
|
|
temporary build directory. The text files have the format used
|
|
|
|
|
by `nix-store --register-validity'. However, the deriver
|
|
|
|
|
fields are left empty. */
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto s = getOr(drv->env, "exportReferencesGraph", "");
|
2021-02-23 08:26:35 +02:00
|
|
|
|
Strings ss = tokenizeString<Strings>(s);
|
|
|
|
|
if (ss.size() % 2 != 0)
|
|
|
|
|
throw BuildError("odd number of tokens in 'exportReferencesGraph': '%1%'", s);
|
|
|
|
|
for (Strings::iterator i = ss.begin(); i != ss.end(); ) {
|
2022-02-25 17:00:00 +02:00
|
|
|
|
auto fileName = *i++;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
static std::regex regex("[A-Za-z_][A-Za-z0-9_.-]*");
|
|
|
|
|
if (!std::regex_match(fileName, regex))
|
|
|
|
|
throw Error("invalid file name '%s' in 'exportReferencesGraph'", fileName);
|
|
|
|
|
|
|
|
|
|
auto storePathS = *i++;
|
|
|
|
|
if (!worker.store.isInStore(storePathS))
|
|
|
|
|
throw BuildError("'exportReferencesGraph' contains a non-store path '%1%'", storePathS);
|
|
|
|
|
auto storePath = worker.store.toStorePath(storePathS).first;
|
|
|
|
|
|
|
|
|
|
/* Write closure info to <fileName>. */
|
|
|
|
|
writeFile(tmpDir + "/" + fileName,
|
|
|
|
|
worker.store.makeValidityRegistration(
|
2021-05-02 18:24:14 +03:00
|
|
|
|
worker.store.exportReferences({storePath}, inputPaths), false, false));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (useChroot) {
|
|
|
|
|
|
|
|
|
|
/* Allow a user-configurable set of directories from the
|
|
|
|
|
host file system. */
|
|
|
|
|
dirsInChroot.clear();
|
|
|
|
|
|
|
|
|
|
for (auto i : settings.sandboxPaths.get()) {
|
|
|
|
|
if (i.empty()) continue;
|
|
|
|
|
bool optional = false;
|
|
|
|
|
if (i[i.size() - 1] == '?') {
|
|
|
|
|
optional = true;
|
|
|
|
|
i.pop_back();
|
|
|
|
|
}
|
|
|
|
|
size_t p = i.find('=');
|
2022-02-25 17:00:00 +02:00
|
|
|
|
if (p == std::string::npos)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
dirsInChroot[i] = {i, optional};
|
|
|
|
|
else
|
2022-02-25 17:00:00 +02:00
|
|
|
|
dirsInChroot[i.substr(0, p)] = {i.substr(p + 1), optional};
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
dirsInChroot[tmpDirInSandbox] = tmpDir;
|
|
|
|
|
|
|
|
|
|
/* Add the closure of store paths to the chroot. */
|
|
|
|
|
StorePathSet closure;
|
|
|
|
|
for (auto & i : dirsInChroot)
|
|
|
|
|
try {
|
|
|
|
|
if (worker.store.isInStore(i.second.source))
|
|
|
|
|
worker.store.computeFSClosure(worker.store.toStorePath(i.second.source).first, closure);
|
|
|
|
|
} catch (InvalidPath & e) {
|
|
|
|
|
} catch (Error & e) {
|
|
|
|
|
e.addTrace({}, "while processing 'sandbox-paths'");
|
|
|
|
|
throw;
|
|
|
|
|
}
|
|
|
|
|
for (auto & i : closure) {
|
|
|
|
|
auto p = worker.store.printStorePath(i);
|
|
|
|
|
dirsInChroot.insert_or_assign(p, p);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
PathSet allowedPaths = settings.allowedImpureHostPrefixes;
|
|
|
|
|
|
|
|
|
|
/* This works like the above, except on a per-derivation level */
|
|
|
|
|
auto impurePaths = parsedDrv->getStringsAttr("__impureHostDeps").value_or(Strings());
|
|
|
|
|
|
|
|
|
|
for (auto & i : impurePaths) {
|
|
|
|
|
bool found = false;
|
|
|
|
|
/* Note: we're not resolving symlinks here to prevent
|
|
|
|
|
giving a non-root user info about inaccessible
|
|
|
|
|
files. */
|
|
|
|
|
Path canonI = canonPath(i);
|
|
|
|
|
/* If only we had a trie to do this more efficiently :) luckily, these are generally going to be pretty small */
|
|
|
|
|
for (auto & a : allowedPaths) {
|
|
|
|
|
Path canonA = canonPath(a);
|
|
|
|
|
if (canonI == canonA || isInDir(canonI, canonA)) {
|
|
|
|
|
found = true;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
if (!found)
|
|
|
|
|
throw Error("derivation '%s' requested impure path '%s', but it was not in allowed-impure-host-deps",
|
|
|
|
|
worker.store.printStorePath(drvPath), i);
|
|
|
|
|
|
2021-05-02 07:30:50 +03:00
|
|
|
|
/* Allow files in __impureHostDeps to be missing; e.g.
|
|
|
|
|
macOS 11+ has no /usr/lib/libSystem*.dylib */
|
|
|
|
|
dirsInChroot[i] = {i, true};
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#if __linux__
|
|
|
|
|
/* Create a temporary directory in which we set up the chroot
|
|
|
|
|
environment using bind-mounts. We put it in the Nix store
|
|
|
|
|
to ensure that we can create hard-links to non-directory
|
|
|
|
|
inputs in the fake Nix store in the chroot (see below). */
|
|
|
|
|
chrootRootDir = worker.store.Store::toRealPath(drvPath) + ".chroot";
|
|
|
|
|
deletePath(chrootRootDir);
|
|
|
|
|
|
|
|
|
|
/* Clean up the chroot directory automatically. */
|
|
|
|
|
autoDelChroot = std::make_shared<AutoDelete>(chrootRootDir);
|
|
|
|
|
|
2023-03-02 16:44:19 +02:00
|
|
|
|
printMsg(lvlChatty, "setting up chroot environment in '%1%'", chrootRootDir);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-11-08 17:03:42 +02:00
|
|
|
|
// FIXME: make this 0700
|
|
|
|
|
if (mkdir(chrootRootDir.c_str(), buildUser && buildUser->getUIDCount() != 1 ? 0755 : 0750) == -1)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw SysError("cannot create '%1%'", chrootRootDir);
|
|
|
|
|
|
2022-11-29 14:10:53 +02:00
|
|
|
|
if (buildUser && chown(chrootRootDir.c_str(), buildUser->getUIDCount() != 1 ? buildUser->getUID() : 0, buildUser->getGID()) == -1)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw SysError("cannot change ownership of '%1%'", chrootRootDir);
|
|
|
|
|
|
|
|
|
|
/* Create a writable /tmp in the chroot. Many builders need
|
|
|
|
|
this. (Of course they should really respect $TMPDIR
|
|
|
|
|
instead.) */
|
|
|
|
|
Path chrootTmpDir = chrootRootDir + "/tmp";
|
|
|
|
|
createDirs(chrootTmpDir);
|
|
|
|
|
chmod_(chrootTmpDir, 01777);
|
|
|
|
|
|
|
|
|
|
/* Create a /etc/passwd with entries for the build user and the
|
|
|
|
|
nobody account. The latter is kind of a hack to support
|
|
|
|
|
Samba-in-QEMU. */
|
|
|
|
|
createDirs(chrootRootDir + "/etc");
|
2023-02-14 14:29:30 +02:00
|
|
|
|
if (parsedDrv->useUidRange())
|
|
|
|
|
chownToBuilder(chrootRootDir + "/etc");
|
2021-04-27 22:06:58 +03:00
|
|
|
|
|
2022-11-08 17:03:42 +02:00
|
|
|
|
if (parsedDrv->useUidRange() && (!buildUser || buildUser->getUIDCount() < 65536))
|
|
|
|
|
throw Error("feature 'uid-range' requires the setting '%s' to be enabled", settings.autoAllocateUids.name);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Declare the build user's group so that programs get a consistent
|
|
|
|
|
view of the system (e.g., "id -gn"). */
|
|
|
|
|
writeFile(chrootRootDir + "/etc/group",
|
|
|
|
|
fmt("root:x:0:\n"
|
|
|
|
|
"nixbld:!:%1%:\n"
|
|
|
|
|
"nogroup:x:65534:\n", sandboxGid()));
|
|
|
|
|
|
|
|
|
|
/* Create /etc/hosts with localhost entry. */
|
2022-03-31 17:06:40 +03:00
|
|
|
|
if (derivationType.isSandboxed())
|
2021-02-23 08:26:35 +02:00
|
|
|
|
writeFile(chrootRootDir + "/etc/hosts", "127.0.0.1 localhost\n::1 localhost\n");
|
|
|
|
|
|
|
|
|
|
/* Make the closure of the inputs available in the chroot,
|
|
|
|
|
rather than the whole Nix store. This prevents any access
|
|
|
|
|
to undeclared dependencies. Directories are bind-mounted,
|
|
|
|
|
while other inputs are hard-linked (since only directories
|
|
|
|
|
can be bind-mounted). !!! As an extra security
|
|
|
|
|
precaution, make the fake Nix store only writable by the
|
|
|
|
|
build user. */
|
|
|
|
|
Path chrootStoreDir = chrootRootDir + worker.store.storeDir;
|
|
|
|
|
createDirs(chrootStoreDir);
|
|
|
|
|
chmod_(chrootStoreDir, 01775);
|
|
|
|
|
|
|
|
|
|
if (buildUser && chown(chrootStoreDir.c_str(), 0, buildUser->getGID()) == -1)
|
|
|
|
|
throw SysError("cannot change ownership of '%1%'", chrootStoreDir);
|
|
|
|
|
|
|
|
|
|
for (auto & i : inputPaths) {
|
|
|
|
|
auto p = worker.store.printStorePath(i);
|
|
|
|
|
Path r = worker.store.toRealPath(p);
|
|
|
|
|
if (S_ISDIR(lstat(r).st_mode))
|
|
|
|
|
dirsInChroot.insert_or_assign(p, r);
|
|
|
|
|
else
|
|
|
|
|
linkOrCopy(r, chrootRootDir + p);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* If we're repairing, checking or rebuilding part of a
|
|
|
|
|
multiple-outputs derivation, it's possible that we're
|
|
|
|
|
rebuilding a path that is in settings.dirsInChroot
|
|
|
|
|
(typically the dependencies of /bin/sh). Throw them
|
|
|
|
|
out. */
|
|
|
|
|
for (auto & i : drv->outputsAndOptPaths(worker.store)) {
|
|
|
|
|
/* If the name isn't known a priori (i.e. floating
|
|
|
|
|
content-addressed derivation), the temporary location we use
|
|
|
|
|
should be fresh. Freshness means it is impossible that the path
|
|
|
|
|
is already in the sandbox, so we don't need to worry about
|
|
|
|
|
removing it. */
|
|
|
|
|
if (i.second.second)
|
|
|
|
|
dirsInChroot.erase(worker.store.printStorePath(*i.second.second));
|
|
|
|
|
}
|
|
|
|
|
|
2022-11-18 11:39:28 +02:00
|
|
|
|
if (cgroup) {
|
|
|
|
|
if (mkdir(cgroup->c_str(), 0755) != 0)
|
|
|
|
|
throw SysError("creating cgroup '%s'", *cgroup);
|
|
|
|
|
chownToBuilder(*cgroup);
|
|
|
|
|
chownToBuilder(*cgroup + "/cgroup.procs");
|
|
|
|
|
chownToBuilder(*cgroup + "/cgroup.threads");
|
|
|
|
|
//chownToBuilder(*cgroup + "/cgroup.subtree_control");
|
2021-04-27 22:06:58 +03:00
|
|
|
|
}
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
#else
|
2022-11-08 18:00:29 +02:00
|
|
|
|
if (parsedDrv->useUidRange())
|
2021-04-27 22:06:58 +03:00
|
|
|
|
throw Error("feature 'uid-range' is not supported on this platform");
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
/* We don't really have any parent prep work to do (yet?)
|
|
|
|
|
All work happens in the child, instead. */
|
|
|
|
|
#else
|
|
|
|
|
throw Error("sandboxing builds is not supported on this platform");
|
|
|
|
|
#endif
|
2021-02-23 08:26:35 +02:00
|
|
|
|
#endif
|
2021-04-27 22:06:58 +03:00
|
|
|
|
} else {
|
2022-11-08 17:03:42 +02:00
|
|
|
|
if (parsedDrv->useUidRange())
|
2021-04-27 22:06:58 +03:00
|
|
|
|
throw Error("feature 'uid-range' is only supported in sandboxed builds");
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (needsHashRewrite() && pathExists(homeDir))
|
|
|
|
|
throw Error("home directory '%1%' exists; please remove it to assure purity of builds without sandboxing", homeDir);
|
|
|
|
|
|
|
|
|
|
if (useChroot && settings.preBuildHook != "" && dynamic_cast<Derivation *>(drv.get())) {
|
2023-03-02 16:44:19 +02:00
|
|
|
|
printMsg(lvlChatty, "executing pre-build hook '%1%'", settings.preBuildHook);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
auto args = useChroot ? Strings({worker.store.printStorePath(drvPath), chrootRootDir}) :
|
|
|
|
|
Strings({ worker.store.printStorePath(drvPath) });
|
|
|
|
|
enum BuildHookState {
|
|
|
|
|
stBegin,
|
|
|
|
|
stExtraChrootDirs
|
|
|
|
|
};
|
|
|
|
|
auto state = stBegin;
|
|
|
|
|
auto lines = runProgram(settings.preBuildHook, false, args);
|
|
|
|
|
auto lastPos = std::string::size_type{0};
|
2022-02-25 17:00:00 +02:00
|
|
|
|
for (auto nlPos = lines.find('\n'); nlPos != std::string::npos;
|
|
|
|
|
nlPos = lines.find('\n', lastPos))
|
|
|
|
|
{
|
|
|
|
|
auto line = lines.substr(lastPos, nlPos - lastPos);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
lastPos = nlPos + 1;
|
|
|
|
|
if (state == stBegin) {
|
|
|
|
|
if (line == "extra-sandbox-paths" || line == "extra-chroot-dirs") {
|
|
|
|
|
state = stExtraChrootDirs;
|
|
|
|
|
} else {
|
|
|
|
|
throw Error("unknown pre-build hook command '%1%'", line);
|
|
|
|
|
}
|
|
|
|
|
} else if (state == stExtraChrootDirs) {
|
|
|
|
|
if (line == "") {
|
|
|
|
|
state = stBegin;
|
|
|
|
|
} else {
|
|
|
|
|
auto p = line.find('=');
|
2022-02-25 17:00:00 +02:00
|
|
|
|
if (p == std::string::npos)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
dirsInChroot[line] = line;
|
|
|
|
|
else
|
2022-02-25 17:00:00 +02:00
|
|
|
|
dirsInChroot[line.substr(0, p)] = line.substr(p + 1);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Fire up a Nix daemon to process recursive Nix calls from the
|
|
|
|
|
builder. */
|
|
|
|
|
if (parsedDrv->getRequiredSystemFeatures().count("recursive-nix"))
|
|
|
|
|
startDaemon();
|
|
|
|
|
|
|
|
|
|
/* Run the builder. */
|
|
|
|
|
printMsg(lvlChatty, "executing builder '%1%'", drv->builder);
|
2022-04-11 07:09:04 +03:00
|
|
|
|
printMsg(lvlChatty, "using builder args '%1%'", concatStringsSep(" ", drv->args));
|
|
|
|
|
for (auto & i : drv->env)
|
|
|
|
|
printMsg(lvlVomit, "setting builder env variable '%1%'='%2%'", i.first, i.second);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Create the log file. */
|
|
|
|
|
Path logFile = openLogFile();
|
|
|
|
|
|
Move pseudoterminal slave open to child
Hopefully this fixes "unexpected EOF" failures on macOS
(#3137, #3605, #7242, #7702).
The problem appears to be that under some circumstances, macOS
discards the output written to the slave side of the
pseudoterminal. Hence the parent never sees the "sandbox initialized"
message from the child, even though it succeeded. The conditions are:
* The child finishes very quickly. That's why this bug is likely to
trigger in nix-env tests, since that uses a builtin builder. Adding
a short sleep before the child exits makes the problem go away.
* The parent has closed its duplicate of the slave file
descriptor. This shouldn't matter, since the child has a duplicate
as well, but it does. E.g. moving the close to the bottom of
startBuilder() makes the problem go away. However, that's not a
solution because it would make Nix hang if the child dies before
sending the "sandbox initialized" message.
* The system is under high load. E.g. "make installcheck -j16" makes
the issue pretty reproducible, while it's very rare under "make
installcheck -j1".
As a fix/workaround, we now open the pseudoterminal slave in the
child, rather than the parent. This removes the second condition
(i.e. the parent no longer needs to close the slave fd) and I haven't
been able to reproduce the "unexpected EOF" with this.
2023-03-14 20:10:59 +02:00
|
|
|
|
/* Create a pseudoterminal to get the output of the builder. */
|
2023-03-15 11:37:39 +02:00
|
|
|
|
builderOut = posix_openpt(O_RDWR | O_NOCTTY);
|
|
|
|
|
if (!builderOut)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw SysError("opening pseudoterminal master");
|
|
|
|
|
|
2021-10-05 12:04:46 +03:00
|
|
|
|
// FIXME: not thread-safe, use ptsname_r
|
2023-03-15 11:41:38 +02:00
|
|
|
|
std::string slaveName = ptsname(builderOut.get());
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (buildUser) {
|
|
|
|
|
if (chmod(slaveName.c_str(), 0600))
|
|
|
|
|
throw SysError("changing mode of pseudoterminal slave");
|
|
|
|
|
|
|
|
|
|
if (chown(slaveName.c_str(), buildUser->getUID(), 0))
|
|
|
|
|
throw SysError("changing owner of pseudoterminal slave");
|
|
|
|
|
}
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
else {
|
2023-03-15 11:37:39 +02:00
|
|
|
|
if (grantpt(builderOut.get()))
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw SysError("granting access to pseudoterminal slave");
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
2023-03-15 11:37:39 +02:00
|
|
|
|
if (unlockpt(builderOut.get()))
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw SysError("unlocking pseudoterminal");
|
|
|
|
|
|
2023-03-20 18:04:57 +02:00
|
|
|
|
/* Open the slave side of the pseudoterminal and use it as stderr. */
|
|
|
|
|
auto openSlave = [&]()
|
|
|
|
|
{
|
|
|
|
|
AutoCloseFD builderOut = open(slaveName.c_str(), O_RDWR | O_NOCTTY);
|
|
|
|
|
if (!builderOut)
|
|
|
|
|
throw SysError("opening pseudoterminal slave");
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2023-03-20 18:04:57 +02:00
|
|
|
|
// Put the pt into raw mode to prevent \n -> \r\n translation.
|
|
|
|
|
struct termios term;
|
|
|
|
|
if (tcgetattr(builderOut.get(), &term))
|
|
|
|
|
throw SysError("getting pseudoterminal attributes");
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2023-03-20 18:04:57 +02:00
|
|
|
|
cfmakeraw(&term);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2023-03-20 18:04:57 +02:00
|
|
|
|
if (tcsetattr(builderOut.get(), TCSANOW, &term))
|
|
|
|
|
throw SysError("putting pseudoterminal into raw mode");
|
|
|
|
|
|
|
|
|
|
if (dup2(builderOut.get(), STDERR_FILENO) == -1)
|
|
|
|
|
throw SysError("cannot pipe standard error into log file");
|
|
|
|
|
};
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-03-08 20:50:46 +02:00
|
|
|
|
buildResult.startTime = time(0);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Fork a child to build the package. */
|
|
|
|
|
|
|
|
|
|
#if __linux__
|
|
|
|
|
if (useChroot) {
|
|
|
|
|
/* Set up private namespaces for the build:
|
|
|
|
|
|
|
|
|
|
- The PID namespace causes the build to start as PID 1.
|
|
|
|
|
Processes outside of the chroot are not visible to those
|
|
|
|
|
on the inside, but processes inside the chroot are
|
|
|
|
|
visible from the outside (though with different PIDs).
|
|
|
|
|
|
|
|
|
|
- The private mount namespace ensures that all the bind
|
|
|
|
|
mounts we do will only show up in this process and its
|
|
|
|
|
children, and will disappear automatically when we're
|
|
|
|
|
done.
|
|
|
|
|
|
|
|
|
|
- The private network namespace ensures that the builder
|
|
|
|
|
cannot talk to the outside world (or vice versa). It
|
|
|
|
|
only has a private loopback interface. (Fixed-output
|
|
|
|
|
derivations are not run in a private network namespace
|
|
|
|
|
to allow functions like fetchurl to work.)
|
|
|
|
|
|
|
|
|
|
- The IPC namespace prevents the builder from communicating
|
|
|
|
|
with outside processes using SysV IPC mechanisms (shared
|
|
|
|
|
memory, message queues, semaphores). It also ensures
|
|
|
|
|
that all IPC objects are destroyed when the builder
|
|
|
|
|
exits.
|
|
|
|
|
|
|
|
|
|
- The UTS namespace ensures that builders see a hostname of
|
|
|
|
|
localhost rather than the actual hostname.
|
|
|
|
|
|
|
|
|
|
We use a helper process to do the clone() to work around
|
|
|
|
|
clone() being broken in multi-threaded programs due to
|
|
|
|
|
at-fork handlers not being run. Note that we use
|
|
|
|
|
CLONE_PARENT to ensure that the real builder is parented to
|
|
|
|
|
us.
|
|
|
|
|
*/
|
|
|
|
|
|
2022-03-31 17:06:40 +03:00
|
|
|
|
if (derivationType.isSandboxed())
|
2021-02-23 08:26:35 +02:00
|
|
|
|
privateNetwork = true;
|
|
|
|
|
|
|
|
|
|
userNamespaceSync.create();
|
|
|
|
|
|
2023-01-25 18:31:27 +02:00
|
|
|
|
usingUserNamespace = userNamespacesSupported();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
Move pseudoterminal slave open to child
Hopefully this fixes "unexpected EOF" failures on macOS
(#3137, #3605, #7242, #7702).
The problem appears to be that under some circumstances, macOS
discards the output written to the slave side of the
pseudoterminal. Hence the parent never sees the "sandbox initialized"
message from the child, even though it succeeded. The conditions are:
* The child finishes very quickly. That's why this bug is likely to
trigger in nix-env tests, since that uses a builtin builder. Adding
a short sleep before the child exits makes the problem go away.
* The parent has closed its duplicate of the slave file
descriptor. This shouldn't matter, since the child has a duplicate
as well, but it does. E.g. moving the close to the bottom of
startBuilder() makes the problem go away. However, that's not a
solution because it would make Nix hang if the child dies before
sending the "sandbox initialized" message.
* The system is under high load. E.g. "make installcheck -j16" makes
the issue pretty reproducible, while it's very rare under "make
installcheck -j1".
As a fix/workaround, we now open the pseudoterminal slave in the
child, rather than the parent. This removes the second condition
(i.e. the parent no longer needs to close the slave fd) and I haven't
been able to reproduce the "unexpected EOF" with this.
2023-03-14 20:10:59 +02:00
|
|
|
|
Pipe sendPid;
|
|
|
|
|
sendPid.create();
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
Pid helper = startProcess([&]() {
|
Move pseudoterminal slave open to child
Hopefully this fixes "unexpected EOF" failures on macOS
(#3137, #3605, #7242, #7702).
The problem appears to be that under some circumstances, macOS
discards the output written to the slave side of the
pseudoterminal. Hence the parent never sees the "sandbox initialized"
message from the child, even though it succeeded. The conditions are:
* The child finishes very quickly. That's why this bug is likely to
trigger in nix-env tests, since that uses a builtin builder. Adding
a short sleep before the child exits makes the problem go away.
* The parent has closed its duplicate of the slave file
descriptor. This shouldn't matter, since the child has a duplicate
as well, but it does. E.g. moving the close to the bottom of
startBuilder() makes the problem go away. However, that's not a
solution because it would make Nix hang if the child dies before
sending the "sandbox initialized" message.
* The system is under high load. E.g. "make installcheck -j16" makes
the issue pretty reproducible, while it's very rare under "make
installcheck -j1".
As a fix/workaround, we now open the pseudoterminal slave in the
child, rather than the parent. This removes the second condition
(i.e. the parent no longer needs to close the slave fd) and I haven't
been able to reproduce the "unexpected EOF" with this.
2023-03-14 20:10:59 +02:00
|
|
|
|
sendPid.readSide.close();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2023-03-20 18:04:57 +02:00
|
|
|
|
/* We need to open the slave early, before
|
|
|
|
|
CLONE_NEWUSER. Otherwise we get EPERM when running as
|
|
|
|
|
root. */
|
|
|
|
|
openSlave();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Drop additional groups here because we can't do it
|
|
|
|
|
after we've created the new user namespace. FIXME:
|
|
|
|
|
this means that if we're not root in the parent
|
|
|
|
|
namespace, we can't drop additional groups; they will
|
|
|
|
|
be mapped to nogroup in the child namespace. There does
|
|
|
|
|
not seem to be a workaround for this. (But who can tell
|
|
|
|
|
from reading user_namespaces(7)?)
|
|
|
|
|
See also https://lwn.net/Articles/621612/. */
|
|
|
|
|
if (getuid() == 0 && setgroups(0, 0) == -1)
|
|
|
|
|
throw SysError("setgroups failed");
|
|
|
|
|
|
2023-02-10 15:44:25 +02:00
|
|
|
|
ProcessOptions options;
|
|
|
|
|
options.cloneFlags = CLONE_NEWPID | CLONE_NEWNS | CLONE_NEWIPC | CLONE_NEWUTS | CLONE_PARENT | SIGCHLD;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (privateNetwork)
|
2023-02-10 15:44:25 +02:00
|
|
|
|
options.cloneFlags |= CLONE_NEWNET;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (usingUserNamespace)
|
2023-02-10 15:44:25 +02:00
|
|
|
|
options.cloneFlags |= CLONE_NEWUSER;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2023-02-10 15:44:25 +02:00
|
|
|
|
pid_t child = startProcess([&]() { runChild(); }, options);
|
2023-01-25 18:31:27 +02:00
|
|
|
|
|
Move pseudoterminal slave open to child
Hopefully this fixes "unexpected EOF" failures on macOS
(#3137, #3605, #7242, #7702).
The problem appears to be that under some circumstances, macOS
discards the output written to the slave side of the
pseudoterminal. Hence the parent never sees the "sandbox initialized"
message from the child, even though it succeeded. The conditions are:
* The child finishes very quickly. That's why this bug is likely to
trigger in nix-env tests, since that uses a builtin builder. Adding
a short sleep before the child exits makes the problem go away.
* The parent has closed its duplicate of the slave file
descriptor. This shouldn't matter, since the child has a duplicate
as well, but it does. E.g. moving the close to the bottom of
startBuilder() makes the problem go away. However, that's not a
solution because it would make Nix hang if the child dies before
sending the "sandbox initialized" message.
* The system is under high load. E.g. "make installcheck -j16" makes
the issue pretty reproducible, while it's very rare under "make
installcheck -j1".
As a fix/workaround, we now open the pseudoterminal slave in the
child, rather than the parent. This removes the second condition
(i.e. the parent no longer needs to close the slave fd) and I haven't
been able to reproduce the "unexpected EOF" with this.
2023-03-14 20:10:59 +02:00
|
|
|
|
writeFull(sendPid.writeSide.get(), fmt("%d\n", child));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
_exit(0);
|
2021-10-06 14:54:59 +03:00
|
|
|
|
});
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
Move pseudoterminal slave open to child
Hopefully this fixes "unexpected EOF" failures on macOS
(#3137, #3605, #7242, #7702).
The problem appears to be that under some circumstances, macOS
discards the output written to the slave side of the
pseudoterminal. Hence the parent never sees the "sandbox initialized"
message from the child, even though it succeeded. The conditions are:
* The child finishes very quickly. That's why this bug is likely to
trigger in nix-env tests, since that uses a builtin builder. Adding
a short sleep before the child exits makes the problem go away.
* The parent has closed its duplicate of the slave file
descriptor. This shouldn't matter, since the child has a duplicate
as well, but it does. E.g. moving the close to the bottom of
startBuilder() makes the problem go away. However, that's not a
solution because it would make Nix hang if the child dies before
sending the "sandbox initialized" message.
* The system is under high load. E.g. "make installcheck -j16" makes
the issue pretty reproducible, while it's very rare under "make
installcheck -j1".
As a fix/workaround, we now open the pseudoterminal slave in the
child, rather than the parent. This removes the second condition
(i.e. the parent no longer needs to close the slave fd) and I haven't
been able to reproduce the "unexpected EOF" with this.
2023-03-14 20:10:59 +02:00
|
|
|
|
sendPid.writeSide.close();
|
|
|
|
|
|
2023-01-25 18:31:27 +02:00
|
|
|
|
if (helper.wait() != 0)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw Error("unable to start build process");
|
|
|
|
|
|
|
|
|
|
userNamespaceSync.readSide = -1;
|
|
|
|
|
|
|
|
|
|
/* Close the write side to prevent runChild() from hanging
|
|
|
|
|
reading from this. */
|
|
|
|
|
Finally cleanup([&]() {
|
|
|
|
|
userNamespaceSync.writeSide = -1;
|
|
|
|
|
});
|
|
|
|
|
|
Move pseudoterminal slave open to child
Hopefully this fixes "unexpected EOF" failures on macOS
(#3137, #3605, #7242, #7702).
The problem appears to be that under some circumstances, macOS
discards the output written to the slave side of the
pseudoterminal. Hence the parent never sees the "sandbox initialized"
message from the child, even though it succeeded. The conditions are:
* The child finishes very quickly. That's why this bug is likely to
trigger in nix-env tests, since that uses a builtin builder. Adding
a short sleep before the child exits makes the problem go away.
* The parent has closed its duplicate of the slave file
descriptor. This shouldn't matter, since the child has a duplicate
as well, but it does. E.g. moving the close to the bottom of
startBuilder() makes the problem go away. However, that's not a
solution because it would make Nix hang if the child dies before
sending the "sandbox initialized" message.
* The system is under high load. E.g. "make installcheck -j16" makes
the issue pretty reproducible, while it's very rare under "make
installcheck -j1".
As a fix/workaround, we now open the pseudoterminal slave in the
child, rather than the parent. This removes the second condition
(i.e. the parent no longer needs to close the slave fd) and I haven't
been able to reproduce the "unexpected EOF" with this.
2023-03-14 20:10:59 +02:00
|
|
|
|
auto ss = tokenizeString<std::vector<std::string>>(readLine(sendPid.readSide.get()));
|
|
|
|
|
assert(ss.size() == 1);
|
|
|
|
|
pid = string2Int<pid_t>(ss[0]).value();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (usingUserNamespace) {
|
|
|
|
|
/* Set the UID/GID mapping of the builder's user namespace
|
|
|
|
|
such that the sandbox user maps to the build user, or to
|
|
|
|
|
the calling user (if build users are disabled). */
|
|
|
|
|
uid_t hostUid = buildUser ? buildUser->getUID() : getuid();
|
|
|
|
|
uid_t hostGid = buildUser ? buildUser->getGID() : getgid();
|
2022-11-08 17:03:42 +02:00
|
|
|
|
uid_t nrIds = buildUser ? buildUser->getUIDCount() : 1;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
writeFile("/proc/" + std::to_string(pid) + "/uid_map",
|
2021-04-27 22:06:58 +03:00
|
|
|
|
fmt("%d %d %d", sandboxUid(), hostUid, nrIds));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-11-08 17:03:42 +02:00
|
|
|
|
if (!buildUser || buildUser->getUIDCount() == 1)
|
2021-04-27 22:06:58 +03:00
|
|
|
|
writeFile("/proc/" + std::to_string(pid) + "/setgroups", "deny");
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
writeFile("/proc/" + std::to_string(pid) + "/gid_map",
|
2021-04-27 22:06:58 +03:00
|
|
|
|
fmt("%d %d %d", sandboxGid(), hostGid, nrIds));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
} else {
|
|
|
|
|
debug("note: not using a user namespace");
|
|
|
|
|
if (!buildUser)
|
|
|
|
|
throw Error("cannot perform a sandboxed build because user namespaces are not enabled; check /proc/sys/user/max_user_namespaces");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Now that we now the sandbox uid, we can write
|
|
|
|
|
/etc/passwd. */
|
|
|
|
|
writeFile(chrootRootDir + "/etc/passwd", fmt(
|
|
|
|
|
"root:x:0:0:Nix build user:%3%:/noshell\n"
|
|
|
|
|
"nixbld:x:%1%:%2%:Nix build user:%3%:/noshell\n"
|
|
|
|
|
"nobody:x:65534:65534:Nobody:/:/noshell\n",
|
|
|
|
|
sandboxUid(), sandboxGid(), settings.sandboxBuildDir));
|
|
|
|
|
|
2021-10-09 01:55:08 +03:00
|
|
|
|
/* Save the mount- and user namespace of the child. We have to do this
|
2021-02-23 08:26:35 +02:00
|
|
|
|
*before* the child does a chroot. */
|
|
|
|
|
sandboxMountNamespace = open(fmt("/proc/%d/ns/mnt", (pid_t) pid).c_str(), O_RDONLY);
|
|
|
|
|
if (sandboxMountNamespace.get() == -1)
|
|
|
|
|
throw SysError("getting sandbox mount namespace");
|
|
|
|
|
|
2022-02-12 17:28:36 +02:00
|
|
|
|
if (usingUserNamespace) {
|
|
|
|
|
sandboxUserNamespace = open(fmt("/proc/%d/ns/user", (pid_t) pid).c_str(), O_RDONLY);
|
|
|
|
|
if (sandboxUserNamespace.get() == -1)
|
|
|
|
|
throw SysError("getting sandbox user namespace");
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-04-27 22:06:58 +03:00
|
|
|
|
/* Move the child into its own cgroup. */
|
2022-11-18 11:39:28 +02:00
|
|
|
|
if (cgroup)
|
|
|
|
|
writeFile(*cgroup + "/cgroup.procs", fmt("%d", (pid_t) pid));
|
2021-04-27 22:06:58 +03:00
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Signal the builder that we've updated its user namespace. */
|
|
|
|
|
writeFull(userNamespaceSync.writeSide.get(), "1");
|
|
|
|
|
|
|
|
|
|
} else
|
|
|
|
|
#endif
|
|
|
|
|
{
|
|
|
|
|
pid = startProcess([&]() {
|
2023-03-20 18:04:57 +02:00
|
|
|
|
openSlave();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
runChild();
|
2021-10-06 14:54:59 +03:00
|
|
|
|
});
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* parent */
|
|
|
|
|
pid.setSeparatePG(true);
|
2023-03-15 11:37:39 +02:00
|
|
|
|
worker.childStarted(shared_from_this(), {builderOut.get()}, true, true);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Check if setting up the build environment failed. */
|
|
|
|
|
std::vector<std::string> msgs;
|
|
|
|
|
while (true) {
|
2022-02-25 17:00:00 +02:00
|
|
|
|
std::string msg = [&]() {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
try {
|
2023-03-15 11:37:39 +02:00
|
|
|
|
return readLine(builderOut.get());
|
2021-02-23 08:26:35 +02:00
|
|
|
|
} catch (Error & e) {
|
2021-09-23 12:42:33 +03:00
|
|
|
|
auto status = pid.wait();
|
|
|
|
|
e.addTrace({}, "while waiting for the build environment for '%s' to initialize (%s, previous messages: %s)",
|
|
|
|
|
worker.store.printStorePath(drvPath),
|
|
|
|
|
statusToString(status),
|
2021-02-23 08:26:35 +02:00
|
|
|
|
concatStringsSep("|", msgs));
|
2021-09-27 15:35:55 +03:00
|
|
|
|
throw;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
}();
|
2022-02-25 17:00:00 +02:00
|
|
|
|
if (msg.substr(0, 1) == "\2") break;
|
|
|
|
|
if (msg.substr(0, 1) == "\1") {
|
2023-03-15 11:37:39 +02:00
|
|
|
|
FdSource source(builderOut.get());
|
2021-02-23 08:26:35 +02:00
|
|
|
|
auto ex = readError(source);
|
|
|
|
|
ex.addTrace({}, "while setting up the build environment");
|
2021-10-11 11:47:02 +03:00
|
|
|
|
throw ex;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
debug("sandbox setup: " + msg);
|
|
|
|
|
msgs.push_back(std::move(msg));
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::initTmpDir() {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* In a sandbox, for determinism, always use the same temporary
|
|
|
|
|
directory. */
|
|
|
|
|
#if __linux__
|
|
|
|
|
tmpDirInSandbox = useChroot ? settings.sandboxBuildDir : tmpDir;
|
|
|
|
|
#else
|
|
|
|
|
tmpDirInSandbox = tmpDir;
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* In non-structured mode, add all bindings specified in the
|
|
|
|
|
derivation via the environment, except those listed in the
|
|
|
|
|
passAsFile attribute. Those are passed as file names pointing
|
|
|
|
|
to temporary files containing the contents. Note that
|
|
|
|
|
passAsFile is ignored in structure mode because it's not
|
|
|
|
|
needed (attributes are not passed through the environment, so
|
|
|
|
|
there is no size constraint). */
|
|
|
|
|
if (!parsedDrv->getStructuredAttrs()) {
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
StringSet passAsFile = tokenizeString<StringSet>(getOr(drv->env, "passAsFile", ""));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
for (auto & i : drv->env) {
|
|
|
|
|
if (passAsFile.find(i.first) == passAsFile.end()) {
|
|
|
|
|
env[i.first] = i.second;
|
|
|
|
|
} else {
|
|
|
|
|
auto hash = hashString(htSHA256, i.first);
|
2022-02-25 17:00:00 +02:00
|
|
|
|
std::string fn = ".attr-" + hash.to_string(Base32, false);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
Path p = tmpDir + "/" + fn;
|
|
|
|
|
writeFile(p, rewriteStrings(i.second, inputRewrites));
|
|
|
|
|
chownToBuilder(p);
|
|
|
|
|
env[i.first + "Path"] = tmpDirInSandbox + "/" + fn;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* For convenience, set an environment pointing to the top build
|
|
|
|
|
directory. */
|
|
|
|
|
env["NIX_BUILD_TOP"] = tmpDirInSandbox;
|
|
|
|
|
|
|
|
|
|
/* Also set TMPDIR and variants to point to this directory. */
|
|
|
|
|
env["TMPDIR"] = env["TEMPDIR"] = env["TMP"] = env["TEMP"] = tmpDirInSandbox;
|
|
|
|
|
|
|
|
|
|
/* Explicitly set PWD to prevent problems with chroot builds. In
|
|
|
|
|
particular, dietlibc cannot figure out the cwd because the
|
|
|
|
|
inode of the current directory doesn't appear in .. (because
|
|
|
|
|
getdents returns the inode of the mount point). */
|
|
|
|
|
env["PWD"] = tmpDirInSandbox;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::initEnv()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
env.clear();
|
|
|
|
|
|
|
|
|
|
/* Most shells initialise PATH to some default (/bin:/usr/bin:...) when
|
|
|
|
|
PATH is not set. We don't want this, so we fill it in with some dummy
|
|
|
|
|
value. */
|
|
|
|
|
env["PATH"] = "/path-not-set";
|
|
|
|
|
|
|
|
|
|
/* Set HOME to a non-existing path to prevent certain programs from using
|
|
|
|
|
/etc/passwd (or NIS, or whatever) to locate the home directory (for
|
|
|
|
|
example, wget looks for ~/.wgetrc). I.e., these tools use /etc/passwd
|
|
|
|
|
if HOME is not set, but they will just assume that the settings file
|
|
|
|
|
they are looking for does not exist if HOME is set but points to some
|
|
|
|
|
non-existing path. */
|
|
|
|
|
env["HOME"] = homeDir;
|
|
|
|
|
|
|
|
|
|
/* Tell the builder where the Nix store is. Usually they
|
|
|
|
|
shouldn't care, but this is useful for purity checking (e.g.,
|
|
|
|
|
the compiler or linker might only want to accept paths to files
|
|
|
|
|
in the store or in the build directory). */
|
|
|
|
|
env["NIX_STORE"] = worker.store.storeDir;
|
|
|
|
|
|
|
|
|
|
/* The maximum number of cores to utilize for parallel building. */
|
2023-03-02 16:44:19 +02:00
|
|
|
|
env["NIX_BUILD_CORES"] = fmt("%d", settings.buildCores);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
initTmpDir();
|
|
|
|
|
|
|
|
|
|
/* Compatibility hack with Nix <= 0.7: if this is a fixed-output
|
|
|
|
|
derivation, tell the builder, so that for instance `fetchurl'
|
|
|
|
|
can skip checking the output. On older Nixes, this environment
|
|
|
|
|
variable won't be set, so `fetchurl' will do the check. */
|
2022-03-18 02:36:52 +02:00
|
|
|
|
if (derivationType.isFixed()) env["NIX_OUTPUT_CHECKED"] = "1";
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* *Only* if this is a fixed-output derivation, propagate the
|
|
|
|
|
values of the environment variables specified in the
|
|
|
|
|
`impureEnvVars' attribute to the builder. This allows for
|
|
|
|
|
instance environment variables for proxy configuration such as
|
|
|
|
|
`http_proxy' to be easily passed to downloaders like
|
|
|
|
|
`fetchurl'. Passing such environment variables from the caller
|
|
|
|
|
to the builder is generally impure, but the output of
|
|
|
|
|
fixed-output derivations is by definition pure (since we
|
|
|
|
|
already know the cryptographic hash of the output). */
|
2022-03-31 17:06:40 +03:00
|
|
|
|
if (!derivationType.isSandboxed()) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
for (auto & i : parsedDrv->getStringsAttr("impureEnvVars").value_or(Strings()))
|
|
|
|
|
env[i] = getEnv(i).value_or("");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Currently structured log messages piggyback on stderr, but we
|
|
|
|
|
may change that in the future. So tell the builder which file
|
|
|
|
|
descriptor to use for that. */
|
|
|
|
|
env["NIX_LOG_FD"] = "2";
|
|
|
|
|
|
|
|
|
|
/* Trigger colored output in various tools. */
|
|
|
|
|
env["TERM"] = "xterm-256color";
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::writeStructuredAttrs()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
2021-06-24 16:06:07 +03:00
|
|
|
|
if (auto structAttrsJson = parsedDrv->prepareStructuredAttrs(worker.store, inputPaths)) {
|
2021-05-13 17:11:56 +03:00
|
|
|
|
auto json = structAttrsJson.value();
|
2021-06-24 16:06:07 +03:00
|
|
|
|
nlohmann::json rewritten;
|
|
|
|
|
for (auto & [i, v] : json["outputs"].get<nlohmann::json::object_t>()) {
|
|
|
|
|
/* The placeholder must have a rewrite, so we use it to cover both the
|
|
|
|
|
cases where we know or don't know the output path ahead of time. */
|
2022-02-25 17:00:00 +02:00
|
|
|
|
rewritten[i] = rewriteStrings((std::string) v, inputRewrites);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2021-06-24 16:06:07 +03:00
|
|
|
|
json["outputs"] = rewritten;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-06-22 21:37:25 +03:00
|
|
|
|
auto jsonSh = writeStructuredAttrsShell(json);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-05-02 18:24:14 +03:00
|
|
|
|
writeFile(tmpDir + "/.attrs.sh", rewriteStrings(jsonSh, inputRewrites));
|
|
|
|
|
chownToBuilder(tmpDir + "/.attrs.sh");
|
2023-03-16 16:21:02 +02:00
|
|
|
|
env["NIX_ATTRS_SH_FILE"] = tmpDirInSandbox + "/.attrs.sh";
|
2021-05-02 18:24:14 +03:00
|
|
|
|
writeFile(tmpDir + "/.attrs.json", rewriteStrings(json.dump(), inputRewrites));
|
|
|
|
|
chownToBuilder(tmpDir + "/.attrs.json");
|
2023-03-16 16:21:02 +02:00
|
|
|
|
env["NIX_ATTRS_JSON_FILE"] = tmpDirInSandbox + "/.attrs.json";
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2021-03-02 05:50:41 +02:00
|
|
|
|
|
2021-04-05 16:48:18 +03:00
|
|
|
|
static StorePath pathPartOfReq(const DerivedPath & req)
|
2021-03-02 05:50:41 +02:00
|
|
|
|
{
|
|
|
|
|
return std::visit(overloaded {
|
2021-10-01 00:31:21 +03:00
|
|
|
|
[&](const DerivedPath::Opaque & bo) {
|
2021-03-02 05:50:41 +02:00
|
|
|
|
return bo.path;
|
|
|
|
|
},
|
2021-10-01 00:31:21 +03:00
|
|
|
|
[&](const DerivedPath::Built & bfd) {
|
2021-03-02 05:50:41 +02:00
|
|
|
|
return bfd.drvPath;
|
|
|
|
|
},
|
2021-04-05 16:24:42 +03:00
|
|
|
|
}, req.raw());
|
2021-03-02 05:50:41 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-04-05 16:48:18 +03:00
|
|
|
|
bool LocalDerivationGoal::isAllowed(const DerivedPath & req)
|
2021-03-02 05:50:41 +02:00
|
|
|
|
{
|
|
|
|
|
return this->isAllowed(pathPartOfReq(req));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
struct RestrictedStoreConfig : virtual LocalFSStoreConfig
|
|
|
|
|
{
|
|
|
|
|
using LocalFSStoreConfig::LocalFSStoreConfig;
|
|
|
|
|
const std::string name() { return "Restricted Store"; }
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
/* A wrapper around LocalStore that only allows building/querying of
|
|
|
|
|
paths that are in the input closures of the build or were added via
|
|
|
|
|
recursive Nix calls. */
|
2022-03-01 20:31:36 +02:00
|
|
|
|
struct RestrictedStore : public virtual RestrictedStoreConfig, public virtual LocalFSStore, public virtual GcStore
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
ref<LocalStore> next;
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
LocalDerivationGoal & goal;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
RestrictedStore(const Params & params, ref<LocalStore> next, LocalDerivationGoal & goal)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
: StoreConfig(params)
|
|
|
|
|
, LocalFSStoreConfig(params)
|
|
|
|
|
, RestrictedStoreConfig(params)
|
|
|
|
|
, Store(params)
|
|
|
|
|
, LocalFSStore(params)
|
|
|
|
|
, next(next), goal(goal)
|
|
|
|
|
{ }
|
|
|
|
|
|
|
|
|
|
Path getRealStoreDir() override
|
|
|
|
|
{ return next->realStoreDir; }
|
|
|
|
|
|
|
|
|
|
std::string getUri() override
|
|
|
|
|
{ return next->getUri(); }
|
|
|
|
|
|
|
|
|
|
StorePathSet queryAllValidPaths() override
|
|
|
|
|
{
|
|
|
|
|
StorePathSet paths;
|
|
|
|
|
for (auto & p : goal.inputPaths) paths.insert(p);
|
|
|
|
|
for (auto & p : goal.addedPaths) paths.insert(p);
|
|
|
|
|
return paths;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void queryPathInfoUncached(const StorePath & path,
|
|
|
|
|
Callback<std::shared_ptr<const ValidPathInfo>> callback) noexcept override
|
|
|
|
|
{
|
|
|
|
|
if (goal.isAllowed(path)) {
|
|
|
|
|
try {
|
|
|
|
|
/* Censor impure information. */
|
|
|
|
|
auto info = std::make_shared<ValidPathInfo>(*next->queryPathInfo(path));
|
|
|
|
|
info->deriver.reset();
|
|
|
|
|
info->registrationTime = 0;
|
|
|
|
|
info->ultimate = false;
|
|
|
|
|
info->sigs.clear();
|
|
|
|
|
callback(info);
|
|
|
|
|
} catch (InvalidPath &) {
|
|
|
|
|
callback(nullptr);
|
|
|
|
|
}
|
|
|
|
|
} else
|
|
|
|
|
callback(nullptr);
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
void queryReferrers(const StorePath & path, StorePathSet & referrers) override
|
|
|
|
|
{ }
|
|
|
|
|
|
|
|
|
|
std::map<std::string, std::optional<StorePath>> queryPartialDerivationOutputMap(const StorePath & path) override
|
|
|
|
|
{
|
|
|
|
|
if (!goal.isAllowed(path))
|
|
|
|
|
throw InvalidPath("cannot query output map for unknown path '%s' in recursive Nix", printStorePath(path));
|
|
|
|
|
return next->queryPartialDerivationOutputMap(path);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
std::optional<StorePath> queryPathFromHashPart(const std::string & hashPart) override
|
|
|
|
|
{ throw Error("queryPathFromHashPart"); }
|
|
|
|
|
|
2022-02-25 17:00:00 +02:00
|
|
|
|
StorePath addToStore(
|
|
|
|
|
std::string_view name,
|
|
|
|
|
const Path & srcPath,
|
|
|
|
|
FileIngestionMethod method,
|
|
|
|
|
HashType hashAlgo,
|
|
|
|
|
PathFilter & filter,
|
|
|
|
|
RepairFlag repair,
|
|
|
|
|
const StorePathSet & references) override
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{ throw Error("addToStore"); }
|
|
|
|
|
|
|
|
|
|
void addToStore(const ValidPathInfo & info, Source & narSource,
|
|
|
|
|
RepairFlag repair = NoRepair, CheckSigsFlag checkSigs = CheckSigs) override
|
|
|
|
|
{
|
|
|
|
|
next->addToStore(info, narSource, repair, checkSigs);
|
|
|
|
|
goal.addDependency(info.path);
|
|
|
|
|
}
|
|
|
|
|
|
2022-02-25 17:00:00 +02:00
|
|
|
|
StorePath addTextToStore(
|
|
|
|
|
std::string_view name,
|
|
|
|
|
std::string_view s,
|
|
|
|
|
const StorePathSet & references,
|
|
|
|
|
RepairFlag repair = NoRepair) override
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
auto path = next->addTextToStore(name, s, references, repair);
|
|
|
|
|
goal.addDependency(path);
|
|
|
|
|
return path;
|
|
|
|
|
}
|
|
|
|
|
|
2022-02-25 17:00:00 +02:00
|
|
|
|
StorePath addToStoreFromDump(
|
|
|
|
|
Source & dump,
|
|
|
|
|
std::string_view name,
|
|
|
|
|
FileIngestionMethod method,
|
|
|
|
|
HashType hashAlgo,
|
|
|
|
|
RepairFlag repair,
|
|
|
|
|
const StorePathSet & references) override
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
2021-10-23 21:30:51 +03:00
|
|
|
|
auto path = next->addToStoreFromDump(dump, name, method, hashAlgo, repair, references);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
goal.addDependency(path);
|
|
|
|
|
return path;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void narFromPath(const StorePath & path, Sink & sink) override
|
|
|
|
|
{
|
|
|
|
|
if (!goal.isAllowed(path))
|
|
|
|
|
throw InvalidPath("cannot dump unknown path '%s' in recursive Nix", printStorePath(path));
|
|
|
|
|
LocalFSStore::narFromPath(path, sink);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void ensurePath(const StorePath & path) override
|
|
|
|
|
{
|
|
|
|
|
if (!goal.isAllowed(path))
|
|
|
|
|
throw InvalidPath("cannot substitute unknown path '%s' in recursive Nix", printStorePath(path));
|
|
|
|
|
/* Nothing to be done; 'path' must already be valid. */
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void registerDrvOutput(const Realisation & info) override
|
|
|
|
|
// XXX: This should probably be allowed as a no-op if the realisation
|
|
|
|
|
// corresponds to an allowed derivation
|
|
|
|
|
{ throw Error("registerDrvOutput"); }
|
|
|
|
|
|
2021-10-27 12:36:51 +03:00
|
|
|
|
void queryRealisationUncached(const DrvOutput & id,
|
|
|
|
|
Callback<std::shared_ptr<const Realisation>> callback) noexcept override
|
2021-02-23 08:26:35 +02:00
|
|
|
|
// XXX: This should probably be allowed if the realisation corresponds to
|
|
|
|
|
// an allowed derivation
|
2021-06-23 18:27:18 +03:00
|
|
|
|
{
|
|
|
|
|
if (!goal.isAllowed(id))
|
2021-10-27 12:36:51 +03:00
|
|
|
|
callback(nullptr);
|
|
|
|
|
next->queryRealisation(id, std::move(callback));
|
2021-06-23 18:27:18 +03:00
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-07-19 16:43:08 +03:00
|
|
|
|
void buildPaths(const std::vector<DerivedPath> & paths, BuildMode buildMode, std::shared_ptr<Store> evalStore) override
|
2022-03-08 20:50:46 +02:00
|
|
|
|
{
|
|
|
|
|
for (auto & result : buildPathsWithResults(paths, buildMode, evalStore))
|
|
|
|
|
if (!result.success())
|
|
|
|
|
result.rethrow();
|
|
|
|
|
}
|
|
|
|
|
|
Make `KeyedBuildResult`, `BuildResult` like before, and fix bug another way
In https://github.com/NixOS/nix/pull/6311#discussion_r834863823, I
realized since derivation goals' wanted outputs can "grow" due to
overlapping dependencies (See `DerivationGoal::addWantedOutputs`, called
by `Worker::makeDerivationGoalCommon`), the previous bug fix had an
unfortunate side effect of causing more pointless rebuilds.
In paticular, we have this situation:
1. Goal made from `DerivedPath::Built { foo, {a} }`.
2. Goal gives on on substituting, starts building.
3. Goal made from `DerivedPath::Built { foo, {b} }`, in fact is just
modified original goal.
4. Though the goal had gotten as far as building, so all outputs were
going to be produced, `addWantedOutputs` no longer knows that and so
the goal is flagged to be restarted.
This might sound far-fetched with input-addressed drvs, where we usually
basically have all our goals "planned out" before we start doing
anything, but with CA derivation goals and especially RFC 92, where *drv
resolution* means goals are created after some building is completed, it
is more likely to happen.
So the first thing to do was restore the clearing of `wantedOutputs` we
used to do, and then filter the outputs in `buildPathsWithResults` to
only get the ones we care about.
But fix also has its own side effect in that the `DerivedPath` in the
`BuildResult` in `DerivationGoal` cannot be trusted; it is merely the
*first* `DerivedPath` for which this goal was originally created.
To remedy this, I made `BuildResult` be like it was before, and instead
made `KeyedBuildResult` be a subclass wit the path. Only
`buildPathsWithResults` returns `KeyedBuildResult`s, everything else
just becomes like it was before, where the "key" is unambiguous from
context.
I think separating the "primary key" field(s) from the other fields is
good practical in general anyways. (I would like to do the same thing
for `ValidPathInfo`.) Among other things, it allows constructions like
`std::map<Key, ThingWithKey>` where doesn't contain duplicate keys and
just precludes the possibility of those duplicate keys being out of
sync.
We might leverage the above someday to overload `buildPathsWithResults`
to take a *set* of return a *map* per the above.
-----
Unfortunately, we need to avoid C++20 strictness on designated
initializers.
(BTW
https://www.open-std.org/jtc1/sc22/wg21/docs/papers/2021/p2287r1.html
this offers some new syntax for this use-case. Hopefully this will be
adopted and we can eventually use it.)
No having that yet, maybe it would be better to not make
`KeyedBuildResult` a subclass to just avoid this.
Co-authored-by: Robert Hensing <roberth@users.noreply.github.com>
2022-03-25 03:26:07 +02:00
|
|
|
|
std::vector<KeyedBuildResult> buildPathsWithResults(
|
2022-03-08 20:50:46 +02:00
|
|
|
|
const std::vector<DerivedPath> & paths,
|
|
|
|
|
BuildMode buildMode = bmNormal,
|
|
|
|
|
std::shared_ptr<Store> evalStore = nullptr) override
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
2021-07-19 16:43:08 +03:00
|
|
|
|
assert(!evalStore);
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (buildMode != bmNormal) throw Error("unsupported build mode");
|
|
|
|
|
|
|
|
|
|
StorePathSet newPaths;
|
2021-06-23 18:27:18 +03:00
|
|
|
|
std::set<Realisation> newRealisations;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-03-02 05:50:41 +02:00
|
|
|
|
for (auto & req : paths) {
|
|
|
|
|
if (!goal.isAllowed(req))
|
2021-04-05 16:24:42 +03:00
|
|
|
|
throw InvalidPath("cannot build '%s' in recursive Nix because path is unknown", req.to_string(*next));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2022-03-08 20:50:46 +02:00
|
|
|
|
auto results = next->buildPathsWithResults(paths, buildMode);
|
|
|
|
|
|
|
|
|
|
for (auto & result : results) {
|
|
|
|
|
for (auto & [outputName, output] : result.builtOutputs) {
|
|
|
|
|
newPaths.insert(output.outPath);
|
|
|
|
|
newRealisations.insert(output);
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
StorePathSet closure;
|
|
|
|
|
next->computeFSClosure(newPaths, closure);
|
|
|
|
|
for (auto & path : closure)
|
|
|
|
|
goal.addDependency(path);
|
2021-06-23 18:27:18 +03:00
|
|
|
|
for (auto & real : Realisation::closure(*next, newRealisations))
|
|
|
|
|
goal.addedDrvOutputs.insert(real.id);
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
|
|
|
|
return results;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
BuildResult buildDerivation(const StorePath & drvPath, const BasicDerivation & drv,
|
|
|
|
|
BuildMode buildMode = bmNormal) override
|
|
|
|
|
{ unsupported("buildDerivation"); }
|
|
|
|
|
|
|
|
|
|
void addTempRoot(const StorePath & path) override
|
|
|
|
|
{ }
|
|
|
|
|
|
|
|
|
|
void addIndirectRoot(const Path & path) override
|
|
|
|
|
{ }
|
|
|
|
|
|
|
|
|
|
Roots findRoots(bool censor) override
|
|
|
|
|
{ return Roots(); }
|
|
|
|
|
|
|
|
|
|
void collectGarbage(const GCOptions & options, GCResults & results) override
|
|
|
|
|
{ }
|
|
|
|
|
|
|
|
|
|
void addSignatures(const StorePath & storePath, const StringSet & sigs) override
|
|
|
|
|
{ unsupported("addSignatures"); }
|
|
|
|
|
|
2021-04-05 16:48:18 +03:00
|
|
|
|
void queryMissing(const std::vector<DerivedPath> & targets,
|
2021-02-23 08:26:35 +02:00
|
|
|
|
StorePathSet & willBuild, StorePathSet & willSubstitute, StorePathSet & unknown,
|
|
|
|
|
uint64_t & downloadSize, uint64_t & narSize) override
|
|
|
|
|
{
|
|
|
|
|
/* This is slightly impure since it leaks information to the
|
|
|
|
|
client about what paths will be built/substituted or are
|
|
|
|
|
already present. Probably not a big deal. */
|
|
|
|
|
|
2021-04-05 16:48:18 +03:00
|
|
|
|
std::vector<DerivedPath> allowed;
|
2021-03-02 05:50:41 +02:00
|
|
|
|
for (auto & req : targets) {
|
|
|
|
|
if (goal.isAllowed(req))
|
|
|
|
|
allowed.emplace_back(req);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
else
|
2021-03-02 05:50:41 +02:00
|
|
|
|
unknown.insert(pathPartOfReq(req));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
next->queryMissing(allowed, willBuild, willSubstitute,
|
|
|
|
|
unknown, downloadSize, narSize);
|
|
|
|
|
}
|
2022-03-08 20:20:39 +02:00
|
|
|
|
|
2022-12-15 22:58:54 +02:00
|
|
|
|
virtual std::optional<std::string> getBuildLogExact(const StorePath & path) override
|
2022-03-08 20:20:39 +02:00
|
|
|
|
{ return std::nullopt; }
|
|
|
|
|
|
|
|
|
|
virtual void addBuildLog(const StorePath & path, std::string_view log) override
|
|
|
|
|
{ unsupported("addBuildLog"); }
|
2022-12-26 22:21:08 +02:00
|
|
|
|
|
|
|
|
|
std::optional<TrustedFlag> isTrustedClient() override
|
|
|
|
|
{ return NotTrusted; }
|
2021-02-23 08:26:35 +02:00
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::startDaemon()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
2023-03-17 16:33:48 +02:00
|
|
|
|
experimentalFeatureSettings.require(Xp::RecursiveNix);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
Store::Params params;
|
|
|
|
|
params["path-info-cache-size"] = "0";
|
|
|
|
|
params["store"] = worker.store.storeDir;
|
2021-02-26 17:20:33 +02:00
|
|
|
|
params["root"] = getLocalStore().rootDir;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
params["state"] = "/no-such-path";
|
|
|
|
|
params["log"] = "/no-such-path";
|
|
|
|
|
auto store = make_ref<RestrictedStore>(params,
|
|
|
|
|
ref<LocalStore>(std::dynamic_pointer_cast<LocalStore>(worker.store.shared_from_this())),
|
|
|
|
|
*this);
|
|
|
|
|
|
|
|
|
|
addedPaths.clear();
|
|
|
|
|
|
|
|
|
|
auto socketName = ".nix-socket";
|
|
|
|
|
Path socketPath = tmpDir + "/" + socketName;
|
|
|
|
|
env["NIX_REMOTE"] = "unix://" + tmpDirInSandbox + "/" + socketName;
|
|
|
|
|
|
|
|
|
|
daemonSocket = createUnixDomainSocket(socketPath, 0600);
|
|
|
|
|
|
|
|
|
|
chownToBuilder(socketPath);
|
|
|
|
|
|
|
|
|
|
daemonThread = std::thread([this, store]() {
|
|
|
|
|
|
|
|
|
|
while (true) {
|
|
|
|
|
|
|
|
|
|
/* Accept a connection. */
|
|
|
|
|
struct sockaddr_un remoteAddr;
|
|
|
|
|
socklen_t remoteAddrLen = sizeof(remoteAddr);
|
|
|
|
|
|
|
|
|
|
AutoCloseFD remote = accept(daemonSocket.get(),
|
|
|
|
|
(struct sockaddr *) &remoteAddr, &remoteAddrLen);
|
|
|
|
|
if (!remote) {
|
2021-08-16 21:03:32 +03:00
|
|
|
|
if (errno == EINTR || errno == EAGAIN) continue;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (errno == EINVAL) break;
|
|
|
|
|
throw SysError("accepting connection");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
closeOnExec(remote.get());
|
|
|
|
|
|
|
|
|
|
debug("received daemon connection");
|
|
|
|
|
|
|
|
|
|
auto workerThread = std::thread([store, remote{std::move(remote)}]() {
|
|
|
|
|
FdSource from(remote.get());
|
|
|
|
|
FdSink to(remote.get());
|
|
|
|
|
try {
|
|
|
|
|
daemon::processConnection(store, from, to,
|
2022-12-26 22:21:08 +02:00
|
|
|
|
NotTrusted, daemon::Recursive);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
debug("terminated daemon connection");
|
|
|
|
|
} catch (SysError &) {
|
|
|
|
|
ignoreException();
|
|
|
|
|
}
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
daemonWorkerThreads.push_back(std::move(workerThread));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
debug("daemon shutting down");
|
|
|
|
|
});
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::stopDaemon()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
if (daemonSocket && shutdown(daemonSocket.get(), SHUT_RDWR) == -1)
|
|
|
|
|
throw SysError("shutting down daemon socket");
|
|
|
|
|
|
|
|
|
|
if (daemonThread.joinable())
|
|
|
|
|
daemonThread.join();
|
|
|
|
|
|
|
|
|
|
// FIXME: should prune worker threads more quickly.
|
|
|
|
|
// FIXME: shutdown the client socket to speed up worker termination.
|
|
|
|
|
for (auto & thread : daemonWorkerThreads)
|
|
|
|
|
thread.join();
|
|
|
|
|
daemonWorkerThreads.clear();
|
|
|
|
|
|
|
|
|
|
daemonSocket = -1;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::addDependency(const StorePath & path)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
if (isAllowed(path)) return;
|
|
|
|
|
|
|
|
|
|
addedPaths.insert(path);
|
|
|
|
|
|
|
|
|
|
/* If we're doing a sandbox build, then we have to make the path
|
|
|
|
|
appear in the sandbox. */
|
|
|
|
|
if (useChroot) {
|
|
|
|
|
|
|
|
|
|
debug("materialising '%s' in the sandbox", worker.store.printStorePath(path));
|
|
|
|
|
|
|
|
|
|
#if __linux__
|
|
|
|
|
|
|
|
|
|
Path source = worker.store.Store::toRealPath(path);
|
|
|
|
|
Path target = chrootRootDir + worker.store.printStorePath(path);
|
|
|
|
|
debug("bind-mounting %s -> %s", target, source);
|
|
|
|
|
|
|
|
|
|
if (pathExists(target))
|
|
|
|
|
throw Error("store path '%s' already exists in the sandbox", worker.store.printStorePath(path));
|
|
|
|
|
|
|
|
|
|
auto st = lstat(source);
|
|
|
|
|
|
|
|
|
|
if (S_ISDIR(st.st_mode)) {
|
|
|
|
|
|
|
|
|
|
/* Bind-mount the path into the sandbox. This requires
|
|
|
|
|
entering its mount namespace, which is not possible
|
|
|
|
|
in multithreaded programs. So we do this in a
|
|
|
|
|
child process.*/
|
|
|
|
|
Pid child(startProcess([&]() {
|
|
|
|
|
|
2021-10-09 01:55:08 +03:00
|
|
|
|
if (usingUserNamespace && (setns(sandboxUserNamespace.get(), 0) == -1))
|
|
|
|
|
throw SysError("entering sandbox user namespace");
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (setns(sandboxMountNamespace.get(), 0) == -1)
|
|
|
|
|
throw SysError("entering sandbox mount namespace");
|
|
|
|
|
|
|
|
|
|
createDirs(target);
|
|
|
|
|
|
|
|
|
|
if (mount(source.c_str(), target.c_str(), "", MS_BIND, 0) == -1)
|
|
|
|
|
throw SysError("bind mount from '%s' to '%s' failed", source, target);
|
|
|
|
|
|
|
|
|
|
_exit(0);
|
|
|
|
|
}));
|
|
|
|
|
|
|
|
|
|
int status = child.wait();
|
|
|
|
|
if (status != 0)
|
|
|
|
|
throw Error("could not add path '%s' to sandbox", worker.store.printStorePath(path));
|
|
|
|
|
|
|
|
|
|
} else
|
|
|
|
|
linkOrCopy(source, target);
|
|
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
throw Error("don't know how to make path '%s' (produced by a recursive Nix call) appear in the sandbox",
|
|
|
|
|
worker.store.printStorePath(path));
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::chownToBuilder(const Path & path)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
if (!buildUser) return;
|
|
|
|
|
if (chown(path.c_str(), buildUser->getUID(), buildUser->getGID()) == -1)
|
|
|
|
|
throw SysError("cannot change ownership of '%1%'", path);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void setupSeccomp()
|
|
|
|
|
{
|
|
|
|
|
#if __linux__
|
|
|
|
|
if (!settings.filterSyscalls) return;
|
|
|
|
|
#if HAVE_SECCOMP
|
|
|
|
|
scmp_filter_ctx ctx;
|
|
|
|
|
|
|
|
|
|
if (!(ctx = seccomp_init(SCMP_ACT_ALLOW)))
|
|
|
|
|
throw SysError("unable to initialize seccomp mode 2");
|
|
|
|
|
|
|
|
|
|
Finally cleanup([&]() {
|
|
|
|
|
seccomp_release(ctx);
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
if (nativeSystem == "x86_64-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_X86) != 0)
|
|
|
|
|
throw SysError("unable to add 32-bit seccomp architecture");
|
|
|
|
|
|
|
|
|
|
if (nativeSystem == "x86_64-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_X32) != 0)
|
|
|
|
|
throw SysError("unable to add X32 seccomp architecture");
|
|
|
|
|
|
|
|
|
|
if (nativeSystem == "aarch64-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_ARM) != 0)
|
|
|
|
|
printError("unable to add ARM seccomp architecture; this may result in spurious build failures if running 32-bit ARM processes");
|
|
|
|
|
|
2022-04-28 03:55:04 +03:00
|
|
|
|
if (nativeSystem == "mips64-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_MIPS) != 0)
|
|
|
|
|
printError("unable to add mips seccomp architecture");
|
|
|
|
|
|
|
|
|
|
if (nativeSystem == "mips64-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_MIPS64N32) != 0)
|
|
|
|
|
printError("unable to add mips64-*abin32 seccomp architecture");
|
|
|
|
|
|
|
|
|
|
if (nativeSystem == "mips64el-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_MIPSEL) != 0)
|
|
|
|
|
printError("unable to add mipsel seccomp architecture");
|
|
|
|
|
|
|
|
|
|
if (nativeSystem == "mips64el-linux" &&
|
|
|
|
|
seccomp_arch_add(ctx, SCMP_ARCH_MIPSEL64N32) != 0)
|
|
|
|
|
printError("unable to add mips64el-*abin32 seccomp architecture");
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Prevent builders from creating setuid/setgid binaries. */
|
|
|
|
|
for (int perm : { S_ISUID, S_ISGID }) {
|
|
|
|
|
if (seccomp_rule_add(ctx, SCMP_ACT_ERRNO(EPERM), SCMP_SYS(chmod), 1,
|
|
|
|
|
SCMP_A1(SCMP_CMP_MASKED_EQ, (scmp_datum_t) perm, (scmp_datum_t) perm)) != 0)
|
|
|
|
|
throw SysError("unable to add seccomp rule");
|
|
|
|
|
|
|
|
|
|
if (seccomp_rule_add(ctx, SCMP_ACT_ERRNO(EPERM), SCMP_SYS(fchmod), 1,
|
|
|
|
|
SCMP_A1(SCMP_CMP_MASKED_EQ, (scmp_datum_t) perm, (scmp_datum_t) perm)) != 0)
|
|
|
|
|
throw SysError("unable to add seccomp rule");
|
|
|
|
|
|
|
|
|
|
if (seccomp_rule_add(ctx, SCMP_ACT_ERRNO(EPERM), SCMP_SYS(fchmodat), 1,
|
|
|
|
|
SCMP_A2(SCMP_CMP_MASKED_EQ, (scmp_datum_t) perm, (scmp_datum_t) perm)) != 0)
|
|
|
|
|
throw SysError("unable to add seccomp rule");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Prevent builders from creating EAs or ACLs. Not all filesystems
|
|
|
|
|
support these, and they're not allowed in the Nix store because
|
|
|
|
|
they're not representable in the NAR serialisation. */
|
|
|
|
|
if (seccomp_rule_add(ctx, SCMP_ACT_ERRNO(ENOTSUP), SCMP_SYS(setxattr), 0) != 0 ||
|
|
|
|
|
seccomp_rule_add(ctx, SCMP_ACT_ERRNO(ENOTSUP), SCMP_SYS(lsetxattr), 0) != 0 ||
|
|
|
|
|
seccomp_rule_add(ctx, SCMP_ACT_ERRNO(ENOTSUP), SCMP_SYS(fsetxattr), 0) != 0)
|
|
|
|
|
throw SysError("unable to add seccomp rule");
|
|
|
|
|
|
|
|
|
|
if (seccomp_attr_set(ctx, SCMP_FLTATR_CTL_NNP, settings.allowNewPrivileges ? 0 : 1) != 0)
|
|
|
|
|
throw SysError("unable to set 'no new privileges' seccomp attribute");
|
|
|
|
|
|
|
|
|
|
if (seccomp_load(ctx) != 0)
|
|
|
|
|
throw SysError("unable to load seccomp BPF program");
|
|
|
|
|
#else
|
|
|
|
|
throw Error(
|
|
|
|
|
"seccomp is not supported on this platform; "
|
|
|
|
|
"you can bypass this error by setting the option 'filter-syscalls' to false, but note that untrusted builds can then create setuid binaries!");
|
|
|
|
|
#endif
|
|
|
|
|
#endif
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::runChild()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
/* Warning: in the child we should absolutely not make any SQLite
|
|
|
|
|
calls! */
|
|
|
|
|
|
2022-10-13 22:35:16 +03:00
|
|
|
|
bool sendException = true;
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
try { /* child */
|
|
|
|
|
|
2023-03-20 19:06:08 +02:00
|
|
|
|
commonChildInit();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
try {
|
|
|
|
|
setupSeccomp();
|
|
|
|
|
} catch (...) {
|
|
|
|
|
if (buildUser) throw;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
bool setUser = true;
|
|
|
|
|
|
|
|
|
|
/* Make the contents of netrc available to builtin:fetchurl
|
|
|
|
|
(which may run under a different uid and/or in a sandbox). */
|
|
|
|
|
std::string netrcData;
|
|
|
|
|
try {
|
|
|
|
|
if (drv->isBuiltin() && drv->builder == "builtin:fetchurl")
|
|
|
|
|
netrcData = readFile(settings.netrcFile);
|
|
|
|
|
} catch (SysError &) { }
|
|
|
|
|
|
|
|
|
|
#if __linux__
|
|
|
|
|
if (useChroot) {
|
|
|
|
|
|
|
|
|
|
userNamespaceSync.writeSide = -1;
|
|
|
|
|
|
|
|
|
|
if (drainFD(userNamespaceSync.readSide.get()) != "1")
|
|
|
|
|
throw Error("user namespace initialisation failed");
|
|
|
|
|
|
|
|
|
|
userNamespaceSync.readSide = -1;
|
|
|
|
|
|
|
|
|
|
if (privateNetwork) {
|
|
|
|
|
|
|
|
|
|
/* Initialise the loopback interface. */
|
|
|
|
|
AutoCloseFD fd(socket(PF_INET, SOCK_DGRAM, IPPROTO_IP));
|
|
|
|
|
if (!fd) throw SysError("cannot open IP socket");
|
|
|
|
|
|
|
|
|
|
struct ifreq ifr;
|
|
|
|
|
strcpy(ifr.ifr_name, "lo");
|
|
|
|
|
ifr.ifr_flags = IFF_UP | IFF_LOOPBACK | IFF_RUNNING;
|
|
|
|
|
if (ioctl(fd.get(), SIOCSIFFLAGS, &ifr) == -1)
|
|
|
|
|
throw SysError("cannot set loopback interface flags");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Set the hostname etc. to fixed values. */
|
|
|
|
|
char hostname[] = "localhost";
|
|
|
|
|
if (sethostname(hostname, sizeof(hostname)) == -1)
|
|
|
|
|
throw SysError("cannot set host name");
|
|
|
|
|
char domainname[] = "(none)"; // kernel default
|
|
|
|
|
if (setdomainname(domainname, sizeof(domainname)) == -1)
|
|
|
|
|
throw SysError("cannot set domain name");
|
|
|
|
|
|
|
|
|
|
/* Make all filesystems private. This is necessary
|
|
|
|
|
because subtrees may have been mounted as "shared"
|
|
|
|
|
(MS_SHARED). (Systemd does this, for instance.) Even
|
|
|
|
|
though we have a private mount namespace, mounting
|
|
|
|
|
filesystems on top of a shared subtree still propagates
|
|
|
|
|
outside of the namespace. Making a subtree private is
|
|
|
|
|
local to the namespace, though, so setting MS_PRIVATE
|
|
|
|
|
does not affect the outside world. */
|
|
|
|
|
if (mount(0, "/", 0, MS_PRIVATE | MS_REC, 0) == -1)
|
|
|
|
|
throw SysError("unable to make '/' private");
|
|
|
|
|
|
|
|
|
|
/* Bind-mount chroot directory to itself, to treat it as a
|
|
|
|
|
different filesystem from /, as needed for pivot_root. */
|
|
|
|
|
if (mount(chrootRootDir.c_str(), chrootRootDir.c_str(), 0, MS_BIND, 0) == -1)
|
|
|
|
|
throw SysError("unable to bind mount '%1%'", chrootRootDir);
|
|
|
|
|
|
|
|
|
|
/* Bind-mount the sandbox's Nix store onto itself so that
|
|
|
|
|
we can mark it as a "shared" subtree, allowing bind
|
|
|
|
|
mounts made in *this* mount namespace to be propagated
|
|
|
|
|
into the child namespace created by the
|
|
|
|
|
unshare(CLONE_NEWNS) call below.
|
|
|
|
|
|
|
|
|
|
Marking chrootRootDir as MS_SHARED causes pivot_root()
|
|
|
|
|
to fail with EINVAL. Don't know why. */
|
|
|
|
|
Path chrootStoreDir = chrootRootDir + worker.store.storeDir;
|
|
|
|
|
|
|
|
|
|
if (mount(chrootStoreDir.c_str(), chrootStoreDir.c_str(), 0, MS_BIND, 0) == -1)
|
|
|
|
|
throw SysError("unable to bind mount the Nix store", chrootStoreDir);
|
|
|
|
|
|
|
|
|
|
if (mount(0, chrootStoreDir.c_str(), 0, MS_SHARED, 0) == -1)
|
|
|
|
|
throw SysError("unable to make '%s' shared", chrootStoreDir);
|
|
|
|
|
|
|
|
|
|
/* Set up a nearly empty /dev, unless the user asked to
|
|
|
|
|
bind-mount the host /dev. */
|
|
|
|
|
Strings ss;
|
|
|
|
|
if (dirsInChroot.find("/dev") == dirsInChroot.end()) {
|
|
|
|
|
createDirs(chrootRootDir + "/dev/shm");
|
|
|
|
|
createDirs(chrootRootDir + "/dev/pts");
|
|
|
|
|
ss.push_back("/dev/full");
|
|
|
|
|
if (worker.store.systemFeatures.get().count("kvm") && pathExists("/dev/kvm"))
|
|
|
|
|
ss.push_back("/dev/kvm");
|
|
|
|
|
ss.push_back("/dev/null");
|
|
|
|
|
ss.push_back("/dev/random");
|
|
|
|
|
ss.push_back("/dev/tty");
|
|
|
|
|
ss.push_back("/dev/urandom");
|
|
|
|
|
ss.push_back("/dev/zero");
|
|
|
|
|
createSymlink("/proc/self/fd", chrootRootDir + "/dev/fd");
|
|
|
|
|
createSymlink("/proc/self/fd/0", chrootRootDir + "/dev/stdin");
|
|
|
|
|
createSymlink("/proc/self/fd/1", chrootRootDir + "/dev/stdout");
|
|
|
|
|
createSymlink("/proc/self/fd/2", chrootRootDir + "/dev/stderr");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Fixed-output derivations typically need to access the
|
|
|
|
|
network, so give them access to /etc/resolv.conf and so
|
|
|
|
|
on. */
|
2022-03-31 17:06:40 +03:00
|
|
|
|
if (!derivationType.isSandboxed()) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
// Only use nss functions to resolve hosts and
|
|
|
|
|
// services. Don’t use it for anything else that may
|
|
|
|
|
// be configured for this system. This limits the
|
|
|
|
|
// potential impurities introduced in fixed-outputs.
|
|
|
|
|
writeFile(chrootRootDir + "/etc/nsswitch.conf", "hosts: files dns\nservices: files\n");
|
|
|
|
|
|
2021-02-27 07:23:14 +02:00
|
|
|
|
/* N.B. it is realistic that these paths might not exist. It
|
|
|
|
|
happens when testing Nix building fixed-output derivations
|
|
|
|
|
within a pure derivation. */
|
2021-07-13 13:04:14 +03:00
|
|
|
|
for (auto & path : { "/etc/resolv.conf", "/etc/services", "/etc/hosts" })
|
2021-02-27 07:23:14 +02:00
|
|
|
|
if (pathExists(path))
|
|
|
|
|
ss.push_back(path);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
for (auto & i : ss) dirsInChroot.emplace(i, i);
|
|
|
|
|
|
|
|
|
|
/* Bind-mount all the directories from the "host"
|
|
|
|
|
filesystem that we want in the chroot
|
|
|
|
|
environment. */
|
|
|
|
|
auto doBind = [&](const Path & source, const Path & target, bool optional = false) {
|
|
|
|
|
debug("bind mounting '%1%' to '%2%'", source, target);
|
|
|
|
|
struct stat st;
|
|
|
|
|
if (stat(source.c_str(), &st) == -1) {
|
|
|
|
|
if (optional && errno == ENOENT)
|
|
|
|
|
return;
|
|
|
|
|
else
|
|
|
|
|
throw SysError("getting attributes of path '%1%'", source);
|
|
|
|
|
}
|
|
|
|
|
if (S_ISDIR(st.st_mode))
|
|
|
|
|
createDirs(target);
|
|
|
|
|
else {
|
|
|
|
|
createDirs(dirOf(target));
|
|
|
|
|
writeFile(target, "");
|
|
|
|
|
}
|
|
|
|
|
if (mount(source.c_str(), target.c_str(), "", MS_BIND | MS_REC, 0) == -1)
|
|
|
|
|
throw SysError("bind mount from '%1%' to '%2%' failed", source, target);
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
for (auto & i : dirsInChroot) {
|
|
|
|
|
if (i.second.source == "/proc") continue; // backwards compatibility
|
2022-06-22 19:21:37 +03:00
|
|
|
|
|
|
|
|
|
#if HAVE_EMBEDDED_SANDBOX_SHELL
|
|
|
|
|
if (i.second.source == "__embedded_sandbox_shell__") {
|
|
|
|
|
static unsigned char sh[] = {
|
|
|
|
|
#include "embedded-sandbox-shell.gen.hh"
|
|
|
|
|
};
|
|
|
|
|
auto dst = chrootRootDir + i.first;
|
|
|
|
|
createDirs(dirOf(dst));
|
|
|
|
|
writeFile(dst, std::string_view((const char *) sh, sizeof(sh)));
|
|
|
|
|
chmod_(dst, 0555);
|
|
|
|
|
} else
|
|
|
|
|
#endif
|
|
|
|
|
doBind(i.second.source, chrootRootDir + i.first, i.second.optional);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Bind a new instance of procfs on /proc. */
|
|
|
|
|
createDirs(chrootRootDir + "/proc");
|
|
|
|
|
if (mount("none", (chrootRootDir + "/proc").c_str(), "proc", 0, 0) == -1)
|
|
|
|
|
throw SysError("mounting /proc");
|
|
|
|
|
|
2021-04-27 22:06:58 +03:00
|
|
|
|
/* Mount sysfs on /sys. */
|
2022-11-08 17:03:42 +02:00
|
|
|
|
if (buildUser && buildUser->getUIDCount() != 1) {
|
2021-04-27 22:06:58 +03:00
|
|
|
|
createDirs(chrootRootDir + "/sys");
|
|
|
|
|
if (mount("none", (chrootRootDir + "/sys").c_str(), "sysfs", 0, 0) == -1)
|
|
|
|
|
throw SysError("mounting /sys");
|
|
|
|
|
}
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Mount a new tmpfs on /dev/shm to ensure that whatever
|
|
|
|
|
the builder puts in /dev/shm is cleaned up automatically. */
|
|
|
|
|
if (pathExists("/dev/shm") && mount("none", (chrootRootDir + "/dev/shm").c_str(), "tmpfs", 0,
|
|
|
|
|
fmt("size=%s", settings.sandboxShmSize).c_str()) == -1)
|
|
|
|
|
throw SysError("mounting /dev/shm");
|
|
|
|
|
|
|
|
|
|
/* Mount a new devpts on /dev/pts. Note that this
|
|
|
|
|
requires the kernel to be compiled with
|
|
|
|
|
CONFIG_DEVPTS_MULTIPLE_INSTANCES=y (which is the case
|
|
|
|
|
if /dev/ptx/ptmx exists). */
|
|
|
|
|
if (pathExists("/dev/pts/ptmx") &&
|
|
|
|
|
!pathExists(chrootRootDir + "/dev/ptmx")
|
|
|
|
|
&& !dirsInChroot.count("/dev/pts"))
|
|
|
|
|
{
|
|
|
|
|
if (mount("none", (chrootRootDir + "/dev/pts").c_str(), "devpts", 0, "newinstance,mode=0620") == 0)
|
|
|
|
|
{
|
|
|
|
|
createSymlink("/dev/pts/ptmx", chrootRootDir + "/dev/ptmx");
|
|
|
|
|
|
|
|
|
|
/* Make sure /dev/pts/ptmx is world-writable. With some
|
|
|
|
|
Linux versions, it is created with permissions 0. */
|
|
|
|
|
chmod_(chrootRootDir + "/dev/pts/ptmx", 0666);
|
|
|
|
|
} else {
|
|
|
|
|
if (errno != EINVAL)
|
|
|
|
|
throw SysError("mounting /dev/pts");
|
|
|
|
|
doBind("/dev/pts", chrootRootDir + "/dev/pts");
|
|
|
|
|
doBind("/dev/ptmx", chrootRootDir + "/dev/ptmx");
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2023-02-17 17:31:55 +02:00
|
|
|
|
/* Make /etc unwritable */
|
|
|
|
|
if (!parsedDrv->useUidRange())
|
|
|
|
|
chmod_(chrootRootDir + "/etc", 0555);
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Unshare this mount namespace. This is necessary because
|
|
|
|
|
pivot_root() below changes the root of the mount
|
|
|
|
|
namespace. This means that the call to setns() in
|
|
|
|
|
addDependency() would hide the host's filesystem,
|
|
|
|
|
making it impossible to bind-mount paths from the host
|
|
|
|
|
Nix store into the sandbox. Therefore, we save the
|
|
|
|
|
pre-pivot_root namespace in
|
|
|
|
|
sandboxMountNamespace. Since we made /nix/store a
|
|
|
|
|
shared subtree above, this allows addDependency() to
|
|
|
|
|
make paths appear in the sandbox. */
|
|
|
|
|
if (unshare(CLONE_NEWNS) == -1)
|
|
|
|
|
throw SysError("unsharing mount namespace");
|
|
|
|
|
|
2021-04-27 22:06:58 +03:00
|
|
|
|
/* Unshare the cgroup namespace. This means
|
|
|
|
|
/proc/self/cgroup will show the child's cgroup as '/'
|
|
|
|
|
rather than whatever it is in the parent. */
|
2022-11-18 11:39:28 +02:00
|
|
|
|
if (cgroup && unshare(CLONE_NEWCGROUP) == -1)
|
2021-04-27 22:06:58 +03:00
|
|
|
|
throw SysError("unsharing cgroup namespace");
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Do the chroot(). */
|
|
|
|
|
if (chdir(chrootRootDir.c_str()) == -1)
|
|
|
|
|
throw SysError("cannot change directory to '%1%'", chrootRootDir);
|
|
|
|
|
|
|
|
|
|
if (mkdir("real-root", 0) == -1)
|
|
|
|
|
throw SysError("cannot create real-root directory");
|
|
|
|
|
|
|
|
|
|
if (pivot_root(".", "real-root") == -1)
|
|
|
|
|
throw SysError("cannot pivot old root directory onto '%1%'", (chrootRootDir + "/real-root"));
|
|
|
|
|
|
|
|
|
|
if (chroot(".") == -1)
|
|
|
|
|
throw SysError("cannot change root directory to '%1%'", chrootRootDir);
|
|
|
|
|
|
|
|
|
|
if (umount2("real-root", MNT_DETACH) == -1)
|
|
|
|
|
throw SysError("cannot unmount real root filesystem");
|
|
|
|
|
|
|
|
|
|
if (rmdir("real-root") == -1)
|
|
|
|
|
throw SysError("cannot remove real-root directory");
|
|
|
|
|
|
|
|
|
|
/* Switch to the sandbox uid/gid in the user namespace,
|
|
|
|
|
which corresponds to the build user or calling user in
|
|
|
|
|
the parent namespace. */
|
|
|
|
|
if (setgid(sandboxGid()) == -1)
|
|
|
|
|
throw SysError("setgid failed");
|
|
|
|
|
if (setuid(sandboxUid()) == -1)
|
|
|
|
|
throw SysError("setuid failed");
|
|
|
|
|
|
|
|
|
|
setUser = false;
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
if (chdir(tmpDirInSandbox.c_str()) == -1)
|
|
|
|
|
throw SysError("changing into '%1%'", tmpDir);
|
|
|
|
|
|
|
|
|
|
/* Close all other file descriptors. */
|
|
|
|
|
closeMostFDs({STDIN_FILENO, STDOUT_FILENO, STDERR_FILENO});
|
|
|
|
|
|
2022-12-23 17:28:26 +02:00
|
|
|
|
setPersonality(drv->platform);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Disable core dumps by default. */
|
|
|
|
|
struct rlimit limit = { 0, RLIM_INFINITY };
|
|
|
|
|
setrlimit(RLIMIT_CORE, &limit);
|
|
|
|
|
|
|
|
|
|
// FIXME: set other limits to deterministic values?
|
|
|
|
|
|
|
|
|
|
/* Fill in the environment. */
|
|
|
|
|
Strings envStrs;
|
|
|
|
|
for (auto & i : env)
|
|
|
|
|
envStrs.push_back(rewriteStrings(i.first + "=" + i.second, inputRewrites));
|
|
|
|
|
|
|
|
|
|
/* If we are running in `build-users' mode, then switch to the
|
|
|
|
|
user we allocated above. Make sure that we drop all root
|
|
|
|
|
privileges. Note that above we have closed all file
|
|
|
|
|
descriptors except std*, so that's safe. Also note that
|
|
|
|
|
setuid() when run as root sets the real, effective and
|
|
|
|
|
saved UIDs. */
|
|
|
|
|
if (setUser && buildUser) {
|
|
|
|
|
/* Preserve supplementary groups of the build user, to allow
|
|
|
|
|
admins to specify groups such as "kvm". */
|
2022-11-22 10:02:17 +02:00
|
|
|
|
auto gids = buildUser->getSupplementaryGIDs();
|
|
|
|
|
if (setgroups(gids.size(), gids.data()) == -1)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw SysError("cannot set supplementary groups of build user");
|
|
|
|
|
|
|
|
|
|
if (setgid(buildUser->getGID()) == -1 ||
|
|
|
|
|
getgid() != buildUser->getGID() ||
|
|
|
|
|
getegid() != buildUser->getGID())
|
|
|
|
|
throw SysError("setgid failed");
|
|
|
|
|
|
|
|
|
|
if (setuid(buildUser->getUID()) == -1 ||
|
|
|
|
|
getuid() != buildUser->getUID() ||
|
|
|
|
|
geteuid() != buildUser->getUID())
|
|
|
|
|
throw SysError("setuid failed");
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Fill in the arguments. */
|
|
|
|
|
Strings args;
|
|
|
|
|
|
2021-09-23 13:07:50 +03:00
|
|
|
|
std::string builder = "invalid";
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (drv->isBuiltin()) {
|
|
|
|
|
;
|
|
|
|
|
}
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
else {
|
|
|
|
|
/* This has to appear before import statements. */
|
|
|
|
|
std::string sandboxProfile = "(version 1)\n";
|
|
|
|
|
|
|
|
|
|
if (useChroot) {
|
|
|
|
|
|
|
|
|
|
/* Lots and lots and lots of file functions freak out if they can't stat their full ancestry */
|
|
|
|
|
PathSet ancestry;
|
|
|
|
|
|
|
|
|
|
/* We build the ancestry before adding all inputPaths to the store because we know they'll
|
|
|
|
|
all have the same parents (the store), and there might be lots of inputs. This isn't
|
|
|
|
|
particularly efficient... I doubt it'll be a bottleneck in practice */
|
|
|
|
|
for (auto & i : dirsInChroot) {
|
|
|
|
|
Path cur = i.first;
|
|
|
|
|
while (cur.compare("/") != 0) {
|
|
|
|
|
cur = dirOf(cur);
|
|
|
|
|
ancestry.insert(cur);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* And we want the store in there regardless of how empty dirsInChroot. We include the innermost
|
|
|
|
|
path component this time, since it's typically /nix/store and we care about that. */
|
|
|
|
|
Path cur = worker.store.storeDir;
|
|
|
|
|
while (cur.compare("/") != 0) {
|
|
|
|
|
ancestry.insert(cur);
|
|
|
|
|
cur = dirOf(cur);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Add all our input paths to the chroot */
|
|
|
|
|
for (auto & i : inputPaths) {
|
|
|
|
|
auto p = worker.store.printStorePath(i);
|
|
|
|
|
dirsInChroot[p] = p;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Violations will go to the syslog if you set this. Unfortunately the destination does not appear to be configurable */
|
|
|
|
|
if (settings.darwinLogSandboxViolations) {
|
|
|
|
|
sandboxProfile += "(deny default)\n";
|
|
|
|
|
} else {
|
|
|
|
|
sandboxProfile += "(deny default (with no-log))\n";
|
|
|
|
|
}
|
|
|
|
|
|
2023-01-04 14:36:07 +02:00
|
|
|
|
sandboxProfile +=
|
|
|
|
|
#include "sandbox-defaults.sb"
|
|
|
|
|
;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-03-31 17:06:40 +03:00
|
|
|
|
if (!derivationType.isSandboxed())
|
2023-01-04 14:36:07 +02:00
|
|
|
|
sandboxProfile +=
|
|
|
|
|
#include "sandbox-network.sb"
|
|
|
|
|
;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Add the output paths we'll use at build-time to the chroot */
|
|
|
|
|
sandboxProfile += "(allow file-read* file-write* process-exec\n";
|
|
|
|
|
for (auto & [_, path] : scratchOutputs)
|
|
|
|
|
sandboxProfile += fmt("\t(subpath \"%s\")\n", worker.store.printStorePath(path));
|
|
|
|
|
|
|
|
|
|
sandboxProfile += ")\n";
|
|
|
|
|
|
|
|
|
|
/* Our inputs (transitive dependencies and any impurities computed above)
|
|
|
|
|
|
|
|
|
|
without file-write* allowed, access() incorrectly returns EPERM
|
|
|
|
|
*/
|
|
|
|
|
sandboxProfile += "(allow file-read* file-write* process-exec\n";
|
|
|
|
|
for (auto & i : dirsInChroot) {
|
|
|
|
|
if (i.first != i.second.source)
|
|
|
|
|
throw Error(
|
|
|
|
|
"can't map '%1%' to '%2%': mismatched impure paths not supported on Darwin",
|
|
|
|
|
i.first, i.second.source);
|
|
|
|
|
|
2022-02-28 16:21:03 +02:00
|
|
|
|
std::string path = i.first;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
struct stat st;
|
|
|
|
|
if (lstat(path.c_str(), &st)) {
|
|
|
|
|
if (i.second.optional && errno == ENOENT)
|
|
|
|
|
continue;
|
|
|
|
|
throw SysError("getting attributes of path '%s", path);
|
|
|
|
|
}
|
|
|
|
|
if (S_ISDIR(st.st_mode))
|
|
|
|
|
sandboxProfile += fmt("\t(subpath \"%s\")\n", path);
|
|
|
|
|
else
|
|
|
|
|
sandboxProfile += fmt("\t(literal \"%s\")\n", path);
|
|
|
|
|
}
|
|
|
|
|
sandboxProfile += ")\n";
|
|
|
|
|
|
|
|
|
|
/* Allow file-read* on full directory hierarchy to self. Allows realpath() */
|
|
|
|
|
sandboxProfile += "(allow file-read*\n";
|
|
|
|
|
for (auto & i : ancestry) {
|
|
|
|
|
sandboxProfile += fmt("\t(literal \"%s\")\n", i);
|
|
|
|
|
}
|
|
|
|
|
sandboxProfile += ")\n";
|
|
|
|
|
|
|
|
|
|
sandboxProfile += additionalSandboxProfile;
|
|
|
|
|
} else
|
2023-01-04 14:36:07 +02:00
|
|
|
|
sandboxProfile +=
|
|
|
|
|
#include "sandbox-minimal.sb"
|
|
|
|
|
;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
debug("Generated sandbox profile:");
|
|
|
|
|
debug(sandboxProfile);
|
|
|
|
|
|
|
|
|
|
Path sandboxFile = tmpDir + "/.sandbox.sb";
|
|
|
|
|
|
|
|
|
|
writeFile(sandboxFile, sandboxProfile);
|
|
|
|
|
|
|
|
|
|
bool allowLocalNetworking = parsedDrv->getBoolAttr("__darwinAllowLocalNetworking");
|
|
|
|
|
|
|
|
|
|
/* The tmpDir in scope points at the temporary build directory for our derivation. Some packages try different mechanisms
|
|
|
|
|
to find temporary directories, so we want to open up a broader place for them to dump their files, if needed. */
|
2023-03-01 21:01:36 +02:00
|
|
|
|
Path globalTmpDir = canonPath(getEnvNonEmpty("TMPDIR").value_or("/tmp"), true);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* They don't like trailing slashes on subpath directives */
|
|
|
|
|
if (globalTmpDir.back() == '/') globalTmpDir.pop_back();
|
|
|
|
|
|
|
|
|
|
if (getEnv("_NIX_TEST_NO_SANDBOX") != "1") {
|
|
|
|
|
builder = "/usr/bin/sandbox-exec";
|
|
|
|
|
args.push_back("sandbox-exec");
|
|
|
|
|
args.push_back("-f");
|
|
|
|
|
args.push_back(sandboxFile);
|
|
|
|
|
args.push_back("-D");
|
|
|
|
|
args.push_back("_GLOBAL_TMP_DIR=" + globalTmpDir);
|
|
|
|
|
if (allowLocalNetworking) {
|
|
|
|
|
args.push_back("-D");
|
2022-02-28 16:21:03 +02:00
|
|
|
|
args.push_back(std::string("_ALLOW_LOCAL_NETWORKING=1"));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
args.push_back(drv->builder);
|
|
|
|
|
} else {
|
2021-09-23 13:07:50 +03:00
|
|
|
|
builder = drv->builder;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
args.push_back(std::string(baseNameOf(drv->builder)));
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
#else
|
|
|
|
|
else {
|
2021-09-23 13:07:50 +03:00
|
|
|
|
builder = drv->builder;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
args.push_back(std::string(baseNameOf(drv->builder)));
|
|
|
|
|
}
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
for (auto & i : drv->args)
|
|
|
|
|
args.push_back(rewriteStrings(i, inputRewrites));
|
|
|
|
|
|
|
|
|
|
/* Indicate that we managed to set up the build environment. */
|
2022-02-25 17:00:00 +02:00
|
|
|
|
writeFull(STDERR_FILENO, std::string("\2\n"));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-10-13 22:35:16 +03:00
|
|
|
|
sendException = false;
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Execute the program. This should not return. */
|
|
|
|
|
if (drv->isBuiltin()) {
|
|
|
|
|
try {
|
|
|
|
|
logger = makeJSONLogger(*logger);
|
|
|
|
|
|
|
|
|
|
BasicDerivation & drv2(*drv);
|
|
|
|
|
for (auto & e : drv2.env)
|
|
|
|
|
e.second = rewriteStrings(e.second, inputRewrites);
|
|
|
|
|
|
|
|
|
|
if (drv->builder == "builtin:fetchurl")
|
|
|
|
|
builtinFetchurl(drv2, netrcData);
|
|
|
|
|
else if (drv->builder == "builtin:buildenv")
|
|
|
|
|
builtinBuildenv(drv2);
|
|
|
|
|
else if (drv->builder == "builtin:unpack-channel")
|
|
|
|
|
builtinUnpackChannel(drv2);
|
|
|
|
|
else
|
2022-02-25 17:00:00 +02:00
|
|
|
|
throw Error("unsupported builtin builder '%1%'", drv->builder.substr(8));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
_exit(0);
|
|
|
|
|
} catch (std::exception & e) {
|
|
|
|
|
writeFull(STDERR_FILENO, e.what() + std::string("\n"));
|
|
|
|
|
_exit(1);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#if __APPLE__
|
|
|
|
|
posix_spawnattr_t attrp;
|
|
|
|
|
|
|
|
|
|
if (posix_spawnattr_init(&attrp))
|
|
|
|
|
throw SysError("failed to initialize builder");
|
|
|
|
|
|
|
|
|
|
if (posix_spawnattr_setflags(&attrp, POSIX_SPAWN_SETEXEC))
|
|
|
|
|
throw SysError("failed to initialize builder");
|
|
|
|
|
|
|
|
|
|
if (drv->platform == "aarch64-darwin") {
|
|
|
|
|
// Unset kern.curproc_arch_affinity so we can escape Rosetta
|
|
|
|
|
int affinity = 0;
|
|
|
|
|
sysctlbyname("kern.curproc_arch_affinity", NULL, NULL, &affinity, sizeof(affinity));
|
|
|
|
|
|
|
|
|
|
cpu_type_t cpu = CPU_TYPE_ARM64;
|
|
|
|
|
posix_spawnattr_setbinpref_np(&attrp, 1, &cpu, NULL);
|
|
|
|
|
} else if (drv->platform == "x86_64-darwin") {
|
|
|
|
|
cpu_type_t cpu = CPU_TYPE_X86_64;
|
|
|
|
|
posix_spawnattr_setbinpref_np(&attrp, 1, &cpu, NULL);
|
|
|
|
|
}
|
|
|
|
|
|
2021-09-23 13:07:50 +03:00
|
|
|
|
posix_spawn(NULL, builder.c_str(), NULL, &attrp, stringsToCharPtrs(args).data(), stringsToCharPtrs(envStrs).data());
|
2021-02-23 08:26:35 +02:00
|
|
|
|
#else
|
2021-09-23 13:07:50 +03:00
|
|
|
|
execve(builder.c_str(), stringsToCharPtrs(args).data(), stringsToCharPtrs(envStrs).data());
|
2021-02-23 08:26:35 +02:00
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
throw SysError("executing '%1%'", drv->builder);
|
|
|
|
|
|
|
|
|
|
} catch (Error & e) {
|
2022-10-13 22:35:16 +03:00
|
|
|
|
if (sendException) {
|
|
|
|
|
writeFull(STDERR_FILENO, "\1\n");
|
|
|
|
|
FdSink sink(STDERR_FILENO);
|
|
|
|
|
sink << e;
|
|
|
|
|
sink.flush();
|
|
|
|
|
} else
|
|
|
|
|
std::cerr << e.msg();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
_exit(1);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2023-04-15 01:18:32 +03:00
|
|
|
|
SingleDrvOutputs LocalDerivationGoal::registerOutputs()
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
/* When using a build hook, the build hook can register the output
|
|
|
|
|
as valid (by doing `nix-store --import'). If so we don't have
|
|
|
|
|
to do anything here.
|
|
|
|
|
|
|
|
|
|
We can only early return when the outputs are known a priori. For
|
|
|
|
|
floating content-addressed derivations this isn't the case.
|
|
|
|
|
*/
|
2022-03-08 20:50:46 +02:00
|
|
|
|
if (hook)
|
|
|
|
|
return DerivationGoal::registerOutputs();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
std::map<std::string, ValidPathInfo> infos;
|
|
|
|
|
|
|
|
|
|
/* Set of inodes seen during calls to canonicalisePathMetaData()
|
|
|
|
|
for this build's outputs. This needs to be shared between
|
|
|
|
|
outputs to allow hard links between outputs. */
|
|
|
|
|
InodesSeen inodesSeen;
|
|
|
|
|
|
|
|
|
|
Path checkSuffix = ".check";
|
|
|
|
|
|
|
|
|
|
std::exception_ptr delayedException;
|
|
|
|
|
|
|
|
|
|
/* The paths that can be referenced are the input closures, the
|
|
|
|
|
output paths, and any paths that have been built via recursive
|
|
|
|
|
Nix calls. */
|
|
|
|
|
StorePathSet referenceablePaths;
|
|
|
|
|
for (auto & p : inputPaths) referenceablePaths.insert(p);
|
|
|
|
|
for (auto & i : scratchOutputs) referenceablePaths.insert(i.second);
|
|
|
|
|
for (auto & p : addedPaths) referenceablePaths.insert(p);
|
|
|
|
|
|
|
|
|
|
/* FIXME `needsHashRewrite` should probably be removed and we get to the
|
|
|
|
|
real reason why we aren't using the chroot dir */
|
|
|
|
|
auto toRealPathChroot = [&](const Path & p) -> Path {
|
|
|
|
|
return useChroot && !needsHashRewrite()
|
|
|
|
|
? chrootRootDir + p
|
|
|
|
|
: worker.store.toRealPath(p);
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
/* Check whether the output paths were created, and make all
|
|
|
|
|
output paths read-only. Then get the references of each output (that we
|
|
|
|
|
might need to register), so we can topologically sort them. For the ones
|
|
|
|
|
that are most definitely already installed, we just store their final
|
|
|
|
|
name so we can also use it in rewrites. */
|
|
|
|
|
StringSet outputsToSort;
|
|
|
|
|
struct AlreadyRegistered { StorePath path; };
|
|
|
|
|
struct PerhapsNeedToRegister { StorePathSet refs; };
|
|
|
|
|
std::map<std::string, std::variant<AlreadyRegistered, PerhapsNeedToRegister>> outputReferencesIfUnregistered;
|
|
|
|
|
std::map<std::string, struct stat> outputStats;
|
|
|
|
|
for (auto & [outputName, _] : drv->outputs) {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto scratchOutput = get(scratchOutputs, outputName);
|
|
|
|
|
if (!scratchOutput)
|
|
|
|
|
throw BuildError(
|
|
|
|
|
"builder for '%s' has no scratch output for '%s'",
|
|
|
|
|
worker.store.printStorePath(drvPath), outputName);
|
|
|
|
|
auto actualPath = toRealPathChroot(worker.store.printStorePath(*scratchOutput));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
outputsToSort.insert(outputName);
|
|
|
|
|
|
|
|
|
|
/* Updated wanted info to remove the outputs we definitely don't need to register */
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto initialOutput = get(initialOutputs, outputName);
|
|
|
|
|
if (!initialOutput)
|
|
|
|
|
throw BuildError(
|
|
|
|
|
"builder for '%s' has no initial output for '%s'",
|
|
|
|
|
worker.store.printStorePath(drvPath), outputName);
|
|
|
|
|
auto & initialInfo = *initialOutput;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Don't register if already valid, and not checking */
|
|
|
|
|
initialInfo.wanted = buildMode == bmCheck
|
|
|
|
|
|| !(initialInfo.known && initialInfo.known->isValid());
|
|
|
|
|
if (!initialInfo.wanted) {
|
|
|
|
|
outputReferencesIfUnregistered.insert_or_assign(
|
|
|
|
|
outputName,
|
|
|
|
|
AlreadyRegistered { .path = initialInfo.known->path });
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
struct stat st;
|
|
|
|
|
if (lstat(actualPath.c_str(), &st) == -1) {
|
|
|
|
|
if (errno == ENOENT)
|
|
|
|
|
throw BuildError(
|
|
|
|
|
"builder for '%s' failed to produce output path for output '%s' at '%s'",
|
|
|
|
|
worker.store.printStorePath(drvPath), outputName, actualPath);
|
|
|
|
|
throw SysError("getting attributes of path '%s'", actualPath);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#ifndef __CYGWIN__
|
|
|
|
|
/* Check that the output is not group or world writable, as
|
|
|
|
|
that means that someone else can have interfered with the
|
|
|
|
|
build. Also, the output should be owned by the build
|
|
|
|
|
user. */
|
|
|
|
|
if ((!S_ISLNK(st.st_mode) && (st.st_mode & (S_IWGRP | S_IWOTH))) ||
|
|
|
|
|
(buildUser && st.st_uid != buildUser->getUID()))
|
|
|
|
|
throw BuildError(
|
|
|
|
|
"suspicious ownership or permission on '%s' for output '%s'; rejecting this build output",
|
|
|
|
|
actualPath, outputName);
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
|
/* Canonicalise first. This ensures that the path we're
|
|
|
|
|
rewriting doesn't contain a hard link to /etc/shadow or
|
|
|
|
|
something like that. */
|
2021-04-27 22:06:58 +03:00
|
|
|
|
canonicalisePathMetaData(
|
|
|
|
|
actualPath,
|
|
|
|
|
buildUser ? std::optional(buildUser->getUIDRange()) : std::nullopt,
|
|
|
|
|
inodesSeen);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-09-24 00:30:29 +03:00
|
|
|
|
bool discardReferences = false;
|
|
|
|
|
if (auto structuredAttrs = parsedDrv->getStructuredAttrs()) {
|
2023-01-03 19:53:01 +02:00
|
|
|
|
if (auto udr = get(*structuredAttrs, "unsafeDiscardReferences")) {
|
2023-03-17 16:33:48 +02:00
|
|
|
|
experimentalFeatureSettings.require(Xp::DiscardReferences);
|
2023-01-03 19:53:01 +02:00
|
|
|
|
if (auto output = get(*udr, outputName)) {
|
|
|
|
|
if (!output->is_boolean())
|
|
|
|
|
throw Error("attribute 'unsafeDiscardReferences.\"%s\"' of derivation '%s' must be a Boolean", outputName, drvPath.to_string());
|
|
|
|
|
discardReferences = output->get<bool>();
|
2022-09-24 00:30:29 +03:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-09-24 00:30:29 +03:00
|
|
|
|
StorePathSet references;
|
|
|
|
|
if (discardReferences)
|
|
|
|
|
debug("discarding references of output '%s'", outputName);
|
|
|
|
|
else {
|
|
|
|
|
debug("scanning for references for output '%s' in temp location '%s'", outputName, actualPath);
|
|
|
|
|
|
|
|
|
|
/* Pass blank Sink as we are not ready to hash data at this stage. */
|
|
|
|
|
NullSink blank;
|
|
|
|
|
references = scanForReferences(blank, actualPath, referenceablePaths);
|
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
outputReferencesIfUnregistered.insert_or_assign(
|
|
|
|
|
outputName,
|
|
|
|
|
PerhapsNeedToRegister { .refs = references });
|
|
|
|
|
outputStats.insert_or_assign(outputName, std::move(st));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
auto sortedOutputNames = topoSort(outputsToSort,
|
|
|
|
|
{[&](const std::string & name) {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto orifu = get(outputReferencesIfUnregistered, name);
|
|
|
|
|
if (!orifu)
|
|
|
|
|
throw BuildError(
|
|
|
|
|
"no output reference for '%s' in build of '%s'",
|
|
|
|
|
name, worker.store.printStorePath(drvPath));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
return std::visit(overloaded {
|
|
|
|
|
/* Since we'll use the already installed versions of these, we
|
|
|
|
|
can treat them as leaves and ignore any references they
|
|
|
|
|
have. */
|
2021-10-01 00:31:21 +03:00
|
|
|
|
[&](const AlreadyRegistered &) { return StringSet {}; },
|
|
|
|
|
[&](const PerhapsNeedToRegister & refs) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
StringSet referencedOutputs;
|
|
|
|
|
/* FIXME build inverted map up front so no quadratic waste here */
|
|
|
|
|
for (auto & r : refs.refs)
|
|
|
|
|
for (auto & [o, p] : scratchOutputs)
|
|
|
|
|
if (r == p)
|
|
|
|
|
referencedOutputs.insert(o);
|
|
|
|
|
return referencedOutputs;
|
|
|
|
|
},
|
2022-05-04 08:44:32 +03:00
|
|
|
|
}, *orifu);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}},
|
|
|
|
|
{[&](const std::string & path, const std::string & parent) {
|
|
|
|
|
// TODO with more -vvvv also show the temporary paths for manual inspection.
|
|
|
|
|
return BuildError(
|
|
|
|
|
"cycle detected in build of '%s' in the references of output '%s' from output '%s'",
|
|
|
|
|
worker.store.printStorePath(drvPath), path, parent);
|
|
|
|
|
}});
|
|
|
|
|
|
|
|
|
|
std::reverse(sortedOutputNames.begin(), sortedOutputNames.end());
|
|
|
|
|
|
2022-03-08 20:50:46 +02:00
|
|
|
|
OutputPathMap finalOutputs;
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
for (auto & outputName : sortedOutputNames) {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto output = get(drv->outputs, outputName);
|
|
|
|
|
auto scratchPath = get(scratchOutputs, outputName);
|
|
|
|
|
assert(output && scratchPath);
|
|
|
|
|
auto actualPath = toRealPathChroot(worker.store.printStorePath(*scratchPath));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
auto finish = [&](StorePath finalStorePath) {
|
|
|
|
|
/* Store the final path */
|
|
|
|
|
finalOutputs.insert_or_assign(outputName, finalStorePath);
|
|
|
|
|
/* The rewrite rule will be used in downstream outputs that refer to
|
|
|
|
|
use. This is why the topological sort is essential to do first
|
|
|
|
|
before this for loop. */
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (*scratchPath != finalStorePath)
|
|
|
|
|
outputRewrites[std::string { scratchPath->hashPart() }] = std::string { finalStorePath.hashPart() };
|
2021-02-23 08:26:35 +02:00
|
|
|
|
};
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto orifu = get(outputReferencesIfUnregistered, outputName);
|
|
|
|
|
assert(orifu);
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
std::optional<StorePathSet> referencesOpt = std::visit(overloaded {
|
2021-10-01 00:31:21 +03:00
|
|
|
|
[&](const AlreadyRegistered & skippedFinalPath) -> std::optional<StorePathSet> {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
finish(skippedFinalPath.path);
|
|
|
|
|
return std::nullopt;
|
|
|
|
|
},
|
2021-10-01 00:31:21 +03:00
|
|
|
|
[&](const PerhapsNeedToRegister & r) -> std::optional<StorePathSet> {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
return r.refs;
|
|
|
|
|
},
|
2022-05-04 08:44:32 +03:00
|
|
|
|
}, *orifu);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (!referencesOpt)
|
|
|
|
|
continue;
|
|
|
|
|
auto references = *referencesOpt;
|
|
|
|
|
|
|
|
|
|
auto rewriteOutput = [&]() {
|
|
|
|
|
/* Apply hash rewriting if necessary. */
|
|
|
|
|
if (!outputRewrites.empty()) {
|
2021-10-01 00:08:14 +03:00
|
|
|
|
debug("rewriting hashes in '%1%'; cross fingers", actualPath);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* FIXME: this is in-memory. */
|
|
|
|
|
StringSink sink;
|
|
|
|
|
dumpPath(actualPath, sink);
|
|
|
|
|
deletePath(actualPath);
|
2022-01-17 23:20:05 +02:00
|
|
|
|
sink.s = rewriteStrings(sink.s, outputRewrites);
|
|
|
|
|
StringSource source(sink.s);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
restorePath(actualPath, source);
|
|
|
|
|
|
|
|
|
|
/* FIXME: set proper permissions in restorePath() so
|
|
|
|
|
we don't have to do another traversal. */
|
2021-04-27 22:06:58 +03:00
|
|
|
|
canonicalisePathMetaData(actualPath, {}, inodesSeen);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
2023-01-06 22:36:05 +02:00
|
|
|
|
auto rewriteRefs = [&]() -> StoreReferences {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* In the CA case, we need the rewritten refs to calculate the
|
|
|
|
|
final path, therefore we look for a *non-rewritten
|
|
|
|
|
self-reference, and use a bool rather try to solve the
|
|
|
|
|
computationally intractable fixed point. */
|
2023-01-06 22:36:05 +02:00
|
|
|
|
StoreReferences res {
|
|
|
|
|
.self = false,
|
2021-02-23 08:26:35 +02:00
|
|
|
|
};
|
|
|
|
|
for (auto & r : references) {
|
|
|
|
|
auto name = r.name();
|
|
|
|
|
auto origHash = std::string { r.hashPart() };
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (r == *scratchPath) {
|
2023-01-06 22:36:05 +02:00
|
|
|
|
res.self = true;
|
2022-05-04 08:44:32 +03:00
|
|
|
|
} else if (auto outputRewrite = get(outputRewrites, origHash)) {
|
|
|
|
|
std::string newRef = *outputRewrite;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
newRef += '-';
|
|
|
|
|
newRef += name;
|
2023-01-06 22:36:05 +02:00
|
|
|
|
res.others.insert(StorePath { newRef });
|
2022-05-04 08:44:32 +03:00
|
|
|
|
} else {
|
2023-01-06 22:36:05 +02:00
|
|
|
|
res.others.insert(r);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
return res;
|
|
|
|
|
};
|
|
|
|
|
|
2022-03-18 00:29:15 +02:00
|
|
|
|
auto newInfoFromCA = [&](const DerivationOutput::CAFloating outputHash) -> ValidPathInfo {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto st = get(outputStats, outputName);
|
|
|
|
|
if (!st)
|
|
|
|
|
throw BuildError(
|
|
|
|
|
"output path %1% without valid stats info",
|
|
|
|
|
actualPath);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (outputHash.method == FileIngestionMethod::Flat) {
|
|
|
|
|
/* The output path should be a regular file without execute permission. */
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (!S_ISREG(st->st_mode) || (st->st_mode & S_IXUSR) != 0)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
throw BuildError(
|
|
|
|
|
"output path '%1%' should be a non-executable regular file "
|
|
|
|
|
"since recursive hashing is not enabled (outputHashMode=flat)",
|
|
|
|
|
actualPath);
|
|
|
|
|
}
|
|
|
|
|
rewriteOutput();
|
|
|
|
|
/* FIXME optimize and deduplicate with addToStore */
|
2022-05-04 08:44:32 +03:00
|
|
|
|
std::string oldHashPart { scratchPath->hashPart() };
|
2021-02-23 08:26:35 +02:00
|
|
|
|
HashModuloSink caSink { outputHash.hashType, oldHashPart };
|
|
|
|
|
switch (outputHash.method) {
|
|
|
|
|
case FileIngestionMethod::Recursive:
|
|
|
|
|
dumpPath(actualPath, caSink);
|
|
|
|
|
break;
|
|
|
|
|
case FileIngestionMethod::Flat:
|
|
|
|
|
readFile(actualPath, caSink);
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
auto got = caSink.finish().first;
|
|
|
|
|
ValidPathInfo newInfo0 {
|
2021-02-27 07:42:13 +02:00
|
|
|
|
worker.store,
|
2023-01-23 19:58:11 +02:00
|
|
|
|
outputPathName(drv->name, outputName),
|
|
|
|
|
FixedOutputInfo {
|
2023-02-28 19:13:43 +02:00
|
|
|
|
.hash = {
|
2023-01-23 19:58:11 +02:00
|
|
|
|
.method = outputHash.method,
|
|
|
|
|
.hash = got,
|
2021-02-27 07:42:13 +02:00
|
|
|
|
},
|
2023-02-28 18:57:20 +02:00
|
|
|
|
.references = rewriteRefs(),
|
2021-02-27 07:42:13 +02:00
|
|
|
|
},
|
2021-10-01 01:36:50 +03:00
|
|
|
|
Hash::dummy,
|
2021-02-23 08:26:35 +02:00
|
|
|
|
};
|
2023-01-06 17:35:20 +02:00
|
|
|
|
if (*scratchPath != newInfo0.path) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
// Also rewrite the output path
|
|
|
|
|
auto source = sinkToSource([&](Sink & nextSink) {
|
|
|
|
|
RewritingSink rsink2(oldHashPart, std::string(newInfo0.path.hashPart()), nextSink);
|
2022-08-19 12:26:26 +03:00
|
|
|
|
dumpPath(actualPath, rsink2);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
rsink2.flush();
|
|
|
|
|
});
|
|
|
|
|
Path tmpPath = actualPath + ".tmp";
|
|
|
|
|
restorePath(tmpPath, *source);
|
|
|
|
|
deletePath(actualPath);
|
|
|
|
|
movePath(tmpPath, actualPath);
|
|
|
|
|
}
|
|
|
|
|
|
2021-05-25 11:29:10 +03:00
|
|
|
|
HashResult narHashAndSize = hashPath(htSHA256, actualPath);
|
2021-10-01 01:36:50 +03:00
|
|
|
|
newInfo0.narHash = narHashAndSize.first;
|
2021-05-25 11:29:10 +03:00
|
|
|
|
newInfo0.narSize = narHashAndSize.second;
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
assert(newInfo0.ca);
|
|
|
|
|
return newInfo0;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
ValidPathInfo newInfo = std::visit(overloaded {
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
2022-03-18 00:29:15 +02:00
|
|
|
|
[&](const DerivationOutput::InputAddressed & output) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* input-addressed case */
|
|
|
|
|
auto requiredFinalPath = output.path;
|
|
|
|
|
/* Preemptively add rewrite rule for final hash, as that is
|
|
|
|
|
what the NAR hash will use rather than normalized-self references */
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (*scratchPath != requiredFinalPath)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
outputRewrites.insert_or_assign(
|
2022-05-04 08:44:32 +03:00
|
|
|
|
std::string { scratchPath->hashPart() },
|
2021-02-23 08:26:35 +02:00
|
|
|
|
std::string { requiredFinalPath.hashPart() });
|
|
|
|
|
rewriteOutput();
|
|
|
|
|
auto narHashAndSize = hashPath(htSHA256, actualPath);
|
|
|
|
|
ValidPathInfo newInfo0 { requiredFinalPath, narHashAndSize.first };
|
|
|
|
|
newInfo0.narSize = narHashAndSize.second;
|
2023-01-14 23:38:43 +02:00
|
|
|
|
auto refs = rewriteRefs();
|
|
|
|
|
newInfo0.references = std::move(refs.others);
|
|
|
|
|
if (refs.self)
|
|
|
|
|
newInfo0.references.insert(newInfo0.path);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
return newInfo0;
|
|
|
|
|
},
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
2022-03-18 00:29:15 +02:00
|
|
|
|
[&](const DerivationOutput::CAFixed & dof) {
|
|
|
|
|
auto newInfo0 = newInfoFromCA(DerivationOutput::CAFloating {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
.method = dof.hash.method,
|
|
|
|
|
.hashType = dof.hash.hash.type,
|
|
|
|
|
});
|
|
|
|
|
|
|
|
|
|
/* Check wanted hash */
|
2021-10-01 00:31:21 +03:00
|
|
|
|
const Hash & wanted = dof.hash.hash;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
assert(newInfo0.ca);
|
2023-03-31 00:12:49 +03:00
|
|
|
|
auto got = newInfo0.ca->getHash();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (wanted != got) {
|
|
|
|
|
/* Throw an error after registering the path as
|
|
|
|
|
valid. */
|
|
|
|
|
worker.hashMismatch = true;
|
|
|
|
|
delayedException = std::make_exception_ptr(
|
|
|
|
|
BuildError("hash mismatch in fixed-output derivation '%s':\n specified: %s\n got: %s",
|
|
|
|
|
worker.store.printStorePath(drvPath),
|
|
|
|
|
wanted.to_string(SRI, true),
|
|
|
|
|
got.to_string(SRI, true)));
|
|
|
|
|
}
|
|
|
|
|
return newInfo0;
|
|
|
|
|
},
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
2022-03-18 00:29:15 +02:00
|
|
|
|
[&](const DerivationOutput::CAFloating & dof) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
return newInfoFromCA(dof);
|
|
|
|
|
},
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
2022-03-18 00:29:15 +02:00
|
|
|
|
[&](const DerivationOutput::Deferred &) -> ValidPathInfo {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
// No derivation should reach that point without having been
|
|
|
|
|
// rewritten first
|
|
|
|
|
assert(false);
|
|
|
|
|
},
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
2022-03-30 17:31:01 +03:00
|
|
|
|
[&](const DerivationOutput::Impure & doi) {
|
|
|
|
|
return newInfoFromCA(DerivationOutput::CAFloating {
|
|
|
|
|
.method = doi.method,
|
|
|
|
|
.hashType = doi.hashType,
|
|
|
|
|
});
|
|
|
|
|
},
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
}, output->raw());
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2021-05-05 22:00:08 +03:00
|
|
|
|
/* FIXME: set proper permissions in restorePath() so
|
|
|
|
|
we don't have to do another traversal. */
|
2021-09-15 19:51:52 +03:00
|
|
|
|
canonicalisePathMetaData(actualPath, {}, inodesSeen);
|
2021-05-05 22:00:08 +03:00
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
/* Calculate where we'll move the output files. In the checking case we
|
|
|
|
|
will leave leave them where they are, for now, rather than move to
|
|
|
|
|
their usual "final destination" */
|
|
|
|
|
auto finalDestPath = worker.store.printStorePath(newInfo.path);
|
|
|
|
|
|
|
|
|
|
/* Lock final output path, if not already locked. This happens with
|
|
|
|
|
floating CA derivations and hash-mismatching fixed-output
|
|
|
|
|
derivations. */
|
|
|
|
|
PathLocks dynamicOutputLock;
|
2021-06-24 16:44:13 +03:00
|
|
|
|
dynamicOutputLock.setDeletion(true);
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto optFixedPath = output->path(worker.store, drv->name, outputName);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
if (!optFixedPath ||
|
|
|
|
|
worker.store.printStorePath(*optFixedPath) != finalDestPath)
|
|
|
|
|
{
|
|
|
|
|
assert(newInfo.ca);
|
|
|
|
|
dynamicOutputLock.lockPaths({worker.store.toRealPath(finalDestPath)});
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Move files, if needed */
|
|
|
|
|
if (worker.store.toRealPath(finalDestPath) != actualPath) {
|
|
|
|
|
if (buildMode == bmRepair) {
|
|
|
|
|
/* Path already exists, need to replace it */
|
|
|
|
|
replaceValidPath(worker.store.toRealPath(finalDestPath), actualPath);
|
|
|
|
|
actualPath = worker.store.toRealPath(finalDestPath);
|
|
|
|
|
} else if (buildMode == bmCheck) {
|
|
|
|
|
/* Path already exists, and we want to compare, so we leave out
|
|
|
|
|
new path in place. */
|
|
|
|
|
} else if (worker.store.isValidPath(newInfo.path)) {
|
|
|
|
|
/* Path already exists because CA path produced by something
|
|
|
|
|
else. No moving needed. */
|
|
|
|
|
assert(newInfo.ca);
|
|
|
|
|
} else {
|
|
|
|
|
auto destPath = worker.store.toRealPath(finalDestPath);
|
2021-06-21 16:17:31 +03:00
|
|
|
|
deletePath(destPath);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
movePath(actualPath, destPath);
|
|
|
|
|
actualPath = destPath;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
auto & localStore = getLocalStore();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (buildMode == bmCheck) {
|
|
|
|
|
|
|
|
|
|
if (!worker.store.isValidPath(newInfo.path)) continue;
|
|
|
|
|
ValidPathInfo oldInfo(*worker.store.queryPathInfo(newInfo.path));
|
|
|
|
|
if (newInfo.narHash != oldInfo.narHash) {
|
|
|
|
|
worker.checkMismatch = true;
|
|
|
|
|
if (settings.runDiffHook || settings.keepFailed) {
|
|
|
|
|
auto dst = worker.store.toRealPath(finalDestPath + checkSuffix);
|
|
|
|
|
deletePath(dst);
|
|
|
|
|
movePath(actualPath, dst);
|
|
|
|
|
|
|
|
|
|
handleDiffHook(
|
|
|
|
|
buildUser ? buildUser->getUID() : getuid(),
|
|
|
|
|
buildUser ? buildUser->getGID() : getgid(),
|
|
|
|
|
finalDestPath, dst, worker.store.printStorePath(drvPath), tmpDir);
|
|
|
|
|
|
|
|
|
|
throw NotDeterministic("derivation '%s' may not be deterministic: output '%s' differs from '%s'",
|
|
|
|
|
worker.store.printStorePath(drvPath), worker.store.toRealPath(finalDestPath), dst);
|
|
|
|
|
} else
|
|
|
|
|
throw NotDeterministic("derivation '%s' may not be deterministic: output '%s' differs",
|
|
|
|
|
worker.store.printStorePath(drvPath), worker.store.toRealPath(finalDestPath));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* Since we verified the build, it's now ultimately trusted. */
|
|
|
|
|
if (!oldInfo.ultimate) {
|
|
|
|
|
oldInfo.ultimate = true;
|
|
|
|
|
localStore.signPathInfo(oldInfo);
|
|
|
|
|
localStore.registerValidPaths({{oldInfo.path, oldInfo}});
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
continue;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* For debugging, print out the referenced and unreferenced paths. */
|
|
|
|
|
for (auto & i : inputPaths) {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (references.count(i))
|
2021-02-23 08:26:35 +02:00
|
|
|
|
debug("referenced input: '%1%'", worker.store.printStorePath(i));
|
2022-05-04 08:44:32 +03:00
|
|
|
|
else
|
|
|
|
|
debug("unreferenced input: '%1%'", worker.store.printStorePath(i));
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2022-09-26 21:55:56 +03:00
|
|
|
|
localStore.optimisePath(actualPath, NoRepair); // FIXME: combine with scanForReferences()
|
|
|
|
|
worker.markContentsGood(newInfo.path);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
newInfo.deriver = drvPath;
|
|
|
|
|
newInfo.ultimate = true;
|
|
|
|
|
localStore.signPathInfo(newInfo);
|
|
|
|
|
|
|
|
|
|
finish(newInfo.path);
|
|
|
|
|
|
|
|
|
|
/* If it's a CA path, register it right away. This is necessary if it
|
|
|
|
|
isn't statically known so that we can safely unlock the path before
|
|
|
|
|
the next iteration */
|
|
|
|
|
if (newInfo.ca)
|
|
|
|
|
localStore.registerValidPaths({{newInfo.path, newInfo}});
|
|
|
|
|
|
|
|
|
|
infos.emplace(outputName, std::move(newInfo));
|
|
|
|
|
}
|
|
|
|
|
|
2021-09-26 16:20:36 +03:00
|
|
|
|
if (buildMode == bmCheck) {
|
2022-03-08 20:50:46 +02:00
|
|
|
|
/* In case of fixed-output derivations, if there are
|
|
|
|
|
mismatches on `--check` an error must be thrown as this is
|
|
|
|
|
also a source for non-determinism. */
|
2021-09-26 16:20:36 +03:00
|
|
|
|
if (delayedException)
|
|
|
|
|
std::rethrow_exception(delayedException);
|
2022-03-08 20:50:46 +02:00
|
|
|
|
return assertPathValidity();
|
2021-09-26 16:20:36 +03:00
|
|
|
|
}
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
/* Apply output checks. */
|
|
|
|
|
checkOutputs(infos);
|
|
|
|
|
|
|
|
|
|
/* Register each output path as valid, and register the sets of
|
|
|
|
|
paths referenced by each of them. If there are cycles in the
|
|
|
|
|
outputs, this will fail. */
|
|
|
|
|
{
|
2021-02-26 17:20:33 +02:00
|
|
|
|
auto & localStore = getLocalStore();
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
ValidPathInfos infos2;
|
|
|
|
|
for (auto & [outputName, newInfo] : infos) {
|
|
|
|
|
infos2.insert_or_assign(newInfo.path, newInfo);
|
|
|
|
|
}
|
|
|
|
|
localStore.registerValidPaths(infos2);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* In case of a fixed-output derivation hash mismatch, throw an
|
|
|
|
|
exception now that we have registered the output as valid. */
|
|
|
|
|
if (delayedException)
|
|
|
|
|
std::rethrow_exception(delayedException);
|
|
|
|
|
|
|
|
|
|
/* If we made it this far, we are sure the output matches the derivation
|
|
|
|
|
(since the delayedException would be a fixed output CA mismatch). That
|
|
|
|
|
means it's safe to link the derivation to the output hash. We must do
|
|
|
|
|
that for floating CA derivations, which otherwise couldn't be cached,
|
|
|
|
|
but it's fine to do in all cases. */
|
2023-04-15 01:18:32 +03:00
|
|
|
|
SingleDrvOutputs builtOutputs;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-03-08 20:50:46 +02:00
|
|
|
|
for (auto & [outputName, newInfo] : infos) {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto oldinfo = get(initialOutputs, outputName);
|
|
|
|
|
assert(oldinfo);
|
2022-03-08 20:50:46 +02:00
|
|
|
|
auto thisRealisation = Realisation {
|
|
|
|
|
.id = DrvOutput {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
oldinfo->outputHash,
|
2022-03-08 20:50:46 +02:00
|
|
|
|
outputName
|
|
|
|
|
},
|
|
|
|
|
.outPath = newInfo.path
|
|
|
|
|
};
|
2023-03-17 16:33:48 +02:00
|
|
|
|
if (experimentalFeatureSettings.isEnabled(Xp::CaDerivations)
|
2022-03-30 17:31:01 +03:00
|
|
|
|
&& drv->type().isPure())
|
|
|
|
|
{
|
2021-03-08 18:32:20 +02:00
|
|
|
|
signRealisation(thisRealisation);
|
2021-03-08 12:56:33 +02:00
|
|
|
|
worker.store.registerDrvOutput(thisRealisation);
|
|
|
|
|
}
|
2023-01-12 01:57:18 +02:00
|
|
|
|
if (wantedOutputs.contains(outputName))
|
2023-04-15 01:18:32 +03:00
|
|
|
|
builtOutputs.emplace(outputName, thisRealisation);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
2022-03-08 20:50:46 +02:00
|
|
|
|
|
|
|
|
|
return builtOutputs;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
2021-03-08 18:32:20 +02:00
|
|
|
|
void LocalDerivationGoal::signRealisation(Realisation & realisation)
|
|
|
|
|
{
|
|
|
|
|
getLocalStore().signRealisation(realisation);
|
|
|
|
|
}
|
|
|
|
|
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
2022-03-08 20:50:46 +02:00
|
|
|
|
void LocalDerivationGoal::checkOutputs(const std::map<std::string, ValidPathInfo> & outputs)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
std::map<Path, const ValidPathInfo &> outputsByPath;
|
|
|
|
|
for (auto & output : outputs)
|
|
|
|
|
outputsByPath.emplace(worker.store.printStorePath(output.second.path), output.second);
|
|
|
|
|
|
|
|
|
|
for (auto & output : outputs) {
|
|
|
|
|
auto & outputName = output.first;
|
|
|
|
|
auto & info = output.second;
|
|
|
|
|
|
|
|
|
|
struct Checks
|
|
|
|
|
{
|
|
|
|
|
bool ignoreSelfRefs = false;
|
|
|
|
|
std::optional<uint64_t> maxSize, maxClosureSize;
|
|
|
|
|
std::optional<Strings> allowedReferences, allowedRequisites, disallowedReferences, disallowedRequisites;
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
/* Compute the closure and closure size of some output. This
|
|
|
|
|
is slightly tricky because some of its references (namely
|
|
|
|
|
other outputs) may not be valid yet. */
|
|
|
|
|
auto getClosure = [&](const StorePath & path)
|
|
|
|
|
{
|
|
|
|
|
uint64_t closureSize = 0;
|
|
|
|
|
StorePathSet pathsDone;
|
|
|
|
|
std::queue<StorePath> pathsLeft;
|
|
|
|
|
pathsLeft.push(path);
|
|
|
|
|
|
|
|
|
|
while (!pathsLeft.empty()) {
|
|
|
|
|
auto path = pathsLeft.front();
|
|
|
|
|
pathsLeft.pop();
|
|
|
|
|
if (!pathsDone.insert(path).second) continue;
|
|
|
|
|
|
|
|
|
|
auto i = outputsByPath.find(worker.store.printStorePath(path));
|
|
|
|
|
if (i != outputsByPath.end()) {
|
|
|
|
|
closureSize += i->second.narSize;
|
2023-01-14 23:38:43 +02:00
|
|
|
|
for (auto & ref : i->second.references)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
pathsLeft.push(ref);
|
|
|
|
|
} else {
|
|
|
|
|
auto info = worker.store.queryPathInfo(path);
|
|
|
|
|
closureSize += info->narSize;
|
2023-01-14 23:38:43 +02:00
|
|
|
|
for (auto & ref : info->references)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
pathsLeft.push(ref);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
return std::make_pair(std::move(pathsDone), closureSize);
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
auto applyChecks = [&](const Checks & checks)
|
|
|
|
|
{
|
|
|
|
|
if (checks.maxSize && info.narSize > *checks.maxSize)
|
|
|
|
|
throw BuildError("path '%s' is too large at %d bytes; limit is %d bytes",
|
|
|
|
|
worker.store.printStorePath(info.path), info.narSize, *checks.maxSize);
|
|
|
|
|
|
|
|
|
|
if (checks.maxClosureSize) {
|
|
|
|
|
uint64_t closureSize = getClosure(info.path).second;
|
|
|
|
|
if (closureSize > *checks.maxClosureSize)
|
|
|
|
|
throw BuildError("closure of path '%s' is too large at %d bytes; limit is %d bytes",
|
|
|
|
|
worker.store.printStorePath(info.path), closureSize, *checks.maxClosureSize);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
auto checkRefs = [&](const std::optional<Strings> & value, bool allowed, bool recursive)
|
|
|
|
|
{
|
|
|
|
|
if (!value) return;
|
|
|
|
|
|
|
|
|
|
/* Parse a list of reference specifiers. Each element must
|
|
|
|
|
either be a store path, or the symbolic name of the output
|
|
|
|
|
of the derivation (such as `out'). */
|
|
|
|
|
StorePathSet spec;
|
|
|
|
|
for (auto & i : *value) {
|
|
|
|
|
if (worker.store.isStorePath(i))
|
|
|
|
|
spec.insert(worker.store.parseStorePath(i));
|
2022-05-04 08:44:32 +03:00
|
|
|
|
else if (auto output = get(outputs, i))
|
|
|
|
|
spec.insert(output->path);
|
|
|
|
|
else
|
|
|
|
|
throw BuildError("derivation contains an illegal reference specifier '%s'", i);
|
2021-02-23 08:26:35 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
auto used = recursive
|
|
|
|
|
? getClosure(info.path).first
|
2023-01-14 23:38:43 +02:00
|
|
|
|
: info.references;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
if (recursive && checks.ignoreSelfRefs)
|
|
|
|
|
used.erase(info.path);
|
|
|
|
|
|
|
|
|
|
StorePathSet badPaths;
|
|
|
|
|
|
|
|
|
|
for (auto & i : used)
|
|
|
|
|
if (allowed) {
|
|
|
|
|
if (!spec.count(i))
|
|
|
|
|
badPaths.insert(i);
|
|
|
|
|
} else {
|
|
|
|
|
if (spec.count(i))
|
|
|
|
|
badPaths.insert(i);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!badPaths.empty()) {
|
2022-02-25 17:00:00 +02:00
|
|
|
|
std::string badPathsStr;
|
2021-02-23 08:26:35 +02:00
|
|
|
|
for (auto & i : badPaths) {
|
|
|
|
|
badPathsStr += "\n ";
|
|
|
|
|
badPathsStr += worker.store.printStorePath(i);
|
|
|
|
|
}
|
|
|
|
|
throw BuildError("output '%s' is not allowed to refer to the following paths:%s",
|
|
|
|
|
worker.store.printStorePath(info.path), badPathsStr);
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
checkRefs(checks.allowedReferences, true, false);
|
|
|
|
|
checkRefs(checks.allowedRequisites, true, true);
|
|
|
|
|
checkRefs(checks.disallowedReferences, false, false);
|
|
|
|
|
checkRefs(checks.disallowedRequisites, false, true);
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
if (auto structuredAttrs = parsedDrv->getStructuredAttrs()) {
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (auto outputChecks = get(*structuredAttrs, "outputChecks")) {
|
|
|
|
|
if (auto output = get(*outputChecks, outputName)) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
Checks checks;
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (auto maxSize = get(*output, "maxSize"))
|
2021-02-23 08:26:35 +02:00
|
|
|
|
checks.maxSize = maxSize->get<uint64_t>();
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
if (auto maxClosureSize = get(*output, "maxClosureSize"))
|
2021-02-23 08:26:35 +02:00
|
|
|
|
checks.maxClosureSize = maxClosureSize->get<uint64_t>();
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
auto get_ = [&](const std::string & name) -> std::optional<Strings> {
|
|
|
|
|
if (auto i = get(*output, name)) {
|
2021-02-23 08:26:35 +02:00
|
|
|
|
Strings res;
|
|
|
|
|
for (auto j = i->begin(); j != i->end(); ++j) {
|
|
|
|
|
if (!j->is_string())
|
|
|
|
|
throw Error("attribute '%s' of derivation '%s' must be a list of strings", name, worker.store.printStorePath(drvPath));
|
|
|
|
|
res.push_back(j->get<std::string>());
|
|
|
|
|
}
|
|
|
|
|
checks.disallowedRequisites = res;
|
|
|
|
|
return res;
|
|
|
|
|
}
|
|
|
|
|
return {};
|
|
|
|
|
};
|
|
|
|
|
|
2022-05-04 08:44:32 +03:00
|
|
|
|
checks.allowedReferences = get_("allowedReferences");
|
|
|
|
|
checks.allowedRequisites = get_("allowedRequisites");
|
|
|
|
|
checks.disallowedReferences = get_("disallowedReferences");
|
|
|
|
|
checks.disallowedRequisites = get_("disallowedRequisites");
|
2021-02-23 08:26:35 +02:00
|
|
|
|
|
|
|
|
|
applyChecks(checks);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
} else {
|
|
|
|
|
// legacy non-structured-attributes case
|
|
|
|
|
Checks checks;
|
|
|
|
|
checks.ignoreSelfRefs = true;
|
|
|
|
|
checks.allowedReferences = parsedDrv->getStringsAttr("allowedReferences");
|
|
|
|
|
checks.allowedRequisites = parsedDrv->getStringsAttr("allowedRequisites");
|
|
|
|
|
checks.disallowedReferences = parsedDrv->getStringsAttr("disallowedReferences");
|
|
|
|
|
checks.disallowedRequisites = parsedDrv->getStringsAttr("disallowedRequisites");
|
|
|
|
|
applyChecks(checks);
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
void LocalDerivationGoal::deleteTmpDir(bool force)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
if (tmpDir != "") {
|
|
|
|
|
/* Don't keep temporary directories for builtins because they
|
|
|
|
|
might have privileged stuff (like a copy of netrc). */
|
|
|
|
|
if (settings.keepFailed && !force && !drv->isBuiltin()) {
|
|
|
|
|
printError("note: keeping build directory '%s'", tmpDir);
|
|
|
|
|
chmod(tmpDir.c_str(), 0755);
|
|
|
|
|
}
|
|
|
|
|
else
|
|
|
|
|
deletePath(tmpDir);
|
|
|
|
|
tmpDir = "";
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
bool LocalDerivationGoal::isReadDesc(int fd)
|
|
|
|
|
{
|
|
|
|
|
return (hook && DerivationGoal::isReadDesc(fd)) ||
|
2023-03-15 11:37:39 +02:00
|
|
|
|
(!hook && fd == builderOut.get());
|
2021-02-26 17:20:33 +02:00
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
StorePath LocalDerivationGoal::makeFallbackPath(std::string_view outputName)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
return worker.store.makeStorePath(
|
|
|
|
|
"rewrite:" + std::string(drvPath.to_string()) + ":name:" + std::string(outputName),
|
|
|
|
|
Hash(htSHA256), outputPathName(drv->name, outputName));
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
2021-02-26 17:20:33 +02:00
|
|
|
|
StorePath LocalDerivationGoal::makeFallbackPath(const StorePath & path)
|
2021-02-23 08:26:35 +02:00
|
|
|
|
{
|
|
|
|
|
return worker.store.makeStorePath(
|
|
|
|
|
"rewrite:" + std::string(drvPath.to_string()) + ":" + std::string(path.to_string()),
|
|
|
|
|
Hash(htSHA256), path.name());
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
}
|