mirror of
https://github.com/privatevoid-net/nix-super.git
synced 2024-11-27 16:16:16 +02:00
462421d345
This provides a pluggable mechanism for defining new fetchers. It adds
a builtin function 'fetchTree' that generalizes existing fetchers like
'fetchGit', 'fetchMercurial' and 'fetchTarball'. 'fetchTree' takes a
set of attributes, e.g.
fetchTree {
type = "git";
url = "https://example.org/repo.git";
ref = "some-branch";
rev = "abcdef...";
}
The existing fetchers are just wrappers around this. Note that the
input attributes to fetchTree are the same as flake input
specifications and flake lock file entries.
All fetchers share a common cache stored in
~/.cache/nix/fetcher-cache-v1.sqlite. This replaces the ad hoc caching
mechanisms in fetchGit and download.cc (e.g. ~/.cache/nix/{tarballs,git-revs*}).
This also adds support for Git worktrees (c169ea5904
).
121 lines
3.2 KiB
C++
121 lines
3.2 KiB
C++
#include "cache.hh"
|
|
#include "sqlite.hh"
|
|
#include "sync.hh"
|
|
#include "store-api.hh"
|
|
|
|
#include <nlohmann/json.hpp>
|
|
|
|
namespace nix::fetchers {
|
|
|
|
static const char * schema = R"sql(
|
|
|
|
create table if not exists Cache (
|
|
input text not null,
|
|
info text not null,
|
|
path text not null,
|
|
immutable integer not null,
|
|
timestamp integer not null,
|
|
primary key (input)
|
|
);
|
|
)sql";
|
|
|
|
struct CacheImpl : Cache
|
|
{
|
|
struct State
|
|
{
|
|
SQLite db;
|
|
SQLiteStmt add, lookup;
|
|
};
|
|
|
|
Sync<State> _state;
|
|
|
|
CacheImpl()
|
|
{
|
|
auto state(_state.lock());
|
|
|
|
auto dbPath = getCacheDir() + "/nix/fetcher-cache-v1.sqlite";
|
|
createDirs(dirOf(dbPath));
|
|
|
|
state->db = SQLite(dbPath);
|
|
state->db.isCache();
|
|
state->db.exec(schema);
|
|
|
|
state->add.create(state->db,
|
|
"insert or replace into Cache(input, info, path, immutable, timestamp) values (?, ?, ?, ?, ?)");
|
|
|
|
state->lookup.create(state->db,
|
|
"select info, path, immutable, timestamp from Cache where input = ?");
|
|
}
|
|
|
|
void add(
|
|
ref<Store> store,
|
|
const Attrs & inAttrs,
|
|
const Attrs & infoAttrs,
|
|
const StorePath & storePath,
|
|
bool immutable) override
|
|
{
|
|
_state.lock()->add.use()
|
|
(attrsToJson(inAttrs).dump())
|
|
(attrsToJson(infoAttrs).dump())
|
|
(store->printStorePath(storePath))
|
|
(immutable)
|
|
(time(0)).exec();
|
|
}
|
|
|
|
std::optional<std::pair<Attrs, StorePath>> lookup(
|
|
ref<Store> store,
|
|
const Attrs & inAttrs) override
|
|
{
|
|
if (auto res = lookupExpired(store, inAttrs)) {
|
|
if (!res->expired)
|
|
return std::make_pair(std::move(res->infoAttrs), std::move(res->storePath));
|
|
debug("ignoring expired cache entry '%s'",
|
|
attrsToJson(inAttrs).dump());
|
|
}
|
|
return {};
|
|
}
|
|
|
|
std::optional<Result> lookupExpired(
|
|
ref<Store> store,
|
|
const Attrs & inAttrs) override
|
|
{
|
|
auto state(_state.lock());
|
|
|
|
auto inAttrsJson = attrsToJson(inAttrs).dump();
|
|
|
|
auto stmt(state->lookup.use()(inAttrsJson));
|
|
if (!stmt.next()) {
|
|
debug("did not find cache entry for '%s'", inAttrsJson);
|
|
return {};
|
|
}
|
|
|
|
auto infoJson = stmt.getStr(0);
|
|
auto storePath = store->parseStorePath(stmt.getStr(1));
|
|
auto immutable = stmt.getInt(2) != 0;
|
|
auto timestamp = stmt.getInt(3);
|
|
|
|
store->addTempRoot(storePath);
|
|
if (!store->isValidPath(storePath)) {
|
|
// FIXME: we could try to substitute 'storePath'.
|
|
debug("ignoring disappeared cache entry '%s'", inAttrsJson);
|
|
return {};
|
|
}
|
|
|
|
debug("using cache entry '%s' -> '%s', '%s'",
|
|
inAttrsJson, infoJson, store->printStorePath(storePath));
|
|
|
|
return Result {
|
|
.expired = !immutable && (settings.tarballTtl.get() == 0 || timestamp + settings.tarballTtl < time(0)),
|
|
.infoAttrs = jsonToAttrs(nlohmann::json::parse(infoJson)),
|
|
.storePath = std::move(storePath)
|
|
};
|
|
}
|
|
};
|
|
|
|
ref<Cache> getCache()
|
|
{
|
|
static auto cache = std::make_shared<CacheImpl>();
|
|
return ref<Cache>(cache);
|
|
}
|
|
|
|
}
|