Compare commits
3 Commits
611255d42c
...
9aee36a0e2
| Author | SHA1 | Date | |
|---|---|---|---|
|
9aee36a0e2
|
|||
|
dcb853ea0a
|
|||
|
2441e10607
|
372
Cargo.lock
generated
372
Cargo.lock
generated
@@ -34,15 +34,6 @@ version = "0.2.21"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "683d7910e743518b0e34f1186f92494becacb047c7b6bf616c96772180fef923"
|
||||
|
||||
[[package]]
|
||||
name = "android_system_properties"
|
||||
version = "0.1.5"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "819e7219dbd41043ac279b19830f2efc897156490d7fd6ea916720117ee66311"
|
||||
dependencies = [
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "anes"
|
||||
version = "0.1.6"
|
||||
@@ -70,28 +61,6 @@ dependencies = [
|
||||
"derive_arbitrary",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "async-stream"
|
||||
version = "0.3.6"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "0b5a71a6f37880a80d1d7f19efd781e4b5de42c88f0722cc13bcb6cc2cfe8476"
|
||||
dependencies = [
|
||||
"async-stream-impl",
|
||||
"futures-core",
|
||||
"pin-project-lite",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "async-stream-impl"
|
||||
version = "0.3.6"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "c7c24de15d275a1ecfd47a380fb4d5ec9bfe0933f309ed5e705b775596a3574d"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "atomic-waker"
|
||||
version = "1.1.2"
|
||||
@@ -126,6 +95,12 @@ dependencies = [
|
||||
"vsimd",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "base64ct"
|
||||
version = "1.8.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "2af50177e190e07a26ab74f8b1efbfe2ef87da2116221318cb1c2e82baf7de06"
|
||||
|
||||
[[package]]
|
||||
name = "bincode"
|
||||
version = "1.3.3"
|
||||
@@ -207,6 +182,17 @@ dependencies = [
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "bstr"
|
||||
version = "1.12.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "63044e1ae8e69f3b5a92c736ca6269b8d12fa7efe39bf34ddb06d102cf0e2cab"
|
||||
dependencies = [
|
||||
"memchr",
|
||||
"regex-automata",
|
||||
"serde",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "bumpalo"
|
||||
version = "3.19.0"
|
||||
@@ -303,7 +289,7 @@ version = "0.6.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "6fac387a98bb7c37292057cffc56d62ecb629900026402633ae9160df93a8766"
|
||||
dependencies = [
|
||||
"nom",
|
||||
"nom 7.1.3",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -324,19 +310,6 @@ version = "0.2.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "613afe47fcd5fac7ccf1db93babcb082c5994d996f20b8b159f2ad1658eb5724"
|
||||
|
||||
[[package]]
|
||||
name = "chrono"
|
||||
version = "0.4.43"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "fac4744fb15ae8337dc853fee7fb3f4e48c0fbaa23d0afe49c447b4fab126118"
|
||||
dependencies = [
|
||||
"iana-time-zone",
|
||||
"js-sys",
|
||||
"num-traits",
|
||||
"wasm-bindgen",
|
||||
"windows-link",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "ciborium"
|
||||
version = "0.2.2"
|
||||
@@ -419,6 +392,12 @@ dependencies = [
|
||||
"error-code",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "const-oid"
|
||||
version = "0.9.6"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "c2459377285ad874054d797f3ccebf984978aa39129f6eafde5cdc8315b612f8"
|
||||
|
||||
[[package]]
|
||||
name = "constant_time_eq"
|
||||
version = "0.3.1"
|
||||
@@ -449,12 +428,6 @@ version = "5.0.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "147be55d677052dabc6b22252d5dd0fd4c29c8c27aa4f2fbef0f94aa003b406f"
|
||||
|
||||
[[package]]
|
||||
name = "core-foundation-sys"
|
||||
version = "0.8.7"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "773648b94d0e5d620f64f280777445740e61fe701025087ec8b57f45c791888b"
|
||||
|
||||
[[package]]
|
||||
name = "core_maths"
|
||||
version = "0.1.1"
|
||||
@@ -580,6 +553,33 @@ dependencies = [
|
||||
"typenum",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "curve25519-dalek"
|
||||
version = "4.1.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "97fb8b7c4503de7d6ae7b42ab72a5a59857b4c937ec27a3d4539dba95b5ab2be"
|
||||
dependencies = [
|
||||
"cfg-if",
|
||||
"cpufeatures",
|
||||
"curve25519-dalek-derive",
|
||||
"digest",
|
||||
"fiat-crypto",
|
||||
"rustc_version",
|
||||
"subtle",
|
||||
"zeroize",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "curve25519-dalek-derive"
|
||||
version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f46882e17999c6cc590af592290432be3bce0428cb0d5f8b6715e4dc7b383eb3"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "data-encoding"
|
||||
version = "2.9.0"
|
||||
@@ -713,6 +713,16 @@ dependencies = [
|
||||
"tokio",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "der"
|
||||
version = "0.7.10"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "e7c1832837b905bbfb5101e07cc24c8deddf52f93225eee6ead5f4d63d53ddcb"
|
||||
dependencies = [
|
||||
"const-oid",
|
||||
"zeroize",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "deranged"
|
||||
version = "0.5.5"
|
||||
@@ -831,6 +841,30 @@ dependencies = [
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "ed25519"
|
||||
version = "2.2.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "115531babc129696a58c64a4fef0a8bf9e9698629fb97e9e40767d235cfbcd53"
|
||||
dependencies = [
|
||||
"pkcs8",
|
||||
"signature",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "ed25519-dalek"
|
||||
version = "2.2.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "70e796c081cee67dc755e1a36a0a172b897fab85fc3f6bc48307991f64e4eca9"
|
||||
dependencies = [
|
||||
"curve25519-dalek",
|
||||
"ed25519",
|
||||
"serde",
|
||||
"sha2",
|
||||
"subtle",
|
||||
"zeroize",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "either"
|
||||
version = "1.15.0"
|
||||
@@ -865,6 +899,18 @@ version = "3.3.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "dea2df4cf52843e0452895c455a1a2cfbb842a1e7329671acf418fdc53ed4c59"
|
||||
|
||||
[[package]]
|
||||
name = "fallible-iterator"
|
||||
version = "0.3.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "2acce4a10f12dc2fb14a218589d4f1f62ef011b2d0cc4b3cb1bba8e94da14649"
|
||||
|
||||
[[package]]
|
||||
name = "fallible-streaming-iterator"
|
||||
version = "0.1.9"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7360491ce676a36bf9bb3c56c1aa791658183a54d2744120f27285738d90465a"
|
||||
|
||||
[[package]]
|
||||
name = "fastrand"
|
||||
version = "2.3.0"
|
||||
@@ -882,6 +928,12 @@ dependencies = [
|
||||
"windows-sys 0.59.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "fiat-crypto"
|
||||
version = "0.2.9"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "28dea519a9695b9977216879a3ebfddf92f1c08c05d984f8996aecd6ecdc811d"
|
||||
|
||||
[[package]]
|
||||
name = "filetime"
|
||||
version = "0.2.26"
|
||||
@@ -1125,6 +1177,15 @@ dependencies = [
|
||||
"foldhash 0.2.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "hashlink"
|
||||
version = "0.10.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7382cf6263419f2d8df38c55d7da83da5c18aef87fc7a7fc1fb1e344edfe14c1"
|
||||
dependencies = [
|
||||
"hashbrown 0.15.5",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "heck"
|
||||
version = "0.5.0"
|
||||
@@ -1262,30 +1323,6 @@ dependencies = [
|
||||
"tracing",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "iana-time-zone"
|
||||
version = "0.1.64"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "33e57f83510bb73707521ebaffa789ec8caf86f9657cad665b092b581d40e9fb"
|
||||
dependencies = [
|
||||
"android_system_properties",
|
||||
"core-foundation-sys",
|
||||
"iana-time-zone-haiku",
|
||||
"js-sys",
|
||||
"log",
|
||||
"wasm-bindgen",
|
||||
"windows-core",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "iana-time-zone-haiku"
|
||||
version = "0.1.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f31827a206f56af32e590ba56d5d2d085f558508192593743f16b2306495269f"
|
||||
dependencies = [
|
||||
"cc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "icu_calendar"
|
||||
version = "2.1.1"
|
||||
@@ -1603,6 +1640,17 @@ dependencies = [
|
||||
"redox_syscall 0.7.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "libsqlite3-sys"
|
||||
version = "0.31.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ad8935b44e7c13394a179a438e0cebba0fe08fe01b54f152e29a93b5cf993fd4"
|
||||
dependencies = [
|
||||
"cc",
|
||||
"pkg-config",
|
||||
"vcpkg",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "linux-raw-sys"
|
||||
version = "0.4.15"
|
||||
@@ -1745,23 +1793,38 @@ dependencies = [
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "nix-daemon"
|
||||
version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "bb28bc02b8ea18d59e15fc8e86ae35850326dc5e4e2dcf17bc659f2fd79f1a08"
|
||||
name = "nix-compat"
|
||||
version = "0.1.0"
|
||||
source = "git+https://git.snix.dev/snix/snix.git#44743974c24f9850a1552d07734f95a54e51dba0"
|
||||
dependencies = [
|
||||
"async-stream",
|
||||
"chrono",
|
||||
"bitflags",
|
||||
"bstr",
|
||||
"bytes",
|
||||
"data-encoding",
|
||||
"ed25519",
|
||||
"ed25519-dalek",
|
||||
"futures",
|
||||
"mimalloc",
|
||||
"nix-compat-derive",
|
||||
"nom 8.0.0",
|
||||
"num_enum",
|
||||
"tap",
|
||||
"thiserror 1.0.69",
|
||||
"pin-project-lite",
|
||||
"sha2",
|
||||
"thiserror 2.0.17",
|
||||
"tokio",
|
||||
"tokio-stream",
|
||||
"tokio-test",
|
||||
"tracing",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "nix-compat-derive"
|
||||
version = "0.1.0"
|
||||
source = "git+https://git.snix.dev/snix/snix.git#44743974c24f9850a1552d07734f95a54e51dba0"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "nix-js"
|
||||
version = "0.1.0"
|
||||
@@ -1778,13 +1841,14 @@ dependencies = [
|
||||
"hex",
|
||||
"itertools 0.14.0",
|
||||
"mimalloc",
|
||||
"nix-daemon",
|
||||
"nix-compat",
|
||||
"nix-js-macros",
|
||||
"nix-nar",
|
||||
"petgraph",
|
||||
"regex",
|
||||
"reqwest",
|
||||
"rnix",
|
||||
"rusqlite",
|
||||
"rustyline",
|
||||
"serde",
|
||||
"serde_json",
|
||||
@@ -1832,6 +1896,15 @@ dependencies = [
|
||||
"minimal-lexical",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "nom"
|
||||
version = "8.0.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "df9761775871bdef83bee530e60050f7e54b1105350d6884eb0fb4f46c2f9405"
|
||||
dependencies = [
|
||||
"memchr",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "nu-ansi-term"
|
||||
version = "0.50.3"
|
||||
@@ -2011,6 +2084,16 @@ version = "0.1.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "8b870d8c151b6f2fb93e84a13146138f05d02ed11c7e7c54f8826aaaf7c9f184"
|
||||
|
||||
[[package]]
|
||||
name = "pkcs8"
|
||||
version = "0.10.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f950b2377845cebe5cf8b5165cb3cc1a5e0fa5cfa3e1f7f55707d8fd82e0a7b7"
|
||||
dependencies = [
|
||||
"der",
|
||||
"spki",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "pkg-config"
|
||||
version = "0.3.32"
|
||||
@@ -2219,6 +2302,9 @@ name = "rand_core"
|
||||
version = "0.6.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ec0be4795e2f6a28069bec0b5ff3e2ac9bafc99e6a9a7dc3547996c5c816922c"
|
||||
dependencies = [
|
||||
"getrandom 0.2.16",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rand_core"
|
||||
@@ -2393,6 +2479,20 @@ dependencies = [
|
||||
"text-size",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rusqlite"
|
||||
version = "0.33.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1c6d5e5acb6f6129fe3f7ba0a7fc77bca1942cb568535e18e7bc40262baf3110"
|
||||
dependencies = [
|
||||
"bitflags",
|
||||
"fallible-iterator",
|
||||
"fallible-streaming-iterator",
|
||||
"hashlink",
|
||||
"libsqlite3-sys",
|
||||
"smallvec",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustc-hash"
|
||||
version = "1.1.0"
|
||||
@@ -2647,6 +2747,15 @@ dependencies = [
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "signature"
|
||||
version = "2.2.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "77549399552de45a898a580c1b41d445bf730df867cc44e6c0233bbc4b8329de"
|
||||
dependencies = [
|
||||
"rand_core 0.6.4",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "simd-adler32"
|
||||
version = "0.3.8"
|
||||
@@ -2693,6 +2802,16 @@ dependencies = [
|
||||
"url",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "spki"
|
||||
version = "0.7.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d91ed6c858b01f942cd56b37a94b3e0a1798290327d1236e4d9cf4eaca44d29d"
|
||||
dependencies = [
|
||||
"base64ct",
|
||||
"der",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "stable_deref_trait"
|
||||
version = "1.2.1"
|
||||
@@ -3048,28 +3167,6 @@ dependencies = [
|
||||
"tokio",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio-stream"
|
||||
version = "0.1.18"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "32da49809aab5c3bc678af03902d4ccddea2a87d028d86392a4b1560c6906c70"
|
||||
dependencies = [
|
||||
"futures-core",
|
||||
"pin-project-lite",
|
||||
"tokio",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio-test"
|
||||
version = "0.4.5"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3f6d24790a10a7af737693a3e8f1d03faef7e6ca0cc99aae5066f533766de545"
|
||||
dependencies = [
|
||||
"futures-core",
|
||||
"tokio",
|
||||
"tokio-stream",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "toml_datetime"
|
||||
version = "0.7.5+spec-1.1.0"
|
||||
@@ -3311,6 +3408,12 @@ version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ba73ea9cf16a25df0c8caa16c51acb937d5712a8429db78a3ee29d5dcacd3a65"
|
||||
|
||||
[[package]]
|
||||
name = "vcpkg"
|
||||
version = "0.2.15"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "accd4ea62f7bb7a82fe23066fb0957d48ef677f6eeb8215f372f52e48bb32426"
|
||||
|
||||
[[package]]
|
||||
name = "version_check"
|
||||
version = "0.9.5"
|
||||
@@ -3497,65 +3600,12 @@ version = "0.4.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "712e227841d057c1ee1cd2fb22fa7e5a5461ae8e48fa2ca79ec42cfc1931183f"
|
||||
|
||||
[[package]]
|
||||
name = "windows-core"
|
||||
version = "0.62.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b8e83a14d34d0623b51dce9581199302a221863196a1dde71a7663a4c2be9deb"
|
||||
dependencies = [
|
||||
"windows-implement",
|
||||
"windows-interface",
|
||||
"windows-link",
|
||||
"windows-result",
|
||||
"windows-strings",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-implement"
|
||||
version = "0.60.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "053e2e040ab57b9dc951b72c264860db7eb3b0200ba345b4e4c3b14f67855ddf"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-interface"
|
||||
version = "0.59.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3f316c4a2570ba26bbec722032c4099d8c8bc095efccdc15688708623367e358"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-link"
|
||||
version = "0.2.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f0805222e57f7521d6a62e36fa9163bc891acd422f971defe97d64e70d0a4fe5"
|
||||
|
||||
[[package]]
|
||||
name = "windows-result"
|
||||
version = "0.4.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7781fa89eaf60850ac3d2da7af8e5242a5ea78d1a11c49bf2910bb5a73853eb5"
|
||||
dependencies = [
|
||||
"windows-link",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-strings"
|
||||
version = "0.5.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7837d08f69c77cf6b07689544538e017c1bfcf57e34b4c0ff58e6c2cd3b37091"
|
||||
dependencies = [
|
||||
"windows-link",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-sys"
|
||||
version = "0.48.0"
|
||||
|
||||
@@ -6,13 +6,13 @@ build = "build.rs"
|
||||
|
||||
[features]
|
||||
default = ["daemon"]
|
||||
daemon = ["dep:tokio", "dep:nix-daemon"]
|
||||
daemon = ["dep:tokio", "dep:nix-compat"]
|
||||
|
||||
[dependencies]
|
||||
mimalloc = "0.1"
|
||||
|
||||
tokio = { version = "1.41", features = ["rt-multi-thread", "sync"], optional = true }
|
||||
nix-daemon = { version = "0.1", optional = true }
|
||||
tokio = { version = "1.41", features = ["rt-multi-thread", "sync", "net", "io-util"], optional = true }
|
||||
nix-compat = { git = "https://git.snix.dev/snix/snix.git", version = "0.1.0", features = ["wire", "async"], optional = true }
|
||||
|
||||
# REPL
|
||||
anyhow = "1.0"
|
||||
@@ -51,6 +51,7 @@ serde = { version = "1.0", features = ["derive"] }
|
||||
serde_json = "1.0"
|
||||
dirs = "5.0"
|
||||
tempfile = "3.24"
|
||||
rusqlite = { version = "0.33", features = ["bundled"] }
|
||||
|
||||
rnix = "0.12"
|
||||
|
||||
|
||||
@@ -158,8 +158,14 @@ export const coerceToString = (
|
||||
|
||||
// Paths coerce to their string value
|
||||
if (isNixPath(v)) {
|
||||
// TODO: Implement copyToStore when needed (copy path to Nix store)
|
||||
// For now, just return the raw path string
|
||||
if (copyToStore) {
|
||||
const pathStr = v.value;
|
||||
const storePath = Deno.core.ops.op_copy_path_to_store(pathStr);
|
||||
if (outContext) {
|
||||
outContext.add(storePath);
|
||||
}
|
||||
return storePath;
|
||||
}
|
||||
return v.value;
|
||||
}
|
||||
|
||||
|
||||
@@ -11,6 +11,7 @@ import { coerceToPath, coerceToString, StringCoercionMode } from "./conversion";
|
||||
import { getPathValue } from "../path";
|
||||
import type { NixStringContext, StringWithContext } from "../string-context";
|
||||
import { mkStringWithContext } from "../string-context";
|
||||
import { isPath } from "./type-check";
|
||||
|
||||
export const importFunc = (path: NixValue): NixValue => {
|
||||
const context: NixStringContext = new Set();
|
||||
@@ -143,8 +144,9 @@ export const fetchTarball = (args: NixValue): string => {
|
||||
|
||||
export const fetchGit = (args: NixValue): NixAttrs => {
|
||||
const forced = force(args);
|
||||
if (typeof forced === "string") {
|
||||
const result: FetchGitResult = Deno.core.ops.op_fetch_git(forced, null, null, false, false, false, null);
|
||||
if (typeof forced === "string" || isPath(forced)) {
|
||||
const path = coerceToPath(forced);
|
||||
const result: FetchGitResult = Deno.core.ops.op_fetch_git(path, null, null, false, false, false, null);
|
||||
return {
|
||||
outPath: result.out_path,
|
||||
rev: result.rev,
|
||||
|
||||
@@ -5,6 +5,7 @@
|
||||
import {
|
||||
HAS_CONTEXT,
|
||||
isNixPath,
|
||||
NixPath,
|
||||
type NixAttrs,
|
||||
type NixBool,
|
||||
type NixFloat,
|
||||
@@ -40,7 +41,7 @@ export const isList = (e: NixValue): e is NixList => Array.isArray(force(e));
|
||||
|
||||
export const isNull = (e: NixValue): e is NixNull => force(e) === null;
|
||||
|
||||
export const isPath = (e: NixValue): boolean => {
|
||||
export const isPath = (e: NixValue): e is NixPath => {
|
||||
const val = force(e);
|
||||
return isNixPath(val);
|
||||
};
|
||||
|
||||
@@ -69,14 +69,16 @@ export const concatStringsWithContext = (parts: NixValue[]): NixString | NixPath
|
||||
}
|
||||
|
||||
// String concatenation mode
|
||||
// Note: firstIsPath is already false at this point, otherwise we would have
|
||||
// returned in the path concatenation branch above
|
||||
const context: NixStringContext = new Set();
|
||||
const strParts: string[] = [];
|
||||
|
||||
for (const part of parts) {
|
||||
// Handle path coercion to string
|
||||
const forced = force(part);
|
||||
if (isNixPath(forced)) {
|
||||
strParts.push(forced.value);
|
||||
const str = coerceToString(forced, StringCoercionMode.Interpolation, true, context);
|
||||
strParts.push(str);
|
||||
} else {
|
||||
const str = coerceToString(forced, StringCoercionMode.Interpolation, false, context);
|
||||
strParts.push(str);
|
||||
|
||||
1
nix-js/runtime-ts/src/types/global.d.ts
vendored
1
nix-js/runtime-ts/src/types/global.d.ts
vendored
@@ -78,6 +78,7 @@ declare global {
|
||||
): string;
|
||||
function op_store_path(path: string): string;
|
||||
function op_to_file(name: string, contents: string, references: string[]): string;
|
||||
function op_copy_path_to_store(path: string): string;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,14 +1,9 @@
|
||||
use anyhow::Result;
|
||||
use nix_js::context::Context;
|
||||
use std::process::exit;
|
||||
use tracing_subscriber::EnvFilter;
|
||||
|
||||
fn main() -> Result<()> {
|
||||
let format = tracing_subscriber::fmt::format().without_time();
|
||||
tracing_subscriber::fmt()
|
||||
.with_env_filter(EnvFilter::from_default_env())
|
||||
.event_format(format)
|
||||
.init();
|
||||
nix_js::logging::init_logging();
|
||||
|
||||
let mut args = std::env::args();
|
||||
if args.len() != 2 {
|
||||
|
||||
@@ -3,14 +3,9 @@ use nix_js::context::Context;
|
||||
use regex::Regex;
|
||||
use rustyline::DefaultEditor;
|
||||
use rustyline::error::ReadlineError;
|
||||
use tracing_subscriber::EnvFilter;
|
||||
|
||||
fn main() -> Result<()> {
|
||||
let format = tracing_subscriber::fmt::format().without_time();
|
||||
tracing_subscriber::fmt()
|
||||
.with_env_filter(EnvFilter::from_default_env())
|
||||
.event_format(format)
|
||||
.init();
|
||||
nix_js::logging::init_logging();
|
||||
|
||||
let mut rl = DefaultEditor::new()?;
|
||||
let mut context = Context::new()?;
|
||||
|
||||
@@ -7,12 +7,12 @@ use crate::ir::*;
|
||||
pub(crate) fn compile(expr: &Ir, ctx: &impl CodegenContext) -> String {
|
||||
let code = expr.compile(ctx);
|
||||
let debug_prefix = if std::env::var("NIX_JS_DEBUG_THUNKS").is_ok() {
|
||||
"Nix.DEBUG_THUNKS.enabled=true,"
|
||||
"Nix.DEBUG_THUNKS.enabled=true;"
|
||||
} else {
|
||||
""
|
||||
};
|
||||
let cur_dir = ctx.get_current_dir().display().to_string().escape_quote();
|
||||
format!("({}currentDir={},{})", debug_prefix, cur_dir, code)
|
||||
format!("(()=>{{{}const currentDir={};return {}}})()", debug_prefix, cur_dir, code)
|
||||
}
|
||||
|
||||
trait Compile<Ctx: CodegenContext> {
|
||||
|
||||
@@ -71,6 +71,7 @@ impl Context {
|
||||
}
|
||||
|
||||
pub fn eval_code(&mut self, expr: &str) -> Result<Value> {
|
||||
tracing::info!("Starting evaluation");
|
||||
self.ctx.current_file = Some(
|
||||
std::env::current_dir()
|
||||
.map_err(|err| {
|
||||
@@ -78,6 +79,8 @@ impl Context {
|
||||
})?
|
||||
.join("__eval__.nix"),
|
||||
);
|
||||
|
||||
tracing::debug!("Compiling code");
|
||||
let code = self.compile_code(expr)?;
|
||||
|
||||
self.runtime
|
||||
@@ -85,6 +88,7 @@ impl Context {
|
||||
.borrow_mut()
|
||||
.put(self.store.clone());
|
||||
|
||||
tracing::debug!("Executing JavaScript");
|
||||
self.runtime
|
||||
.eval(format!("Nix.force({code})"), CtxPtr::new(&mut self.ctx))
|
||||
}
|
||||
@@ -192,18 +196,20 @@ impl Ctx {
|
||||
}
|
||||
|
||||
fn compile_code(&mut self, expr: &str) -> Result<String> {
|
||||
tracing::debug!("Parsing Nix expression");
|
||||
let root = rnix::Root::parse(expr);
|
||||
if !root.errors().is_empty() {
|
||||
return Err(Error::parse_error(root.errors().iter().join("; ")));
|
||||
}
|
||||
|
||||
#[allow(clippy::unwrap_used)]
|
||||
// Always `Some` since there is no parse error
|
||||
let root = self
|
||||
.downgrade_ctx()
|
||||
.downgrade(root.tree().expr().unwrap())?;
|
||||
|
||||
tracing::debug!("Generating JavaScript code");
|
||||
let code = compile(self.get_ir(root), self);
|
||||
tracing::debug!("generated code: {}", &code);
|
||||
tracing::debug!("Generated code: {}", &code);
|
||||
Ok(code)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,15 +1,18 @@
|
||||
use deno_core::op2;
|
||||
use deno_core::OpState;
|
||||
use serde::Serialize;
|
||||
use tracing::debug;
|
||||
use tracing::{debug, info, warn};
|
||||
|
||||
mod archive;
|
||||
pub(crate) mod cache;
|
||||
mod download;
|
||||
mod git;
|
||||
mod hg;
|
||||
mod metadata_cache;
|
||||
|
||||
pub use cache::FetcherCache;
|
||||
pub use download::Downloader;
|
||||
pub use metadata_cache::MetadataCache;
|
||||
|
||||
use crate::runtime::NixError;
|
||||
use crate::nar;
|
||||
@@ -51,31 +54,69 @@ pub struct FetchHgResult {
|
||||
#[op2]
|
||||
#[serde]
|
||||
pub fn op_fetch_url(
|
||||
state: &mut OpState,
|
||||
#[string] url: String,
|
||||
#[string] expected_hash: Option<String>,
|
||||
#[string] name: Option<String>,
|
||||
executable: bool,
|
||||
) -> Result<FetchUrlResult, NixError> {
|
||||
debug!("fetchurl: {}", url);
|
||||
let cache = FetcherCache::new().map_err(|e| NixError::from(e.to_string()))?;
|
||||
let downloader = Downloader::new();
|
||||
use crate::store::StoreBackend;
|
||||
use std::sync::Arc;
|
||||
|
||||
let _span = tracing::info_span!("op_fetch_url", url = %url).entered();
|
||||
info!("fetchurl started");
|
||||
|
||||
let file_name =
|
||||
name.unwrap_or_else(|| url.rsplit('/').next().unwrap_or("download").to_string());
|
||||
|
||||
if let Some(ref hash) = expected_hash
|
||||
&& let Some(cached) = cache.get_url(&url, hash)
|
||||
let metadata_cache =
|
||||
MetadataCache::new(3600).map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
let input = serde_json::json!({
|
||||
"type": "file",
|
||||
"url": url,
|
||||
"name": file_name,
|
||||
"executable": executable,
|
||||
});
|
||||
|
||||
if let Some(cached_entry) = metadata_cache
|
||||
.lookup(&input)
|
||||
.map_err(|e| NixError::from(e.to_string()))?
|
||||
{
|
||||
return Ok(FetchUrlResult {
|
||||
store_path: cached.to_string_lossy().to_string(),
|
||||
hash: hash.clone(),
|
||||
});
|
||||
let cached_hash = cached_entry
|
||||
.info
|
||||
.get("hash")
|
||||
.and_then(|v| v.as_str())
|
||||
.unwrap_or("");
|
||||
|
||||
if let Some(ref expected) = expected_hash {
|
||||
let normalized_expected = normalize_hash(expected);
|
||||
if cached_hash != normalized_expected {
|
||||
warn!("Cached hash mismatch, re-fetching");
|
||||
} else {
|
||||
info!("Cache hit");
|
||||
return Ok(FetchUrlResult {
|
||||
store_path: cached_entry.store_path.clone(),
|
||||
hash: cached_hash.to_string(),
|
||||
});
|
||||
}
|
||||
} else {
|
||||
info!("Cache hit (no hash check)");
|
||||
return Ok(FetchUrlResult {
|
||||
store_path: cached_entry.store_path.clone(),
|
||||
hash: cached_hash.to_string(),
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
info!("Cache miss, downloading");
|
||||
let downloader = Downloader::new();
|
||||
let data = downloader
|
||||
.download(&url)
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
info!(bytes = data.len(), "Download complete");
|
||||
|
||||
let hash = crate::nix_hash::sha256_hex(&String::from_utf8_lossy(&data));
|
||||
|
||||
if let Some(ref expected) = expected_hash {
|
||||
@@ -88,12 +129,35 @@ pub fn op_fetch_url(
|
||||
}
|
||||
}
|
||||
|
||||
let store_path = cache
|
||||
.put_url(&url, &hash, &data, &file_name, executable)
|
||||
let store = state.borrow::<Arc<StoreBackend>>();
|
||||
let store_path = store
|
||||
.as_store()
|
||||
.add_to_store(&file_name, &data, false, vec![])
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
info!(store_path = %store_path, "Added to store");
|
||||
|
||||
#[cfg(unix)]
|
||||
if executable {
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
if let Ok(metadata) = std::fs::metadata(&store_path) {
|
||||
let mut perms = metadata.permissions();
|
||||
perms.set_mode(0o755);
|
||||
let _ = std::fs::set_permissions(&store_path, perms);
|
||||
}
|
||||
}
|
||||
|
||||
let info = serde_json::json!({
|
||||
"hash": hash,
|
||||
"url": url,
|
||||
});
|
||||
|
||||
metadata_cache
|
||||
.add(&input, &info, &store_path, true)
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
Ok(FetchUrlResult {
|
||||
store_path: store_path.to_string_lossy().to_string(),
|
||||
store_path,
|
||||
hash,
|
||||
})
|
||||
}
|
||||
@@ -101,63 +165,73 @@ pub fn op_fetch_url(
|
||||
#[op2]
|
||||
#[serde]
|
||||
pub fn op_fetch_tarball(
|
||||
state: &mut OpState,
|
||||
#[string] url: String,
|
||||
#[string] expected_hash: Option<String>,
|
||||
#[string] expected_nar_hash: Option<String>,
|
||||
#[string] name: Option<String>,
|
||||
) -> Result<FetchTarballResult, NixError> {
|
||||
debug!(
|
||||
"fetchTarball: url={}, expected_hash={:?}, expected_nar_hash={:?}",
|
||||
url, expected_hash, expected_nar_hash
|
||||
);
|
||||
let cache = FetcherCache::new().map_err(|e| NixError::from(e.to_string()))?;
|
||||
let downloader = Downloader::new();
|
||||
use crate::store::StoreBackend;
|
||||
use std::sync::Arc;
|
||||
|
||||
let _span = tracing::info_span!("op_fetch_tarball", url = %url).entered();
|
||||
info!("fetchTarball started");
|
||||
|
||||
let dir_name = name.unwrap_or_else(|| "source".to_string());
|
||||
let metadata_cache =
|
||||
MetadataCache::new(3600).map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
// Try cache lookup with narHash if provided
|
||||
if let Some(ref nar_hash) = expected_nar_hash {
|
||||
let normalized = normalize_hash(nar_hash);
|
||||
debug!("fetchTarball: normalized nar_hash={}", normalized);
|
||||
if let Some(cached) = cache.get_extracted_tarball(&url, &normalized) {
|
||||
debug!("fetchTarball: cache HIT (with expected nar_hash)");
|
||||
// Need to compute tarball hash if not cached
|
||||
let tarball_hash = expected_hash
|
||||
.as_ref()
|
||||
.map(|h| normalize_hash(h))
|
||||
.unwrap_or_default();
|
||||
let input = serde_json::json!({
|
||||
"type": "tarball",
|
||||
"url": url,
|
||||
"name": dir_name,
|
||||
});
|
||||
|
||||
if let Some(cached_entry) = metadata_cache
|
||||
.lookup(&input)
|
||||
.map_err(|e| NixError::from(e.to_string()))?
|
||||
{
|
||||
let cached_nar_hash = cached_entry
|
||||
.info
|
||||
.get("nar_hash")
|
||||
.and_then(|v| v.as_str())
|
||||
.unwrap_or("");
|
||||
let cached_tarball_hash = cached_entry
|
||||
.info
|
||||
.get("tarball_hash")
|
||||
.and_then(|v| v.as_str())
|
||||
.unwrap_or("");
|
||||
|
||||
if let Some(ref expected_nar) = expected_nar_hash {
|
||||
let normalized_expected = normalize_hash(expected_nar);
|
||||
if cached_nar_hash == normalized_expected {
|
||||
info!("Cache hit");
|
||||
return Ok(FetchTarballResult {
|
||||
store_path: cached_entry.store_path.clone(),
|
||||
hash: cached_tarball_hash.to_string(),
|
||||
nar_hash: cached_nar_hash.to_string(),
|
||||
});
|
||||
}
|
||||
} else if !cached_entry.is_expired(3600) {
|
||||
info!("Cache hit (no hash check)");
|
||||
return Ok(FetchTarballResult {
|
||||
store_path: cached.to_string_lossy().to_string(),
|
||||
hash: tarball_hash,
|
||||
nar_hash: normalized,
|
||||
store_path: cached_entry.store_path.clone(),
|
||||
hash: cached_tarball_hash.to_string(),
|
||||
nar_hash: cached_nar_hash.to_string(),
|
||||
});
|
||||
}
|
||||
debug!("fetchTarball: cache MISS, downloading...");
|
||||
} else if let Some((cached, cached_nar_hash)) = cache.get_extracted_tarball_by_url(&url) {
|
||||
debug!(
|
||||
"fetchTarball: cache HIT (by URL, nar_hash={})",
|
||||
cached_nar_hash
|
||||
);
|
||||
let tarball_hash = expected_hash
|
||||
.as_ref()
|
||||
.map(|h| normalize_hash(h))
|
||||
.unwrap_or_default();
|
||||
return Ok(FetchTarballResult {
|
||||
store_path: cached.to_string_lossy().to_string(),
|
||||
hash: tarball_hash,
|
||||
nar_hash: cached_nar_hash,
|
||||
});
|
||||
}
|
||||
debug!("fetchTarball: cache MISS, downloading...");
|
||||
|
||||
info!("Cache miss, downloading tarball");
|
||||
let downloader = Downloader::new();
|
||||
let data = downloader
|
||||
.download(&url)
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
// Compute tarball hash (hash of the archive file itself)
|
||||
info!(bytes = data.len(), "Download complete");
|
||||
|
||||
let tarball_hash = crate::nix_hash::sha256_hex(&String::from_utf8_lossy(&data));
|
||||
|
||||
// Verify tarball hash if provided
|
||||
if let Some(ref expected) = expected_hash {
|
||||
let normalized_expected = normalize_hash(expected);
|
||||
if tarball_hash != normalized_expected {
|
||||
@@ -168,23 +242,24 @@ pub fn op_fetch_tarball(
|
||||
}
|
||||
}
|
||||
|
||||
let temp_dir = tempfile::tempdir().map_err(|e| NixError::from(e.to_string()))?;
|
||||
let extracted_path = archive::extract_archive(&data, temp_dir.path())
|
||||
info!("Extracting tarball");
|
||||
let cache = FetcherCache::new().map_err(|e| NixError::from(e.to_string()))?;
|
||||
let (extracted_path, _temp_dir) = cache
|
||||
.extract_tarball_to_temp(&data)
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
// Compute NAR hash (hash of the extracted content)
|
||||
info!("Computing NAR hash");
|
||||
let nar_hash =
|
||||
nar::compute_nar_hash(&extracted_path).map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
debug!(
|
||||
"fetchTarball: computed tarball_hash={}, nar_hash={}",
|
||||
tarball_hash, nar_hash
|
||||
tarball_hash = %tarball_hash,
|
||||
nar_hash = %nar_hash,
|
||||
"Hash computation complete"
|
||||
);
|
||||
|
||||
// Verify NAR hash if provided
|
||||
if let Some(ref expected) = expected_nar_hash {
|
||||
let normalized_expected = normalize_hash(expected);
|
||||
|
||||
if nar_hash != normalized_expected {
|
||||
return Err(NixError::from(format!(
|
||||
"NAR hash mismatch for '{}': expected {}, got {}",
|
||||
@@ -193,12 +268,28 @@ pub fn op_fetch_tarball(
|
||||
}
|
||||
}
|
||||
|
||||
let store_path = cache
|
||||
.put_tarball_from_extracted(&url, &nar_hash, &extracted_path, &dir_name)
|
||||
info!("Adding to store");
|
||||
let store = state.borrow::<Arc<StoreBackend>>();
|
||||
let store_path = store
|
||||
.as_store()
|
||||
.add_to_store_from_path(&dir_name, &extracted_path, vec![])
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
info!(store_path = %store_path, "Added to store");
|
||||
|
||||
let info = serde_json::json!({
|
||||
"tarball_hash": tarball_hash,
|
||||
"nar_hash": nar_hash,
|
||||
"url": url,
|
||||
});
|
||||
|
||||
let immutable = expected_nar_hash.is_some();
|
||||
metadata_cache
|
||||
.add(&input, &info, &store_path, immutable)
|
||||
.map_err(|e| NixError::from(e.to_string()))?;
|
||||
|
||||
Ok(FetchTarballResult {
|
||||
store_path: store_path.to_string_lossy().to_string(),
|
||||
store_path,
|
||||
hash: tarball_hash,
|
||||
nar_hash,
|
||||
})
|
||||
@@ -207,6 +298,7 @@ pub fn op_fetch_tarball(
|
||||
#[op2]
|
||||
#[serde]
|
||||
pub fn op_fetch_git(
|
||||
state: &mut OpState,
|
||||
#[string] url: String,
|
||||
#[string] git_ref: Option<String>,
|
||||
#[string] rev: Option<String>,
|
||||
@@ -215,12 +307,19 @@ pub fn op_fetch_git(
|
||||
all_refs: bool,
|
||||
#[string] name: Option<String>,
|
||||
) -> Result<FetchGitResult, NixError> {
|
||||
debug!("fetchGit: {} (ref: {:?}, rev: {:?})", url, git_ref, rev);
|
||||
use crate::store::StoreBackend;
|
||||
use std::sync::Arc;
|
||||
|
||||
let _span = tracing::info_span!("op_fetch_git", url = %url).entered();
|
||||
info!("fetchGit started");
|
||||
let cache = FetcherCache::new().map_err(|e| NixError::from(e.to_string()))?;
|
||||
let dir_name = name.unwrap_or_else(|| "source".to_string());
|
||||
|
||||
let store = state.borrow::<Arc<StoreBackend>>();
|
||||
|
||||
git::fetch_git(
|
||||
&cache,
|
||||
store.as_store(),
|
||||
&url,
|
||||
git_ref.as_deref(),
|
||||
rev.as_deref(),
|
||||
|
||||
@@ -1,17 +1,12 @@
|
||||
use std::fs::{self, File};
|
||||
use std::io::Write;
|
||||
use std::fs;
|
||||
use std::path::{Path, PathBuf};
|
||||
|
||||
use serde::{Deserialize, Serialize};
|
||||
use tracing::debug;
|
||||
|
||||
use super::archive::ArchiveError;
|
||||
|
||||
#[derive(Debug)]
|
||||
pub enum CacheError {
|
||||
Io(std::io::Error),
|
||||
Archive(ArchiveError),
|
||||
Json(serde_json::Error),
|
||||
}
|
||||
|
||||
impl std::fmt::Display for CacheError {
|
||||
@@ -19,7 +14,6 @@ impl std::fmt::Display for CacheError {
|
||||
match self {
|
||||
CacheError::Io(e) => write!(f, "I/O error: {}", e),
|
||||
CacheError::Archive(e) => write!(f, "Archive error: {}", e),
|
||||
CacheError::Json(e) => write!(f, "JSON error: {}", e),
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -38,24 +32,11 @@ impl From<ArchiveError> for CacheError {
|
||||
}
|
||||
}
|
||||
|
||||
impl From<serde_json::Error> for CacheError {
|
||||
fn from(e: serde_json::Error) -> Self {
|
||||
CacheError::Json(e)
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug)]
|
||||
pub struct FetcherCache {
|
||||
base_dir: PathBuf,
|
||||
}
|
||||
|
||||
#[derive(Serialize, Deserialize)]
|
||||
struct CacheMetadata {
|
||||
url: String,
|
||||
hash: String,
|
||||
name: String,
|
||||
}
|
||||
|
||||
impl FetcherCache {
|
||||
pub fn new() -> Result<Self, std::io::Error> {
|
||||
let base_dir = dirs::cache_dir()
|
||||
@@ -68,16 +49,15 @@ impl FetcherCache {
|
||||
Ok(Self { base_dir })
|
||||
}
|
||||
|
||||
fn url_cache_dir(&self) -> PathBuf {
|
||||
self.base_dir.join("url")
|
||||
}
|
||||
|
||||
fn tarball_cache_dir(&self) -> PathBuf {
|
||||
self.base_dir.join("tarball")
|
||||
pub fn make_store_path(&self, hash: &str, name: &str) -> PathBuf {
|
||||
let short_hash = &hash[..32.min(hash.len())];
|
||||
self.base_dir
|
||||
.join("store")
|
||||
.join(format!("{}-{}", short_hash, name))
|
||||
}
|
||||
|
||||
fn git_cache_dir(&self) -> PathBuf {
|
||||
self.base_dir.join("git")
|
||||
self.base_dir.join("gitv3")
|
||||
}
|
||||
|
||||
fn hg_cache_dir(&self) -> PathBuf {
|
||||
@@ -88,277 +68,6 @@ impl FetcherCache {
|
||||
crate::nix_hash::sha256_hex(url)
|
||||
}
|
||||
|
||||
pub fn get_url(&self, url: &str, expected_hash: &str) -> Option<PathBuf> {
|
||||
let cache_dir = self.url_cache_dir();
|
||||
let key = Self::hash_key(url);
|
||||
let meta_path = cache_dir.join(format!("{}.meta", key));
|
||||
let data_path = cache_dir.join(format!("{}.data", key));
|
||||
|
||||
if !meta_path.exists() || !data_path.exists() {
|
||||
return None;
|
||||
}
|
||||
|
||||
let meta: CacheMetadata =
|
||||
serde_json::from_str(&fs::read_to_string(&meta_path).ok()?).ok()?;
|
||||
|
||||
if meta.hash == expected_hash {
|
||||
let store_path = self.make_store_path(&meta.hash, &meta.name);
|
||||
if store_path.exists() {
|
||||
Some(store_path)
|
||||
} else {
|
||||
None
|
||||
}
|
||||
} else {
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
pub fn put_url(
|
||||
&self,
|
||||
url: &str,
|
||||
hash: &str,
|
||||
data: &[u8],
|
||||
name: &str,
|
||||
executable: bool,
|
||||
) -> Result<PathBuf, std::io::Error> {
|
||||
let cache_dir = self.url_cache_dir();
|
||||
fs::create_dir_all(&cache_dir)?;
|
||||
|
||||
let key = Self::hash_key(url);
|
||||
let meta_path = cache_dir.join(format!("{}.meta", key));
|
||||
let data_path = cache_dir.join(format!("{}.data", key));
|
||||
|
||||
let mut file = File::create(&data_path)?;
|
||||
file.write_all(data)?;
|
||||
|
||||
#[cfg(unix)]
|
||||
if executable {
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
let mut perms = fs::metadata(&data_path)?.permissions();
|
||||
perms.set_mode(0o755);
|
||||
fs::set_permissions(&data_path, perms)?;
|
||||
}
|
||||
|
||||
let meta = CacheMetadata {
|
||||
url: url.to_string(),
|
||||
hash: hash.to_string(),
|
||||
name: name.to_string(),
|
||||
};
|
||||
fs::write(&meta_path, serde_json::to_string(&meta)?)?;
|
||||
|
||||
let store_path = self.make_store_path(hash, name);
|
||||
if !store_path.exists() {
|
||||
fs::create_dir_all(store_path.parent().unwrap_or(&store_path))?;
|
||||
fs::copy(&data_path, &store_path)?;
|
||||
|
||||
#[cfg(unix)]
|
||||
if executable {
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
let mut perms = fs::metadata(&store_path)?.permissions();
|
||||
perms.set_mode(0o755);
|
||||
fs::set_permissions(&store_path, perms)?;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(store_path)
|
||||
}
|
||||
|
||||
pub fn get_tarball(&self, url: &str, expected_hash: &str) -> Option<PathBuf> {
|
||||
let cache_dir = self.tarball_cache_dir();
|
||||
let key = Self::hash_key(url);
|
||||
let meta_path = cache_dir.join(&key).join(".meta");
|
||||
let data_dir = cache_dir.join(&key);
|
||||
|
||||
debug!("get_tarball: url={}, expected_hash={}", url, expected_hash);
|
||||
|
||||
if !meta_path.exists() || !data_dir.exists() {
|
||||
debug!("get_tarball: cache miss - meta or data dir not found");
|
||||
return None;
|
||||
}
|
||||
|
||||
let meta: CacheMetadata =
|
||||
serde_json::from_str(&fs::read_to_string(&meta_path).ok()?).ok()?;
|
||||
|
||||
debug!("get_tarball: cached hash={}, name={}", meta.hash, meta.name);
|
||||
|
||||
if meta.hash == expected_hash {
|
||||
let store_path = self.make_store_path(&meta.hash, &meta.name);
|
||||
debug!(
|
||||
"get_tarball: hash match, checking store_path={}",
|
||||
store_path.display()
|
||||
);
|
||||
if store_path.exists() {
|
||||
debug!("get_tarball: HIT - returning store path");
|
||||
Some(store_path)
|
||||
} else {
|
||||
debug!("get_tarball: store path doesn't exist");
|
||||
None
|
||||
}
|
||||
} else {
|
||||
debug!(
|
||||
"get_tarball: hash mismatch (cached={}, expected={})",
|
||||
meta.hash, expected_hash
|
||||
);
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
pub fn put_tarball(
|
||||
&self,
|
||||
url: &str,
|
||||
hash: &str,
|
||||
data: &[u8],
|
||||
name: &str,
|
||||
) -> Result<PathBuf, CacheError> {
|
||||
let cache_dir = self.tarball_cache_dir();
|
||||
let key = Self::hash_key(url);
|
||||
let extract_dir = cache_dir.join(&key);
|
||||
|
||||
fs::create_dir_all(&extract_dir)?;
|
||||
|
||||
let extracted_path = super::archive::extract_archive(data, &extract_dir)?;
|
||||
|
||||
let meta = CacheMetadata {
|
||||
url: url.to_string(),
|
||||
hash: hash.to_string(),
|
||||
name: name.to_string(),
|
||||
};
|
||||
fs::write(extract_dir.join(".meta"), serde_json::to_string(&meta)?)?;
|
||||
|
||||
let store_path = self.make_store_path(hash, name);
|
||||
if !store_path.exists() {
|
||||
fs::create_dir_all(store_path.parent().unwrap_or(&store_path))?;
|
||||
copy_dir_recursive(&extracted_path, &store_path)?;
|
||||
}
|
||||
|
||||
Ok(store_path)
|
||||
}
|
||||
|
||||
pub fn get_extracted_tarball(&self, url: &str, expected_nar_hash: &str) -> Option<PathBuf> {
|
||||
let cache_dir = self.tarball_cache_dir();
|
||||
let key = Self::hash_key(url);
|
||||
let cache_entry_dir = cache_dir.join(&key);
|
||||
let meta_path = cache_entry_dir.join(".meta");
|
||||
let cached_content = cache_entry_dir.join("content");
|
||||
|
||||
debug!(
|
||||
"get_extracted_tarball: url={}, expected_nar_hash={}",
|
||||
url, expected_nar_hash
|
||||
);
|
||||
|
||||
if !meta_path.exists() || !cached_content.exists() {
|
||||
debug!("get_extracted_tarball: cache miss - meta or content dir not found");
|
||||
return None;
|
||||
}
|
||||
|
||||
let meta: CacheMetadata =
|
||||
serde_json::from_str(&fs::read_to_string(&meta_path).ok()?).ok()?;
|
||||
|
||||
debug!(
|
||||
"get_extracted_tarball: cached hash={}, name={}",
|
||||
meta.hash, meta.name
|
||||
);
|
||||
|
||||
if meta.hash == expected_nar_hash {
|
||||
let store_path = self.make_store_path(&meta.hash, &meta.name);
|
||||
debug!(
|
||||
"get_extracted_tarball: hash match, checking store_path={}",
|
||||
store_path.display()
|
||||
);
|
||||
if store_path.exists() {
|
||||
debug!("get_extracted_tarball: HIT - returning store path");
|
||||
Some(store_path)
|
||||
} else {
|
||||
debug!("get_extracted_tarball: store path doesn't exist");
|
||||
None
|
||||
}
|
||||
} else {
|
||||
debug!(
|
||||
"get_extracted_tarball: hash mismatch (cached={}, expected={})",
|
||||
meta.hash, expected_nar_hash
|
||||
);
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
pub fn get_extracted_tarball_by_url(&self, url: &str) -> Option<(PathBuf, String)> {
|
||||
let cache_dir = self.tarball_cache_dir();
|
||||
let key = Self::hash_key(url);
|
||||
let cache_entry_dir = cache_dir.join(&key);
|
||||
let meta_path = cache_entry_dir.join(".meta");
|
||||
let cached_content = cache_entry_dir.join("content");
|
||||
|
||||
debug!("get_extracted_tarball_by_url: url={}", url);
|
||||
|
||||
if !meta_path.exists() || !cached_content.exists() {
|
||||
debug!("get_extracted_tarball_by_url: cache miss - meta or content dir not found");
|
||||
return None;
|
||||
}
|
||||
|
||||
let meta: CacheMetadata =
|
||||
serde_json::from_str(&fs::read_to_string(&meta_path).ok()?).ok()?;
|
||||
|
||||
debug!(
|
||||
"get_extracted_tarball_by_url: cached hash={}, name={}",
|
||||
meta.hash, meta.name
|
||||
);
|
||||
|
||||
let store_path = self.make_store_path(&meta.hash, &meta.name);
|
||||
if store_path.exists() {
|
||||
debug!("get_extracted_tarball_by_url: HIT - returning store path and hash");
|
||||
Some((store_path, meta.hash))
|
||||
} else {
|
||||
debug!("get_extracted_tarball_by_url: store path doesn't exist");
|
||||
None
|
||||
}
|
||||
}
|
||||
|
||||
pub fn put_tarball_from_extracted(
|
||||
&self,
|
||||
url: &str,
|
||||
hash: &str,
|
||||
extracted_path: &Path,
|
||||
name: &str,
|
||||
) -> Result<PathBuf, CacheError> {
|
||||
let cache_dir = self.tarball_cache_dir();
|
||||
let key = Self::hash_key(url);
|
||||
let cache_entry_dir = cache_dir.join(&key);
|
||||
|
||||
debug!(
|
||||
"put_tarball_from_extracted: url={}, hash={}, name={}",
|
||||
url, hash, name
|
||||
);
|
||||
|
||||
fs::create_dir_all(&cache_entry_dir)?;
|
||||
|
||||
let cached_content = cache_entry_dir.join("content");
|
||||
if !cached_content.exists() {
|
||||
copy_dir_recursive(extracted_path, &cached_content)?;
|
||||
}
|
||||
|
||||
let meta = CacheMetadata {
|
||||
url: url.to_string(),
|
||||
hash: hash.to_string(),
|
||||
name: name.to_string(),
|
||||
};
|
||||
fs::write(cache_entry_dir.join(".meta"), serde_json::to_string(&meta)?)?;
|
||||
|
||||
let store_path = self.make_store_path(hash, name);
|
||||
debug!(
|
||||
"put_tarball_from_extracted: store_path={}",
|
||||
store_path.display()
|
||||
);
|
||||
if !store_path.exists() {
|
||||
fs::create_dir_all(store_path.parent().unwrap_or(&store_path))?;
|
||||
copy_dir_recursive(extracted_path, &store_path)?;
|
||||
debug!("put_tarball_from_extracted: copied to store");
|
||||
} else {
|
||||
debug!("put_tarball_from_extracted: store path already exists");
|
||||
}
|
||||
|
||||
Ok(store_path)
|
||||
}
|
||||
|
||||
pub fn get_git_bare(&self, url: &str) -> PathBuf {
|
||||
let key = Self::hash_key(url);
|
||||
self.git_cache_dir().join(key)
|
||||
@@ -369,15 +78,14 @@ impl FetcherCache {
|
||||
self.hg_cache_dir().join(key)
|
||||
}
|
||||
|
||||
pub fn make_store_path(&self, hash: &str, name: &str) -> PathBuf {
|
||||
let short_hash = &hash[..32.min(hash.len())];
|
||||
self.base_dir
|
||||
.join("store")
|
||||
.join(format!("{}-{}", short_hash, name))
|
||||
pub fn extract_tarball_to_temp(&self, data: &[u8]) -> Result<(PathBuf, tempfile::TempDir), CacheError> {
|
||||
let temp_dir = tempfile::tempdir()?;
|
||||
let extracted_path = super::archive::extract_archive(data, temp_dir.path())?;
|
||||
Ok((extracted_path, temp_dir))
|
||||
}
|
||||
}
|
||||
|
||||
fn copy_dir_recursive(src: &Path, dst: &Path) -> Result<(), std::io::Error> {
|
||||
pub fn copy_dir_recursive(src: &Path, dst: &Path) -> Result<(), std::io::Error> {
|
||||
fs::create_dir_all(dst)?;
|
||||
|
||||
for entry in fs::read_dir(src)? {
|
||||
|
||||
@@ -4,10 +4,12 @@ use std::process::Command;
|
||||
|
||||
use super::FetchGitResult;
|
||||
use super::cache::FetcherCache;
|
||||
use crate::store::Store;
|
||||
|
||||
#[allow(clippy::too_many_arguments)]
|
||||
pub fn fetch_git(
|
||||
cache: &FetcherCache,
|
||||
store: &dyn Store,
|
||||
url: &str,
|
||||
git_ref: Option<&str>,
|
||||
rev: Option<&str>,
|
||||
@@ -25,7 +27,15 @@ pub fn fetch_git(
|
||||
}
|
||||
|
||||
let target_rev = resolve_rev(&bare_repo, git_ref, rev)?;
|
||||
let checkout_dir = checkout_rev(&bare_repo, &target_rev, submodules, name, cache)?;
|
||||
|
||||
let temp_dir = tempfile::tempdir()?;
|
||||
let checkout_dir = checkout_rev_to_temp(&bare_repo, &target_rev, submodules, temp_dir.path())?;
|
||||
|
||||
let nar_hash = crate::nar::compute_nar_hash(&checkout_dir)
|
||||
.map_err(|e| GitError::NarHashError(e.to_string()))?;
|
||||
|
||||
let store_path = store.add_to_store_from_path(name, &checkout_dir, vec![])
|
||||
.map_err(|e| GitError::StoreError(e.to_string()))?;
|
||||
|
||||
let rev_count = get_rev_count(&bare_repo, &target_rev)?;
|
||||
let last_modified = get_last_modified(&bare_repo, &target_rev)?;
|
||||
@@ -38,14 +48,14 @@ pub fn fetch_git(
|
||||
};
|
||||
|
||||
Ok(FetchGitResult {
|
||||
out_path: checkout_dir.to_string_lossy().to_string(),
|
||||
out_path: store_path,
|
||||
rev: target_rev,
|
||||
short_rev,
|
||||
rev_count,
|
||||
last_modified,
|
||||
last_modified_date,
|
||||
submodules,
|
||||
nar_hash: None,
|
||||
nar_hash: Some(nar_hash),
|
||||
})
|
||||
}
|
||||
|
||||
@@ -127,20 +137,13 @@ fn resolve_rev(
|
||||
Ok(String::from_utf8_lossy(&output.stdout).trim().to_string())
|
||||
}
|
||||
|
||||
fn checkout_rev(
|
||||
fn checkout_rev_to_temp(
|
||||
bare_repo: &PathBuf,
|
||||
rev: &str,
|
||||
submodules: bool,
|
||||
name: &str,
|
||||
cache: &FetcherCache,
|
||||
temp_path: &std::path::Path,
|
||||
) -> Result<PathBuf, GitError> {
|
||||
let hash = crate::nix_hash::sha256_hex(&format!("{}:{}", bare_repo.display(), rev));
|
||||
let checkout_dir = cache.make_store_path(&hash, name);
|
||||
|
||||
if checkout_dir.exists() {
|
||||
return Ok(checkout_dir);
|
||||
}
|
||||
|
||||
let checkout_dir = temp_path.join("checkout");
|
||||
fs::create_dir_all(&checkout_dir)?;
|
||||
|
||||
let output = Command::new("git")
|
||||
@@ -283,6 +286,8 @@ impl<T> Pipe for T {}
|
||||
pub enum GitError {
|
||||
IoError(std::io::Error),
|
||||
CommandFailed { operation: String, message: String },
|
||||
NarHashError(String),
|
||||
StoreError(String),
|
||||
}
|
||||
|
||||
impl std::fmt::Display for GitError {
|
||||
@@ -292,6 +297,8 @@ impl std::fmt::Display for GitError {
|
||||
GitError::CommandFailed { operation, message } => {
|
||||
write!(f, "Git {} failed: {}", operation, message)
|
||||
}
|
||||
GitError::NarHashError(e) => write!(f, "NAR hash error: {}", e),
|
||||
GitError::StoreError(e) => write!(f, "Store error: {}", e),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
215
nix-js/src/fetcher/metadata_cache.rs
Normal file
215
nix-js/src/fetcher/metadata_cache.rs
Normal file
@@ -0,0 +1,215 @@
|
||||
use rusqlite::{params, Connection, OptionalExtension};
|
||||
use serde::{Deserialize, Serialize};
|
||||
use serde_json;
|
||||
use std::path::PathBuf;
|
||||
use std::time::{SystemTime, UNIX_EPOCH};
|
||||
|
||||
#[derive(Debug)]
|
||||
pub enum CacheError {
|
||||
Database(rusqlite::Error),
|
||||
Json(serde_json::Error),
|
||||
}
|
||||
|
||||
impl std::fmt::Display for CacheError {
|
||||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||||
match self {
|
||||
CacheError::Database(e) => write!(f, "Database error: {}", e),
|
||||
CacheError::Json(e) => write!(f, "JSON error: {}", e),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
impl std::error::Error for CacheError {}
|
||||
|
||||
impl From<rusqlite::Error> for CacheError {
|
||||
fn from(e: rusqlite::Error) -> Self {
|
||||
CacheError::Database(e)
|
||||
}
|
||||
}
|
||||
|
||||
impl From<serde_json::Error> for CacheError {
|
||||
fn from(e: serde_json::Error) -> Self {
|
||||
CacheError::Json(e)
|
||||
}
|
||||
}
|
||||
|
||||
#[derive(Debug, Clone, Serialize, Deserialize)]
|
||||
pub struct CacheEntry {
|
||||
pub input: serde_json::Value,
|
||||
pub info: serde_json::Value,
|
||||
pub store_path: String,
|
||||
pub immutable: bool,
|
||||
pub timestamp: u64,
|
||||
}
|
||||
|
||||
impl CacheEntry {
|
||||
pub fn is_expired(&self, ttl_seconds: u64) -> bool {
|
||||
if self.immutable {
|
||||
return false;
|
||||
}
|
||||
|
||||
if ttl_seconds == 0 {
|
||||
return false;
|
||||
}
|
||||
|
||||
let now = SystemTime::now()
|
||||
.duration_since(UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_secs();
|
||||
|
||||
now > self.timestamp + ttl_seconds
|
||||
}
|
||||
}
|
||||
|
||||
pub struct MetadataCache {
|
||||
conn: Connection,
|
||||
ttl_seconds: u64,
|
||||
}
|
||||
|
||||
impl MetadataCache {
|
||||
pub fn new(ttl_seconds: u64) -> Result<Self, CacheError> {
|
||||
let cache_dir = dirs::cache_dir()
|
||||
.unwrap_or_else(|| PathBuf::from("/tmp"))
|
||||
.join("nix-js");
|
||||
|
||||
std::fs::create_dir_all(&cache_dir)
|
||||
.map_err(|e| CacheError::Database(rusqlite::Error::ToSqlConversionFailure(Box::new(e))))?;
|
||||
|
||||
let db_path = cache_dir.join("fetcher-cache.sqlite");
|
||||
let conn = Connection::open(db_path)?;
|
||||
|
||||
conn.execute(
|
||||
"CREATE TABLE IF NOT EXISTS cache (
|
||||
input TEXT NOT NULL PRIMARY KEY,
|
||||
info TEXT NOT NULL,
|
||||
store_path TEXT NOT NULL,
|
||||
immutable INTEGER NOT NULL,
|
||||
timestamp INTEGER NOT NULL
|
||||
)",
|
||||
[],
|
||||
)?;
|
||||
|
||||
Ok(Self { conn, ttl_seconds })
|
||||
}
|
||||
|
||||
pub fn lookup(&self, input: &serde_json::Value) -> Result<Option<CacheEntry>, CacheError> {
|
||||
let input_str = serde_json::to_string(input)?;
|
||||
|
||||
let entry: Option<(String, String, String, i64, i64)> = self
|
||||
.conn
|
||||
.query_row(
|
||||
"SELECT input, info, store_path, immutable, timestamp FROM cache WHERE input = ?1",
|
||||
params![input_str],
|
||||
|row| {
|
||||
Ok((
|
||||
row.get(0)?,
|
||||
row.get(1)?,
|
||||
row.get(2)?,
|
||||
row.get(3)?,
|
||||
row.get(4)?,
|
||||
))
|
||||
},
|
||||
)
|
||||
.optional()?;
|
||||
|
||||
match entry {
|
||||
Some((input_json, info_json, store_path, immutable, timestamp)) => {
|
||||
let entry = CacheEntry {
|
||||
input: serde_json::from_str(&input_json)?,
|
||||
info: serde_json::from_str(&info_json)?,
|
||||
store_path,
|
||||
immutable: immutable != 0,
|
||||
timestamp: timestamp as u64,
|
||||
};
|
||||
|
||||
if entry.is_expired(self.ttl_seconds) {
|
||||
Ok(None)
|
||||
} else {
|
||||
Ok(Some(entry))
|
||||
}
|
||||
}
|
||||
None => Ok(None),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn lookup_expired(
|
||||
&self,
|
||||
input: &serde_json::Value,
|
||||
) -> Result<Option<CacheEntry>, CacheError> {
|
||||
let input_str = serde_json::to_string(input)?;
|
||||
|
||||
let entry: Option<(String, String, String, i64, i64)> = self
|
||||
.conn
|
||||
.query_row(
|
||||
"SELECT input, info, store_path, immutable, timestamp FROM cache WHERE input = ?1",
|
||||
params![input_str],
|
||||
|row| {
|
||||
Ok((
|
||||
row.get(0)?,
|
||||
row.get(1)?,
|
||||
row.get(2)?,
|
||||
row.get(3)?,
|
||||
row.get(4)?,
|
||||
))
|
||||
},
|
||||
)
|
||||
.optional()?;
|
||||
|
||||
match entry {
|
||||
Some((input_json, info_json, store_path, immutable, timestamp)) => Ok(Some(
|
||||
CacheEntry {
|
||||
input: serde_json::from_str(&input_json)?,
|
||||
info: serde_json::from_str(&info_json)?,
|
||||
store_path,
|
||||
immutable: immutable != 0,
|
||||
timestamp: timestamp as u64,
|
||||
},
|
||||
)),
|
||||
None => Ok(None),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn add(
|
||||
&self,
|
||||
input: &serde_json::Value,
|
||||
info: &serde_json::Value,
|
||||
store_path: &str,
|
||||
immutable: bool,
|
||||
) -> Result<(), CacheError> {
|
||||
let input_str = serde_json::to_string(input)?;
|
||||
let info_str = serde_json::to_string(info)?;
|
||||
let timestamp = SystemTime::now()
|
||||
.duration_since(UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_secs();
|
||||
|
||||
self.conn.execute(
|
||||
"INSERT OR REPLACE INTO cache (input, info, store_path, immutable, timestamp)
|
||||
VALUES (?1, ?2, ?3, ?4, ?5)",
|
||||
params![
|
||||
input_str,
|
||||
info_str,
|
||||
store_path,
|
||||
if immutable { 1 } else { 0 },
|
||||
timestamp as i64
|
||||
],
|
||||
)?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
||||
pub fn update_timestamp(&self, input: &serde_json::Value) -> Result<(), CacheError> {
|
||||
let input_str = serde_json::to_string(input)?;
|
||||
let timestamp = SystemTime::now()
|
||||
.duration_since(UNIX_EPOCH)
|
||||
.unwrap()
|
||||
.as_secs();
|
||||
|
||||
self.conn.execute(
|
||||
"UPDATE cache SET timestamp = ?1 WHERE input = ?2",
|
||||
params![timestamp as i64, input_str],
|
||||
)?;
|
||||
|
||||
Ok(())
|
||||
}
|
||||
}
|
||||
@@ -2,6 +2,7 @@
|
||||
|
||||
pub mod context;
|
||||
pub mod error;
|
||||
pub mod logging;
|
||||
pub mod value;
|
||||
|
||||
mod codegen;
|
||||
|
||||
63
nix-js/src/logging.rs
Normal file
63
nix-js/src/logging.rs
Normal file
@@ -0,0 +1,63 @@
|
||||
use std::env;
|
||||
use std::io::IsTerminal;
|
||||
use tracing_subscriber::{fmt, layer::SubscriberExt, util::SubscriberInitExt, EnvFilter, Layer};
|
||||
|
||||
pub fn init_logging() {
|
||||
let is_terminal = std::io::stderr().is_terminal();
|
||||
let show_time = env::var("NIX_JS_LOG_TIME")
|
||||
.map(|v| v == "1" || v.to_lowercase() == "true")
|
||||
.unwrap_or(false);
|
||||
|
||||
let filter = EnvFilter::from_default_env();
|
||||
let fmt_layer = fmt::layer()
|
||||
.with_target(true)
|
||||
.with_thread_ids(false)
|
||||
.with_thread_names(false)
|
||||
.with_file(false)
|
||||
.with_line_number(false)
|
||||
.with_ansi(is_terminal)
|
||||
.with_level(true);
|
||||
|
||||
let fmt_layer = if show_time {
|
||||
fmt_layer
|
||||
.with_timer(fmt::time::uptime())
|
||||
.boxed()
|
||||
} else {
|
||||
fmt_layer.without_time().boxed()
|
||||
};
|
||||
|
||||
tracing_subscriber::registry()
|
||||
.with(filter)
|
||||
.with(fmt_layer)
|
||||
.init();
|
||||
}
|
||||
|
||||
#[macro_export]
|
||||
macro_rules! trace_span {
|
||||
($name:expr) => {
|
||||
tracing::trace_span!($name)
|
||||
};
|
||||
($name:expr, $($field:tt)*) => {
|
||||
tracing::trace_span!($name, $($field)*)
|
||||
};
|
||||
}
|
||||
|
||||
#[macro_export]
|
||||
macro_rules! debug_span {
|
||||
($name:expr) => {
|
||||
tracing::debug_span!($name)
|
||||
};
|
||||
($name:expr, $($field:tt)*) => {
|
||||
tracing::debug_span!($name, $($field)*)
|
||||
};
|
||||
}
|
||||
|
||||
#[macro_export]
|
||||
macro_rules! info_span {
|
||||
($name:expr) => {
|
||||
tracing::info_span!($name)
|
||||
};
|
||||
($name:expr, $($field:tt)*) => {
|
||||
tracing::info_span!($name, $($field)*)
|
||||
};
|
||||
}
|
||||
@@ -34,6 +34,7 @@ fn runtime_extension<Ctx: RuntimeContext>() -> Extension {
|
||||
op_add_path(),
|
||||
op_store_path(),
|
||||
op_to_file(),
|
||||
op_copy_path_to_store(),
|
||||
];
|
||||
ops.extend(crate::fetcher::register_ops());
|
||||
|
||||
@@ -80,6 +81,7 @@ fn op_import<Ctx: RuntimeContext>(
|
||||
state: &mut OpState,
|
||||
#[string] path: String,
|
||||
) -> std::result::Result<String, NixError> {
|
||||
let _span = tracing::info_span!("op_import", path = %path).entered();
|
||||
let ctx = state.borrow_mut::<Ctx>();
|
||||
|
||||
let current_dir = ctx.get_current_dir();
|
||||
@@ -91,10 +93,12 @@ fn op_import<Ctx: RuntimeContext>(
|
||||
absolute_path.push("default.nix")
|
||||
}
|
||||
|
||||
tracing::info!("Importing file: {}", absolute_path.display());
|
||||
|
||||
let content = std::fs::read_to_string(&absolute_path)
|
||||
.map_err(|e| format!("Failed to read {}: {}", absolute_path.display(), e))?;
|
||||
|
||||
tracing::debug!("compiling file: {}", absolute_path.display());
|
||||
tracing::debug!("Compiling file");
|
||||
ctx.set_current_file(absolute_path);
|
||||
|
||||
Ok(ctx.compile_code(&content).map_err(|err| err.to_string())?)
|
||||
@@ -117,13 +121,24 @@ fn op_resolve_path(
|
||||
#[string] current_dir: String,
|
||||
#[string] path: String,
|
||||
) -> std::result::Result<String, NixError> {
|
||||
let _span = tracing::debug_span!("op_resolve_path").entered();
|
||||
tracing::debug!(current_dir, path);
|
||||
|
||||
// If already absolute, return as-is
|
||||
if path.starts_with('/') {
|
||||
return Ok(path);
|
||||
}
|
||||
|
||||
// Resolve relative path against current file directory (or CWD)
|
||||
let current_dir = PathBuf::from(current_dir).join(&path);
|
||||
let current_dir = if !path.starts_with("~/") {
|
||||
let mut dir = PathBuf::from(current_dir);
|
||||
dir.push(path);
|
||||
dir
|
||||
} else {
|
||||
let mut dir = std::env::home_dir()
|
||||
.ok_or("home dir not defined")?;
|
||||
dir.push(&path[2..]);
|
||||
dir
|
||||
};
|
||||
let mut normalized = PathBuf::new();
|
||||
for component in current_dir.components() {
|
||||
match component {
|
||||
@@ -136,6 +151,7 @@ fn op_resolve_path(
|
||||
Component::Normal(c) => normalized.push(c),
|
||||
}
|
||||
}
|
||||
tracing::debug!(normalized = normalized.display().to_string());
|
||||
Ok(normalized.to_string_lossy().to_string())
|
||||
}
|
||||
|
||||
@@ -251,8 +267,11 @@ fn op_add_path(
|
||||
}
|
||||
|
||||
let store = state.borrow::<Arc<StoreBackend>>();
|
||||
let store_dir = store.as_store().get_store_dir();
|
||||
let store_path = crate::nix_hash::make_store_path(store_dir, "source", &computed_hash, &computed_name);
|
||||
|
||||
let store_path = store
|
||||
.as_store()
|
||||
.add_to_store_from_path(&computed_name, path_obj, vec![])
|
||||
.map_err(|e| NixError::from(format!("failed to add path to store: {}", e)))?;
|
||||
|
||||
Ok(store_path)
|
||||
}
|
||||
@@ -298,7 +317,7 @@ fn op_store_path(
|
||||
state: &mut OpState,
|
||||
#[string] path: String,
|
||||
) -> std::result::Result<String, NixError> {
|
||||
use crate::store::{validate_store_path, StoreBackend};
|
||||
use crate::store::{StoreBackend, validate_store_path};
|
||||
use std::sync::Arc;
|
||||
|
||||
let store = state.borrow::<Arc<StoreBackend>>();
|
||||
@@ -334,6 +353,37 @@ fn op_to_file(
|
||||
Ok(store_path)
|
||||
}
|
||||
|
||||
#[deno_core::op2]
|
||||
#[string]
|
||||
fn op_copy_path_to_store(
|
||||
state: &mut OpState,
|
||||
#[string] path: String,
|
||||
) -> std::result::Result<String, NixError> {
|
||||
use crate::store::StoreBackend;
|
||||
use std::path::Path;
|
||||
use std::sync::Arc;
|
||||
|
||||
let path_obj = Path::new(&path);
|
||||
|
||||
if !path_obj.exists() {
|
||||
return Err(NixError::from(format!("path '{}' does not exist", path)));
|
||||
}
|
||||
|
||||
let name = path_obj
|
||||
.file_name()
|
||||
.and_then(|n| n.to_str())
|
||||
.unwrap_or("source")
|
||||
.to_string();
|
||||
|
||||
let store = state.borrow::<Arc<StoreBackend>>();
|
||||
let store_path = store
|
||||
.as_store()
|
||||
.add_to_store_from_path(&name, path_obj, vec![])
|
||||
.map_err(|e| NixError::from(format!("failed to copy path to store: {}", e)))?;
|
||||
|
||||
Ok(store_path)
|
||||
}
|
||||
|
||||
pub(crate) struct Runtime<Ctx: RuntimeContext> {
|
||||
js_runtime: JsRuntime,
|
||||
is_thunk_symbol: v8::Global<v8::Symbol>,
|
||||
|
||||
@@ -2,8 +2,6 @@ mod config;
|
||||
mod error;
|
||||
mod validation;
|
||||
|
||||
use std::path::Path;
|
||||
|
||||
pub use config::{StoreConfig, StoreMode};
|
||||
pub use validation::validate_store_path;
|
||||
|
||||
@@ -24,6 +22,13 @@ pub trait Store: Send + Sync {
|
||||
references: Vec<String>,
|
||||
) -> Result<String>;
|
||||
|
||||
fn add_to_store_from_path(
|
||||
&self,
|
||||
name: &str,
|
||||
source_path: &std::path::Path,
|
||||
references: Vec<String>,
|
||||
) -> Result<String>;
|
||||
|
||||
fn add_text_to_store(
|
||||
&self,
|
||||
name: &str,
|
||||
|
||||
@@ -1,7 +1,13 @@
|
||||
use std::io::{Error as IoError, ErrorKind as IoErrorKind, Result as IoResult};
|
||||
use std::path::Path;
|
||||
use std::sync::Arc;
|
||||
|
||||
use nix_daemon::{Progress as _, Store as _, nix};
|
||||
use nix_compat::nix_daemon::types::{AddToStoreNarRequest, UnkeyedValidPathInfo};
|
||||
use nix_compat::nix_daemon::worker_protocol::{ClientSettings, Operation};
|
||||
use nix_compat::store_path::StorePath;
|
||||
use nix_compat::wire::ProtocolVersion;
|
||||
use nix_compat::wire::de::{NixRead, NixReader};
|
||||
use nix_compat::wire::ser::{NixSerialize, NixWrite, NixWriter, NixWriterBuilder};
|
||||
use tokio::io::{AsyncReadExt, AsyncWriteExt, ReadHalf, WriteHalf, split};
|
||||
use tokio::net::UnixStream;
|
||||
use tokio::sync::Mutex;
|
||||
|
||||
@@ -11,7 +17,7 @@ use super::Store;
|
||||
|
||||
pub struct DaemonStore {
|
||||
runtime: tokio::runtime::Runtime,
|
||||
store: Arc<Mutex<nix::DaemonStore<UnixStream>>>,
|
||||
connection: NixDaemonConnection,
|
||||
}
|
||||
|
||||
impl DaemonStore {
|
||||
@@ -19,25 +25,21 @@ impl DaemonStore {
|
||||
let runtime = tokio::runtime::Runtime::new()
|
||||
.map_err(|e| Error::internal(format!("Failed to create tokio runtime: {}", e)))?;
|
||||
|
||||
let socket_str = socket_path
|
||||
.to_str()
|
||||
.ok_or_else(|| Error::internal("Invalid socket path: not UTF-8".to_string()))?;
|
||||
|
||||
let store = runtime.block_on(async {
|
||||
nix_daemon::nix::DaemonStore::builder()
|
||||
.connect_unix(socket_str)
|
||||
let connection = runtime.block_on(async {
|
||||
NixDaemonConnection::connect(socket_path)
|
||||
.await
|
||||
.map_err(|e| {
|
||||
Error::internal(format!(
|
||||
"Failed to connect to nix-daemon at {}: {}",
|
||||
socket_str, e
|
||||
socket_path.display(),
|
||||
e
|
||||
))
|
||||
})
|
||||
})?;
|
||||
|
||||
Ok(Self {
|
||||
runtime,
|
||||
store: Arc::new(Mutex::new(store)),
|
||||
connection,
|
||||
})
|
||||
}
|
||||
|
||||
@@ -56,10 +58,8 @@ impl Store for DaemonStore {
|
||||
|
||||
fn is_valid_path(&self, path: &str) -> Result<bool> {
|
||||
self.block_on(async {
|
||||
let mut store = self.store.lock().await;
|
||||
store
|
||||
self.connection
|
||||
.is_valid_path(path)
|
||||
.result()
|
||||
.await
|
||||
.map_err(|e| Error::internal(format!("Daemon error in is_valid_path: {}", e)))
|
||||
})
|
||||
@@ -67,8 +67,7 @@ impl Store for DaemonStore {
|
||||
|
||||
fn ensure_path(&self, path: &str) -> Result<()> {
|
||||
self.block_on(async {
|
||||
let mut store = self.store.lock().await;
|
||||
store.ensure_path(path).result().await.map_err(|e| {
|
||||
self.connection.ensure_path(path).await.map_err(|e| {
|
||||
Error::eval_error(
|
||||
format!(
|
||||
"builtins.storePath: path '{}' is not valid in nix store: {}",
|
||||
@@ -87,34 +86,156 @@ impl Store for DaemonStore {
|
||||
recursive: bool,
|
||||
references: Vec<String>,
|
||||
) -> Result<String> {
|
||||
let temp_dir = tempfile::tempdir()
|
||||
.map_err(|e| Error::internal(format!("Failed to create temp dir: {}", e)))?;
|
||||
let content_path = temp_dir.path().join(name);
|
||||
std::fs::write(&content_path, content)
|
||||
.map_err(|e| Error::internal(format!("Failed to write content: {}", e)))?;
|
||||
use nix_compat::nix_daemon::types::AddToStoreNarRequest;
|
||||
use nix_compat::nixhash::{CAHash, NixHash};
|
||||
use nix_compat::store_path::{StorePath, build_ca_path};
|
||||
use sha2::{Digest, Sha256};
|
||||
use std::fs;
|
||||
use tempfile::NamedTempFile;
|
||||
|
||||
let cam_str = if recursive {
|
||||
"fixed:r:sha256"
|
||||
let temp_file = NamedTempFile::new()
|
||||
.map_err(|e| Error::internal(format!("Failed to create temp file: {}", e)))?;
|
||||
fs::write(temp_file.path(), content)
|
||||
.map_err(|e| Error::internal(format!("Failed to write temp file: {}", e)))?;
|
||||
|
||||
let nar_data = crate::nar::pack_nar(temp_file.path())?;
|
||||
|
||||
let nar_hash_hex = {
|
||||
let mut hasher = Sha256::new();
|
||||
hasher.update(&nar_data);
|
||||
hex::encode(hasher.finalize())
|
||||
};
|
||||
|
||||
let nar_hash_bytes = hex::decode(&nar_hash_hex)
|
||||
.map_err(|e| Error::internal(format!("Invalid nar hash: {}", e)))?;
|
||||
let mut nar_hash_arr = [0u8; 32];
|
||||
nar_hash_arr.copy_from_slice(&nar_hash_bytes);
|
||||
|
||||
let ca_hash = if recursive {
|
||||
CAHash::Nar(NixHash::Sha256(nar_hash_arr))
|
||||
} else {
|
||||
"fixed:sha256"
|
||||
let mut content_hasher = Sha256::new();
|
||||
content_hasher.update(content);
|
||||
let content_hash = content_hasher.finalize();
|
||||
let mut content_hash_arr = [0u8; 32];
|
||||
content_hash_arr.copy_from_slice(&content_hash);
|
||||
CAHash::Flat(NixHash::Sha256(content_hash_arr))
|
||||
};
|
||||
|
||||
let ref_store_paths: std::result::Result<Vec<StorePath<String>>, _> = references
|
||||
.iter()
|
||||
.map(|r| StorePath::<String>::from_absolute_path(r.as_bytes()))
|
||||
.collect();
|
||||
let ref_store_paths = ref_store_paths
|
||||
.map_err(|e| Error::internal(format!("Invalid reference path: {}", e)))?;
|
||||
|
||||
let store_path: StorePath<String> =
|
||||
build_ca_path(name, &ca_hash, references.clone(), false)
|
||||
.map_err(|e| Error::internal(format!("Failed to build store path: {}", e)))?;
|
||||
|
||||
let store_path_str = store_path.to_absolute_path();
|
||||
|
||||
if self.is_valid_path(&store_path_str)? {
|
||||
return Ok(store_path_str);
|
||||
}
|
||||
|
||||
let request = AddToStoreNarRequest {
|
||||
path: store_path,
|
||||
deriver: None,
|
||||
nar_hash: unsafe {
|
||||
std::mem::transmute::<[u8; 32], nix_compat::nix_daemon::types::NarHash>(
|
||||
nar_hash_arr,
|
||||
)
|
||||
},
|
||||
references: ref_store_paths,
|
||||
registration_time: 0,
|
||||
nar_size: nar_data.len() as u64,
|
||||
ultimate: false,
|
||||
signatures: vec![],
|
||||
ca: Some(ca_hash),
|
||||
repair: false,
|
||||
dont_check_sigs: false,
|
||||
};
|
||||
|
||||
self.block_on(async {
|
||||
let mut store = self.store.lock().await;
|
||||
let (store_path, _path_info) = store
|
||||
.add_to_store(
|
||||
name,
|
||||
cam_str,
|
||||
references,
|
||||
false,
|
||||
content_path.as_os_str().as_encoded_bytes(),
|
||||
)
|
||||
.result()
|
||||
self.connection
|
||||
.add_to_store_nar(request, &nar_data)
|
||||
.await
|
||||
.map_err(|e| Error::internal(format!("Daemon error in add_to_store: {}", e)))?;
|
||||
.map_err(|e| Error::internal(format!("Failed to add to store: {}", e)))
|
||||
})?;
|
||||
|
||||
Ok(store_path)
|
||||
})
|
||||
Ok(store_path_str)
|
||||
}
|
||||
|
||||
fn add_to_store_from_path(
|
||||
&self,
|
||||
name: &str,
|
||||
source_path: &std::path::Path,
|
||||
references: Vec<String>,
|
||||
) -> Result<String> {
|
||||
use nix_compat::nix_daemon::types::AddToStoreNarRequest;
|
||||
use nix_compat::nixhash::{CAHash, NixHash};
|
||||
use nix_compat::store_path::{StorePath, build_ca_path};
|
||||
use sha2::{Digest, Sha256};
|
||||
|
||||
let nar_data = crate::nar::pack_nar(source_path)?;
|
||||
|
||||
let nar_hash_hex = {
|
||||
let mut hasher = Sha256::new();
|
||||
hasher.update(&nar_data);
|
||||
hex::encode(hasher.finalize())
|
||||
};
|
||||
|
||||
let nar_hash_bytes = hex::decode(&nar_hash_hex)
|
||||
.map_err(|e| Error::internal(format!("Invalid nar hash: {}", e)))?;
|
||||
let mut nar_hash_arr = [0u8; 32];
|
||||
nar_hash_arr.copy_from_slice(&nar_hash_bytes);
|
||||
|
||||
let ca_hash = CAHash::Nar(NixHash::Sha256(nar_hash_arr));
|
||||
|
||||
let ref_store_paths: std::result::Result<Vec<StorePath<String>>, _> = references
|
||||
.iter()
|
||||
.map(|r| StorePath::<String>::from_absolute_path(r.as_bytes()))
|
||||
.collect();
|
||||
let ref_store_paths = ref_store_paths
|
||||
.map_err(|e| Error::internal(format!("Invalid reference path: {}", e)))?;
|
||||
|
||||
let store_path: StorePath<String> =
|
||||
build_ca_path(name, &ca_hash, references.clone(), false)
|
||||
.map_err(|e| Error::internal(format!("Failed to build store path: {}", e)))?;
|
||||
|
||||
let store_path_str = store_path.to_absolute_path();
|
||||
|
||||
if self.is_valid_path(&store_path_str)? {
|
||||
return Ok(store_path_str);
|
||||
}
|
||||
|
||||
let request = AddToStoreNarRequest {
|
||||
path: store_path,
|
||||
deriver: None,
|
||||
nar_hash: unsafe {
|
||||
std::mem::transmute::<[u8; 32], nix_compat::nix_daemon::types::NarHash>(
|
||||
nar_hash_arr,
|
||||
)
|
||||
},
|
||||
references: ref_store_paths,
|
||||
registration_time: 0,
|
||||
nar_size: nar_data.len() as u64,
|
||||
ultimate: false,
|
||||
signatures: vec![],
|
||||
ca: Some(ca_hash),
|
||||
repair: false,
|
||||
dont_check_sigs: false,
|
||||
};
|
||||
|
||||
self.block_on(async {
|
||||
self.connection
|
||||
.add_to_store_nar(request, &nar_data)
|
||||
.await
|
||||
.map_err(|e| Error::internal(format!("Failed to add to store: {}", e)))
|
||||
})?;
|
||||
|
||||
Ok(store_path_str)
|
||||
}
|
||||
|
||||
fn add_text_to_store(
|
||||
@@ -123,28 +244,559 @@ impl Store for DaemonStore {
|
||||
content: &str,
|
||||
references: Vec<String>,
|
||||
) -> Result<String> {
|
||||
self.block_on(async {
|
||||
let mut store = self.store.lock().await;
|
||||
let (store_path, _) = store
|
||||
.add_to_store(name, "text:sha256", references, false, content.as_bytes())
|
||||
.result()
|
||||
.await
|
||||
.map_err(|e| {
|
||||
Error::internal(format!("Daemon error in add_text_to_store: {}", e))
|
||||
})?;
|
||||
use nix_compat::nix_daemon::types::AddToStoreNarRequest;
|
||||
use nix_compat::nixhash::CAHash;
|
||||
use nix_compat::store_path::{StorePath, build_text_path};
|
||||
use sha2::{Digest, Sha256};
|
||||
use std::fs;
|
||||
use tempfile::NamedTempFile;
|
||||
|
||||
Ok(store_path)
|
||||
})
|
||||
let temp_file = NamedTempFile::new()
|
||||
.map_err(|e| Error::internal(format!("Failed to create temp file: {}", e)))?;
|
||||
fs::write(temp_file.path(), content.as_bytes())
|
||||
.map_err(|e| Error::internal(format!("Failed to write temp file: {}", e)))?;
|
||||
|
||||
let nar_data = crate::nar::pack_nar(temp_file.path())?;
|
||||
|
||||
let nar_hash_hex = {
|
||||
let mut hasher = Sha256::new();
|
||||
hasher.update(&nar_data);
|
||||
hex::encode(hasher.finalize())
|
||||
};
|
||||
|
||||
let nar_hash_bytes = hex::decode(&nar_hash_hex)
|
||||
.map_err(|e| Error::internal(format!("Invalid nar hash: {}", e)))?;
|
||||
let mut nar_hash_arr = [0u8; 32];
|
||||
nar_hash_arr.copy_from_slice(&nar_hash_bytes);
|
||||
|
||||
let content_hash = {
|
||||
let mut hasher = Sha256::new();
|
||||
hasher.update(content.as_bytes());
|
||||
hasher.finalize().into()
|
||||
};
|
||||
|
||||
let ref_store_paths: std::result::Result<Vec<StorePath<String>>, _> = references
|
||||
.iter()
|
||||
.map(|r| StorePath::<String>::from_absolute_path(r.as_bytes()))
|
||||
.collect();
|
||||
let ref_store_paths = ref_store_paths
|
||||
.map_err(|e| Error::internal(format!("Invalid reference path: {}", e)))?;
|
||||
|
||||
let store_path: StorePath<String> = build_text_path(name, content, references.clone())
|
||||
.map_err(|e| Error::internal(format!("Failed to build text store path: {}", e)))?;
|
||||
|
||||
let store_path_str = store_path.to_absolute_path();
|
||||
|
||||
if self.is_valid_path(&store_path_str)? {
|
||||
return Ok(store_path_str);
|
||||
}
|
||||
|
||||
let request = AddToStoreNarRequest {
|
||||
path: store_path,
|
||||
deriver: None,
|
||||
nar_hash: unsafe {
|
||||
std::mem::transmute::<[u8; 32], nix_compat::nix_daemon::types::NarHash>(
|
||||
nar_hash_arr,
|
||||
)
|
||||
},
|
||||
references: ref_store_paths,
|
||||
registration_time: 0,
|
||||
nar_size: nar_data.len() as u64,
|
||||
ultimate: false,
|
||||
signatures: vec![],
|
||||
ca: Some(CAHash::Text(content_hash)),
|
||||
repair: false,
|
||||
dont_check_sigs: false,
|
||||
};
|
||||
|
||||
self.block_on(async {
|
||||
self.connection
|
||||
.add_to_store_nar(request, &nar_data)
|
||||
.await
|
||||
.map_err(|e| Error::internal(format!("Failed to add text to store: {}", e)))
|
||||
})?;
|
||||
|
||||
Ok(store_path_str)
|
||||
}
|
||||
|
||||
fn make_fixed_output_path(
|
||||
&self,
|
||||
_hash_algo: &str,
|
||||
hash_algo: &str,
|
||||
hash: &str,
|
||||
_hash_mode: &str,
|
||||
hash_mode: &str,
|
||||
name: &str,
|
||||
) -> Result<String> {
|
||||
let short_hash = &hash[..32.min(hash.len())];
|
||||
Ok(format!("/nix/store/{}-{}", short_hash, name))
|
||||
use nix_compat::nixhash::{CAHash, NixHash};
|
||||
use nix_compat::store_path::build_ca_path;
|
||||
|
||||
let nix_hash = match hash_algo {
|
||||
"sha256" => {
|
||||
let hash_bytes = hex::decode(hash)
|
||||
.map_err(|e| Error::internal(format!("Invalid hash hex: {}", e)))?;
|
||||
if hash_bytes.len() != 32 {
|
||||
return Err(Error::internal(format!(
|
||||
"Invalid sha256 hash length: expected 32, got {}",
|
||||
hash_bytes.len()
|
||||
)));
|
||||
}
|
||||
let mut arr = [0u8; 32];
|
||||
arr.copy_from_slice(&hash_bytes);
|
||||
NixHash::Sha256(arr)
|
||||
}
|
||||
_ => {
|
||||
return Err(Error::internal(format!(
|
||||
"Unsupported hash algorithm: {}",
|
||||
hash_algo
|
||||
)));
|
||||
}
|
||||
};
|
||||
|
||||
let ca_hash = if hash_mode == "r" {
|
||||
CAHash::Nar(nix_hash)
|
||||
} else {
|
||||
CAHash::Flat(nix_hash)
|
||||
};
|
||||
|
||||
let store_path: nix_compat::store_path::StorePath<String> =
|
||||
build_ca_path(name, &ca_hash, Vec::<String>::new(), false)
|
||||
.map_err(|e| Error::internal(format!("Failed to build store path: {}", e)))?;
|
||||
|
||||
Ok(store_path.to_absolute_path())
|
||||
}
|
||||
}
|
||||
|
||||
const PROTOCOL_VERSION: ProtocolVersion = ProtocolVersion::from_parts(1, 37);
|
||||
|
||||
// Protocol magic numbers (from nix-compat worker_protocol.rs)
|
||||
const WORKER_MAGIC_1: u64 = 0x6e697863; // "nixc"
|
||||
const WORKER_MAGIC_2: u64 = 0x6478696f; // "dxio"
|
||||
const STDERR_LAST: u64 = 0x616c7473; // "alts"
|
||||
const STDERR_ERROR: u64 = 0x63787470; // "cxtp"
|
||||
|
||||
/// Performs the client handshake with a nix-daemon server
|
||||
///
|
||||
/// This is the client-side counterpart to `server_handshake_client`.
|
||||
/// It exchanges magic numbers, negotiates protocol version, and sends client settings.
|
||||
async fn client_handshake<RW>(
|
||||
conn: &mut RW,
|
||||
client_settings: &ClientSettings,
|
||||
) -> IoResult<ProtocolVersion>
|
||||
where
|
||||
RW: AsyncReadExt + AsyncWriteExt + Unpin,
|
||||
{
|
||||
// 1. Send magic number 1
|
||||
conn.write_u64_le(WORKER_MAGIC_1).await?;
|
||||
|
||||
// 2. Receive magic number 2
|
||||
let magic2 = conn.read_u64_le().await?;
|
||||
if magic2 != WORKER_MAGIC_2 {
|
||||
return Err(IoError::new(
|
||||
IoErrorKind::InvalidData,
|
||||
format!("Invalid magic number from server: {}", magic2),
|
||||
));
|
||||
}
|
||||
|
||||
// 3. Receive server protocol version
|
||||
let server_version_raw = conn.read_u64_le().await?;
|
||||
let server_version: ProtocolVersion = server_version_raw.try_into().map_err(|e| {
|
||||
IoError::new(
|
||||
IoErrorKind::InvalidData,
|
||||
format!("Invalid protocol version: {}", e),
|
||||
)
|
||||
})?;
|
||||
|
||||
// 4. Send our protocol version
|
||||
conn.write_u64_le(PROTOCOL_VERSION.into()).await?;
|
||||
|
||||
// Pick the minimum version
|
||||
let protocol_version = std::cmp::min(PROTOCOL_VERSION, server_version);
|
||||
|
||||
// 5. Send obsolete fields based on protocol version
|
||||
if protocol_version.minor() >= 14 {
|
||||
// CPU affinity (obsolete, send 0)
|
||||
conn.write_u64_le(0).await?;
|
||||
}
|
||||
|
||||
if protocol_version.minor() >= 11 {
|
||||
// Reserve space (obsolete, send 0)
|
||||
conn.write_u64_le(0).await?;
|
||||
}
|
||||
|
||||
if protocol_version.minor() >= 33 {
|
||||
// Read Nix version string
|
||||
let version_len = conn.read_u64_le().await? as usize;
|
||||
let mut version_bytes = vec![0u8; version_len];
|
||||
conn.read_exact(&mut version_bytes).await?;
|
||||
// Padding
|
||||
let padding = (8 - (version_len % 8)) % 8;
|
||||
if padding > 0 {
|
||||
let mut pad = vec![0u8; padding];
|
||||
conn.read_exact(&mut pad).await?;
|
||||
}
|
||||
}
|
||||
|
||||
if protocol_version.minor() >= 35 {
|
||||
// Read trust level
|
||||
let _trust = conn.read_u64_le().await?;
|
||||
}
|
||||
|
||||
// 6. Read STDERR_LAST
|
||||
let stderr_last = conn.read_u64_le().await?;
|
||||
if stderr_last != STDERR_LAST {
|
||||
return Err(IoError::new(
|
||||
IoErrorKind::InvalidData,
|
||||
format!("Expected STDERR_LAST, got: {}", stderr_last),
|
||||
));
|
||||
}
|
||||
|
||||
// 7. Send SetOptions operation with client settings
|
||||
conn.write_u64_le(Operation::SetOptions.into()).await?;
|
||||
conn.flush().await?;
|
||||
|
||||
// Serialize client settings
|
||||
let mut settings_buf = Vec::new();
|
||||
{
|
||||
let mut writer = NixWriterBuilder::default()
|
||||
.set_version(protocol_version)
|
||||
.build(&mut settings_buf);
|
||||
writer.write_value(client_settings).await?;
|
||||
writer.flush().await?;
|
||||
}
|
||||
conn.write_all(&settings_buf).await?;
|
||||
conn.flush().await?;
|
||||
|
||||
// 8. Read response to SetOptions
|
||||
let response = conn.read_u64_le().await?;
|
||||
if response != STDERR_LAST {
|
||||
return Err(IoError::new(
|
||||
IoErrorKind::InvalidData,
|
||||
format!("Expected STDERR_LAST after SetOptions, got: {}", response),
|
||||
));
|
||||
}
|
||||
|
||||
Ok(protocol_version)
|
||||
}
|
||||
|
||||
/// Low-level Nix Daemon client
|
||||
///
|
||||
/// This struct manages communication with a nix-daemon using the wire protocol.
|
||||
/// It is NOT thread-safe and should be wrapped in a Mutex for concurrent access.
|
||||
pub struct NixDaemonClient {
|
||||
protocol_version: ProtocolVersion,
|
||||
reader: NixReader<ReadHalf<UnixStream>>,
|
||||
writer: NixWriter<WriteHalf<UnixStream>>,
|
||||
}
|
||||
|
||||
impl NixDaemonClient {
|
||||
/// Connect to a nix-daemon at the given Unix socket path
|
||||
pub async fn connect(socket_path: &Path) -> IoResult<Self> {
|
||||
let stream = UnixStream::connect(socket_path).await?;
|
||||
Self::from_stream(stream).await
|
||||
}
|
||||
|
||||
/// Create a client from an existing Unix stream
|
||||
pub async fn from_stream(mut stream: UnixStream) -> IoResult<Self> {
|
||||
let client_settings = ClientSettings::default();
|
||||
|
||||
// Perform handshake
|
||||
let protocol_version = client_handshake(&mut stream, &client_settings).await?;
|
||||
|
||||
// Split stream into reader and writer
|
||||
let (read_half, write_half) = split(stream);
|
||||
|
||||
let reader = NixReader::builder()
|
||||
.set_version(protocol_version)
|
||||
.build(read_half);
|
||||
|
||||
let writer = NixWriterBuilder::default()
|
||||
.set_version(protocol_version)
|
||||
.build(write_half);
|
||||
|
||||
Ok(Self {
|
||||
protocol_version,
|
||||
reader,
|
||||
writer,
|
||||
})
|
||||
}
|
||||
|
||||
/// Execute an operation that returns a typed result
|
||||
///
|
||||
/// This is the main method for implementing protocol operations:
|
||||
/// 1. Send operation code
|
||||
/// 2. Send operation parameters
|
||||
/// 3. Receive response or error
|
||||
async fn execute<T>(&mut self, operation: Operation) -> IoResult<T>
|
||||
where
|
||||
T: nix_compat::wire::de::NixDeserialize,
|
||||
{
|
||||
// Send operation
|
||||
self.writer.write_value(&operation).await?;
|
||||
self.writer.flush().await?;
|
||||
|
||||
self.read_response().await
|
||||
}
|
||||
|
||||
/// Execute an operation with a single parameter
|
||||
async fn execute_with<P, T>(&mut self, operation: Operation, param: &P) -> IoResult<T>
|
||||
where
|
||||
P: NixSerialize + Send,
|
||||
T: nix_compat::wire::de::NixDeserialize,
|
||||
{
|
||||
// Send operation
|
||||
self.writer.write_value(&operation).await?;
|
||||
// Send parameter
|
||||
self.writer.write_value(param).await?;
|
||||
self.writer.flush().await?;
|
||||
|
||||
self.read_response().await
|
||||
}
|
||||
|
||||
/// Read a response from the daemon
|
||||
///
|
||||
/// The daemon sends either:
|
||||
/// - STDERR_LAST followed by the result
|
||||
/// - STDERR_ERROR followed by an error message
|
||||
async fn read_response<T>(&mut self) -> IoResult<T>
|
||||
where
|
||||
T: nix_compat::wire::de::NixDeserialize,
|
||||
{
|
||||
loop {
|
||||
let msg = self.reader.read_number().await?;
|
||||
|
||||
if msg == STDERR_LAST {
|
||||
// Success, read the actual response
|
||||
let result: T = self.reader.read_value().await?;
|
||||
return Ok(result);
|
||||
} else if msg == STDERR_ERROR {
|
||||
// IoError, read error message
|
||||
// The error is sent as a NixIoError struct, but we just read the message
|
||||
let error_msg: String = self.reader.read_value().await?;
|
||||
return Err(IoError::other(error_msg));
|
||||
} else {
|
||||
// Other STDERR_* codes (logging, etc.) - for now, we ignore them
|
||||
// Read and discard the associated data
|
||||
let _data: String = self.reader.read_value().await?;
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Check if a path is valid in the store
|
||||
pub async fn is_valid_path(&mut self, path: &str) -> IoResult<bool> {
|
||||
let store_path = StorePath::<String>::from_absolute_path(path.as_bytes())
|
||||
.map_err(|e| IoError::new(IoErrorKind::InvalidInput, e.to_string()))?;
|
||||
|
||||
self.execute_with(Operation::IsValidPath, &store_path).await
|
||||
}
|
||||
|
||||
/// Query information about a store path
|
||||
pub async fn query_path_info(&mut self, path: &str) -> IoResult<Option<UnkeyedValidPathInfo>> {
|
||||
let store_path = StorePath::<String>::from_absolute_path(path.as_bytes())
|
||||
.map_err(|e| IoError::new(IoErrorKind::InvalidInput, e.to_string()))?;
|
||||
|
||||
// QueryPathInfo returns Option<UnkeyedValidPathInfo> which is serialized
|
||||
// as a bool followed by the value if true
|
||||
self.writer.write_value(&Operation::QueryPathInfo).await?;
|
||||
self.writer.write_value(&store_path).await?;
|
||||
self.writer.flush().await?;
|
||||
|
||||
// Read response - it's serialized as bool + optional value
|
||||
loop {
|
||||
let msg = self.reader.read_number().await?;
|
||||
if msg == STDERR_LAST {
|
||||
let has_value: bool = self.reader.read_value().await?;
|
||||
if has_value {
|
||||
// Manually deserialize UnkeyedValidPathInfo
|
||||
use nix_compat::narinfo::Signature;
|
||||
use nix_compat::nixhash::CAHash;
|
||||
|
||||
let deriver = self.reader.read_value().await?;
|
||||
let nar_hash: String = self.reader.read_value().await?;
|
||||
let references = self.reader.read_value().await?;
|
||||
let registration_time = self.reader.read_value().await?;
|
||||
let nar_size = self.reader.read_value().await?;
|
||||
let ultimate = self.reader.read_value().await?;
|
||||
let signatures: Vec<Signature<String>> = self.reader.read_value().await?;
|
||||
let ca: Option<CAHash> = self.reader.read_value().await?;
|
||||
|
||||
let value = UnkeyedValidPathInfo {
|
||||
deriver,
|
||||
nar_hash,
|
||||
references,
|
||||
registration_time,
|
||||
nar_size,
|
||||
ultimate,
|
||||
signatures,
|
||||
ca,
|
||||
};
|
||||
return Ok(Some(value));
|
||||
} else {
|
||||
return Ok(None);
|
||||
}
|
||||
} else if msg == STDERR_ERROR {
|
||||
let error_msg: String = self.reader.read_value().await?;
|
||||
return Err(IoError::other(error_msg));
|
||||
} else {
|
||||
let _data: String = self.reader.read_value().await?;
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Ensure a path is available in the store
|
||||
pub async fn ensure_path(&mut self, path: &str) -> IoResult<()> {
|
||||
let store_path = StorePath::<String>::from_absolute_path(path.as_bytes())
|
||||
.map_err(|e| IoError::new(IoErrorKind::InvalidInput, e.to_string()))?;
|
||||
|
||||
// EnsurePath returns void (no value)
|
||||
self.writer.write_value(&Operation::EnsurePath).await?;
|
||||
self.writer.write_value(&store_path).await?;
|
||||
self.writer.flush().await?;
|
||||
|
||||
// Read response - expect STDERR_LAST with no value
|
||||
loop {
|
||||
let msg = self.reader.read_number().await?;
|
||||
if msg == STDERR_LAST {
|
||||
return Ok(());
|
||||
} else if msg == STDERR_ERROR {
|
||||
let error_msg: String = self.reader.read_value().await?;
|
||||
return Err(IoError::other(error_msg));
|
||||
} else {
|
||||
let _data: String = self.reader.read_value().await?;
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Query which paths are valid
|
||||
pub async fn query_valid_paths(&mut self, paths: Vec<String>) -> IoResult<Vec<String>> {
|
||||
let store_paths: IoResult<Vec<StorePath<String>>> = paths
|
||||
.iter()
|
||||
.map(|p| {
|
||||
StorePath::<String>::from_absolute_path(p.as_bytes())
|
||||
.map_err(|e| IoError::new(IoErrorKind::InvalidInput, e.to_string()))
|
||||
})
|
||||
.collect();
|
||||
let store_paths = store_paths?;
|
||||
|
||||
// Send operation
|
||||
self.writer.write_value(&Operation::QueryValidPaths).await?;
|
||||
|
||||
// Manually serialize the request since QueryValidPaths doesn't impl NixSerialize
|
||||
// QueryValidPaths = { paths: Vec<StorePath>, substitute: bool }
|
||||
self.writer.write_value(&store_paths).await?;
|
||||
|
||||
// For protocol >= 1.27, send substitute flag
|
||||
if self.protocol_version.minor() >= 27 {
|
||||
self.writer.write_value(&false).await?;
|
||||
}
|
||||
|
||||
self.writer.flush().await?;
|
||||
|
||||
let result: Vec<StorePath<String>> = self.read_response().await?;
|
||||
Ok(result.into_iter().map(|p| p.to_absolute_path()).collect())
|
||||
}
|
||||
|
||||
/// Add a NAR to the store
|
||||
pub async fn add_to_store_nar(
|
||||
&mut self,
|
||||
request: AddToStoreNarRequest,
|
||||
nar_data: &[u8],
|
||||
) -> IoResult<()> {
|
||||
tracing::debug!(
|
||||
"add_to_store_nar: path={}, nar_size={}",
|
||||
request.path.to_absolute_path(),
|
||||
request.nar_size,
|
||||
);
|
||||
|
||||
self.writer.write_value(&Operation::AddToStoreNar).await?;
|
||||
|
||||
self.writer.write_value(&request.path).await?;
|
||||
self.writer.write_value(&request.deriver).await?;
|
||||
|
||||
let nar_hash_hex = hex::encode(request.nar_hash.as_ref());
|
||||
self.writer.write_value(&nar_hash_hex).await?;
|
||||
|
||||
self.writer.write_value(&request.references).await?;
|
||||
self.writer.write_value(&request.registration_time).await?;
|
||||
self.writer.write_value(&request.nar_size).await?;
|
||||
self.writer.write_value(&request.ultimate).await?;
|
||||
self.writer.write_value(&request.signatures).await?;
|
||||
self.writer.write_value(&request.ca).await?;
|
||||
self.writer.write_value(&request.repair).await?;
|
||||
self.writer.write_value(&request.dont_check_sigs).await?;
|
||||
|
||||
if self.protocol_version.minor() >= 23 {
|
||||
self.writer.write_number(nar_data.len() as u64).await?;
|
||||
self.writer.write_all(nar_data).await?;
|
||||
self.writer.write_number(0u64).await?;
|
||||
} else {
|
||||
self.writer.write_slice(nar_data).await?;
|
||||
}
|
||||
|
||||
self.writer.flush().await?;
|
||||
|
||||
loop {
|
||||
let msg = self.reader.read_number().await?;
|
||||
|
||||
if msg == STDERR_LAST {
|
||||
return Ok(());
|
||||
} else if msg == STDERR_ERROR {
|
||||
let error_msg: String = self.reader.read_value().await?;
|
||||
return Err(IoError::other(error_msg));
|
||||
} else {
|
||||
let _data: String = self.reader.read_value().await?;
|
||||
continue;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Thread-safe wrapper around NixDaemonClient
|
||||
pub struct NixDaemonConnection {
|
||||
client: Mutex<NixDaemonClient>,
|
||||
}
|
||||
|
||||
impl NixDaemonConnection {
|
||||
/// Connect to a nix-daemon at the given socket path
|
||||
pub async fn connect(socket_path: &Path) -> IoResult<Self> {
|
||||
let client = NixDaemonClient::connect(socket_path).await?;
|
||||
Ok(Self {
|
||||
client: Mutex::new(client),
|
||||
})
|
||||
}
|
||||
|
||||
/// Check if a path is valid in the store
|
||||
pub async fn is_valid_path(&self, path: &str) -> IoResult<bool> {
|
||||
let mut client = self.client.lock().await;
|
||||
client.is_valid_path(path).await
|
||||
}
|
||||
|
||||
/// Query information about a store path
|
||||
pub async fn query_path_info(&self, path: &str) -> IoResult<Option<UnkeyedValidPathInfo>> {
|
||||
let mut client = self.client.lock().await;
|
||||
client.query_path_info(path).await
|
||||
}
|
||||
|
||||
/// Ensure a path is available in the store
|
||||
pub async fn ensure_path(&self, path: &str) -> IoResult<()> {
|
||||
let mut client = self.client.lock().await;
|
||||
client.ensure_path(path).await
|
||||
}
|
||||
|
||||
/// Query which paths are valid
|
||||
pub async fn query_valid_paths(&self, paths: Vec<String>) -> IoResult<Vec<String>> {
|
||||
let mut client = self.client.lock().await;
|
||||
client.query_valid_paths(paths).await
|
||||
}
|
||||
|
||||
/// Add a NAR to the store
|
||||
pub async fn add_to_store_nar(
|
||||
&self,
|
||||
request: AddToStoreNarRequest,
|
||||
nar_data: &[u8],
|
||||
) -> IoResult<()> {
|
||||
let mut client = self.client.lock().await;
|
||||
client.add_to_store_nar(request, nar_data).await
|
||||
}
|
||||
}
|
||||
|
||||
@@ -74,6 +74,35 @@ impl Store for SimulatedStore {
|
||||
Ok(store_path.to_string_lossy().to_string())
|
||||
}
|
||||
|
||||
fn add_to_store_from_path(
|
||||
&self,
|
||||
name: &str,
|
||||
source_path: &Path,
|
||||
_references: Vec<String>,
|
||||
) -> Result<String> {
|
||||
use crate::fetcher::cache::copy_dir_recursive;
|
||||
|
||||
let nar_hash = crate::nar::compute_nar_hash(source_path)
|
||||
.map_err(|e| Error::internal(format!("Failed to compute NAR hash: {}", e)))?;
|
||||
|
||||
let store_path = self.cache.make_store_path(&nar_hash, name);
|
||||
|
||||
if !store_path.exists() {
|
||||
fs::create_dir_all(&store_path)
|
||||
.map_err(|e| Error::internal(format!("Failed to create store directory: {}", e)))?;
|
||||
|
||||
if source_path.is_dir() {
|
||||
copy_dir_recursive(source_path, &store_path)
|
||||
.map_err(|e| Error::internal(format!("Failed to copy to store: {}", e)))?;
|
||||
} else {
|
||||
fs::copy(source_path, &store_path)
|
||||
.map_err(|e| Error::internal(format!("Failed to copy to store: {}", e)))?;
|
||||
}
|
||||
}
|
||||
|
||||
Ok(store_path.to_string_lossy().to_string())
|
||||
}
|
||||
|
||||
fn add_text_to_store(
|
||||
&self,
|
||||
name: &str,
|
||||
|
||||
Reference in New Issue
Block a user