god forsaken http signatures
This commit is contained in:
parent
bb26926edb
commit
b91da3c4ab
20 changed files with 1606 additions and 550 deletions
241
Cargo.lock
generated
241
Cargo.lock
generated
|
@ -188,6 +188,12 @@ version = "0.22.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "9475866fec1451be56a3c2400fd081ff546538961565ccb5b7142cbd22bc7a51"
|
checksum = "9475866fec1451be56a3c2400fd081ff546538961565ccb5b7142cbd22bc7a51"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "base64ct"
|
||||||
|
version = "1.6.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "8c3c1a368f70d6cf7302d78f8f7093da241fb8e8807c05cc9e51a125895a6d5b"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "bincode"
|
name = "bincode"
|
||||||
version = "2.0.0-rc.3"
|
version = "2.0.0-rc.3"
|
||||||
|
@ -239,12 +245,27 @@ version = "2.5.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "cf4b9d6a944f767f8e5e0db018570623c85f3d925ac718db4e06d0187adb21c1"
|
checksum = "cf4b9d6a944f767f8e5e0db018570623c85f3d925ac718db4e06d0187adb21c1"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "block-buffer"
|
||||||
|
version = "0.10.4"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "3078c7629b62d3f0439517fa394996acacc5cbc91c5a20d8c658e77abd503a71"
|
||||||
|
dependencies = [
|
||||||
|
"generic-array",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "bumpalo"
|
name = "bumpalo"
|
||||||
version = "3.16.0"
|
version = "3.16.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "79296716171880943b8470b5f8d03aa55eb2e645a4874bdbb28adb49162e012c"
|
checksum = "79296716171880943b8470b5f8d03aa55eb2e645a4874bdbb28adb49162e012c"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "byteorder"
|
||||||
|
version = "1.5.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "1fd0f2584146f6f2ef48085050886acf353beff7305ebd1ae69500e27c67f64b"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "bytes"
|
name = "bytes"
|
||||||
version = "1.6.0"
|
version = "1.6.0"
|
||||||
|
@ -381,6 +402,12 @@ version = "1.0.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "acbf1af155f9b9ef647e42cdc158db4b64a1b61f743629225fde6f3e0be2a7c7"
|
checksum = "acbf1af155f9b9ef647e42cdc158db4b64a1b61f743629225fde6f3e0be2a7c7"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "const-oid"
|
||||||
|
version = "0.9.6"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "c2459377285ad874054d797f3ccebf984978aa39129f6eafde5cdc8315b612f8"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "convert_case"
|
name = "convert_case"
|
||||||
version = "0.4.0"
|
version = "0.4.0"
|
||||||
|
@ -403,6 +430,25 @@ version = "0.8.6"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "06ea2b9bc92be3c2baa9334a323ebca2d6f074ff852cd1d7b11064035cd3868f"
|
checksum = "06ea2b9bc92be3c2baa9334a323ebca2d6f074ff852cd1d7b11064035cd3868f"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "cpufeatures"
|
||||||
|
version = "0.2.12"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "53fe5e26ff1b7aef8bca9c6080520cfb8d9333c7568e1829cef191a9723e5504"
|
||||||
|
dependencies = [
|
||||||
|
"libc",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "crypto-common"
|
||||||
|
version = "0.1.6"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "1bfb12502f3fc46cca1bb51ac28df9d618d813cdc3d2f25b9fe775a34af26bb3"
|
||||||
|
dependencies = [
|
||||||
|
"generic-array",
|
||||||
|
"typenum",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "csv"
|
name = "csv"
|
||||||
version = "1.3.0"
|
version = "1.3.0"
|
||||||
|
@ -424,6 +470,17 @@ dependencies = [
|
||||||
"memchr",
|
"memchr",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "der"
|
||||||
|
version = "0.7.9"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "f55bf8e7b65898637379c1b74eb1551107c8294ed26d855ceb9fd1a09cfc9bc0"
|
||||||
|
dependencies = [
|
||||||
|
"const-oid",
|
||||||
|
"pem-rfc7468",
|
||||||
|
"zeroize",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "derive_more"
|
name = "derive_more"
|
||||||
version = "0.99.17"
|
version = "0.99.17"
|
||||||
|
@ -437,6 +494,17 @@ dependencies = [
|
||||||
"syn 1.0.109",
|
"syn 1.0.109",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "digest"
|
||||||
|
version = "0.10.7"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "9ed9a281f7bc9b7576e61468ba615a66a5c8cfdff42420a70aa82701a3b1e292"
|
||||||
|
dependencies = [
|
||||||
|
"block-buffer",
|
||||||
|
"const-oid",
|
||||||
|
"crypto-common",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "either"
|
name = "either"
|
||||||
version = "1.11.0"
|
version = "1.11.0"
|
||||||
|
@ -478,8 +546,12 @@ checksum = "658bd65b1cf4c852a3cc96f18a8ce7b5640f6b703f905c7d74532294c2a63984"
|
||||||
name = "fetch"
|
name = "fetch"
|
||||||
version = "0.0.0"
|
version = "0.0.0"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
|
"base64 0.22.0",
|
||||||
|
"chrono",
|
||||||
"derive_more",
|
"derive_more",
|
||||||
|
"http",
|
||||||
"reqwest",
|
"reqwest",
|
||||||
|
"rsa",
|
||||||
"serde_json",
|
"serde_json",
|
||||||
"sigh",
|
"sigh",
|
||||||
]
|
]
|
||||||
|
@ -553,6 +625,16 @@ dependencies = [
|
||||||
"pin-utils",
|
"pin-utils",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "generic-array"
|
||||||
|
version = "0.14.7"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "85649ca51fd72272d7821adaf274ad91c288277713d9c18820d8499a7ff69e9a"
|
||||||
|
dependencies = [
|
||||||
|
"typenum",
|
||||||
|
"version_check",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "getrandom"
|
name = "getrandom"
|
||||||
version = "0.2.14"
|
version = "0.2.14"
|
||||||
|
@ -805,6 +887,9 @@ name = "lazy_static"
|
||||||
version = "1.4.0"
|
version = "1.4.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "e2abad23fbc42b3700f2f279844dc832adb2b2eb069b2df918f455c4e18cc646"
|
checksum = "e2abad23fbc42b3700f2f279844dc832adb2b2eb069b2df918f455c4e18cc646"
|
||||||
|
dependencies = [
|
||||||
|
"spin",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "lazycell"
|
name = "lazycell"
|
||||||
|
@ -828,6 +913,12 @@ dependencies = [
|
||||||
"windows-targets 0.52.5",
|
"windows-targets 0.52.5",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "libm"
|
||||||
|
version = "0.2.8"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "4ec2a862134d2a7d32d7983ddcdd1c4923530833c9f2ea1a44fc5fa473989058"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "librocksdb-sys"
|
name = "librocksdb-sys"
|
||||||
version = "0.17.0+9.0.0"
|
version = "0.17.0+9.0.0"
|
||||||
|
@ -968,6 +1059,43 @@ dependencies = [
|
||||||
"minimal-lexical",
|
"minimal-lexical",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "num-bigint-dig"
|
||||||
|
version = "0.8.4"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "dc84195820f291c7697304f3cbdadd1cb7199c0efc917ff5eafd71225c136151"
|
||||||
|
dependencies = [
|
||||||
|
"byteorder",
|
||||||
|
"lazy_static",
|
||||||
|
"libm",
|
||||||
|
"num-integer",
|
||||||
|
"num-iter",
|
||||||
|
"num-traits",
|
||||||
|
"rand",
|
||||||
|
"smallvec",
|
||||||
|
"zeroize",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "num-integer"
|
||||||
|
version = "0.1.46"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "7969661fd2958a5cb096e56c8e1ad0444ac2bbcd0061bd28660485a44879858f"
|
||||||
|
dependencies = [
|
||||||
|
"num-traits",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "num-iter"
|
||||||
|
version = "0.1.44"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "d869c01cc0c455284163fd0092f1f93835385ccab5a98a0dcc497b2f8bf055a9"
|
||||||
|
dependencies = [
|
||||||
|
"autocfg",
|
||||||
|
"num-integer",
|
||||||
|
"num-traits",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "num-traits"
|
name = "num-traits"
|
||||||
version = "0.2.18"
|
version = "0.2.18"
|
||||||
|
@ -975,6 +1103,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "da0df0e5185db44f69b44f26786fe401b6c293d1907744beaa7fa62b2e5a517a"
|
checksum = "da0df0e5185db44f69b44f26786fe401b6c293d1907744beaa7fa62b2e5a517a"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"autocfg",
|
"autocfg",
|
||||||
|
"libm",
|
||||||
]
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
|
@ -1069,6 +1198,15 @@ dependencies = [
|
||||||
"windows-targets 0.48.5",
|
"windows-targets 0.48.5",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "pem-rfc7468"
|
||||||
|
version = "0.7.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "88b39c9bfcfc231068454382784bb460aae594343fb030d46e9f50a645418412"
|
||||||
|
dependencies = [
|
||||||
|
"base64ct",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "percent-encoding"
|
name = "percent-encoding"
|
||||||
version = "2.3.1"
|
version = "2.3.1"
|
||||||
|
@ -1107,6 +1245,27 @@ version = "0.1.0"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "8b870d8c151b6f2fb93e84a13146138f05d02ed11c7e7c54f8826aaaf7c9f184"
|
checksum = "8b870d8c151b6f2fb93e84a13146138f05d02ed11c7e7c54f8826aaaf7c9f184"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "pkcs1"
|
||||||
|
version = "0.7.5"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "c8ffb9f10fa047879315e6625af03c164b16962a5368d724ed16323b68ace47f"
|
||||||
|
dependencies = [
|
||||||
|
"der",
|
||||||
|
"pkcs8",
|
||||||
|
"spki",
|
||||||
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "pkcs8"
|
||||||
|
version = "0.10.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "f950b2377845cebe5cf8b5165cb3cc1a5e0fa5cfa3e1f7f55707d8fd82e0a7b7"
|
||||||
|
dependencies = [
|
||||||
|
"der",
|
||||||
|
"spki",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "pkg-config"
|
name = "pkg-config"
|
||||||
version = "0.3.30"
|
version = "0.3.30"
|
||||||
|
@ -1277,6 +1436,27 @@ dependencies = [
|
||||||
"librocksdb-sys",
|
"librocksdb-sys",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "rsa"
|
||||||
|
version = "0.9.6"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "5d0e5124fcb30e76a7e79bfee683a2746db83784b86289f6251b54b7950a0dfc"
|
||||||
|
dependencies = [
|
||||||
|
"const-oid",
|
||||||
|
"digest",
|
||||||
|
"num-bigint-dig",
|
||||||
|
"num-integer",
|
||||||
|
"num-traits",
|
||||||
|
"pkcs1",
|
||||||
|
"pkcs8",
|
||||||
|
"rand_core",
|
||||||
|
"sha2",
|
||||||
|
"signature",
|
||||||
|
"spki",
|
||||||
|
"subtle",
|
||||||
|
"zeroize",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "rustc-demangle"
|
name = "rustc-demangle"
|
||||||
version = "0.1.23"
|
version = "0.1.23"
|
||||||
|
@ -1446,6 +1626,17 @@ dependencies = [
|
||||||
"tokio",
|
"tokio",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "sha2"
|
||||||
|
version = "0.10.8"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "793db75ad2bcafc3ffa7c68b215fee268f537982cd901d132f89c6343f3a3dc8"
|
||||||
|
dependencies = [
|
||||||
|
"cfg-if",
|
||||||
|
"cpufeatures",
|
||||||
|
"digest",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "shlex"
|
name = "shlex"
|
||||||
version = "1.3.0"
|
version = "1.3.0"
|
||||||
|
@ -1474,6 +1665,16 @@ dependencies = [
|
||||||
"libc",
|
"libc",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "signature"
|
||||||
|
version = "2.2.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "77549399552de45a898a580c1b41d445bf730df867cc44e6c0233bbc4b8329de"
|
||||||
|
dependencies = [
|
||||||
|
"digest",
|
||||||
|
"rand_core",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "slab"
|
name = "slab"
|
||||||
version = "0.4.9"
|
version = "0.4.9"
|
||||||
|
@ -1499,6 +1700,22 @@ dependencies = [
|
||||||
"windows-sys 0.52.0",
|
"windows-sys 0.52.0",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "spin"
|
||||||
|
version = "0.5.2"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "6e63cff320ae2c57904679ba7cb63280a3dc4613885beafb148ee7bf9aa9042d"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "spki"
|
||||||
|
version = "0.7.3"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "d91ed6c858b01f942cd56b37a94b3e0a1798290327d1236e4d9cf4eaca44d29d"
|
||||||
|
dependencies = [
|
||||||
|
"base64ct",
|
||||||
|
"der",
|
||||||
|
]
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "store"
|
name = "store"
|
||||||
version = "0.0.0"
|
version = "0.0.0"
|
||||||
|
@ -1519,6 +1736,12 @@ version = "0.11.1"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "7da8b5736845d9f2fcb837ea5d9e2628564b3b043a70948a3f0b778838c5fb4f"
|
checksum = "7da8b5736845d9f2fcb837ea5d9e2628564b3b043a70948a3f0b778838c5fb4f"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "subtle"
|
||||||
|
version = "2.5.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "81cdd64d312baedb58e21336b31bc043b77e01cc99033ce76ef539f78e965ebc"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "syn"
|
name = "syn"
|
||||||
version = "1.0.109"
|
version = "1.0.109"
|
||||||
|
@ -1738,6 +1961,12 @@ version = "0.2.5"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "e421abadd41a4225275504ea4d6566923418b7f05506fbc9c0fe86ba7396114b"
|
checksum = "e421abadd41a4225275504ea4d6566923418b7f05506fbc9c0fe86ba7396114b"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "typenum"
|
||||||
|
version = "1.17.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "42ff0bf0c66b8238c6f3b578df37d0b7848e55df8577b3f74f92a69acceeb825"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "ulid"
|
name = "ulid"
|
||||||
version = "1.1.2"
|
version = "1.1.2"
|
||||||
|
@ -1799,6 +2028,12 @@ version = "0.2.15"
|
||||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
checksum = "accd4ea62f7bb7a82fe23066fb0957d48ef677f6eeb8215f372f52e48bb32426"
|
checksum = "accd4ea62f7bb7a82fe23066fb0957d48ef677f6eeb8215f372f52e48bb32426"
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "version_check"
|
||||||
|
version = "0.9.4"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "49874b5167b65d7193b8aba1567f5c7d93d001cafc34600cee003eda787e483f"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "virtue"
|
name = "virtue"
|
||||||
version = "0.0.13"
|
version = "0.0.13"
|
||||||
|
@ -2095,6 +2330,12 @@ dependencies = [
|
||||||
"windows-sys 0.48.0",
|
"windows-sys 0.48.0",
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[[package]]
|
||||||
|
name = "zeroize"
|
||||||
|
version = "1.7.0"
|
||||||
|
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||||
|
checksum = "525b4ec142c6b68a2d10f01f7bbf6755599ca3f81ea53b8431b7dd348f5fdb2d"
|
||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "zstd-sys"
|
name = "zstd-sys"
|
||||||
version = "2.0.10+zstd.1.5.6"
|
version = "2.0.10+zstd.1.5.6"
|
||||||
|
|
|
@ -1,38 +1,38 @@
|
||||||
|
#![feature(iterator_try_collect)]
|
||||||
use puppy::{
|
use puppy::{
|
||||||
mixin_ap_actor,
|
actor::Actor,
|
||||||
model::{schema, Bite, FollowRequest, Follows, Profile, Username},
|
data::{Bite, Profile},
|
||||||
post::Author,
|
post::Author,
|
||||||
store::{self, Error, OK},
|
store::util::IterExt as _,
|
||||||
Key, Store,
|
Context,
|
||||||
};
|
};
|
||||||
|
|
||||||
fn main() -> store::Result<()> {
|
fn main() -> puppy::Result<()> {
|
||||||
// Store::nuke(".state")?;
|
// puppy::store::Store::nuke(".state")?;
|
||||||
let db = Store::open(".state", schema())?;
|
puppy::context(|cx| {
|
||||||
|
let db = cx.store();
|
||||||
println!("creating actors");
|
println!("creating actors");
|
||||||
let riley = get_or_create_actor(&db, "riley")?;
|
let riley = get_or_create_actor(&cx, "riley")?;
|
||||||
let linen = get_or_create_actor(&db, "linen")?;
|
let linen = get_or_create_actor(&cx, "linen")?;
|
||||||
// db.run(|tx| {
|
|
||||||
// mixin_ap_actor(tx, riley, "test.pup.riley.lgbt")?;
|
|
||||||
// mixin_ap_actor(tx, linen, "test.pup.riley.lgbt")?;
|
|
||||||
// OK
|
|
||||||
// })?;
|
|
||||||
if true {
|
if true {
|
||||||
println!("creating posts");
|
println!("creating posts");
|
||||||
puppy::post::create_post(&db, riley, "@linen <3")?;
|
puppy::post::create_post(&cx, riley.key, "@linen <3")?;
|
||||||
puppy::post::create_post(&db, linen, "@riley <3")?;
|
puppy::post::create_post(&cx, linen.key, "@riley <3")?;
|
||||||
}
|
}
|
||||||
|
|
||||||
if true {
|
if true {
|
||||||
println!("making riley follow linen");
|
println!("making riley follow linen");
|
||||||
if !db.exists::<Follows>(riley, linen)? {
|
if !riley.follows(&cx, &linen)? {
|
||||||
println!("follow relation does not exist yet");
|
println!("follow relation does not exist yet");
|
||||||
if !db.exists::<FollowRequest>(riley, linen)? {
|
if let Some(req) = linen
|
||||||
println!("no pending follow request; creating");
|
.pending_requests(&cx)
|
||||||
puppy::follows::request(&db, riley, linen)?;
|
.find_ok(|r| r.origin == riley.key)?
|
||||||
} else {
|
{
|
||||||
println!("accepting the pending follow request");
|
println!("accepting the pending follow request");
|
||||||
puppy::follows::accept(&db, riley, linen)?;
|
linen.do_accept_request(&cx, req)?;
|
||||||
|
} else {
|
||||||
|
println!("no pending follow request; creating");
|
||||||
|
riley.do_follow_request(&cx, &linen)?;
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
println!("riley already follows linen");
|
println!("riley already follows linen");
|
||||||
|
@ -43,47 +43,47 @@ fn main() -> store::Result<()> {
|
||||||
for post in puppy::post::fetch_timeline(&db, .., None)?.posts() {
|
for post in puppy::post::fetch_timeline(&db, .., None)?.posts() {
|
||||||
let Author { ref handle, .. } = post.author;
|
let Author { ref handle, .. } = post.author;
|
||||||
let content = post.content.content.as_ref().unwrap();
|
let content = post.content.content.as_ref().unwrap();
|
||||||
println!("- {} by {handle}:\n{content}", post.id)
|
println!("- {:?} by {handle}:\n{content}", post.id)
|
||||||
}
|
}
|
||||||
|
|
||||||
println!("\nLinen's followers:");
|
println!("\nLinen's followers:");
|
||||||
for id in puppy::follows::followers_of(&db, linen)? {
|
for id in linen.followers(&cx).try_collect::<Vec<_>>()? {
|
||||||
let Profile { account_name, .. } = db.get_mixin(id)?.unwrap();
|
let Profile { account_name, .. } = db.get_mixin(id)?.unwrap();
|
||||||
println!("- @{account_name} ({id})");
|
println!("- @{account_name} ({id})");
|
||||||
}
|
}
|
||||||
|
|
||||||
println!("\nRiley's following:");
|
println!("\nRiley's following:");
|
||||||
for id in puppy::follows::following_of(&db, riley)? {
|
for id in riley.following(&cx).try_collect::<Vec<_>>()? {
|
||||||
let Profile { account_name, .. } = db.get_mixin(id)?.unwrap();
|
let Profile { account_name, .. } = db.get_mixin(id)?.unwrap();
|
||||||
println!("- @{account_name} ({id})");
|
println!("- @{account_name} ({id})");
|
||||||
}
|
}
|
||||||
|
|
||||||
if false {
|
if false {
|
||||||
println!("Biting riley");
|
println!("Biting riley");
|
||||||
puppy::bites::bite_actor(&db, linen, riley).unwrap();
|
linen.do_bite(&cx, &riley)?;
|
||||||
for Bite { id, biter, .. } in puppy::bites::bites_on(&db, riley).unwrap() {
|
for Bite { id, biter, .. } in riley.bites_suffered(&cx).try_collect::<Vec<_>>()? {
|
||||||
let Profile { account_name, .. } = db.get_mixin(biter)?.unwrap();
|
let Profile { account_name, .. } = db.get_mixin(biter)?.unwrap();
|
||||||
println!("riley was bitten by @{account_name} at {}", id.timestamp());
|
println!("riley was bitten by @{account_name} at {}", id.timestamp());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
store::OK
|
Ok(())
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
fn get_or_create_actor(db: &Store, username: &str) -> Result<Key, Error> {
|
fn get_or_create_actor(cx: &Context<'_>, username: &str) -> puppy::Result<Actor> {
|
||||||
let user = db.lookup(Username(username.to_string()));
|
let user = Actor::by_username(cx, username)?;
|
||||||
match user {
|
match user {
|
||||||
Ok(Some(key)) => {
|
Some(key) => {
|
||||||
println!("found '{username}' ({key})");
|
println!("found '{username}' ({key:?})");
|
||||||
Ok(key)
|
Ok(key)
|
||||||
}
|
}
|
||||||
Ok(None) => {
|
None => {
|
||||||
println!("'{username}' doesn't exist yet, creating");
|
println!("'{username}' doesn't exist yet, creating");
|
||||||
let r = puppy::create_local_actor(&db, username, "test.pup.riley.lgbt");
|
let r = puppy::actor::create_local(cx, username);
|
||||||
if let Ok(ref key) = r {
|
if let Ok(ref key) = r {
|
||||||
println!("created '{username}' with key {key}");
|
println!("created '{username}' with key {key:?}");
|
||||||
}
|
}
|
||||||
r
|
r
|
||||||
}
|
}
|
||||||
Err(e) => Err(e),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -1,16 +1,78 @@
|
||||||
use axum::{extract::Path, routing::get, Json, Router};
|
#![feature(try_blocks)]
|
||||||
use puppy::{get_local_ap_object, model::schema, Key, Store};
|
use std::collections::HashMap;
|
||||||
|
|
||||||
|
use axum::{
|
||||||
|
extract::{Path, Query},
|
||||||
|
response::{AppendHeaders, IntoResponse as _},
|
||||||
|
routing::get,
|
||||||
|
Json, Router,
|
||||||
|
};
|
||||||
|
use puppy::{
|
||||||
|
actor::Actor,
|
||||||
|
context,
|
||||||
|
data::{PrivateKey, PublicKey},
|
||||||
|
get_local_ap_object, Key,
|
||||||
|
};
|
||||||
|
use serde_json::json;
|
||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() {
|
async fn main() {
|
||||||
let db = Store::open(".state", schema()).unwrap();
|
let app = Router::new()
|
||||||
let app = Router::new().route(
|
.route(
|
||||||
"/o/:ulid",
|
"/o/:ulid",
|
||||||
get(|Path(raw_object_id): Path<String>| async move {
|
get(
|
||||||
|
|Path(raw_object_id): Path<String>, req: axum::extract::Request| async move {
|
||||||
|
eprintln!("req: {req:?}");
|
||||||
|
context::<_, puppy::Error>(|cx| try {
|
||||||
let object_id = raw_object_id.parse::<Key>().unwrap();
|
let object_id = raw_object_id.parse::<Key>().unwrap();
|
||||||
let obj = get_local_ap_object(&db, object_id).unwrap().to_json_ld();
|
let obj = get_local_ap_object(&cx, object_id).unwrap().to_json_ld();
|
||||||
Json(obj)
|
(
|
||||||
|
AppendHeaders([("content-type", "application/activity+json")]),
|
||||||
|
Json(obj).into_response(),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
.unwrap()
|
||||||
|
},
|
||||||
|
),
|
||||||
|
)
|
||||||
|
.route(
|
||||||
|
"/proxy",
|
||||||
|
get(|Query(q): Query<HashMap<String, String>>| async move {
|
||||||
|
let (key_pem, key_id) = context::<_, puppy::Error>(|cx| try {
|
||||||
|
let actor = Actor::by_username(&cx, "riley")?.unwrap();
|
||||||
|
let (private, public) = cx
|
||||||
|
.store()
|
||||||
|
.get_mixin_many::<(PrivateKey, PublicKey)>(actor.key)?;
|
||||||
|
(private.key_pem, public.key_id)
|
||||||
|
})
|
||||||
|
.unwrap();
|
||||||
|
puppy::fetch::resolve(&key_pem, &key_id, &q["target"])
|
||||||
|
.await
|
||||||
|
.unwrap()
|
||||||
}),
|
}),
|
||||||
|
)
|
||||||
|
.route(
|
||||||
|
"/.well-known/webfinger",
|
||||||
|
get(
|
||||||
|
|Query(q): Query<HashMap<String, String>>, req: axum::extract::Request| async move {
|
||||||
|
eprintln!("req: {req:?}");
|
||||||
|
let Some(rest) = q["resource"].strip_prefix("acct:") else {
|
||||||
|
panic!("{q:?}");
|
||||||
|
};
|
||||||
|
if rest == "riley@test.piss-on.me" {
|
||||||
|
Json(json!({
|
||||||
|
"subject": "acct:riley@test.piss-on.me",
|
||||||
|
"links": [{
|
||||||
|
"rel": "self",
|
||||||
|
"type": "application/activity+json",
|
||||||
|
"href": "https://test.piss-on.me/o/01HWG4BQJR23TWF12KVPBBP1HG",
|
||||||
|
}],
|
||||||
|
}))
|
||||||
|
} else {
|
||||||
|
panic!("{rest:?}")
|
||||||
|
}
|
||||||
|
},
|
||||||
|
),
|
||||||
);
|
);
|
||||||
let sock = tokio::net::TcpListener::bind("0.0.0.0:1312").await.unwrap();
|
let sock = tokio::net::TcpListener::bind("0.0.0.0:1312").await.unwrap();
|
||||||
axum::serve(sock, app).await.unwrap();
|
axum::serve(sock, app).await.unwrap();
|
||||||
|
|
|
@ -10,3 +10,7 @@ reqwest = "*"
|
||||||
sigh = "*"
|
sigh = "*"
|
||||||
serde_json = "*"
|
serde_json = "*"
|
||||||
derive_more = "*"
|
derive_more = "*"
|
||||||
|
http = "*"
|
||||||
|
chrono = "*"
|
||||||
|
base64 = "*"
|
||||||
|
rsa = { version = "*", features = ["sha2"] }
|
||||||
|
|
|
@ -1,36 +1,112 @@
|
||||||
|
#![feature(iter_intersperse)]
|
||||||
|
use chrono::Utc;
|
||||||
use derive_more::{Display, From, Into};
|
use derive_more::{Display, From, Into};
|
||||||
|
use reqwest::Body;
|
||||||
use serde_json::{json, Value};
|
use serde_json::{json, Value};
|
||||||
|
use sigh::{
|
||||||
|
alg::{Hs2019, RsaSha256},
|
||||||
|
Key as _, SigningConfig,
|
||||||
|
};
|
||||||
|
|
||||||
|
use crate::keys::{HS2019, RSA_SHA256};
|
||||||
|
|
||||||
pub enum Object {
|
pub enum Object {
|
||||||
Activity(Activity),
|
Activity(Activity),
|
||||||
Actor(Actor),
|
Actor(Actor),
|
||||||
Object { id: Id },
|
Object {
|
||||||
|
id: Id,
|
||||||
|
kind: String,
|
||||||
|
content: Option<String>,
|
||||||
|
summary: Option<String>,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Object {
|
impl Object {
|
||||||
pub fn to_json(self) -> Value {
|
pub fn id(&self) -> &Id {
|
||||||
match self {
|
match self {
|
||||||
Object::Activity(_) => todo!(),
|
Object::Activity(a) => &a.id,
|
||||||
Object::Actor(a) => json!({
|
Object::Actor(a) => &a.id,
|
||||||
"id": a.id.to_string(),
|
Object::Object { id, .. } => id,
|
||||||
"inbox": a.inbox.to_string(),
|
|
||||||
"preferredUsername": a.account_name,
|
|
||||||
"name": a.display_name,
|
|
||||||
}),
|
|
||||||
Object::Object { id } => json!({
|
|
||||||
"id": id.to_string()
|
|
||||||
}),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub fn to_json_ld(self) -> Value {
|
pub fn to_json_ld(&self) -> Value {
|
||||||
let mut json = self.to_json();
|
match self {
|
||||||
json["@context"] = json!([]);
|
Object::Activity(a) => a.to_json_ld(),
|
||||||
json
|
Object::Actor(a) => a.to_json_ld(),
|
||||||
|
Object::Object { id, kind, content, summary } => json!({
|
||||||
|
"id": id.to_string(),
|
||||||
|
"type": kind,
|
||||||
|
"content": content,
|
||||||
|
"summary": summary,
|
||||||
|
}),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct Activity {
|
pub struct Activity<T = String> {
|
||||||
pub id: Id,
|
pub id: Id,
|
||||||
|
pub actor: Id,
|
||||||
|
pub object: Box<Object>,
|
||||||
|
pub kind: T,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<K> Activity<K>
|
||||||
|
where
|
||||||
|
K: ToString,
|
||||||
|
{
|
||||||
|
pub fn to_json_ld(&self) -> Value {
|
||||||
|
json!({
|
||||||
|
"@context": [
|
||||||
|
"https://www.w3.org/ns/activitystreams",
|
||||||
|
{ "Bite": "https://ns.mia.jetzt/as#Bite" },
|
||||||
|
],
|
||||||
|
"id": self.id.to_string(),
|
||||||
|
"actor": self.actor.to_string(),
|
||||||
|
"type": self.kind.to_string(),
|
||||||
|
"object": self.object.to_json_ld()
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Deliver an [`Activity`] to a particular `inbox`.
|
||||||
|
pub async fn deliver(private_key_pem: &str, key_id: &str, activity: Activity, inbox: &str) -> () {
|
||||||
|
let body = serde_json::to_string_pretty(&activity.to_json_ld()).unwrap();
|
||||||
|
let mut req = http::Request::post(inbox)
|
||||||
|
.header("content-type", "application/activity+json")
|
||||||
|
.header("user-agent", "ActivityPuppy/0.0.0 (delivery)")
|
||||||
|
.header("date", Utc::now().to_rfc3339())
|
||||||
|
.body(body)
|
||||||
|
.unwrap();
|
||||||
|
let key = sigh::PrivateKey::from_pem(private_key_pem.as_bytes()).unwrap();
|
||||||
|
SigningConfig::new(RsaSha256, &key, key_id)
|
||||||
|
.sign(&mut req)
|
||||||
|
.unwrap();
|
||||||
|
reqwest::Client::new()
|
||||||
|
.execute(req.try_into().unwrap())
|
||||||
|
.await
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn resolve(private_key_pem: &str, key_id: &str, target: &str) -> reqwest::Result<String> {
|
||||||
|
const EMPTY_DIGEST: &str = "sha-256=47DEQpj8HBSa+/TImW+5JCeuQeRkm5NMpJWZG3hSuFU=";
|
||||||
|
// Sun, 06 Nov 1994 08:49:37 GMT
|
||||||
|
const RFC_822: &str = "%a, %d %b %Y %H:%M:%S GMT";
|
||||||
|
|
||||||
|
let date = Utc::now().format(RFC_822).to_string();
|
||||||
|
let mut req = http::Request::get(target)
|
||||||
|
.header("accept", "application/json")
|
||||||
|
.header("user-agent", "ActivityPuppy/0.0.0 (resolver)")
|
||||||
|
.header("date", date)
|
||||||
|
// Empty body
|
||||||
|
.body(Body::default())
|
||||||
|
.unwrap();
|
||||||
|
// hs2019 works with masto
|
||||||
|
keys::sign(&mut req, HS2019, private_key_pem, key_id).unwrap();
|
||||||
|
reqwest::Client::new()
|
||||||
|
.execute(req.try_into().unwrap())
|
||||||
|
.await?
|
||||||
|
.text()
|
||||||
|
.await
|
||||||
}
|
}
|
||||||
|
|
||||||
/// An actor is an entity capable of producing Takes.
|
/// An actor is an entity capable of producing Takes.
|
||||||
|
@ -43,7 +119,238 @@ pub struct Actor {
|
||||||
pub account_name: String,
|
pub account_name: String,
|
||||||
/// Note: this maps to the `name` property.
|
/// Note: this maps to the `name` property.
|
||||||
pub display_name: Option<String>,
|
pub display_name: Option<String>,
|
||||||
|
/// Public counterpart to the signing key used to sign activities
|
||||||
|
/// generated by the actor.
|
||||||
|
pub public_key: PublicKey,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Actor {
|
||||||
|
pub fn to_json_ld(&self) -> Value {
|
||||||
|
json!({
|
||||||
|
"@context": [
|
||||||
|
"https://www.w3.org/ns/activitystreams",
|
||||||
|
"https://w3id.org/security/v1",
|
||||||
|
],
|
||||||
|
"id": self.id.to_string(),
|
||||||
|
"inbox": self.inbox.to_string(),
|
||||||
|
"outbox": self.inbox.to_string().replace("inbox", "outbox"),
|
||||||
|
"preferredUsername": self.account_name,
|
||||||
|
"name": self.display_name,
|
||||||
|
"type": "Person",
|
||||||
|
"publicKey": {
|
||||||
|
"id": self.public_key.key_id.to_string(),
|
||||||
|
"publicKeyPem": self.public_key.public_key_pem,
|
||||||
|
"owner": self.id.to_string(),
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct PublicKey {
|
||||||
|
pub key_id: Id,
|
||||||
|
pub owner: Id,
|
||||||
|
pub public_key_pem: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Display, From, Into, Debug, Clone)]
|
#[derive(Display, From, Into, Debug, Clone)]
|
||||||
pub struct Id(String);
|
pub struct Id(String);
|
||||||
|
|
||||||
|
pub mod keys {
|
||||||
|
//! Cryptography and such.
|
||||||
|
|
||||||
|
use chrono::{TimeDelta, Utc};
|
||||||
|
use http::{HeaderValue, Method, Request};
|
||||||
|
use reqwest::Body;
|
||||||
|
use rsa::{pkcs1v15::SigningKey, pkcs8::DecodePrivateKey as _, sha2};
|
||||||
|
use sigh::{
|
||||||
|
Key as _,
|
||||||
|
alg::{Algorithm as _, RsaSha256},
|
||||||
|
};
|
||||||
|
|
||||||
|
pub struct RawPrivateKey {
|
||||||
|
inner: sigh::PrivateKey,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl RawPrivateKey {
|
||||||
|
pub fn to_pem(&self) -> String {
|
||||||
|
self.inner.to_pem().unwrap()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct RawPublicKey {
|
||||||
|
inner: sigh::PublicKey,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl RawPublicKey {
|
||||||
|
pub fn to_pem(&self) -> String {
|
||||||
|
self.inner.to_pem().unwrap()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Generate a private and public keypair.
|
||||||
|
pub fn gen_keypair() -> (RawPrivateKey, RawPublicKey) {
|
||||||
|
let (private, public) = RsaSha256.generate_keys().unwrap();
|
||||||
|
(RawPrivateKey { inner: private }, RawPublicKey {
|
||||||
|
inner: public,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The algorithm to sign with.
|
||||||
|
#[derive(PartialEq, Debug)]
|
||||||
|
pub struct Algorithm(&'static str);
|
||||||
|
|
||||||
|
pub const RSA_SHA256: Algorithm = Algorithm("rsa-sha256");
|
||||||
|
pub const HS2019: Algorithm = Algorithm("hs2019");
|
||||||
|
|
||||||
|
struct Signature<'k> {
|
||||||
|
key_id: &'k str,
|
||||||
|
alg: Algorithm,
|
||||||
|
components: Vec<Component>,
|
||||||
|
created: String,
|
||||||
|
expires: String,
|
||||||
|
/// Base64-encoded signature.
|
||||||
|
signature_encoded: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Sign `req`.
|
||||||
|
pub fn sign(
|
||||||
|
req: &mut Request<Body>,
|
||||||
|
alg: Algorithm,
|
||||||
|
private_key: &str,
|
||||||
|
key_id: &str,
|
||||||
|
) -> Result<(), String> {
|
||||||
|
// NOTE: Rough translation of https://nest.pijul.org/ez/ActivityPuppy:main/XXPS2UOWSWD2Y.ZJAAA
|
||||||
|
let pieces = gather_pieces(&req)?;
|
||||||
|
let signing_string = make_signing_string(&pieces);
|
||||||
|
let signature = create(&signing_string, alg, private_key, key_id, pieces)?;
|
||||||
|
req.headers_mut().insert("signature", render(signature));
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Gather all the bits from a `Request`.
|
||||||
|
fn gather_pieces(req: &Request<Body>) -> Result<Vec<Component>, &'static str> {
|
||||||
|
let target = {
|
||||||
|
let method = req.method().as_str().to_lowercase();
|
||||||
|
let path = req.uri().path();
|
||||||
|
format!("{method} {path}")
|
||||||
|
};
|
||||||
|
|
||||||
|
let created = Utc::now();
|
||||||
|
let expires = created + TimeDelta::minutes(5);
|
||||||
|
|
||||||
|
let mut components = vec![
|
||||||
|
("(request-target)", target),
|
||||||
|
("(created)", created.timestamp().to_string()),
|
||||||
|
("(expires)", expires.timestamp().to_string()),
|
||||||
|
("host", req.uri().host().unwrap().to_owned()),
|
||||||
|
];
|
||||||
|
|
||||||
|
if [Method::POST, Method::PUT, Method::PATCH].contains(req.method()) {
|
||||||
|
let digest = req
|
||||||
|
.headers()
|
||||||
|
.get("digest")
|
||||||
|
.map(|v| v.to_str().unwrap().to_string())
|
||||||
|
.ok_or("digest header is required for POST, PUT and PATCH requests")?;
|
||||||
|
components.push(("digest", digest));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(components)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn make_signing_string(pieces: &[Component]) -> String {
|
||||||
|
pieces
|
||||||
|
.iter()
|
||||||
|
.map(|(k, v)| format!("{k}: {v}"))
|
||||||
|
.intersperse("\n".to_string())
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
type Component = (&'static str, String);
|
||||||
|
|
||||||
|
/// Sign the `signing_string`.
|
||||||
|
fn create<'s>(
|
||||||
|
signing_string: &'s str,
|
||||||
|
alg: Algorithm,
|
||||||
|
key_pem: &str,
|
||||||
|
key_url: &'s str,
|
||||||
|
components: Vec<Component>,
|
||||||
|
) -> Result<Signature<'s>, String> {
|
||||||
|
let created = components
|
||||||
|
.iter()
|
||||||
|
.find_map(|(k, v)| if *k == "(created)" { Some(v) } else { None })
|
||||||
|
.cloned()
|
||||||
|
.unwrap();
|
||||||
|
let expires = components
|
||||||
|
.iter()
|
||||||
|
.find_map(|(k, v)| if *k == "(expires)" { Some(v) } else { None })
|
||||||
|
.cloned()
|
||||||
|
.unwrap();
|
||||||
|
// We regardless of the algorithm, we produce RSA-SHA256 signatures, because this is broadly compatible
|
||||||
|
// with everything.
|
||||||
|
let signature = sign_rsa_sha256(signing_string, key_pem)?;
|
||||||
|
let encoded = base64(signature);
|
||||||
|
Ok(Signature {
|
||||||
|
signature_encoded: encoded,
|
||||||
|
key_id: key_url,
|
||||||
|
components,
|
||||||
|
created,
|
||||||
|
expires,
|
||||||
|
alg,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
/// `rsa-sha256` is created using an rsa key and a sha256 hash.
|
||||||
|
fn sign_rsa_sha256(signing_string: &str, key_pem: &str) -> Result<Vec<u8>, String> {
|
||||||
|
use rsa::{
|
||||||
|
signature::{Signer as _, SignatureEncoding as _},
|
||||||
|
RsaPrivateKey,
|
||||||
|
};
|
||||||
|
let rsa = RsaPrivateKey::from_pkcs8_pem(key_pem).map_err(|e| e.to_string())?;
|
||||||
|
let key = SigningKey::<sha2::Sha256>::new(rsa);
|
||||||
|
let buf = key.sign(signing_string.as_bytes()).to_vec();
|
||||||
|
Ok(buf)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn base64(buf: Vec<u8>) -> String {
|
||||||
|
use base64::Engine as _;
|
||||||
|
// STANDARD works on masto, url safe does not.
|
||||||
|
base64::prelude::BASE64_STANDARD.encode(buf)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Format the signature.
|
||||||
|
fn render(sig: Signature<'_>) -> HeaderValue {
|
||||||
|
let headers = sig
|
||||||
|
.components
|
||||||
|
.iter()
|
||||||
|
// We only need the header names.
|
||||||
|
.map(|(name, _)| name.as_ref())
|
||||||
|
// Names need to be space-separated.
|
||||||
|
.intersperse(" ")
|
||||||
|
// Equivalent to `fold ["a", "b"]` in haskell
|
||||||
|
.collect::<String>();
|
||||||
|
|
||||||
|
#[rustfmt::skip]
|
||||||
|
let data = [
|
||||||
|
("keyId", sig.key_id),
|
||||||
|
("created", &sig.created),
|
||||||
|
("expires", &sig.expires),
|
||||||
|
("algorithm", sig.alg.0),
|
||||||
|
("headers", &headers),
|
||||||
|
("signature", &sig.signature_encoded),
|
||||||
|
];
|
||||||
|
|
||||||
|
// Ok, now let's put it all together.
|
||||||
|
data.into_iter()
|
||||||
|
// Step 1: all the values need to be surrounded by quotes
|
||||||
|
.map(|(k, v)| (k, format!(r#""{v}""#)))
|
||||||
|
// Step 2. join each pair together
|
||||||
|
.map(|(k, v)| format!("{k}={v}"))
|
||||||
|
// Step 3. comma separate everything
|
||||||
|
.intersperse(", ".to_string())
|
||||||
|
// Step 4. fold the entire thing into one
|
||||||
|
.collect::<String>()
|
||||||
|
// Then, it needs to become a header value
|
||||||
|
.try_into()
|
||||||
|
.expect("signature formatting should give a correct header value")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
6
lib/puppy/clippy.toml
Normal file
6
lib/puppy/clippy.toml
Normal file
|
@ -0,0 +1,6 @@
|
||||||
|
disallowed-methods = [
|
||||||
|
{ path = "puppy::context", reason = "for external use only (interferes with `test_context`)" }
|
||||||
|
]
|
||||||
|
disallowed-types = [
|
||||||
|
{ path = "store::Result", reason = "generates confusing docs; parameterize puppy::Result by puppy::StoreError instead." }
|
||||||
|
]
|
|
@ -1,28 +0,0 @@
|
||||||
//! The most essential feature of any social network.
|
|
||||||
|
|
||||||
use store::{Arrow, Key, Store};
|
|
||||||
|
|
||||||
/// *Bites you*
|
|
||||||
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
|
||||||
pub struct Bite {
|
|
||||||
#[identity]
|
|
||||||
pub id: Key,
|
|
||||||
#[origin]
|
|
||||||
pub biter: Key,
|
|
||||||
#[target]
|
|
||||||
pub victim: Key,
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn bite_actor(db: &Store, biter: Key, victim: Key) -> store::Result<Key> {
|
|
||||||
db.run(|tx| {
|
|
||||||
let id = Key::gen();
|
|
||||||
// TODO: Create an `Activity` arrow with the same ID.
|
|
||||||
tx.create(Bite { id, biter, victim })?;
|
|
||||||
Ok(id)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Who has bitten `victim`?
|
|
||||||
pub fn bites_on(db: &Store, victim: Key) -> store::Result<Vec<Bite>> {
|
|
||||||
db.incoming::<Bite>(victim).try_collect()
|
|
||||||
}
|
|
70
lib/puppy/src/context.rs
Normal file
70
lib/puppy/src/context.rs
Normal file
|
@ -0,0 +1,70 @@
|
||||||
|
use std::sync::OnceLock;
|
||||||
|
|
||||||
|
use store::{Key, Store, Transaction};
|
||||||
|
|
||||||
|
use crate::{config::Config, data::schema, Error, Result};
|
||||||
|
|
||||||
|
/// The context of a running ActivityPuppy.
|
||||||
|
///
|
||||||
|
/// This type provides access to the data store and configuration.
|
||||||
|
pub struct Context<'c> {
|
||||||
|
pub(crate) config: &'c Config,
|
||||||
|
pub(crate) db: Store,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Context<'_> {
|
||||||
|
/// Do a data store [transaction][store::Transaction].
|
||||||
|
pub fn run<T>(&self, f: impl FnOnce(&Transaction<'_>) -> Result<T>) -> Result<T> {
|
||||||
|
self.db.run(f)
|
||||||
|
}
|
||||||
|
/// Access the store directly.
|
||||||
|
pub fn store(&self) -> &Store {
|
||||||
|
&self.db
|
||||||
|
}
|
||||||
|
/// Access the configuration.
|
||||||
|
pub fn config(&self) -> &Config {
|
||||||
|
self.config
|
||||||
|
}
|
||||||
|
/// Create an ActivityPub object ID from a key.
|
||||||
|
pub fn mk_url(&self, key: Key) -> String {
|
||||||
|
format!("https://{}/o/{key}", self.config.ap_domain)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The store, which we initialize only once this way.
|
||||||
|
///
|
||||||
|
/// This makes it so we don't have to thread everything down everywhere, we can just access the context
|
||||||
|
/// when we need it.
|
||||||
|
static STORE: OnceLock<Store> = OnceLock::new();
|
||||||
|
|
||||||
|
/// Load the puppy [`Context`] from anywhere!
|
||||||
|
///
|
||||||
|
/// This gives you access to the data store and the configuration, without having to thread it through every place.
|
||||||
|
// WARNING: don't use this within this crate. there's a clippy lint.
|
||||||
|
pub fn context<T, E>(f: impl FnOnce(Context<'_>) -> Result<T, E>) -> Result<T, E>
|
||||||
|
where
|
||||||
|
E: From<Error>,
|
||||||
|
{
|
||||||
|
let cfg = Config {
|
||||||
|
ap_domain: String::from("test.piss-on.me"),
|
||||||
|
wf_domain: String::from("test.piss-on.me"),
|
||||||
|
};
|
||||||
|
let db = STORE
|
||||||
|
.get_or_try_init(|| Store::open(".state", schema()))
|
||||||
|
.map_err(Error::Store)?
|
||||||
|
.clone();
|
||||||
|
f(Context { db, config: &cfg })
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
use store::types::Schema;
|
||||||
|
|
||||||
|
/// Load a context for running tests in.
|
||||||
|
#[cfg(test)]
|
||||||
|
pub fn test_context<T>(
|
||||||
|
config: Config,
|
||||||
|
schema: Schema,
|
||||||
|
test: impl FnOnce(Context<'_>) -> Result<T>,
|
||||||
|
) -> Result<T> {
|
||||||
|
Store::test(schema, |db| test(Context { config: &config, db }))
|
||||||
|
}
|
215
lib/puppy/src/data.rs
Normal file
215
lib/puppy/src/data.rs
Normal file
|
@ -0,0 +1,215 @@
|
||||||
|
//! Datas for the data store!
|
||||||
|
//!
|
||||||
|
//! This module contains the definitions for the data store.
|
||||||
|
|
||||||
|
use bincode::{Decode, Encode};
|
||||||
|
use derive_more::Display;
|
||||||
|
use store::{types::Schema, Alias, Arrow, Key, Mixin};
|
||||||
|
|
||||||
|
/// *Bites you*
|
||||||
|
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
||||||
|
pub struct Bite {
|
||||||
|
#[identity]
|
||||||
|
pub id: Key,
|
||||||
|
#[origin]
|
||||||
|
pub biter: Key,
|
||||||
|
#[target]
|
||||||
|
pub victim: Key,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Properties of ActivityPub objects.
|
||||||
|
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
||||||
|
pub struct Object {
|
||||||
|
/// The ActivityPub id of the object.
|
||||||
|
#[index]
|
||||||
|
pub id: Id,
|
||||||
|
/// What kind of object is it?
|
||||||
|
pub kind: ObjectKind,
|
||||||
|
/// Whether or not the object resides on this server or on another one.
|
||||||
|
pub local: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Allows case analysis on the type of ActivityPub objects.
|
||||||
|
#[derive(Encode, Decode, Debug, Clone)]
|
||||||
|
pub enum ObjectKind {
|
||||||
|
Actor,
|
||||||
|
Activity(ActivityKind),
|
||||||
|
Notelike(String),
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The type of an activity.
|
||||||
|
#[derive(Encode, Decode, Debug, Clone)]
|
||||||
|
pub enum ActivityKind {
|
||||||
|
/// Used for posting stuff!
|
||||||
|
Create = 0,
|
||||||
|
/// Represents a follow request.
|
||||||
|
Follow = 1,
|
||||||
|
/// Used to signal that a follow request was accepted.
|
||||||
|
Accept = 2,
|
||||||
|
/// Used to reject a follow request.
|
||||||
|
Reject = 3,
|
||||||
|
/// See [`bites`](crate::bites).
|
||||||
|
Bite = 4,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
||||||
|
pub struct Channel {
|
||||||
|
pub inbox: String,
|
||||||
|
// TODO: add public key here
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A predicate; `follower` "follows" `followed`.
|
||||||
|
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
||||||
|
pub struct Follows {
|
||||||
|
#[origin]
|
||||||
|
pub follower: Key,
|
||||||
|
#[target]
|
||||||
|
pub followed: Key,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// An instance of a request from some `origin` user to follow a `target` user.
|
||||||
|
///
|
||||||
|
/// This should not be used to determine whether two actors are following each other. For that, use
|
||||||
|
/// [`Follows`], a basic arrow for exactly this purpose. *This* arrow is used to identify specific
|
||||||
|
/// instances of *requests*, and serves mostly as a historical reference and for synchronizing with
|
||||||
|
/// other servers.
|
||||||
|
///
|
||||||
|
/// Used to represent a `Follow` activity.
|
||||||
|
///
|
||||||
|
/// Mixins always present for the `id`:
|
||||||
|
///
|
||||||
|
/// - [`Status`], carrying the status of the request.
|
||||||
|
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
||||||
|
pub struct FollowRequest {
|
||||||
|
/// The unique ID of this particular request.
|
||||||
|
#[identity]
|
||||||
|
pub id: Key,
|
||||||
|
/// The "follower", the user that made the request.
|
||||||
|
pub origin: Key,
|
||||||
|
/// The one the request is made to.
|
||||||
|
pub target: Key,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The status of a [`FollowRequest`].
|
||||||
|
///
|
||||||
|
/// Valid state transitions:
|
||||||
|
///
|
||||||
|
/// ```text
|
||||||
|
/// ┌──────────────▶ Rejected
|
||||||
|
/// │
|
||||||
|
/// │
|
||||||
|
/// │
|
||||||
|
///
|
||||||
|
/// None ─────────▶ Pending ────────▶ Accepted
|
||||||
|
///
|
||||||
|
/// │ │
|
||||||
|
/// │ │
|
||||||
|
/// │ │
|
||||||
|
/// ▼ │
|
||||||
|
/// Withdrawn ◀────────────┘
|
||||||
|
/// ```
|
||||||
|
///
|
||||||
|
/// In addition, a follow request will be deleted if either endpoint is removed from the graph.
|
||||||
|
#[derive(Mixin, Encode, Decode, Eq, PartialEq, Clone)]
|
||||||
|
pub enum Status {
|
||||||
|
/// The follow request was previously pending or accepted, but since withdrawn.
|
||||||
|
///
|
||||||
|
/// This can happen when someone cancels their follow request or unfollows the target.
|
||||||
|
Withdrawn,
|
||||||
|
/// The follow request was accepted.
|
||||||
|
Accepted,
|
||||||
|
/// The follow request was denied.
|
||||||
|
Rejected,
|
||||||
|
/// The follow request is still under review.
|
||||||
|
Pending,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// An ActivityPub ID, used to look up remote objects by their canonical URL.
|
||||||
|
#[derive(Alias, Encode, Decode, Clone, PartialEq, Eq, Debug, Hash, Display)]
|
||||||
|
pub struct Id(pub String);
|
||||||
|
|
||||||
|
/// A "profile" in the social media sense.
|
||||||
|
///
|
||||||
|
/// Contains all presentation information about someone making posts.
|
||||||
|
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
||||||
|
pub struct Profile {
|
||||||
|
/// How many posts has this user made?
|
||||||
|
pub post_count: usize,
|
||||||
|
/// The name used for the profile's handle.
|
||||||
|
#[index] // <- currently doesnt do anything but i have an idea
|
||||||
|
pub account_name: Username,
|
||||||
|
/// The name displayed above their posts.
|
||||||
|
pub display_name: Option<String>,
|
||||||
|
/// The "bio", a freeform "about me" field.
|
||||||
|
pub about_string: Option<String>,
|
||||||
|
/// Arbitrary custom metadata fields.
|
||||||
|
pub about_fields: Vec<(String, String)>,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A unique name for an actor that is part of their "handle".
|
||||||
|
#[derive(Alias, Encode, Decode, Clone, PartialEq, Eq, Debug, Hash, Display)]
|
||||||
|
pub struct Username(pub String);
|
||||||
|
|
||||||
|
/// The relation that `author` has constructed and published `object`.
|
||||||
|
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
||||||
|
pub struct AuthorOf {
|
||||||
|
#[origin]
|
||||||
|
pub author: Key,
|
||||||
|
#[target]
|
||||||
|
pub object: Key,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The contents of a post.
|
||||||
|
#[derive(Mixin, Encode, Decode, Debug, Clone, Default)]
|
||||||
|
pub struct Content {
|
||||||
|
/// Main post body.
|
||||||
|
pub content: Option<String>,
|
||||||
|
/// Content warning for the post.
|
||||||
|
pub warning: Option<String>,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A public key used for verifying requests.
|
||||||
|
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
||||||
|
pub struct PublicKey {
|
||||||
|
pub key_id: String,
|
||||||
|
pub key_pem: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// A private key for signing requests with.
|
||||||
|
#[derive(Mixin, Encode, Decode, Clone)]
|
||||||
|
pub struct PrivateKey {
|
||||||
|
pub key_pem: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Represents a `Create` activity.
|
||||||
|
#[derive(Arrow)]
|
||||||
|
pub struct Create {
|
||||||
|
#[identity]
|
||||||
|
pub id: Key,
|
||||||
|
#[origin]
|
||||||
|
pub actor: Key,
|
||||||
|
#[target]
|
||||||
|
pub object: Key,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Construct the schema.
|
||||||
|
pub fn schema() -> Schema {
|
||||||
|
Schema::new()
|
||||||
|
// Mixins
|
||||||
|
.has::<Profile>()
|
||||||
|
.has::<Content>()
|
||||||
|
.has::<Status>()
|
||||||
|
.has::<Object>()
|
||||||
|
.has::<Channel>()
|
||||||
|
.has::<PrivateKey>()
|
||||||
|
.has::<PublicKey>()
|
||||||
|
// Aliases
|
||||||
|
.has::<Username>()
|
||||||
|
.has::<Id>()
|
||||||
|
// Arrows
|
||||||
|
.has::<Bite>()
|
||||||
|
.has::<FollowRequest>()
|
||||||
|
.has::<AuthorOf>()
|
||||||
|
.has::<Follows>()
|
||||||
|
.has::<Create>()
|
||||||
|
}
|
|
@ -1,234 +0,0 @@
|
||||||
//! Follow requests and related stuff.
|
|
||||||
|
|
||||||
use bincode::{Decode, Encode};
|
|
||||||
use store::{util::IterExt, Arrow, Error, Key, Mixin, Store, OK};
|
|
||||||
|
|
||||||
/// A predicate; `follower` "follows" `followed`.
|
|
||||||
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
|
||||||
pub struct Follows {
|
|
||||||
#[origin]
|
|
||||||
pub follower: Key,
|
|
||||||
#[target]
|
|
||||||
pub followed: Key,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// An instance of a request from some `origin` user to follow a `target` user.
|
|
||||||
///
|
|
||||||
/// This should not be used to determine whether two actors are following each other. For that, use
|
|
||||||
/// [`Follows`], a basic arrow for exactly this purpose. *This* arrow is used to identify specific
|
|
||||||
/// instances of *requests*, and serves mostly as a historical reference and for synchronizing with
|
|
||||||
/// other servers.
|
|
||||||
///
|
|
||||||
/// Mixins always present for the `id`:
|
|
||||||
///
|
|
||||||
/// - [`Status`], carrying the status of the request.
|
|
||||||
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
|
||||||
pub struct FollowRequest {
|
|
||||||
/// The unique ID of this particular request.
|
|
||||||
#[identity]
|
|
||||||
pub id: Key,
|
|
||||||
/// The "follower", the user that made the request.
|
|
||||||
pub origin: Key,
|
|
||||||
/// The one the request is made to.
|
|
||||||
pub target: Key,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl FollowRequest {
|
|
||||||
/// Determine if this follow request is pending.
|
|
||||||
pub fn is_pending(&self, db: &Store) -> store::Result<bool> {
|
|
||||||
// The status is stored as a mixin, so we need to get it.
|
|
||||||
let Some(st) = db.get_mixin::<Status>(self.id)? else {
|
|
||||||
// If we don't have a status for a follow request, something is borked.
|
|
||||||
return Err(Error::Missing);
|
|
||||||
};
|
|
||||||
// If the status of the follow request is pending, it can't also be true that the follows
|
|
||||||
// relation already exists.
|
|
||||||
debug_assert! {
|
|
||||||
!(st == Status::Pending)
|
|
||||||
|| db.exists::<Follows>(self.origin, self.target).map(|x| !x)?,
|
|
||||||
"fr.is_pending -> !(fr.origin follows fr.target)"
|
|
||||||
};
|
|
||||||
Ok(st == Status::Pending)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// The status of a [`FollowRequest`].
|
|
||||||
///
|
|
||||||
/// Valid state transitions:
|
|
||||||
///
|
|
||||||
/// ```text
|
|
||||||
/// ┌──────────────▶ Rejected
|
|
||||||
/// │
|
|
||||||
/// │
|
|
||||||
/// │
|
|
||||||
///
|
|
||||||
/// None ─────────▶ Pending ────────▶ Accepted
|
|
||||||
///
|
|
||||||
/// │ │
|
|
||||||
/// │ │
|
|
||||||
/// │ │
|
|
||||||
/// ▼ │
|
|
||||||
/// Withdrawn ◀────────────┘
|
|
||||||
/// ```
|
|
||||||
///
|
|
||||||
/// In addition, a follow request will be deleted if either endpoint is removed from the graph.
|
|
||||||
#[derive(Mixin, Encode, Decode, Eq, PartialEq, Clone)]
|
|
||||||
pub enum Status {
|
|
||||||
/// The follow request was previously pending or accepted, but since withdrawn.
|
|
||||||
///
|
|
||||||
/// This can happen when someone cancels their follow request or unfollows the target.
|
|
||||||
Withdrawn,
|
|
||||||
/// The follow request was accepted.
|
|
||||||
Accepted,
|
|
||||||
/// The follow request was denied.
|
|
||||||
Rejected,
|
|
||||||
/// The follow request is still under review.
|
|
||||||
Pending,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Request to follow another actor.
|
|
||||||
pub fn request(db: &Store, requester: Key, target: Key) -> store::Result<FollowRequest> {
|
|
||||||
db.run(|tx| {
|
|
||||||
let req = FollowRequest {
|
|
||||||
id: Key::gen(),
|
|
||||||
origin: requester,
|
|
||||||
target,
|
|
||||||
};
|
|
||||||
tx.create(req)?;
|
|
||||||
tx.add_mixin(req.id, Status::Pending)?;
|
|
||||||
Ok(req)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Accept the open follow request from `requester` to `target`, if one exists.
|
|
||||||
pub fn accept(db: &Store, requester: Key, target: Key) -> store::Result<()> {
|
|
||||||
db.run(|tx| {
|
|
||||||
// TODO: This logic is a little broken but it'll do for now. i'll fix it later.
|
|
||||||
let fr = tx
|
|
||||||
.between::<FollowRequest>(requester, target)
|
|
||||||
.filter(|fr| fr.as_ref().is_ok_and(|f| f.target == target))
|
|
||||||
// We'll want the latest one, because that one was inserted last so it'll be the most
|
|
||||||
// recent
|
|
||||||
.last()
|
|
||||||
.ok_or_else(|| Error::Missing)??;
|
|
||||||
// Only apply the update if the last follow request is still in a pending state.
|
|
||||||
if let Some(Status::Pending) = db.get_mixin(fr.id)? {
|
|
||||||
tx.update(fr.id, |_| Status::Accepted)?;
|
|
||||||
tx.create(Follows {
|
|
||||||
follower: requester,
|
|
||||||
followed: target,
|
|
||||||
})?;
|
|
||||||
}
|
|
||||||
OK
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
pub fn reject(db: &Store, request: Key) -> store::Result<()> {
|
|
||||||
db.run(|tx| {
|
|
||||||
tx.update(request, |_| Status::Rejected)?;
|
|
||||||
OK
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
/// List all pending follow requests for a user.
|
|
||||||
pub fn list_pending(db: &Store, target: Key) -> store::Result<Vec<FollowRequest>> {
|
|
||||||
db.incoming::<FollowRequest>(target)
|
|
||||||
.filter_bind_results(|req| Ok(if req.is_pending(db)? { Some(req) } else { None }))
|
|
||||||
.collect()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Get all actors followed by `actor`.
|
|
||||||
pub fn following_of(db: &Store, actor: Key) -> store::Result<Vec<Key>> {
|
|
||||||
db.outgoing::<Follows>(actor)
|
|
||||||
.map_ok(|a| a.followed)
|
|
||||||
.collect()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Get all actors following `actor`.
|
|
||||||
pub fn followers_of(db: &Store, actor: Key) -> store::Result<Vec<Key>> {
|
|
||||||
db.incoming::<Follows>(actor)
|
|
||||||
.map_ok(|a| a.follower)
|
|
||||||
.collect()
|
|
||||||
}
|
|
||||||
|
|
||||||
#[cfg(test)]
|
|
||||||
mod tests {
|
|
||||||
use store::{Key, Store, OK};
|
|
||||||
|
|
||||||
use crate::{
|
|
||||||
create_actor,
|
|
||||||
model::{schema, FollowRequest, Follows},
|
|
||||||
};
|
|
||||||
|
|
||||||
fn make_test_actors(db: &Store) -> store::Result<(Key, Key)> {
|
|
||||||
let alice = create_actor(&db, "alice")?;
|
|
||||||
let bob = create_actor(&db, "bob")?;
|
|
||||||
eprintln!("alice={alice}, bob={bob}");
|
|
||||||
Ok((alice, bob))
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn create_fr() -> store::Result<()> {
|
|
||||||
Store::test(schema(), |db| {
|
|
||||||
let (alice, bob) = make_test_actors(&db)?;
|
|
||||||
super::request(&db, alice, bob)?;
|
|
||||||
assert!(
|
|
||||||
db.exists::<FollowRequest>(alice, bob)?,
|
|
||||||
"(alice -> bob) ∈ follow-requested"
|
|
||||||
);
|
|
||||||
assert!(
|
|
||||||
!db.exists::<Follows>(alice, bob)?,
|
|
||||||
"(alice -> bob) ∉ follows"
|
|
||||||
);
|
|
||||||
let pending_for_bob = super::list_pending(&db, bob)?
|
|
||||||
.into_iter()
|
|
||||||
.map(|fr| fr.origin)
|
|
||||||
.collect::<Vec<_>>();
|
|
||||||
assert_eq!(pending_for_bob, vec![alice], "bob.pending = {{alice}}");
|
|
||||||
OK
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn accept_fr() -> store::Result<()> {
|
|
||||||
Store::test(schema(), |db| {
|
|
||||||
let (alice, bob) = make_test_actors(&db)?;
|
|
||||||
super::request(&db, alice, bob)?;
|
|
||||||
super::accept(&db, alice, bob)?;
|
|
||||||
|
|
||||||
assert!(
|
|
||||||
db.exists::<Follows>(alice, bob)?,
|
|
||||||
"(alice -> bob) ∈ follows"
|
|
||||||
);
|
|
||||||
assert!(
|
|
||||||
!db.exists::<Follows>(bob, alice)?,
|
|
||||||
"(bob -> alice) ∉ follows"
|
|
||||||
);
|
|
||||||
|
|
||||||
let pending_for_bob = super::list_pending(&db, bob)?;
|
|
||||||
assert!(pending_for_bob.is_empty(), "bob.pending = ∅");
|
|
||||||
|
|
||||||
let followers_of_bob = super::followers_of(&db, bob)?;
|
|
||||||
assert_eq!(followers_of_bob, vec![alice], "bob.followers = {{alice}}");
|
|
||||||
|
|
||||||
OK
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn listing_follow_relations() -> store::Result<()> {
|
|
||||||
Store::test(schema(), |db| {
|
|
||||||
let (alice, bob) = make_test_actors(&db)?;
|
|
||||||
super::request(&db, alice, bob)?;
|
|
||||||
super::accept(&db, alice, bob)?;
|
|
||||||
|
|
||||||
let followers_of_bob = super::followers_of(&db, bob)?;
|
|
||||||
assert_eq!(followers_of_bob, vec![alice], "bob.followers = {{alice}}");
|
|
||||||
|
|
||||||
let following_of_alice = super::following_of(&db, alice)?;
|
|
||||||
assert_eq!(following_of_alice, vec![bob], "alice.following = {{bob}}");
|
|
||||||
|
|
||||||
OK
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
230
lib/puppy/src/interact.rs
Normal file
230
lib/puppy/src/interact.rs
Normal file
|
@ -0,0 +1,230 @@
|
||||||
|
//! Interactions between actors.
|
||||||
|
|
||||||
|
use store::{util::IterExt as _, Key, StoreError};
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
actor::Actor,
|
||||||
|
data::{FollowRequest, Bite, Status, Follows},
|
||||||
|
Context, Error, Result,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// Interactions with other objects.
|
||||||
|
impl Actor {
|
||||||
|
/// Create a [`Bite`].
|
||||||
|
pub fn bite(&self, victim: &Actor) -> Bite {
|
||||||
|
Bite {
|
||||||
|
victim: victim.key,
|
||||||
|
biter: self.key,
|
||||||
|
id: Key::gen(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/// Construct a [`FollowRequest`].
|
||||||
|
pub fn follow_request(&self, target: &Actor) -> FollowRequest {
|
||||||
|
FollowRequest {
|
||||||
|
origin: self.key,
|
||||||
|
target: target.key,
|
||||||
|
id: Key::gen(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
/// Makes `biter` bite `victim` and inserts the records into the database.
|
||||||
|
pub fn do_bite(&self, cx: &Context, victim: &Actor) -> Result<Bite> {
|
||||||
|
let bite = self.bite(victim);
|
||||||
|
cx.run(|tx| {
|
||||||
|
tx.create(bite)?;
|
||||||
|
Ok(bite)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
/// Creates a follow request from `self` to `target`.
|
||||||
|
pub fn do_follow_request(&self, cx: &Context, target: &Actor) -> Result<FollowRequest> {
|
||||||
|
cx.run(|tx| {
|
||||||
|
let req = self.follow_request(target);
|
||||||
|
tx.create(req)?;
|
||||||
|
tx.add_mixin(req.id, Status::Pending)?;
|
||||||
|
Ok(req)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
/// Accept a follow request.
|
||||||
|
pub fn do_accept_request(&self, cx: &Context, req: FollowRequest) -> Result<()> {
|
||||||
|
debug_assert! {
|
||||||
|
self.key == req.target,
|
||||||
|
"only the target of a follow request may accept it"
|
||||||
|
};
|
||||||
|
cx.run(|tx| try {
|
||||||
|
let fr = tx
|
||||||
|
.between::<FollowRequest>(req.origin, req.target)
|
||||||
|
// Get the one that is equal to `req`.
|
||||||
|
.filter(|fr| fr.as_ref().is_ok_and(|f| f.id == req.id))
|
||||||
|
.last()
|
||||||
|
.unwrap()?;
|
||||||
|
// Only apply the update if the follow request is still in a pending state.
|
||||||
|
if let Some(Status::Pending) = tx.get_mixin(fr.id)? {
|
||||||
|
tx.update(fr.id, |_| Status::Accepted)?;
|
||||||
|
tx.create(Follows {
|
||||||
|
follower: req.origin,
|
||||||
|
followed: req.target,
|
||||||
|
})?;
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
/// Reject a follow request.
|
||||||
|
pub fn do_reject_request(&self, cx: &Context, req: FollowRequest) -> Result<()> {
|
||||||
|
debug_assert! {
|
||||||
|
self.key == req.target,
|
||||||
|
"only the target of a follow request may accept it"
|
||||||
|
};
|
||||||
|
cx.run(|tx| try {
|
||||||
|
tx.update(req.id, |_| Status::Rejected)?;
|
||||||
|
})
|
||||||
|
}
|
||||||
|
/// Get all pending follow request for `self`.
|
||||||
|
pub fn pending_requests<'c>(
|
||||||
|
&self,
|
||||||
|
cx: &'c Context,
|
||||||
|
) -> impl Iterator<Item = Result<FollowRequest>> + 'c {
|
||||||
|
cx.store()
|
||||||
|
.incoming::<FollowRequest>(self.key)
|
||||||
|
.map_err(Error::Store)
|
||||||
|
.filter_bind_results(|req| Ok(if req.is_pending(cx)? { Some(req) } else { None }))
|
||||||
|
}
|
||||||
|
/// Get all nodes `self` is following.
|
||||||
|
pub fn following<'c>(&self, cx: &'c Context) -> impl Iterator<Item = Result<Key>> + 'c {
|
||||||
|
cx.store()
|
||||||
|
.outgoing::<Follows>(self.key)
|
||||||
|
.map_err(Error::Store)
|
||||||
|
.map_ok(|a| a.followed)
|
||||||
|
}
|
||||||
|
/// Get all followers of `self`.
|
||||||
|
pub fn followers<'c>(&self, cx: &'c Context) -> impl Iterator<Item = Result<Key>> + 'c {
|
||||||
|
cx.store()
|
||||||
|
.incoming::<Follows>(self.key)
|
||||||
|
.map_err(Error::Store)
|
||||||
|
.map_ok(|a| a.follower)
|
||||||
|
}
|
||||||
|
/// List all specific times `self` was bitten.
|
||||||
|
pub fn bites_suffered<'c>(&self, cx: &'c Context) -> impl Iterator<Item = Result<Bite>> + 'c {
|
||||||
|
cx.store().incoming::<Bite>(self.key).map_err(Error::Store)
|
||||||
|
}
|
||||||
|
/// Check whether `self` follows `other`.
|
||||||
|
pub fn follows(&self, cx: &Context, other: &Actor) -> Result<bool> {
|
||||||
|
try { cx.db.exists::<Follows>(self.key, other.key)? }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl FollowRequest {
|
||||||
|
/// Determine if this follow request is pending.
|
||||||
|
pub fn is_pending(&self, cx: &Context) -> Result<bool> {
|
||||||
|
// The status is stored as a mixin, so we need to get it.
|
||||||
|
let Some(st) = cx.db.get_mixin::<Status>(self.id)? else {
|
||||||
|
// If we don't have a status for a follow request, something is borked.
|
||||||
|
return Err(StoreError::Missing.into());
|
||||||
|
};
|
||||||
|
// If the status of the follow request is pending, it can't also be true that the follows
|
||||||
|
// relation already exists.
|
||||||
|
debug_assert! {
|
||||||
|
!(st == Status::Pending)
|
||||||
|
|| cx.db.exists::<Follows>(self.origin, self.target).map(|x| !x)?,
|
||||||
|
"fr.is_pending -> !(fr.origin follows fr.target)"
|
||||||
|
};
|
||||||
|
Ok(st == Status::Pending)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use store::util::IterExt as _;
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
actor::{create_local as create_actor, Actor},
|
||||||
|
config::Config,
|
||||||
|
data::{schema, FollowRequest, Follows},
|
||||||
|
test_context, Context, Result,
|
||||||
|
};
|
||||||
|
|
||||||
|
fn make_test_actors(cx: &Context) -> Result<(Actor, Actor)> {
|
||||||
|
let alice = create_actor(&cx, "alice")?;
|
||||||
|
let bob = create_actor(&cx, "bob")?;
|
||||||
|
eprintln!("alice={alice:?}, bob={bob:?}");
|
||||||
|
Ok((alice, bob))
|
||||||
|
}
|
||||||
|
fn test_config() -> Config {
|
||||||
|
Config {
|
||||||
|
ap_domain: String::from("unit-test.puppy.gay"),
|
||||||
|
wf_domain: String::from("unit-test.puppy.gay"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_fr() -> Result<()> {
|
||||||
|
test_context(test_config(), schema(), |cx| try {
|
||||||
|
let db = cx.store();
|
||||||
|
let (alice, bob) = make_test_actors(&cx)?;
|
||||||
|
alice.do_follow_request(&cx, &bob)?;
|
||||||
|
assert!(
|
||||||
|
db.exists::<FollowRequest>(alice.key, bob.key)?,
|
||||||
|
"(alice -> bob) ∈ follow-requested"
|
||||||
|
);
|
||||||
|
assert!(
|
||||||
|
!db.exists::<Follows>(alice.key, bob.key)?,
|
||||||
|
"(alice -> bob) ∉ follows"
|
||||||
|
);
|
||||||
|
let pending_for_bob = bob
|
||||||
|
.pending_requests(&cx)
|
||||||
|
.map_ok(|fr| fr.origin)
|
||||||
|
.try_collect::<Vec<_>>()?;
|
||||||
|
assert_eq!(pending_for_bob, vec![alice.key], "bob.pending = {{alice}}");
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn accept_fr() -> Result<()> {
|
||||||
|
test_context(test_config(), schema(), |cx| {
|
||||||
|
let db = cx.store();
|
||||||
|
let (alice, bob) = make_test_actors(&cx)?;
|
||||||
|
let req = alice.do_follow_request(&cx, &bob)?;
|
||||||
|
bob.do_accept_request(&cx, req)?;
|
||||||
|
|
||||||
|
assert!(
|
||||||
|
db.exists::<Follows>(alice.key, bob.key)?,
|
||||||
|
"(alice -> bob) ∈ follows"
|
||||||
|
);
|
||||||
|
assert!(
|
||||||
|
!db.exists::<Follows>(bob.key, alice.key)?,
|
||||||
|
"(bob -> alice) ∉ follows"
|
||||||
|
);
|
||||||
|
|
||||||
|
let pending_for_bob: Vec<_> = bob.pending_requests(&cx).try_collect()?;
|
||||||
|
assert!(pending_for_bob.is_empty(), "bob.pending = ∅");
|
||||||
|
|
||||||
|
let followers_of_bob: Vec<_> = bob.followers(&cx).try_collect()?;
|
||||||
|
assert_eq!(
|
||||||
|
followers_of_bob,
|
||||||
|
vec![alice.key],
|
||||||
|
"bob.followers = {{alice}}"
|
||||||
|
);
|
||||||
|
Ok(())
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn listing_follow_relations() -> Result<()> {
|
||||||
|
test_context(test_config(), schema(), |cx| try {
|
||||||
|
let (alice, bob) = make_test_actors(&cx)?;
|
||||||
|
let req = alice.do_follow_request(&cx, &bob)?;
|
||||||
|
bob.do_accept_request(&cx, req)?;
|
||||||
|
|
||||||
|
let followers_of_bob: Vec<_> = bob.followers(&cx).try_collect()?;
|
||||||
|
assert_eq!(
|
||||||
|
followers_of_bob,
|
||||||
|
vec![alice.key],
|
||||||
|
"bob.followers = {{alice}}"
|
||||||
|
);
|
||||||
|
|
||||||
|
let following_of_alice: Vec<_> = alice.following(&cx).try_collect()?;
|
||||||
|
assert_eq!(
|
||||||
|
following_of_alice,
|
||||||
|
vec![bob.key],
|
||||||
|
"alice.following = {{bob}}"
|
||||||
|
);
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
|
@ -1,138 +1,128 @@
|
||||||
#![feature(iterator_try_collect, try_blocks)]
|
// Working with result types is such a bitch without these.
|
||||||
use model::{Channel, Id, Object, ObjectKind, Profile, Username};
|
#![feature(iterator_try_collect, try_blocks, once_cell_try)]
|
||||||
pub use store::{self, Key, Store};
|
// Cause an error if someone tries to call [`context`] from within this crate. If we need one,
|
||||||
use store::{Error, Transaction, OK};
|
// it must be passed in as a parameter. Getting a hold of a context is not our job. We need to
|
||||||
|
// forbid internal code from calling the function because it is stateful, and that can cause
|
||||||
|
// spooky action at a distance within tests.
|
||||||
|
//
|
||||||
|
// Ideally this would be enforced by the compiler, and we *can* enforce it at the type level,
|
||||||
|
// but that would make every type signature ever 100x more complicated, so we're not doing it.
|
||||||
|
#![deny(clippy::disallowed_methods, clippy::disallowed_types)]
|
||||||
|
|
||||||
pub mod model {
|
pub use context::{context, Context};
|
||||||
use bincode::{Decode, Encode};
|
#[cfg(test)]
|
||||||
use derive_more::Display;
|
pub use context::test_context;
|
||||||
use store::{types::Schema, Alias, Mixin};
|
|
||||||
|
|
||||||
use crate::follows::Status;
|
use data::{ActivityKind, Channel, Content, Create, Id, Object, ObjectKind, Profile, PublicKey};
|
||||||
pub use crate::{
|
|
||||||
bites::Bite,
|
|
||||||
follows::{FollowRequest, Follows},
|
|
||||||
post::{AuthorOf, Content},
|
|
||||||
};
|
|
||||||
|
|
||||||
/// A "profile" in the social media sense.
|
pub use store::{self, Key, StoreError};
|
||||||
///
|
pub use fetch;
|
||||||
/// Contains all presentation information about someone making posts.
|
|
||||||
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
|
||||||
pub struct Profile {
|
|
||||||
/// How many posts has this user made?
|
|
||||||
pub post_count: usize,
|
|
||||||
/// The name used for the profile's handle.
|
|
||||||
#[index] // <- currently doesnt do anything but i have an idea
|
|
||||||
pub account_name: Username,
|
|
||||||
/// The name displayed above their posts.
|
|
||||||
pub display_name: Option<String>,
|
|
||||||
/// The "bio", a freeform "about me" field.
|
|
||||||
pub about_string: Option<String>,
|
|
||||||
/// Arbitrary custom metadata fields.
|
|
||||||
pub about_fields: Vec<(String, String)>,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Properties of ActivityPub objects.
|
mod context;
|
||||||
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
pub mod data;
|
||||||
pub struct Object {
|
|
||||||
#[index]
|
|
||||||
pub id: Id,
|
|
||||||
pub kind: ObjectKind,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Allows case analysis on the type of ActivityPub objects.
|
pub mod post;
|
||||||
#[derive(Encode, Decode, Debug, Clone)]
|
mod interact;
|
||||||
pub enum ObjectKind {
|
|
||||||
Actor,
|
|
||||||
Activity(ActivityKind),
|
|
||||||
Notelike(String),
|
|
||||||
}
|
|
||||||
|
|
||||||
/// The type of an activity.
|
|
||||||
#[derive(Encode, Decode, Debug, Clone)]
|
|
||||||
pub enum ActivityKind {
|
|
||||||
/// Used for posting stuff!
|
|
||||||
Create = 0,
|
|
||||||
/// Represents a follow request.
|
|
||||||
Follow = 1,
|
|
||||||
/// Used to signal that a follow request was accepted.
|
|
||||||
Accept = 2,
|
|
||||||
/// Used to reject a follow request.
|
|
||||||
Reject = 3,
|
|
||||||
/// See [`bites`](crate::bites).
|
|
||||||
Bite = 4,
|
|
||||||
}
|
|
||||||
|
|
||||||
#[derive(Mixin, Encode, Decode, Debug, Clone)]
|
|
||||||
pub struct Channel {
|
|
||||||
pub inbox: String,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// An ActivityPub ID, used to look up remote objects by their canonical URL.
|
|
||||||
#[derive(Alias, Encode, Decode, Clone, PartialEq, Eq, Debug, Hash, Display)]
|
|
||||||
pub struct Id(pub String);
|
|
||||||
|
|
||||||
/// A unique name for an actor that is part of their "handle".
|
|
||||||
#[derive(Alias, Encode, Decode, Clone, PartialEq, Eq, Debug, Hash, Display)]
|
|
||||||
pub struct Username(pub String);
|
|
||||||
|
|
||||||
/// Construct the schema.
|
|
||||||
pub fn schema() -> Schema {
|
|
||||||
Schema::new()
|
|
||||||
// Mixins
|
|
||||||
.has::<Profile>()
|
|
||||||
.has::<Content>()
|
|
||||||
.has::<Status>()
|
|
||||||
.has::<Object>()
|
|
||||||
.has::<Channel>()
|
|
||||||
// Aliases
|
|
||||||
.has::<Username>()
|
|
||||||
.has::<Id>()
|
|
||||||
// Arrows
|
|
||||||
.has::<Bite>()
|
|
||||||
.has::<FollowRequest>()
|
|
||||||
.has::<AuthorOf>()
|
|
||||||
.has::<Follows>()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Retrieve an ActivityPub object from the database.
|
/// Retrieve an ActivityPub object from the database.
|
||||||
///
|
///
|
||||||
/// Fails with `Error::Missing` if the required properties are not present.
|
/// Fails with `Error::Missing` if the required properties are not present.
|
||||||
pub fn get_local_ap_object(db: &Store, key: Key) -> store::Result<fetch::Object> {
|
pub fn get_local_ap_object(cx: &Context<'_>, key: Key) -> Result<fetch::Object> {
|
||||||
let Some(obj) = db.get_mixin::<Object>(key)? else {
|
let Some(obj) = cx.db.get_mixin::<Object>(key)? else {
|
||||||
return Err(Error::Missing);
|
// We need this data in order to determine the object type. If the passed key does not
|
||||||
|
// have this data, it must not be an ActivityPub object.
|
||||||
|
return Err(Error::MissingData { node: key, prop: "Object" });
|
||||||
};
|
};
|
||||||
match obj.kind {
|
match obj.kind {
|
||||||
ObjectKind::Actor => {
|
ObjectKind::Actor => {
|
||||||
let Some(Profile { account_name, display_name, .. }) = db.get_mixin(key)? else {
|
let Some(Profile { account_name, display_name, .. }) = cx.db.get_mixin(key)? else {
|
||||||
return Err(Error::Missing);
|
return Err(Error::MissingData { node: key, prop: "Profile" });
|
||||||
};
|
};
|
||||||
let Some(Channel { inbox }) = db.get_mixin(key)? else {
|
let Some(Channel { inbox }) = cx.db.get_mixin(key)? else {
|
||||||
return Err(Error::Missing);
|
return Err(Error::MissingData { node: key, prop: "Channel" });
|
||||||
|
};
|
||||||
|
let Some(PublicKey { key_id, key_pem }) = cx.db.get_mixin(key)? else {
|
||||||
|
return Err(Error::MissingData { node: key, prop: "PublicKey" });
|
||||||
};
|
};
|
||||||
Ok(fetch::Object::Actor(fetch::Actor {
|
Ok(fetch::Object::Actor(fetch::Actor {
|
||||||
id: obj.id.0.into(),
|
id: obj.id.0.clone().into(),
|
||||||
inbox: inbox.into(),
|
inbox: inbox.into(),
|
||||||
account_name: account_name.0,
|
account_name: account_name.0,
|
||||||
display_name,
|
display_name,
|
||||||
|
public_key: fetch::PublicKey {
|
||||||
|
owner: obj.id.0.into(),
|
||||||
|
key_id: key_id.into(),
|
||||||
|
public_key_pem: key_pem,
|
||||||
|
},
|
||||||
}))
|
}))
|
||||||
}
|
}
|
||||||
ObjectKind::Activity(_) => {
|
ObjectKind::Activity(ActivityKind::Create) => {
|
||||||
todo!()
|
let Some(Create { object, actor, .. }) = cx.db.get_arrow(key)? else {
|
||||||
|
panic!("expected a `Create`");
|
||||||
|
};
|
||||||
|
let Id(actor) = cx.db.get_alias(actor)?.unwrap();
|
||||||
|
Ok(fetch::Object::Activity(fetch::Activity {
|
||||||
|
id: obj.id.0.into(),
|
||||||
|
actor: actor.into(),
|
||||||
|
object: Box::new(get_local_ap_object(cx, object)?),
|
||||||
|
kind: String::from("Create"),
|
||||||
|
}))
|
||||||
}
|
}
|
||||||
ObjectKind::Notelike(_) => todo!(),
|
ObjectKind::Notelike(kind) => {
|
||||||
|
let Some(Content { content, warning, .. }) = cx.db.get_mixin(key)? else {
|
||||||
|
panic!()
|
||||||
|
};
|
||||||
|
Ok(fetch::Object::Object {
|
||||||
|
id: obj.id.0.clone().into(),
|
||||||
|
summary: warning,
|
||||||
|
content,
|
||||||
|
kind,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
_ => todo!(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Create a fresh local actor.
|
pub mod actor {
|
||||||
pub fn create_local_actor(db: &Store, domain: &str, username: impl ToString) -> store::Result<Key> {
|
use store::{Key, StoreError, Transaction};
|
||||||
|
|
||||||
|
use crate::{
|
||||||
|
data::{Channel, Id, Object, ObjectKind, PrivateKey, Profile, PublicKey, Username},
|
||||||
|
Context, Error, Result,
|
||||||
|
};
|
||||||
|
|
||||||
|
/// A reference to an actor.
|
||||||
|
#[derive(Clone, Eq, PartialEq, Debug)]
|
||||||
|
pub struct Actor {
|
||||||
|
/// The key identifying the actor in the data store.
|
||||||
|
pub key: Key,
|
||||||
|
local: bool,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Actor {
|
||||||
|
/// Get a local actor from the store by their username.
|
||||||
|
pub fn by_username(cx: &Context, username: impl ToString) -> Result<Option<Actor>> {
|
||||||
|
let maybe_key = cx
|
||||||
|
.store()
|
||||||
|
.lookup(Username(username.to_string()))
|
||||||
|
.map_err(Error::Store)?;
|
||||||
|
// For now, we only have local actors.
|
||||||
|
Ok(maybe_key.map(|key| Actor { key, local: true }))
|
||||||
|
}
|
||||||
|
/// Check whether the actor is local or not.
|
||||||
|
pub fn is_local(self) -> bool {
|
||||||
|
self.local
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a fresh local actor.
|
||||||
|
pub fn create_local(cx: &Context, username: impl ToString) -> Result<Actor> {
|
||||||
let key = Key::gen();
|
let key = Key::gen();
|
||||||
db.run(|tx| {
|
cx.run(|tx| {
|
||||||
let username: Username = username.to_string().into();
|
let username: Username = username.to_string().into();
|
||||||
// Federation stuff
|
// Federation stuff
|
||||||
mixin_ap_actor(tx, key, domain)?;
|
mixin_ap_actor(tx, key, &cx.config.ap_domain, true)?;
|
||||||
|
mixin_priv_key(tx, key, &cx.config.ap_domain)?;
|
||||||
// Social properties
|
// Social properties
|
||||||
tx.add_alias(key, username.clone())?;
|
tx.add_alias(key, username.clone())?;
|
||||||
tx.add_mixin(key, Profile {
|
tx.add_mixin(key, Profile {
|
||||||
|
@ -142,19 +132,65 @@ pub fn create_local_actor(db: &Store, domain: &str, username: impl ToString) ->
|
||||||
about_string: None,
|
about_string: None,
|
||||||
about_fields: Vec::new(),
|
about_fields: Vec::new(),
|
||||||
})?;
|
})?;
|
||||||
Ok(key)
|
Ok(Actor { key, local: true })
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Add properties related to ActivityPub actors to a vertex.
|
/// Add properties related to ActivityPub actors to a vertex.
|
||||||
pub fn mixin_ap_actor(tx: &Transaction<'_>, vertex: Key, domain: &str) -> store::Result<()> {
|
pub fn mixin_ap_actor(
|
||||||
let id = Id(format!("http://{domain}/o/{vertex}"));
|
tx: &Transaction<'_>,
|
||||||
|
vertex: Key,
|
||||||
|
domain: &str,
|
||||||
|
local: bool,
|
||||||
|
) -> Result<(), StoreError> {
|
||||||
|
let id = Id(format!("https://{domain}/o/{vertex}"));
|
||||||
tx.add_alias(vertex, id.clone())?;
|
tx.add_alias(vertex, id.clone())?;
|
||||||
tx.add_mixin(vertex, Channel { inbox: format!("{id}/inbox") })?;
|
tx.add_mixin(vertex, Channel { inbox: format!("{id}/inbox") })?;
|
||||||
tx.add_mixin(vertex, Object { id, kind: ObjectKind::Actor })?;
|
tx.add_mixin(vertex, Object {
|
||||||
OK
|
kind: ObjectKind::Actor,
|
||||||
|
local,
|
||||||
|
id,
|
||||||
|
})?;
|
||||||
|
store::OK
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Generate and attach a public/private key pair to the vertex.
|
||||||
|
pub fn mixin_priv_key(
|
||||||
|
tx: &Transaction<'_>,
|
||||||
|
vertex: Key,
|
||||||
|
domain: &str,
|
||||||
|
) -> Result<(), StoreError> {
|
||||||
|
let key_id = format!("https://{domain}/o/{vertex}#sig-key");
|
||||||
|
let (private, public) = fetch::keys::gen_keypair();
|
||||||
|
tx.add_mixin(vertex, PublicKey {
|
||||||
|
key_pem: public.to_pem(),
|
||||||
|
key_id,
|
||||||
|
})?;
|
||||||
|
tx.add_mixin(vertex, PrivateKey {
|
||||||
|
key_pem: dbg!(private.to_pem()),
|
||||||
|
})?;
|
||||||
|
store::OK
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub mod bites;
|
pub type Result<T, E = Error> = std::result::Result<T, E>;
|
||||||
pub mod post;
|
|
||||||
pub mod follows;
|
#[derive(derive_more::From, Debug)]
|
||||||
|
pub enum Error {
|
||||||
|
/// An error internal to the store.
|
||||||
|
Store(StoreError),
|
||||||
|
/// Expected `node` to have some property that it doesn't have.
|
||||||
|
MissingData {
|
||||||
|
/// The node that is missing the data.
|
||||||
|
node: Key,
|
||||||
|
/// Name of the thing it is missing.
|
||||||
|
prop: &'static str,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
pub mod config {
|
||||||
|
pub struct Config {
|
||||||
|
pub ap_domain: String,
|
||||||
|
pub wf_domain: String,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
|
@ -2,20 +2,66 @@
|
||||||
|
|
||||||
use std::ops::RangeBounds;
|
use std::ops::RangeBounds;
|
||||||
|
|
||||||
use bincode::{Decode, Encode};
|
|
||||||
use chrono::{DateTime, Utc};
|
use chrono::{DateTime, Utc};
|
||||||
use either::Either::{Left, Right};
|
use either::Either::{Left, Right};
|
||||||
use store::{util::IterExt as _, Arrow, Error, Key, Mixin, Result, Store, Transaction};
|
use store::{util::IterExt as _, Key, Store, StoreError, Transaction};
|
||||||
|
|
||||||
use crate::model::Profile;
|
use crate::{
|
||||||
|
data::{self, AuthorOf, Content, Id, Object, ObjectKind, PrivateKey, Profile, PublicKey},
|
||||||
|
Context,
|
||||||
|
};
|
||||||
|
|
||||||
/// The contents of a post.
|
#[derive(Clone, Debug)]
|
||||||
#[derive(Mixin, Encode, Decode, Debug, Clone, Default)]
|
pub struct Post {
|
||||||
pub struct Content {
|
pub key: Key,
|
||||||
/// Main post body.
|
}
|
||||||
pub content: Option<String>,
|
|
||||||
/// Content warning for the post.
|
impl Post {
|
||||||
pub warning: Option<String>,
|
pub async fn federate(&self, cx: &Context<'_>) -> crate::Result<()> {
|
||||||
|
use fetch::{Activity, Object};
|
||||||
|
let post @ Object::Object { .. } = crate::get_local_ap_object(cx, self.key)? else {
|
||||||
|
todo!()
|
||||||
|
};
|
||||||
|
let author = cx
|
||||||
|
.db
|
||||||
|
.incoming::<AuthorOf>(self.key)
|
||||||
|
.map_ok(|a| a.author)
|
||||||
|
.next()
|
||||||
|
.unwrap()?;
|
||||||
|
let author_id = cx.db.get_alias::<data::Id>(author)?.unwrap();
|
||||||
|
let activity = Activity {
|
||||||
|
id: post.id().clone().into(),
|
||||||
|
object: Box::new(post),
|
||||||
|
kind: String::from("Create"),
|
||||||
|
actor: author_id.0.into(),
|
||||||
|
};
|
||||||
|
let (private, public) = cx.db.get_mixin_many::<(PrivateKey, PublicKey)>(author)?;
|
||||||
|
// Insert the activity in the database.
|
||||||
|
cx.run(|tx| try {
|
||||||
|
let activity_key = Key::gen();
|
||||||
|
let id: data::Id = cx.mk_url(activity_key).into();
|
||||||
|
tx.add_mixin(activity_key, data::Object {
|
||||||
|
id: id.clone(),
|
||||||
|
kind: ObjectKind::Activity(data::ActivityKind::Create),
|
||||||
|
local: true,
|
||||||
|
})?;
|
||||||
|
tx.create(data::Create {
|
||||||
|
id: activity_key,
|
||||||
|
actor: author,
|
||||||
|
object: self.key,
|
||||||
|
})?;
|
||||||
|
tx.add_alias(activity_key, id)?;
|
||||||
|
})?;
|
||||||
|
// Send the requests.
|
||||||
|
fetch::deliver(
|
||||||
|
&private.key_pem,
|
||||||
|
&public.key_id,
|
||||||
|
activity,
|
||||||
|
"https://crimew.gay/users/ezri/inbox",
|
||||||
|
)
|
||||||
|
.await;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl From<&str> for Content {
|
impl From<&str> for Content {
|
||||||
|
@ -33,20 +79,11 @@ impl From<String> for Content {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/// The relation that `author` has constructed and published `object`.
|
|
||||||
#[derive(Arrow, Debug, PartialEq, Eq, Clone, Copy)]
|
|
||||||
pub struct AuthorOf {
|
|
||||||
#[origin]
|
|
||||||
pub author: Key,
|
|
||||||
#[target]
|
|
||||||
pub object: Key,
|
|
||||||
}
|
|
||||||
|
|
||||||
/// A piece of content posted by someone.
|
/// A piece of content posted by someone.
|
||||||
#[derive(Clone, Debug)]
|
#[derive(Clone, Debug)]
|
||||||
pub struct Post {
|
pub struct PostData {
|
||||||
/// The post's internal ID.
|
/// The post's internal ID.
|
||||||
pub id: Key,
|
pub id: Post,
|
||||||
/// The actual post contents.
|
/// The actual post contents.
|
||||||
pub content: Content,
|
pub content: Content,
|
||||||
/// Metadata about the post's author.
|
/// Metadata about the post's author.
|
||||||
|
@ -73,12 +110,12 @@ pub struct Timeline {
|
||||||
/// Discrete events that can be displayed to a user as part of a timeline.
|
/// Discrete events that can be displayed to a user as part of a timeline.
|
||||||
#[derive(Debug)]
|
#[derive(Debug)]
|
||||||
enum Item {
|
enum Item {
|
||||||
Post(Post),
|
Post(PostData),
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Item {
|
impl Item {
|
||||||
/// Get the timeline item if it is a [`Post`].
|
/// Get the timeline item if it is a [`Post`].
|
||||||
pub fn as_post(&self) -> Option<&Post> {
|
pub fn as_post(&self) -> Option<&PostData> {
|
||||||
match self {
|
match self {
|
||||||
Item::Post(ref post) => Some(post),
|
Item::Post(ref post) => Some(post),
|
||||||
}
|
}
|
||||||
|
@ -87,7 +124,7 @@ impl Item {
|
||||||
|
|
||||||
impl Timeline {
|
impl Timeline {
|
||||||
/// Get all the posts in the timeline.
|
/// Get all the posts in the timeline.
|
||||||
pub fn posts(&self) -> impl Iterator<Item = &Post> {
|
pub fn posts(&self) -> impl Iterator<Item = &PostData> {
|
||||||
self.items.iter().filter_map(|x| x.as_post())
|
self.items.iter().filter_map(|x| x.as_post())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -97,7 +134,7 @@ pub fn fetch_timeline(
|
||||||
db: &Store,
|
db: &Store,
|
||||||
time_range: impl RangeBounds<DateTime<Utc>>,
|
time_range: impl RangeBounds<DateTime<Utc>>,
|
||||||
limit: Option<usize>,
|
limit: Option<usize>,
|
||||||
) -> Result<Timeline> {
|
) -> Result<Timeline, StoreError> {
|
||||||
let posts = db.run(|tx| {
|
let posts = db.run(|tx| {
|
||||||
// Get all post content entries (the argument passed here is a range of chrono datetimes).
|
// Get all post content entries (the argument passed here is a range of chrono datetimes).
|
||||||
let iter = tx.range::<Content>(time_range);
|
let iter = tx.range::<Content>(time_range);
|
||||||
|
@ -107,28 +144,28 @@ pub fn fetch_timeline(
|
||||||
};
|
};
|
||||||
// Then, we're gonna map each of them to their author, and get the profile information needed to
|
// Then, we're gonna map each of them to their author, and get the profile information needed to
|
||||||
// render the post (mostly display name and handle).
|
// render the post (mostly display name and handle).
|
||||||
iter.bind_results(|(id, content)| {
|
iter.bind_results(|(key, content)| try {
|
||||||
// Take the first author. There is nothing stopping a post from having multiple authors, but
|
// Take the first author. There is nothing stopping a post from having multiple authors, but
|
||||||
// let's take it one step at a time.
|
// let's take it one step at a time.
|
||||||
let (author, Some(Profile { display_name, account_name, .. })) = tx
|
let (author, Some(Profile { display_name, account_name, .. })) = tx
|
||||||
.join_on(|a: AuthorOf| a.author, tx.incoming(id))?
|
.join_on(|a: AuthorOf| a.author, tx.incoming(key))?
|
||||||
.swap_remove(0)
|
.swap_remove(0)
|
||||||
else {
|
else {
|
||||||
// We expect all posts to have at least one author, so we should complain if there is one
|
// We expect all posts to have at least one author, so we should complain if there is one
|
||||||
// that doesn't (for now). For robustness, the `.collect()` down there should be replaced
|
// that doesn't (for now). For robustness, the `.collect()` down there should be replaced
|
||||||
// with a strategy where we log a warning instead of failing, but in the current state of
|
// with a strategy where we log a warning instead of failing, but in the current state of
|
||||||
// the project, failing fast is a good thing.
|
// the project, failing fast is a good thing.
|
||||||
return Err(Error::Missing);
|
return Err(StoreError::Missing);
|
||||||
};
|
};
|
||||||
Ok(Item::Post(Post {
|
Item::Post(PostData {
|
||||||
id,
|
id: Post { key },
|
||||||
author: Author {
|
author: Author {
|
||||||
id: author,
|
id: author,
|
||||||
handle: format!("@{account_name}"),
|
handle: format!("@{account_name}"),
|
||||||
display_name: display_name.unwrap_or(account_name.0),
|
display_name: display_name.unwrap_or(account_name.0),
|
||||||
},
|
},
|
||||||
content,
|
content,
|
||||||
}))
|
})
|
||||||
})
|
})
|
||||||
.collect()
|
.collect()
|
||||||
})?;
|
})?;
|
||||||
|
@ -136,8 +173,22 @@ pub fn fetch_timeline(
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Create a new post.
|
/// Create a new post.
|
||||||
pub fn create_post(db: &Store, author: Key, content: impl Into<Content>) -> store::Result<Key> {
|
pub fn create_post(cx: &Context, author: Key, content: impl Into<Content>) -> crate::Result<Post> {
|
||||||
db.run(|tx| mixin_post(tx, Key::gen(), author, content))
|
let content = content.into();
|
||||||
|
cx.run(|tx| {
|
||||||
|
let key = Key::gen();
|
||||||
|
// Local stuff
|
||||||
|
mixin_post(tx, key, author, content)?;
|
||||||
|
// Federation stuff
|
||||||
|
let id = Id(format!("https://{}/o/{key}", cx.config.ap_domain));
|
||||||
|
tx.add_alias(key, id.clone())?;
|
||||||
|
tx.add_mixin(key, Object {
|
||||||
|
kind: ObjectKind::Notelike("Note".to_string()),
|
||||||
|
local: true,
|
||||||
|
id,
|
||||||
|
})?;
|
||||||
|
Ok(Post { key })
|
||||||
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Add a post's mixins and predicates to an existing `node`.
|
/// Add a post's mixins and predicates to an existing `node`.
|
||||||
|
@ -146,7 +197,7 @@ pub fn mixin_post(
|
||||||
node: Key,
|
node: Key,
|
||||||
author: Key,
|
author: Key,
|
||||||
content: impl Into<Content>,
|
content: impl Into<Content>,
|
||||||
) -> store::Result<Key> {
|
) -> Result<Key, StoreError> {
|
||||||
tx.update::<Profile>(author, |mut profile| {
|
tx.update::<Profile>(author, |mut profile| {
|
||||||
profile.post_count += 1;
|
profile.post_count += 1;
|
||||||
profile
|
profile
|
||||||
|
@ -156,7 +207,7 @@ pub fn mixin_post(
|
||||||
Ok(node)
|
Ok(node)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn list_posts_by_author(db: &Store, author: Key) -> store::Result<Vec<(Key, Content)>> {
|
pub fn list_posts_by_author(db: &Store, author: Key) -> Result<Vec<(Key, Content)>, StoreError> {
|
||||||
db.run(|tx| {
|
db.run(|tx| {
|
||||||
let posts = tx
|
let posts = tx
|
||||||
.join_on(|a: AuthorOf| a.object, tx.outgoing(author))?
|
.join_on(|a: AuthorOf| a.object, tx.outgoing(author))?
|
||||||
|
|
|
@ -45,7 +45,7 @@ use super::{
|
||||||
types::{ArrowSpec, DataType},
|
types::{ArrowSpec, DataType},
|
||||||
Batch, Store, Transaction,
|
Batch, Store, Transaction,
|
||||||
};
|
};
|
||||||
use crate::{util::IterExt as _, Key, Result};
|
use crate::{internal::Context as _, util::IterExt as _, Key, Result};
|
||||||
|
|
||||||
/// A directed edge.
|
/// A directed edge.
|
||||||
///
|
///
|
||||||
|
@ -122,6 +122,18 @@ impl Store {
|
||||||
{
|
{
|
||||||
op::between::<A>(self, a, b).map_ok(A::from)
|
op::between::<A>(self, a, b).map_ok(A::from)
|
||||||
}
|
}
|
||||||
|
/// Construct the arrow from its identifier.
|
||||||
|
pub fn get_arrow<A>(&self, key: Key) -> Result<Option<A>>
|
||||||
|
where
|
||||||
|
A: Arrow<Kind = Multi>,
|
||||||
|
{
|
||||||
|
let arrow = self
|
||||||
|
.open(crate::types::MULTIEDGE_HEADERS)
|
||||||
|
.get(key)?
|
||||||
|
.map(|v| Key::split(v.as_ref()))
|
||||||
|
.map(|(origin, target)| A::from(Multi { origin, target, identity: key }));
|
||||||
|
Ok(arrow)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Transaction<'_> {
|
impl Transaction<'_> {
|
||||||
|
|
|
@ -5,7 +5,7 @@ use std::sync::Arc;
|
||||||
use rocksdb::{BoundColumnFamily, IteratorMode};
|
use rocksdb::{BoundColumnFamily, IteratorMode};
|
||||||
|
|
||||||
pub use self::cx::{Context, Query, Write};
|
pub use self::cx::{Context, Query, Write};
|
||||||
use crate::{util::IterExt as _, Error, Result};
|
use crate::{util::IterExt as _, Result, StoreError};
|
||||||
|
|
||||||
/// An internal interface to a specific keyspace that exposes basic hashmap-esque operations
|
/// An internal interface to a specific keyspace that exposes basic hashmap-esque operations
|
||||||
/// on that keyspace, generic over whether the source of the data is a [`Transaction`] or a
|
/// on that keyspace, generic over whether the source of the data is a [`Transaction`] or a
|
||||||
|
@ -41,13 +41,13 @@ where
|
||||||
Ok((ref k, _)) => k.starts_with(&t),
|
Ok((ref k, _)) => k.starts_with(&t),
|
||||||
_ => true,
|
_ => true,
|
||||||
})
|
})
|
||||||
.map_err(Error::Internal)
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
/// List all pairs in the keyspace.
|
/// List all pairs in the keyspace.
|
||||||
pub fn list(&self) -> impl Iterator<Item = Result<(Box<[u8]>, Box<[u8]>)>> + 'db {
|
pub fn list(&self) -> impl Iterator<Item = Result<(Box<[u8]>, Box<[u8]>)>> + 'db {
|
||||||
self.context
|
self.context
|
||||||
.full_iterator(&self.cf, IteratorMode::Start)
|
.full_iterator(&self.cf, IteratorMode::Start)
|
||||||
.map_err(Error::Internal)
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
/// Execute a range scan
|
/// Execute a range scan
|
||||||
pub fn range<const N: usize>(
|
pub fn range<const N: usize>(
|
||||||
|
@ -68,7 +68,7 @@ where
|
||||||
Ok((ref k, _)) => k.as_ref() < &upper,
|
Ok((ref k, _)) => k.as_ref() < &upper,
|
||||||
_ => true,
|
_ => true,
|
||||||
})
|
})
|
||||||
.map_err(Error::Internal)
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
/// Join all the keys to their values in this keyspace.
|
/// Join all the keys to their values in this keyspace.
|
||||||
///
|
///
|
||||||
|
@ -106,7 +106,7 @@ mod cx {
|
||||||
};
|
};
|
||||||
|
|
||||||
use super::Keyspace;
|
use super::Keyspace;
|
||||||
use crate::{util::IterExt as _, Backend, Batch, Error, Result, Store, Transaction, OK};
|
use crate::{util::IterExt as _, Backend, Batch, Result, Store, StoreError, Transaction, OK};
|
||||||
|
|
||||||
/// A context for executing database operations.
|
/// A context for executing database operations.
|
||||||
pub trait Context {
|
pub trait Context {
|
||||||
|
@ -173,7 +173,9 @@ mod cx {
|
||||||
cf: &impl AsColumnFamilyRef,
|
cf: &impl AsColumnFamilyRef,
|
||||||
key: impl AsRef<[u8]>,
|
key: impl AsRef<[u8]>,
|
||||||
) -> Result<Option<DBPinnableSlice<'a>>> {
|
) -> Result<Option<DBPinnableSlice<'a>>> {
|
||||||
self.inner.get_pinned_cf(cf, key).map_err(Error::Internal)
|
self.inner
|
||||||
|
.get_pinned_cf(cf, key)
|
||||||
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn prefix_iterator<'a>(
|
fn prefix_iterator<'a>(
|
||||||
|
@ -199,7 +201,7 @@ mod cx {
|
||||||
self.inner
|
self.inner
|
||||||
.multi_get_cf(keys)
|
.multi_get_cf(keys)
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.map_err(Error::Internal)
|
.map_err(StoreError::Internal)
|
||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -222,7 +224,9 @@ mod cx {
|
||||||
cf: &impl AsColumnFamilyRef,
|
cf: &impl AsColumnFamilyRef,
|
||||||
key: impl AsRef<[u8]>,
|
key: impl AsRef<[u8]>,
|
||||||
) -> Result<Option<DBPinnableSlice<'a>>> {
|
) -> Result<Option<DBPinnableSlice<'a>>> {
|
||||||
self.inner.get_pinned_cf(cf, key).map_err(Error::Internal)
|
self.inner
|
||||||
|
.get_pinned_cf(cf, key)
|
||||||
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn prefix_iterator<'a>(
|
fn prefix_iterator<'a>(
|
||||||
|
@ -248,14 +252,14 @@ mod cx {
|
||||||
self.inner
|
self.inner
|
||||||
.multi_get_cf(keys)
|
.multi_get_cf(keys)
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.map_err(Error::Internal)
|
.map_err(StoreError::Internal)
|
||||||
.collect()
|
.collect()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Write for Transaction<'_> {
|
impl Write for Transaction<'_> {
|
||||||
fn delete(&self, cf: &impl AsColumnFamilyRef, key: impl AsRef<[u8]>) -> Result<()> {
|
fn delete(&self, cf: &impl AsColumnFamilyRef, key: impl AsRef<[u8]>) -> Result<()> {
|
||||||
self.inner.delete_cf(cf, key).map_err(Error::Internal)
|
self.inner.delete_cf(cf, key).map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn put(
|
fn put(
|
||||||
|
@ -264,7 +268,9 @@ mod cx {
|
||||||
key: impl AsRef<[u8]>,
|
key: impl AsRef<[u8]>,
|
||||||
val: impl AsRef<[u8]>,
|
val: impl AsRef<[u8]>,
|
||||||
) -> Result<()> {
|
) -> Result<()> {
|
||||||
self.inner.put_cf(cf, key, val).map_err(Error::Internal)
|
self.inner
|
||||||
|
.put_cf(cf, key, val)
|
||||||
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -6,7 +6,7 @@ use std::{
|
||||||
use chrono::{DateTime, Utc};
|
use chrono::{DateTime, Utc};
|
||||||
use ulid::Ulid;
|
use ulid::Ulid;
|
||||||
|
|
||||||
use crate::Error;
|
use crate::StoreError;
|
||||||
|
|
||||||
/// A unique identifier for vertices in the database.
|
/// A unique identifier for vertices in the database.
|
||||||
#[derive(Clone, Copy, Eq, PartialEq, Ord, PartialOrd, Hash)]
|
#[derive(Clone, Copy, Eq, PartialEq, Ord, PartialOrd, Hash)]
|
||||||
|
@ -57,12 +57,12 @@ impl AsRef<[u8]> for Key {
|
||||||
}
|
}
|
||||||
|
|
||||||
impl FromStr for Key {
|
impl FromStr for Key {
|
||||||
type Err = crate::Error;
|
type Err = crate::StoreError;
|
||||||
|
|
||||||
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
fn from_str(s: &str) -> Result<Self, Self::Err> {
|
||||||
s.parse::<Ulid>()
|
s.parse::<Ulid>()
|
||||||
.map(|x| Key(x.to_bytes()))
|
.map(|x| Key(x.to_bytes()))
|
||||||
.map_err(|err| Error::BadKey(err))
|
.map_err(|err| StoreError::BadKey(err))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -78,7 +78,7 @@ impl Store {
|
||||||
/// changes are committed.
|
/// changes are committed.
|
||||||
pub fn run<T, E>(&self, f: impl FnOnce(&Transaction<'_>) -> Result<T, E>) -> Result<T, E>
|
pub fn run<T, E>(&self, f: impl FnOnce(&Transaction<'_>) -> Result<T, E>) -> Result<T, E>
|
||||||
where
|
where
|
||||||
E: From<Error>,
|
E: From<StoreError>,
|
||||||
{
|
{
|
||||||
let tx = Transaction {
|
let tx = Transaction {
|
||||||
inner: self.inner.transaction(),
|
inner: self.inner.transaction(),
|
||||||
|
@ -90,7 +90,7 @@ impl Store {
|
||||||
} else {
|
} else {
|
||||||
tx.inner.commit()
|
tx.inner.commit()
|
||||||
} {
|
} {
|
||||||
return Err(E::from(Error::Internal(e)));
|
return Err(E::from(StoreError::Internal(e)));
|
||||||
}
|
}
|
||||||
r
|
r
|
||||||
}
|
}
|
||||||
|
@ -123,7 +123,7 @@ impl Store {
|
||||||
/// Delete the main data store in `state_dir` if it exists.
|
/// Delete the main data store in `state_dir` if it exists.
|
||||||
pub fn nuke(state_dir: impl AsRef<Path>) -> Result<()> {
|
pub fn nuke(state_dir: impl AsRef<Path>) -> Result<()> {
|
||||||
Backend::destroy(&Options::default(), state_dir.as_ref().join(STORE_NAME))
|
Backend::destroy(&Options::default(), state_dir.as_ref().join(STORE_NAME))
|
||||||
.map_err(Error::Internal)
|
.map_err(StoreError::Internal)
|
||||||
}
|
}
|
||||||
/// Open a store that lives until `f` returns, for testing.
|
/// Open a store that lives until `f` returns, for testing.
|
||||||
pub fn test<T>(schema: Schema, f: impl FnOnce(Store) -> T) -> T {
|
pub fn test<T>(schema: Schema, f: impl FnOnce(Store) -> T) -> T {
|
||||||
|
@ -136,11 +136,11 @@ impl Store {
|
||||||
pub const OK: Result<()> = Ok(());
|
pub const OK: Result<()> = Ok(());
|
||||||
|
|
||||||
/// Results from this component.
|
/// Results from this component.
|
||||||
pub type Result<T, E = Error> = std::result::Result<T, E>;
|
pub type Result<T, E = StoreError> = std::result::Result<T, E>;
|
||||||
|
|
||||||
/// Errors from the data store.
|
/// Errors from the data store.
|
||||||
#[derive(From, Debug)]
|
#[derive(From, Debug)]
|
||||||
pub enum Error {
|
pub enum StoreError {
|
||||||
/// The requested value was expected to exist in a particular keyspace, but does not actually
|
/// The requested value was expected to exist in a particular keyspace, but does not actually
|
||||||
/// exist there. This can occur on updates for example.
|
/// exist there. This can occur on updates for example.
|
||||||
Missing,
|
Missing,
|
||||||
|
|
|
@ -7,7 +7,7 @@ use super::{
|
||||||
types::{DataType, MixinSpec},
|
types::{DataType, MixinSpec},
|
||||||
Batch, Store, Transaction,
|
Batch, Store, Transaction,
|
||||||
};
|
};
|
||||||
use crate::{util::IterExt as _, Error, Key, Result};
|
use crate::{util::IterExt as _, Key, Result, StoreError};
|
||||||
|
|
||||||
/// Mixins are the simplest pieces of data in the store.
|
/// Mixins are the simplest pieces of data in the store.
|
||||||
pub trait Mixin: DataType<Type = MixinSpec> + Encode + Decode {}
|
pub trait Mixin: DataType<Type = MixinSpec> + Encode + Decode {}
|
||||||
|
@ -53,6 +53,13 @@ impl Store {
|
||||||
{
|
{
|
||||||
op::join_on(self, iter)
|
op::join_on(self, iter)
|
||||||
}
|
}
|
||||||
|
/// Get multiple mixins associated with the same key.
|
||||||
|
pub fn get_mixin_many<T>(&self, key: Key) -> Result<T>
|
||||||
|
where
|
||||||
|
T: GetMany,
|
||||||
|
{
|
||||||
|
T::get(self, key)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Transaction<'_> {
|
impl Transaction<'_> {
|
||||||
|
@ -88,7 +95,7 @@ impl Transaction<'_> {
|
||||||
M: Mixin,
|
M: Mixin,
|
||||||
{
|
{
|
||||||
if op::has_mixin::<M>(self, node)? {
|
if op::has_mixin::<M>(self, node)? {
|
||||||
return Err(Error::Conflict);
|
return Err(StoreError::Conflict);
|
||||||
} else {
|
} else {
|
||||||
op::add_mixin::<M>(self, node, mixin)
|
op::add_mixin::<M>(self, node, mixin)
|
||||||
}
|
}
|
||||||
|
@ -136,6 +143,12 @@ impl Batch {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Getting tuples of stuff.
|
||||||
|
pub trait GetMany: Sized {
|
||||||
|
#[doc(hidden)]
|
||||||
|
fn get(cx: &impl crate::internal::Query, key: Key) -> Result<Self>;
|
||||||
|
}
|
||||||
|
|
||||||
mod op {
|
mod op {
|
||||||
use std::ops::{Bound, RangeBounds};
|
use std::ops::{Bound, RangeBounds};
|
||||||
|
|
||||||
|
@ -143,7 +156,64 @@ mod op {
|
||||||
use either::Either;
|
use either::Either;
|
||||||
|
|
||||||
use super::Mixin;
|
use super::Mixin;
|
||||||
use crate::{internal::*, util::IterExt as _, Error, Key, Result};
|
use crate::{internal::*, util::IterExt as _, Key, Result, StoreError};
|
||||||
|
|
||||||
|
impl<A, B> super::GetMany for (A, B)
|
||||||
|
where
|
||||||
|
A: Mixin,
|
||||||
|
B: Mixin,
|
||||||
|
{
|
||||||
|
fn get(cx: &impl Query, key: Key) -> Result<Self> {
|
||||||
|
let ksps = [cx.open(A::SPEC.keyspace).cf, cx.open(B::SPEC.keyspace).cf];
|
||||||
|
let mut vec = cx.multi_get(ksps.iter().map(|c| (&*c, key)));
|
||||||
|
let b = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
let a = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
Ok((decode(a)?, decode(b)?))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<A, B, C> super::GetMany for (A, B, C)
|
||||||
|
where
|
||||||
|
A: Mixin,
|
||||||
|
B: Mixin,
|
||||||
|
C: Mixin,
|
||||||
|
{
|
||||||
|
fn get(cx: &impl Query, key: Key) -> Result<Self> {
|
||||||
|
let ksps = [
|
||||||
|
cx.open(A::SPEC.keyspace).cf,
|
||||||
|
cx.open(B::SPEC.keyspace).cf,
|
||||||
|
cx.open(C::SPEC.keyspace).cf,
|
||||||
|
];
|
||||||
|
let mut vec = cx.multi_get(ksps.iter().map(|c| (&*c, key)));
|
||||||
|
let c = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
let b = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
let a = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
Ok((decode(a)?, decode(b)?, decode(c)?))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<A, B, C, D> super::GetMany for (A, B, C, D)
|
||||||
|
where
|
||||||
|
A: Mixin,
|
||||||
|
B: Mixin,
|
||||||
|
C: Mixin,
|
||||||
|
D: Mixin,
|
||||||
|
{
|
||||||
|
fn get(cx: &impl Query, key: Key) -> Result<Self> {
|
||||||
|
let ksps = [
|
||||||
|
cx.open(A::SPEC.keyspace).cf,
|
||||||
|
cx.open(B::SPEC.keyspace).cf,
|
||||||
|
cx.open(C::SPEC.keyspace).cf,
|
||||||
|
cx.open(D::SPEC.keyspace).cf,
|
||||||
|
];
|
||||||
|
let mut vec = cx.multi_get(ksps.iter().map(|c| (&*c, key)));
|
||||||
|
let d = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
let c = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
let b = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
let a = vec.pop().ok_or(StoreError::Missing)??.unwrap();
|
||||||
|
Ok((decode(a)?, decode(b)?, decode(c)?, decode(d)?))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
pub fn update<M>(
|
pub fn update<M>(
|
||||||
cx: &(impl Query + Write),
|
cx: &(impl Query + Write),
|
||||||
|
@ -165,7 +235,7 @@ mod op {
|
||||||
// [^1]: https://github.com/facebook/rocksdb/blob/9d37408f9af15c7a1ae42f9b94d06b27d98a011a/include/rocksdb/options.h#L128
|
// [^1]: https://github.com/facebook/rocksdb/blob/9d37408f9af15c7a1ae42f9b94d06b27d98a011a/include/rocksdb/options.h#L128
|
||||||
let tree = cx.open(M::SPEC.keyspace);
|
let tree = cx.open(M::SPEC.keyspace);
|
||||||
match tree.get(node.as_ref())? {
|
match tree.get(node.as_ref())? {
|
||||||
None => Err(Error::Missing),
|
None => Err(StoreError::Missing),
|
||||||
Some(buf) => {
|
Some(buf) => {
|
||||||
let new = decode(buf).map(update).and_then(encode)?;
|
let new = decode(buf).map(update).and_then(encode)?;
|
||||||
tree.set(node, new)
|
tree.set(node, new)
|
||||||
|
@ -237,7 +307,7 @@ mod op {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(super) fn encode(data: impl bincode::Encode) -> Result<Vec<u8>> {
|
pub(super) fn encode(data: impl bincode::Encode) -> Result<Vec<u8>> {
|
||||||
bincode::encode_to_vec(data, bincode::config::standard()).map_err(Error::Encoding)
|
bincode::encode_to_vec(data, bincode::config::standard()).map_err(StoreError::Encoding)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub(super) fn decode<T>(data: impl AsRef<[u8]>) -> Result<T>
|
pub(super) fn decode<T>(data: impl AsRef<[u8]>) -> Result<T>
|
||||||
|
@ -245,7 +315,7 @@ mod op {
|
||||||
T: bincode::Decode,
|
T: bincode::Decode,
|
||||||
{
|
{
|
||||||
bincode::decode_from_slice(data.as_ref(), bincode::config::standard())
|
bincode::decode_from_slice(data.as_ref(), bincode::config::standard())
|
||||||
.map_err(Error::Decoding)
|
.map_err(StoreError::Decoding)
|
||||||
.map(|(v, _)| v)
|
.map(|(v, _)| v)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -56,6 +56,14 @@ pub trait IterExt: Iterator + Sized {
|
||||||
{
|
{
|
||||||
self.filter_map(move |r| r.and_then(|x| f(x)).transpose())
|
self.filter_map(move |r| r.and_then(|x| f(x)).transpose())
|
||||||
}
|
}
|
||||||
|
/// Like [`Iterator::find`].
|
||||||
|
fn find_ok<'a, I, E>(mut self, mut f: impl FnMut(&I) -> bool) -> Result<Option<I>, E>
|
||||||
|
where
|
||||||
|
Self: Iterator<Item = Result<I, E>> + 'a,
|
||||||
|
{
|
||||||
|
self.find(move |r| r.as_ref().is_ok_and(|x| f(x)))
|
||||||
|
.transpose()
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<I> IterExt for I where I: Iterator {}
|
impl<I> IterExt for I where I: Iterator {}
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
unstable_features = true
|
unstable_features = true
|
||||||
overflow_delimited_expr = true
|
overflow_delimited_expr = true
|
||||||
group_imports = "StdExternalCrate"
|
|
||||||
use_field_init_shorthand = true
|
use_field_init_shorthand = true
|
||||||
reorder_modules = false
|
reorder_modules = false
|
||||||
|
reorder_imports = false
|
||||||
struct_lit_width = 30
|
struct_lit_width = 30
|
||||||
|
|
Loading…
Reference in a new issue