mirror of
https://github.com/RichieCahill/dotfiles.git
synced 2026-04-21 14:49:10 -04:00
Compare commits
535 Commits
feature/se
...
claude/add
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
56b0b6651e | ||
| f627a5ac6e | |||
| a5e7d97213 | |||
| 1419deb3c6 | |||
| 1f06692696 | |||
| 8f8177f36e | |||
| 8534edc285 | |||
| 73b28a855b | |||
| 0c0810a06b | |||
| 239bef975a | |||
| 2577b791f7 | |||
| b4d9562591 | |||
| 66f972ac2b | |||
| aca756f479 | |||
| 7f59f7f7ac | |||
|
|
70864c620f | ||
|
|
304f1c8433 | ||
| 1b5a036061 | |||
| 42330ec186 | |||
| 3f4373d1f6 | |||
| cc73dfc467 | |||
| 976c3f9d3e | |||
| 2661127426 | |||
| 1b3e6725ea | |||
| 7d2fbaea43 | |||
| a19b1c7e60 | |||
| 76da6cbc54 | |||
| c83bbe2c24 | |||
| 7611a3b2df | |||
| aec5e3e22b | |||
| 4e3273d5ec | |||
| b5ee7c2dc2 | |||
| 958b06ecf0 | |||
| 71ad8ab29e | |||
| 852759c510 | |||
| d684d5d62c | |||
| f1e394565d | |||
| 754ced4822 | |||
| 5b054dfc8f | |||
| 663833d4fa | |||
| 433ec9a38e | |||
| 3a3267ee9a | |||
| 0497a50a43 | |||
| 6365dd8067 | |||
| a6fbbd245f | |||
| 7ad321e5e2 | |||
| 14338e34df | |||
| c73aa5c98a | |||
| f762f12bd2 | |||
| ab5df442c6 | |||
|
|
f11c9bed58 | ||
|
|
ab2d8dbd51 | ||
|
|
42ede19472 | ||
|
|
f4f33eacc4 | ||
|
|
51f6cd23ad | ||
| 3dadb145b7 | |||
| 75a67294ea | |||
| 58b25f2e89 | |||
| 568bf8dd38 | |||
| 82851eb287 | |||
| b7bce0bcb9 | |||
| 583af965ad | |||
| ec80bf1c5f | |||
| bd490334f5 | |||
| e893ea0f57 | |||
| 18f149b831 | |||
| 69f5b87e5f | |||
| 66acc010ca | |||
| e8f3a563be | |||
| 8f1d765cad | |||
| 4f0ba687c4 | |||
|
|
27891c3903 | ||
| ccdc61b4dd | |||
| 1d732bf41c | |||
| 13ba118cfc | |||
| 47c6f42d2f | |||
|
|
ff9dcde5d9 | ||
| 7de800b519 | |||
| 55767ad555 | |||
| c262ff9048 | |||
|
|
9abac2978a | ||
| 70d20e55d2 | |||
| f038f248a1 | |||
| af828fc9c4 | |||
| 4d121ae9f9 | |||
| 959d599ff9 | |||
| d470243fdd | |||
| d96c93fa17 | |||
| 6bea380e3d | |||
| 56c933c8cb | |||
| e7dae1eb4b | |||
| 17ebe50ac9 | |||
| 97b35ce27b | |||
|
|
595579fe8b | ||
| fcfbce4e16 | |||
| 80af3377e6 | |||
| 557c1a4d5d | |||
| 89e37249af | |||
|
|
ccd523b4d0 | ||
| 606035432b | |||
| 4d2f6831e3 | |||
| 86e72d1da0 | |||
| 139727bf50 | |||
| 88c2f1b139 | |||
| e75a3ef9c6 | |||
| 258f918794 | |||
| cf4635922e | |||
| 0615ece46a | |||
| 8afa4fce6c | |||
| 8bbcd37933 | |||
| 037b2f9cf7 | |||
| 7dbc4c248f | |||
| 08dffc6f6d | |||
| 0109167b10 | |||
| b87f6b0b34 | |||
| 35376c3fca | |||
| 0c218f2551 | |||
| d0b66496a1 | |||
| 5101da4914 | |||
| 393545868f | |||
| 6bb7904782 | |||
| 59147834f7 | |||
|
|
52235239d0 | ||
|
|
9e43c3e8b8 | ||
| 156d624d81 | |||
| 9a7cf03a00 | |||
| 6299d42f75 | |||
| e6472b2cf5 | |||
| 41d3a8fe1a | |||
|
|
e6ac8f8021 | ||
| 0f8f6f96d6 | |||
| 4cb4bd6f3d | |||
| c046710258 | |||
| 7f9fbe3602 | |||
|
|
8ee3b4d6e5 | ||
| 18b7fb2d60 | |||
| 2f1fa5c750 | |||
| 164d0dd59e | |||
| d4459643ab | |||
| c09dba0c37 | |||
| 409f376166 | |||
| a9a6e1f932 | |||
| 6472f07a88 | |||
|
|
51c79f6b40 | ||
| b0d5147296 | |||
| c56082b516 | |||
| 34b728c88f | |||
| 5697458bad | |||
| 276c2ac74b | |||
| 69e5aa20d5 | |||
| 3d1f773fa5 | |||
| 14dd1fe52e | |||
| 30fe41ea1b | |||
| 3a17c5514d | |||
| c6586db91e | |||
| 81b199373e | |||
| a957e23041 | |||
| 52389f729d | |||
| cc2a609f52 | |||
| ca4693a1ba | |||
| 90e5e0855d | |||
| e339667c2b | |||
| 85540ee920 | |||
| 3be1b8aa8f | |||
| 7c56954cda | |||
| 290f972346 | |||
| 72c3ccfb6d | |||
| 9630633ff5 | |||
| 8c83f306b2 | |||
| 5b4609dc3b | |||
| d1be25c6e8 | |||
| 31910586d2 | |||
| b8dfd0852a | |||
| 6ce622e93e | |||
| 55e652a51d | |||
| b5455a5483 | |||
| 8baf388061 | |||
| 7ffb7b4a37 | |||
| eb04f4a56d | |||
| 5b8e543226 | |||
| da48f62195 | |||
| 60f2ab1039 | |||
| c1de454005 | |||
| 391e37b746 | |||
|
|
27565173d4 | ||
| 0c0ed92cb4 | |||
|
|
cc9996d6fa | ||
| 102f36eb1b | |||
| 9ec988729b | |||
| 4e3c25afb4 | |||
| 0d482aca4b | |||
|
|
c624781d84 | ||
| f4996b71e4 | |||
| 58a29214d3 | |||
| c4171b56b5 | |||
| d6d48516ea | |||
| ae882ba578 | |||
| 100b8145e8 | |||
| e99cd8e54a | |||
| de9348432c | |||
| b1fa596f37 | |||
| 908bccb8dc | |||
| b8cc9c5772 | |||
| d62076a900 | |||
| 0c9bd40659 | |||
| f713b8d4fa | |||
| ddba7d1068 | |||
| 41aad90140 | |||
| 76cd6e1188 | |||
| 20ef02b0cc | |||
| c0e9f3f937 | |||
| 9e0a2810f5 | |||
| 5c488422a1 | |||
| 9d43704b64 | |||
| d5bc6e9c6e | |||
| 17cebe1a82 | |||
| f02a866b19 | |||
| 65c2bed046 | |||
| 26cf123357 | |||
| a7c0a58c9a | |||
| e89fb9fae1 | |||
| 44feda70c1 | |||
| 1bfdbfd785 | |||
| 6a09bc66b6 | |||
| 65fca5c8a4 | |||
| a6e2334999 | |||
| c5981e0e6c | |||
| 825672a450 | |||
| d2db0de371 | |||
| 8142582e4a | |||
| 3038e1c704 | |||
| 18de5bc12c | |||
| 4fa2141461 | |||
| 626bd70d67 | |||
| 8ed7eda020 | |||
| e9ae9478bf | |||
| c1ce7e0ac4 | |||
| d040b06869 | |||
| 04c41c6ac0 | |||
| 298adcce87 | |||
| ef25153c84 | |||
| 9416bbd00b | |||
| b8200af6d5 | |||
| afb62b97d1 | |||
| cf75f3d75a | |||
| 0f8a594545 | |||
| db37eb2f9e | |||
| 534d9110e2 | |||
| 86a1cac42c | |||
| d3452dfab5 | |||
| acab92ac9c | |||
|
|
f22a9d107a | ||
| e5e0f883b0 | |||
| 04bf6f2038 | |||
| d35ba60c69 | |||
| 1e85635e89 | |||
| 6423192ee7 | |||
|
|
a33aba3afc | ||
|
|
d4d481e4b2 | ||
| f092348736 | |||
| e6c3ae0bee | |||
| d1f4f21521 | |||
| 00a5536208 | |||
| b554325b13 | |||
| 357168695c | |||
| fc31447591 | |||
| eea620aa2f | |||
| 82d463bfd6 | |||
| ca8c9925ad | |||
| 45a31e6b4d | |||
| c4be520190 | |||
| 28cd7f3f6f | |||
|
|
879885dc3b | ||
| 5e03efa1e8 | |||
| 2ed08501fe | |||
| c36624cc55 | |||
| f074344ac8 | |||
|
|
b773eb2db7 | ||
|
|
6efcc9add1 | ||
| bff561946f | |||
| 1ec05d13da | |||
| 4686a85bcd | |||
| 3434b32fbe | |||
| 98d9efcde3 | |||
|
|
4286f39177 | ||
| 0fe439ceaf | |||
| 7a3c2026b3 | |||
| 0c9ce78c20 | |||
| c10a76babc | |||
| 113ca9c99a | |||
| 29f51bf116 | |||
| 9f3a2b2a4b | |||
|
|
8a4d021541 | ||
| 5225bf1732 | |||
| 7a5dea1c36 | |||
| 0956ea6f58 | |||
| 5d643e0cd1 | |||
| ed3805a89e | |||
| ce8de13734 | |||
| 68b7a2f80d | |||
| 62c3aa69fe | |||
| 470be2b761 | |||
| e1c5bd0f84 | |||
| 786b275f7c | |||
| 324721ff8d | |||
| 7137435703 | |||
| a7b336a7de | |||
| 529b03525b | |||
| d0364bdaad | |||
| 45ddf8bc54 | |||
|
|
6913c7046e | ||
| c4d2b0b8d4 | |||
| 27e939459b | |||
| 501a838be9 | |||
| 2325249687 | |||
|
|
45951f6525 | ||
| fd56fa66f0 | |||
| 1314298c0b | |||
| f0eee80c2d | |||
| c20b6d1da2 | |||
| 34a59f966d | |||
| a677046330 | |||
| 111afa1c6b | |||
| c5f2805e74 | |||
| 9a61d06f08 | |||
| 559ca7a45e | |||
| 03b636eb3a | |||
| a7f5d3c71d | |||
|
|
3bbf8dc7a6 | ||
|
|
1cd4084ec8 | ||
|
|
e65b4b696a | ||
|
|
20a4a8c2fc | ||
| f4348c2ab5 | |||
| 6eab8497ba | |||
|
|
da5cdb8f05 | ||
| 8aa9eea322 | |||
| febe7c7e53 | |||
| b1ca58b2f4 | |||
|
|
7ad4ccd5ca | ||
| c936501afb | |||
| 239d7833f6 | |||
| 8fb6ae41b9 | |||
| 80e0b03463 | |||
| 747e2700ed | |||
| 472f11e5b6 | |||
| d75493997e | |||
| d711983ac7 | |||
| 1aabb2b112 | |||
|
|
8c09b8a78f | ||
| 0fe99d9d4e | |||
| 148722be43 | |||
| be2421e3dc | |||
| 2eb6c43b49 | |||
| 897e06f622 | |||
| 4e99c54c12 | |||
| 3abd04ec5e | |||
| a6a0fd0727 | |||
| f23d381895 | |||
| 65335d1d38 | |||
|
|
7b318e038a | ||
| 2fe3ad1f68 | |||
| fa27adb4fe | |||
| ac7aec4efd | |||
| 87bebda02e | |||
| 43276a693a | |||
| 84504c68b7 | |||
|
|
45267b5c59 | ||
|
|
bd964a8390 | ||
|
|
86e90440c4 | ||
|
|
2b743fa1b8 | ||
| eb69646603 | |||
| ccfde41209 | |||
| eb90251671 | |||
| a5ccedbb5d | |||
| 7dbf67a2d3 | |||
| c349af60ce | |||
| 3fcc3447b1 | |||
| fc166e9c1a | |||
| e902608c6c | |||
| 59a794169e | |||
| fa618f8849 | |||
| 84e469897e | |||
| 4694d75a62 | |||
| ca2081bf93 | |||
| fa3f14785c | |||
| aa2836c452 | |||
| 5b750043ce | |||
| da721a32f3 | |||
| fd172638ec | |||
| 9dd32c7c06 | |||
|
|
96ce4ccb4a | ||
|
|
f448da92c7 | ||
| 8d14e85a5e | |||
| 171dcb6273 | |||
| 8576cb2113 | |||
| d0077f00f5 | |||
| 0ea628ddb7 | |||
| 4feba720fc | |||
| 81ffea08e3 | |||
| 61db4659dd | |||
| 0a50d94af4 | |||
| 27c59192c8 | |||
|
|
5a1078fec0 | ||
| 39ed05a73a | |||
|
|
68202fdd15 | ||
| 9a64d9cbfb | |||
| 1d0d46d22d | |||
| 1d1bfae8ee | |||
|
|
8d6328ce57 | ||
| 02db3d2eb5 | |||
| bdf9d267e8 | |||
| 4db29ae280 | |||
| 1b97fa7dda | |||
| 9248337e55 | |||
| f77c341e2b | |||
| f7f0d02ef9 | |||
| e723ab9e86 | |||
| 5182bf18bc | |||
| 11883f981e | |||
| 6d41287eec | |||
| 433c52309b | |||
| e1bbadfcba | |||
| 868f782523 | |||
| 434c80adac | |||
| 6c15bbe0a3 | |||
| cbdefc0c52 | |||
| 3fc9d9d614 | |||
| aa4a11c0c3 | |||
| 300aa1f032 | |||
| c12bd0c7ca | |||
| f43a34f37c | |||
| 33c2718d40 | |||
| 78500e6122 | |||
| 9459d4797e | |||
| bfbcb70327 | |||
| b95e553093 | |||
| 8668603404 | |||
| 60c3e3db20 | |||
| d3aa49f9e4 | |||
| f691015920 | |||
| 81ab03f3f1 | |||
| ba1170720c | |||
| 4316c8a52e | |||
| a036467131 | |||
| 62c26b351c | |||
| cd5ad0dcae | |||
| fa38cc6321 | |||
| e8d971340b | |||
| 30f2af87be | |||
| 8522b77285 | |||
| f1a48819a5 | |||
| eff58f08ad | |||
| a24c4fd756 | |||
| f10ab4a9af | |||
| 7cc831a2d0 | |||
| 91690734ba | |||
|
|
b97e81e96e | ||
| 0d680dc68d | |||
| 81867eb312 | |||
| 00852c019d | |||
|
|
b53f4ec8d8 | ||
| 64a4223560 | |||
| 88d3d6311f | |||
|
|
bccc569650 | ||
| 633a48af7b | |||
| 799794ebe5 | |||
| 7b07266c44 | |||
| 60ce45a08d | |||
| 12b6c1e279 | |||
| 12b4fe92da | |||
| d076183777 | |||
| 7a454ca740 | |||
| 8dd892a88f | |||
| 7f3bb43d39 | |||
| 793d06683e | |||
| bb1d92252f | |||
| f7a95f32cd | |||
| ea24fb9865 | |||
| c39f8c4a7e | |||
| d7fd0796ea | |||
| 01d2bd04b2 | |||
| 040a44b608 | |||
|
|
a3a132b41c | ||
| c0a97c6ed8 | |||
| 83dad24351 | |||
| 1a1cc41265 | |||
| 0457265bd7 | |||
| 3c2476336b | |||
| 27ec9d9204 | |||
| 0f094bfacb | |||
| 14a312dd0f | |||
| c87e653ad7 | |||
| e88fc08d91 | |||
| 2e26035693 | |||
| f877b15d35 | |||
| 7736c36eee | |||
| 78159e1d90 | |||
| a39d3f0c51 | |||
| 7995d383ca | |||
| a7e8494d8a | |||
|
|
17599aea9d | ||
| ae5981dcfa | |||
| 1770cb87d3 | |||
| 61b548e11c | |||
| 76da8cd379 | |||
| 0e4cb002c8 | |||
| ce27dcfa39 | |||
|
|
6dae76c51e | ||
| 92a87c01d3 | |||
| bca98cfadb | |||
| 91ebec054c | |||
| 6ae6bc0dd6 | |||
| 0a1e517468 | |||
| f852d9b9fb | |||
| 7ad195ff24 | |||
| 4cadc374b1 | |||
| 56456ab375 | |||
| 17b19967e8 | |||
| 9ff57ba6f3 | |||
| 573031e4c9 | |||
| 6e2703f852 | |||
|
|
a6b6d66b9f | ||
| bdfe36f919 | |||
| b3be9dd2c8 | |||
| d991f94d7e | |||
| 2a12b80dfa | |||
| d783b0d03d | |||
| 4ae7b9ab4d | |||
| 0f788a1901 | |||
| 20ed83e80a | |||
| 1ea95f9518 | |||
| b1e3ce2afa | |||
| 71e6fa377c | |||
| 15234fa2bb | |||
| cdf26994a3 |
10
.github/workflows/build_systems.yml
vendored
10
.github/workflows/build_systems.yml
vendored
@@ -1,9 +1,11 @@
|
|||||||
name: build_jeeves
|
name: build_systems
|
||||||
on:
|
on:
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
pull_request:
|
pull_request:
|
||||||
push:
|
push:
|
||||||
branches: [main]
|
branches: [main]
|
||||||
|
schedule:
|
||||||
|
- cron: "0 22 * * *"
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
build:
|
build:
|
||||||
@@ -13,10 +15,14 @@ jobs:
|
|||||||
matrix:
|
matrix:
|
||||||
system:
|
system:
|
||||||
- "bob"
|
- "bob"
|
||||||
|
- "brain"
|
||||||
- "jeeves"
|
- "jeeves"
|
||||||
- "muninn"
|
- "leviathan"
|
||||||
- "rhapsody-in-green"
|
- "rhapsody-in-green"
|
||||||
|
continue-on-error: true
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- name: Build default package
|
- name: Build default package
|
||||||
run: "nixos-rebuild build --flake ./#${{ matrix.system }}"
|
run: "nixos-rebuild build --flake ./#${{ matrix.system }}"
|
||||||
|
- name: copy to nix-cache
|
||||||
|
run: nix copy --accept-flake-config --to unix:///host-nix/var/nix/daemon-socket/socket .#nixosConfigurations.${{ matrix.system }}.config.system.build.toplevel
|
||||||
|
|||||||
30
.github/workflows/fix_eval_warnings.yml
vendored
Normal file
30
.github/workflows/fix_eval_warnings.yml
vendored
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
name: fix_eval_warnings
|
||||||
|
on:
|
||||||
|
workflow_run:
|
||||||
|
workflows: ["build_systems"]
|
||||||
|
types: [completed]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
check-warnings:
|
||||||
|
if: >-
|
||||||
|
github.event.workflow_run.conclusion != 'cancelled' &&
|
||||||
|
github.event.workflow_run.head_branch == 'main' &&
|
||||||
|
(github.event.workflow_run.event == 'push' || github.event.workflow_run.event == 'schedule')
|
||||||
|
runs-on: self-hosted
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
pull-requests: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Fix eval warnings
|
||||||
|
env:
|
||||||
|
GH_TOKEN: ${{ secrets.GH_TOKEN_FOR_UPDATES }}
|
||||||
|
run: >-
|
||||||
|
nix develop .#devShells.x86_64-linux.default -c
|
||||||
|
python -m python.eval_warnings.main
|
||||||
|
--run-id "${{ github.event.workflow_run.id }}"
|
||||||
|
--repo "${{ github.repository }}"
|
||||||
|
--ollama-url "${{ secrets.OLLAMA_URL }}"
|
||||||
|
--run-url "${{ github.event.workflow_run.html_url }}"
|
||||||
29
.github/workflows/merge_flake_lock_update.yml
vendored
Normal file
29
.github/workflows/merge_flake_lock_update.yml
vendored
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
name: merge_flake_lock_update
|
||||||
|
on:
|
||||||
|
workflow_dispatch:
|
||||||
|
schedule:
|
||||||
|
- cron: "0 2 * * 6"
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
merge:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: write
|
||||||
|
pull-requests: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: merge_flake_lock_update
|
||||||
|
run: |
|
||||||
|
pr_number=$(gh pr list --state open --author RichieCahill --label flake_lock_update --json number --jq '.[0].number')
|
||||||
|
echo "pr_number=$pr_number" >> $GITHUB_ENV
|
||||||
|
if [ -n "$pr_number" ]; then
|
||||||
|
gh pr merge "$pr_number" --rebase
|
||||||
|
else
|
||||||
|
echo "No open PR found with label flake_lock_update"
|
||||||
|
fi
|
||||||
|
env:
|
||||||
|
GITHUB_TOKEN: ${{ secrets.GH_TOKEN_FOR_UPDATES }}
|
||||||
19
.github/workflows/pytest.yml
vendored
Normal file
19
.github/workflows/pytest.yml
vendored
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
name: pytest
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
pull_request:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
merge_group:
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
pytest:
|
||||||
|
runs-on: self-hosted
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
- name: Run tests
|
||||||
|
run: nix develop .#devShells.x86_64-linux.default -c pytest tests
|
||||||
15
.github/workflows/treefmt.yml
vendored
Normal file
15
.github/workflows/treefmt.yml
vendored
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
name: treefmt
|
||||||
|
on:
|
||||||
|
workflow_dispatch:
|
||||||
|
pull_request:
|
||||||
|
push:
|
||||||
|
branches: [main]
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
treefmt:
|
||||||
|
name: nix fmt
|
||||||
|
runs-on: self-hosted
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
- name: runs treefmt
|
||||||
|
run: "treefmt --ci"
|
||||||
4
.github/workflows/update-flake-lock.yml
vendored
4
.github/workflows/update-flake-lock.yml
vendored
@@ -2,7 +2,7 @@ name: update-flake-lock
|
|||||||
on:
|
on:
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
schedule:
|
schedule:
|
||||||
- cron: "0 0 * * *"
|
- cron: "0 0 * * 6"
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
lockfile:
|
lockfile:
|
||||||
@@ -15,7 +15,9 @@ jobs:
|
|||||||
- name: Update flake.lock
|
- name: Update flake.lock
|
||||||
uses: DeterminateSystems/update-flake-lock@main
|
uses: DeterminateSystems/update-flake-lock@main
|
||||||
with:
|
with:
|
||||||
|
token: ${{ secrets.GH_TOKEN_FOR_UPDATES }}
|
||||||
pr-title: "Update flake.lock"
|
pr-title: "Update flake.lock"
|
||||||
pr-labels: |
|
pr-labels: |
|
||||||
dependencies
|
dependencies
|
||||||
automated
|
automated
|
||||||
|
flake_lock_update
|
||||||
|
|||||||
8
.gitignore
vendored
8
.gitignore
vendored
@@ -162,4 +162,10 @@ cython_debug/
|
|||||||
#.idea/
|
#.idea/
|
||||||
|
|
||||||
test.*
|
test.*
|
||||||
secrets.*
|
|
||||||
|
# syncthing
|
||||||
|
.stfolder
|
||||||
|
|
||||||
|
# Frontend build output
|
||||||
|
frontend/dist/
|
||||||
|
frontend/node_modules/
|
||||||
|
|||||||
22
.sops.yaml
Normal file
22
.sops.yaml
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
# Generate AGE keys from SSH keys with:
|
||||||
|
# ssh-keygen -A
|
||||||
|
# nix-shell -p ssh-to-age --run 'cat /etc/ssh/ssh_host_ed25519_key.pub | ssh-to-age'
|
||||||
|
keys:
|
||||||
|
- &admin_richie age1u8zj599elqqvcmhxn8zuwrufsz8w8w366d3ayrljjejljt2q45kq8mxw9c # cspell:disable-line
|
||||||
|
|
||||||
|
- &system_bob age1q47vup0tjhulkg7d6xwmdsgrw64h4ax3la3evzqpxyy4adsmk9fs56qz3y # cspell:disable-line
|
||||||
|
- &system_brain age1jhf7vm0005j60mjq63696frrmjhpy8kpc2d66mw044lqap5mjv4snmwvwm # cspell:disable-line
|
||||||
|
- &system_jeeves age13lmqgc3jvkyah5e3vcwmj4s5wsc2akctcga0lpc0x8v8du3fxprqp4ldkv # cspell:disable-line
|
||||||
|
- &system_leviathan age1l272y8udvg60z7edgje42fu49uwt4x2gxn5zvywssnv9h2krms8s094m4k # cspell:disable-line
|
||||||
|
- &system_rhapsody age1ufnewppysaq2wwcl4ugngjz8pfzc5a35yg7luq0qmuqvctajcycs5lf6k4 # cspell:disable-line
|
||||||
|
|
||||||
|
creation_rules:
|
||||||
|
- path_regex: users/secrets\.yaml$
|
||||||
|
key_groups:
|
||||||
|
- age:
|
||||||
|
- *admin_richie
|
||||||
|
- *system_bob
|
||||||
|
- *system_brain
|
||||||
|
- *system_jeeves
|
||||||
|
- *system_leviathan
|
||||||
|
- *system_rhapsody
|
||||||
98
.vscode/settings.json
vendored
98
.vscode/settings.json
vendored
@@ -2,11 +2,15 @@
|
|||||||
"cSpell.words": [
|
"cSpell.words": [
|
||||||
"aboutwelcome",
|
"aboutwelcome",
|
||||||
"acltype",
|
"acltype",
|
||||||
|
"addopts",
|
||||||
"addstr",
|
"addstr",
|
||||||
"advplyr",
|
"advplyr",
|
||||||
"ahci",
|
"ahci",
|
||||||
"aioesphomeapi",
|
"aioesphomeapi",
|
||||||
|
"aiounifi",
|
||||||
"alsa",
|
"alsa",
|
||||||
|
"apiclient",
|
||||||
|
"apscheduler",
|
||||||
"archlinux",
|
"archlinux",
|
||||||
"ashift",
|
"ashift",
|
||||||
"asrouter",
|
"asrouter",
|
||||||
@@ -17,11 +21,14 @@
|
|||||||
"auditd",
|
"auditd",
|
||||||
"autofetch",
|
"autofetch",
|
||||||
"autologin",
|
"autologin",
|
||||||
|
"automations",
|
||||||
"autopull",
|
"autopull",
|
||||||
"autotrim",
|
"autotrim",
|
||||||
|
"autoupdate",
|
||||||
"azuretools",
|
"azuretools",
|
||||||
"bantime",
|
"bantime",
|
||||||
"bazarr",
|
"bazarr",
|
||||||
|
"bgwriter",
|
||||||
"binhex",
|
"binhex",
|
||||||
"bitwarden",
|
"bitwarden",
|
||||||
"blkdiscard",
|
"blkdiscard",
|
||||||
@@ -32,20 +39,27 @@
|
|||||||
"captivedetect",
|
"captivedetect",
|
||||||
"cgroupdriver",
|
"cgroupdriver",
|
||||||
"charliermarsh",
|
"charliermarsh",
|
||||||
|
"Checkpointing",
|
||||||
"cloudflared",
|
"cloudflared",
|
||||||
|
"codellama",
|
||||||
"codezombiech",
|
"codezombiech",
|
||||||
"compactmode",
|
"compactmode",
|
||||||
"Compat",
|
"Compat",
|
||||||
"contentblocking",
|
"contentblocking",
|
||||||
"cookiebanners",
|
"cookiebanners",
|
||||||
|
"createdb",
|
||||||
|
"createrole",
|
||||||
"crlite",
|
"crlite",
|
||||||
"cryptsetup",
|
"cryptsetup",
|
||||||
|
"cuda",
|
||||||
"darkreader",
|
"darkreader",
|
||||||
|
"datagrip",
|
||||||
"datareporting",
|
"datareporting",
|
||||||
"davidanson",
|
"davidanson",
|
||||||
"dconf",
|
"dconf",
|
||||||
"dearrow",
|
"dearrow",
|
||||||
"debugpy",
|
"debugpy",
|
||||||
|
"deepseek",
|
||||||
"dialout",
|
"dialout",
|
||||||
"diffie",
|
"diffie",
|
||||||
"direnv",
|
"direnv",
|
||||||
@@ -53,23 +67,28 @@
|
|||||||
"dnodesize",
|
"dnodesize",
|
||||||
"dotfiles",
|
"dotfiles",
|
||||||
"drawio",
|
"drawio",
|
||||||
|
"duckdns",
|
||||||
"eamodio",
|
"eamodio",
|
||||||
|
"ehci",
|
||||||
|
"emerg",
|
||||||
"endlessh",
|
"endlessh",
|
||||||
"errorlens",
|
"errorlens",
|
||||||
"esbenp",
|
"esbenp",
|
||||||
"esphome",
|
"esphome",
|
||||||
"extest",
|
"extest",
|
||||||
|
"fadvise",
|
||||||
|
"fastfetch",
|
||||||
"fastforwardteam",
|
"fastforwardteam",
|
||||||
"FASTFOX",
|
"FASTFOX",
|
||||||
"ffmpegthumbnailer",
|
"ffmpegthumbnailer",
|
||||||
"filebot",
|
"filebot",
|
||||||
"filebrowser",
|
|
||||||
"fileroller",
|
"fileroller",
|
||||||
"findbar",
|
"findbar",
|
||||||
"Fira",
|
"Fira",
|
||||||
"fmask",
|
"fmask",
|
||||||
"fontconfig",
|
"fontconfig",
|
||||||
"formfill",
|
"formfill",
|
||||||
|
"forwardfor",
|
||||||
"foxundermoon",
|
"foxundermoon",
|
||||||
"FULLSCREEN",
|
"FULLSCREEN",
|
||||||
"fwupd",
|
"fwupd",
|
||||||
@@ -79,9 +98,11 @@
|
|||||||
"getch",
|
"getch",
|
||||||
"getmaxyx",
|
"getmaxyx",
|
||||||
"ghdeploy",
|
"ghdeploy",
|
||||||
|
"gitea",
|
||||||
"globalprivacycontrol",
|
"globalprivacycontrol",
|
||||||
"gparted",
|
"gparted",
|
||||||
"gtts",
|
"gtts",
|
||||||
|
"gutenprint",
|
||||||
"hass",
|
"hass",
|
||||||
"healthreport",
|
"healthreport",
|
||||||
"Heatsink",
|
"Heatsink",
|
||||||
@@ -91,24 +112,34 @@
|
|||||||
"hmac",
|
"hmac",
|
||||||
"homeassistant",
|
"homeassistant",
|
||||||
"HPKP",
|
"HPKP",
|
||||||
|
"hplip",
|
||||||
"htmlaboutaddons",
|
"htmlaboutaddons",
|
||||||
|
"httpchk",
|
||||||
"hurlenko",
|
"hurlenko",
|
||||||
"hwloc",
|
"hwloc",
|
||||||
|
"ical",
|
||||||
|
"ignorelist",
|
||||||
|
"improv",
|
||||||
"INITDB",
|
"INITDB",
|
||||||
"iocharset",
|
"iocharset",
|
||||||
"ioit",
|
"ioit",
|
||||||
"iperf",
|
"iperf",
|
||||||
"isal",
|
"isal",
|
||||||
|
"jellyfin",
|
||||||
"jnoortheen",
|
"jnoortheen",
|
||||||
"jsbc",
|
"jsbc",
|
||||||
"kagi",
|
"kagi",
|
||||||
|
"keyformat",
|
||||||
|
"keylocation",
|
||||||
"kuma",
|
"kuma",
|
||||||
|
"lazer",
|
||||||
"levelname",
|
"levelname",
|
||||||
"libglvnd",
|
"libglvnd",
|
||||||
"libmysqlclient",
|
"libmysqlclient",
|
||||||
"libsodium",
|
"libsodium",
|
||||||
"libssh",
|
"libssh",
|
||||||
"libvirtd",
|
"libvirtd",
|
||||||
|
"llms",
|
||||||
"localtime",
|
"localtime",
|
||||||
"louislam",
|
"louislam",
|
||||||
"lsnew",
|
"lsnew",
|
||||||
@@ -116,43 +147,56 @@
|
|||||||
"lynis",
|
"lynis",
|
||||||
"mangohud",
|
"mangohud",
|
||||||
"markdownlint",
|
"markdownlint",
|
||||||
|
"maxconn",
|
||||||
|
"maxpages",
|
||||||
"maxretry",
|
"maxretry",
|
||||||
"maxtime",
|
"maxtime",
|
||||||
"mechatroner",
|
"mechatroner",
|
||||||
"mediainfo",
|
"mediainfo",
|
||||||
|
"mixtral",
|
||||||
"mklabel",
|
"mklabel",
|
||||||
"mkpart",
|
"mkpart",
|
||||||
|
"modbus",
|
||||||
|
"modbuss",
|
||||||
"modesetting",
|
"modesetting",
|
||||||
"mountpoint",
|
"mountpoint",
|
||||||
"mountpoints",
|
"mountpoints",
|
||||||
"mousewheel",
|
"mousewheel",
|
||||||
|
"mqtt",
|
||||||
"mtxr",
|
"mtxr",
|
||||||
"muninn",
|
"mypy",
|
||||||
"ncdu",
|
"ncdu",
|
||||||
"nemo",
|
"nemo",
|
||||||
"neofetch",
|
|
||||||
"nerdfonts",
|
"nerdfonts",
|
||||||
"netdev",
|
"netdev",
|
||||||
"netdevs",
|
"netdevs",
|
||||||
"Networkd",
|
"Networkd",
|
||||||
"networkmanager",
|
"networkmanager",
|
||||||
"newtabpage",
|
"newtabpage",
|
||||||
|
"nixfmt",
|
||||||
"nixos",
|
"nixos",
|
||||||
"nixpkgs",
|
"nixpkgs",
|
||||||
"nmap",
|
"nmap",
|
||||||
"noauto",
|
"noauto",
|
||||||
|
"nodev",
|
||||||
"noecho",
|
"noecho",
|
||||||
"nonsponsored",
|
"nonsponsored",
|
||||||
"Noto",
|
"Noto",
|
||||||
|
"nprt",
|
||||||
"nvme",
|
"nvme",
|
||||||
"OCSP",
|
"OCSP",
|
||||||
"oderwat",
|
"oderwat",
|
||||||
|
"ollama",
|
||||||
|
"ondemand",
|
||||||
"oneshot",
|
"oneshot",
|
||||||
|
"openwakeword",
|
||||||
"optimise",
|
"optimise",
|
||||||
"optoutstudies",
|
"optoutstudies",
|
||||||
"overalljails",
|
"overalljails",
|
||||||
"overscroll",
|
"overscroll",
|
||||||
"overseerr",
|
"overseerr",
|
||||||
|
"paho",
|
||||||
|
"partitionwise",
|
||||||
"pbmode",
|
"pbmode",
|
||||||
"pciutils",
|
"pciutils",
|
||||||
"pcscd",
|
"pcscd",
|
||||||
@@ -160,16 +204,17 @@
|
|||||||
"peerconnection",
|
"peerconnection",
|
||||||
"PESKYFOX",
|
"PESKYFOX",
|
||||||
"PGID",
|
"PGID",
|
||||||
"photoprism",
|
|
||||||
"pipewire",
|
"pipewire",
|
||||||
"pkgs",
|
"pkgs",
|
||||||
"plugdev",
|
"plugdev",
|
||||||
"poppler",
|
"poppler",
|
||||||
"posixacl",
|
"posixacl",
|
||||||
|
"powertop",
|
||||||
"primarycache",
|
"primarycache",
|
||||||
"prismlauncher",
|
"prismlauncher",
|
||||||
"privatebrowsing",
|
"privatebrowsing",
|
||||||
"PRIVOXY",
|
"PRIVOXY",
|
||||||
|
"protontricks",
|
||||||
"prowlarr",
|
"prowlarr",
|
||||||
"proxychains",
|
"proxychains",
|
||||||
"prusa",
|
"prusa",
|
||||||
@@ -177,21 +222,30 @@
|
|||||||
"PUID",
|
"PUID",
|
||||||
"pulseaudio",
|
"pulseaudio",
|
||||||
"punycode",
|
"punycode",
|
||||||
|
"pychromecast",
|
||||||
|
"pydocstyle",
|
||||||
|
"pyfakefs",
|
||||||
"pylance",
|
"pylance",
|
||||||
|
"pylint",
|
||||||
"pymetno",
|
"pymetno",
|
||||||
"qbit",
|
"pymodbus",
|
||||||
"qbittorrent",
|
"pyopenweathermap",
|
||||||
"qbittorrentvpn",
|
"pyownet",
|
||||||
"qbitvpn",
|
"pytest",
|
||||||
|
"qalculate",
|
||||||
"quicksuggest",
|
"quicksuggest",
|
||||||
"radarr",
|
"radarr",
|
||||||
"readahead",
|
"readahead",
|
||||||
|
"receiveencrypted",
|
||||||
|
"recordsize",
|
||||||
"Redistributable",
|
"Redistributable",
|
||||||
"referer",
|
"referer",
|
||||||
"REFERERS",
|
"REFERERS",
|
||||||
"relatime",
|
"relatime",
|
||||||
"Rhosts",
|
"Rhosts",
|
||||||
"ripgrep",
|
"ripgrep",
|
||||||
|
"roboto",
|
||||||
|
"rokuecp",
|
||||||
"routable",
|
"routable",
|
||||||
"rspace",
|
"rspace",
|
||||||
"rtkit",
|
"rtkit",
|
||||||
@@ -200,8 +254,10 @@
|
|||||||
"schemeless",
|
"schemeless",
|
||||||
"scrollback",
|
"scrollback",
|
||||||
"SECUREFOX",
|
"SECUREFOX",
|
||||||
|
"sessionmaker",
|
||||||
"sessionstore",
|
"sessionstore",
|
||||||
"shellcheck",
|
"shellcheck",
|
||||||
|
"signalbot",
|
||||||
"signon",
|
"signon",
|
||||||
"Signons",
|
"Signons",
|
||||||
"skia",
|
"skia",
|
||||||
@@ -211,8 +267,10 @@
|
|||||||
"socialtracking",
|
"socialtracking",
|
||||||
"sonarr",
|
"sonarr",
|
||||||
"sponsorblock",
|
"sponsorblock",
|
||||||
|
"sqlalchemy",
|
||||||
"sqltools",
|
"sqltools",
|
||||||
"ssdp",
|
"ssdp",
|
||||||
|
"SSHOPTS",
|
||||||
"stdenv",
|
"stdenv",
|
||||||
"subresource",
|
"subresource",
|
||||||
"substituters",
|
"substituters",
|
||||||
@@ -220,19 +278,28 @@
|
|||||||
"sysstat",
|
"sysstat",
|
||||||
"tabmanager",
|
"tabmanager",
|
||||||
"tamasfe",
|
"tamasfe",
|
||||||
|
"TCPIP",
|
||||||
|
"testdisk",
|
||||||
"tiktok",
|
"tiktok",
|
||||||
"timonwong",
|
"timonwong",
|
||||||
|
"titlebar",
|
||||||
"tmmworkshop",
|
"tmmworkshop",
|
||||||
"Tmpfs",
|
"Tmpfs",
|
||||||
"topsites",
|
"topsites",
|
||||||
"topstories",
|
"topstories",
|
||||||
"torrenting",
|
"treefmt",
|
||||||
"twimg",
|
"twimg",
|
||||||
|
"typedmonarchmoney",
|
||||||
|
"typer",
|
||||||
"uaccess",
|
"uaccess",
|
||||||
|
"ubiquiti",
|
||||||
"ublock",
|
"ublock",
|
||||||
|
"uiprotect",
|
||||||
"uitour",
|
"uitour",
|
||||||
|
"unifi",
|
||||||
"unrar",
|
"unrar",
|
||||||
"unsubmitted",
|
"unsubmitted",
|
||||||
|
"uptimekuma",
|
||||||
"urlbar",
|
"urlbar",
|
||||||
"urlclassifier",
|
"urlclassifier",
|
||||||
"usbhid",
|
"usbhid",
|
||||||
@@ -240,16 +307,22 @@
|
|||||||
"useragent",
|
"useragent",
|
||||||
"usernamehw",
|
"usernamehw",
|
||||||
"userprefs",
|
"userprefs",
|
||||||
|
"vaninventory",
|
||||||
"vfat",
|
"vfat",
|
||||||
|
"victron",
|
||||||
"virt",
|
"virt",
|
||||||
"virtualisation",
|
"virtualisation",
|
||||||
"vpnpromourl",
|
"vpnpromourl",
|
||||||
|
"wakeonlan",
|
||||||
"webchannel",
|
"webchannel",
|
||||||
"WEBRTC",
|
"WEBRTC",
|
||||||
"WEBUI",
|
"WEBUI",
|
||||||
|
"wireplumber",
|
||||||
"wireshark",
|
"wireshark",
|
||||||
"Workqueues",
|
"Workqueues",
|
||||||
"xattr",
|
"xattr",
|
||||||
|
"xcursorgen",
|
||||||
|
"xdist",
|
||||||
"xhci",
|
"xhci",
|
||||||
"yazi",
|
"yazi",
|
||||||
"yubikey",
|
"yubikey",
|
||||||
@@ -261,5 +334,10 @@
|
|||||||
"zoxide",
|
"zoxide",
|
||||||
"zram",
|
"zram",
|
||||||
"zstd"
|
"zstd"
|
||||||
]
|
],
|
||||||
|
"python-envs.defaultEnvManager": "ms-python.python:system",
|
||||||
|
"python-envs.pythonProjects": [],
|
||||||
|
"python.testing.pytestArgs": ["tests"],
|
||||||
|
"python.testing.unittestEnabled": false,
|
||||||
|
"python.testing.pytestEnabled": true
|
||||||
}
|
}
|
||||||
|
|||||||
12
AGENTS.md
Normal file
12
AGENTS.md
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
## Dev environment tips
|
||||||
|
|
||||||
|
- use treefmt to format all files
|
||||||
|
- make python code ruff compliant
|
||||||
|
- use pytest to test python code
|
||||||
|
- always use the minimum amount of complexity
|
||||||
|
- if judgment calls are easy to reverse make them. if not ask me first
|
||||||
|
- Match existing code style.
|
||||||
|
- Use builtin helpers getenv() over os.environ.get.
|
||||||
|
- Prefer single-purpose functions over “do everything” helpers.
|
||||||
|
- Avoid compatibility branches like PG_USER and POSTGRESQL_URL unless requested.
|
||||||
|
- Keep helpers only if reused or they simplify the code otherwise inline.
|
||||||
25
README.md
25
README.md
@@ -1 +1,26 @@
|
|||||||
# dotfiles
|
# dotfiles
|
||||||
|
|
||||||
|
<!-- LINE-COUNT-START -->
|
||||||
|
This repo has **20,055** lines of technical debt.
|
||||||
|
|
||||||
|
| File Type | Lines | Percentage |
|
||||||
|
|-----------|------:|-----------:|
|
||||||
|
| .py | 11,441 | 57.0% |
|
||||||
|
| .nix | 4,471 | 22.3% |
|
||||||
|
| .yaml | 1,121 | 5.6% |
|
||||||
|
| .html | 1,009 | 5.0% |
|
||||||
|
| .json | 555 | 2.8% |
|
||||||
|
| .yml | 479 | 2.4% |
|
||||||
|
| .toml | 290 | 1.4% |
|
||||||
|
| .css | 212 | 1.1% |
|
||||||
|
| .gitignore | 199 | 1.0% |
|
||||||
|
| .md | 75 | 0.4% |
|
||||||
|
| .cfg | 73 | 0.4% |
|
||||||
|
| .sh | 48 | 0.2% |
|
||||||
|
| .mako | 36 | 0.2% |
|
||||||
|
| .LICENSE | 21 | 0.1% |
|
||||||
|
| .conf | 17 | 0.1% |
|
||||||
|
| .Gemfile | 4 | 0.0% |
|
||||||
|
| .svg | 3 | 0.0% |
|
||||||
|
| .new | 1 | 0.0% |
|
||||||
|
<!-- LINE-COUNT-END -->
|
||||||
|
|||||||
6
build.sh
6
build.sh
@@ -1,6 +0,0 @@
|
|||||||
#!/bin/sh
|
|
||||||
|
|
||||||
nixos-rebuild build --flake /home/richie/projects/dotfiles#bob
|
|
||||||
nixos-rebuild build --flake /home/richie/projects/dotfiles#jeeves
|
|
||||||
nixos-rebuild build --flake /home/richie/projects/dotfiles#muninn
|
|
||||||
nixos-rebuild build --flake /home/richie/projects/dotfiles#rhapsody-in-green
|
|
||||||
@@ -8,6 +8,7 @@
|
|||||||
{
|
{
|
||||||
imports = [
|
imports = [
|
||||||
inputs.home-manager.nixosModules.home-manager
|
inputs.home-manager.nixosModules.home-manager
|
||||||
|
inputs.sops-nix.nixosModules.sops
|
||||||
./fail2ban.nix
|
./fail2ban.nix
|
||||||
./fonts.nix
|
./fonts.nix
|
||||||
./libs.nix
|
./libs.nix
|
||||||
@@ -22,7 +23,7 @@
|
|||||||
boot = {
|
boot = {
|
||||||
tmp.useTmpfs = true;
|
tmp.useTmpfs = true;
|
||||||
kernelPackages = lib.mkDefault pkgs.linuxPackages_6_12;
|
kernelPackages = lib.mkDefault pkgs.linuxPackages_6_12;
|
||||||
zfs.package = lib.mkDefault pkgs.zfs;
|
zfs.package = lib.mkDefault pkgs.zfs_2_4;
|
||||||
};
|
};
|
||||||
|
|
||||||
hardware.enableRedistributableFirmware = true;
|
hardware.enableRedistributableFirmware = true;
|
||||||
@@ -30,23 +31,44 @@
|
|||||||
home-manager = {
|
home-manager = {
|
||||||
useGlobalPkgs = true;
|
useGlobalPkgs = true;
|
||||||
useUserPackages = true;
|
useUserPackages = true;
|
||||||
extraSpecialArgs = {inherit inputs outputs;};
|
extraSpecialArgs = { inherit inputs outputs; };
|
||||||
|
backupFileExtension = "backup";
|
||||||
};
|
};
|
||||||
|
|
||||||
nixpkgs = {
|
nixpkgs = {
|
||||||
overlays = builtins.attrValues outputs.overlays;
|
overlays = builtins.attrValues outputs.overlays;
|
||||||
config = {
|
config.allowUnfree = true;
|
||||||
allowUnfree = true;
|
};
|
||||||
|
|
||||||
|
services = {
|
||||||
|
# firmware update
|
||||||
|
fwupd.enable = true;
|
||||||
|
|
||||||
|
snapshot_manager = {
|
||||||
|
enable = lib.mkDefault true;
|
||||||
|
PYTHONPATH = "${inputs.self}/";
|
||||||
|
};
|
||||||
|
|
||||||
|
zfs = {
|
||||||
|
trim.enable = lib.mkDefault true;
|
||||||
|
autoScrub.enable = lib.mkDefault true;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
services.fwupd.enable = true;
|
powerManagement.powertop.enable = lib.mkDefault true;
|
||||||
|
|
||||||
programs.zsh.enable = true;
|
programs.zsh.enable = true;
|
||||||
|
|
||||||
security.auditd.enable = lib.mkDefault true;
|
security = {
|
||||||
|
auditd.enable = lib.mkDefault true;
|
||||||
|
sudo-rs = {
|
||||||
|
enable = true;
|
||||||
|
execWheelOnly = true;
|
||||||
|
};
|
||||||
|
sudo.enable = false;
|
||||||
|
};
|
||||||
|
|
||||||
users.mutableUsers = lib.mkDefault true;
|
users.mutableUsers = lib.mkDefault false;
|
||||||
|
|
||||||
zramSwap = {
|
zramSwap = {
|
||||||
enable = lib.mkDefault true;
|
enable = lib.mkDefault true;
|
||||||
|
|||||||
@@ -1,4 +1,10 @@
|
|||||||
{ lib, pkgs, ... }:
|
{ lib, pkgs, ... }:
|
||||||
|
let
|
||||||
|
libPath = pkgs.lib.makeLibraryPath [
|
||||||
|
pkgs.zlib
|
||||||
|
pkgs.stdenv.cc.cc.lib
|
||||||
|
];
|
||||||
|
in
|
||||||
{
|
{
|
||||||
programs.nix-ld = {
|
programs.nix-ld = {
|
||||||
enable = lib.mkDefault true;
|
enable = lib.mkDefault true;
|
||||||
@@ -15,6 +21,7 @@
|
|||||||
libxml2
|
libxml2
|
||||||
openssl
|
openssl
|
||||||
stdenv.cc.cc
|
stdenv.cc.cc
|
||||||
|
stdenv.cc.cc.lib
|
||||||
systemd
|
systemd
|
||||||
util-linux
|
util-linux
|
||||||
xz
|
xz
|
||||||
@@ -23,4 +30,9 @@
|
|||||||
zstd
|
zstd
|
||||||
];
|
];
|
||||||
};
|
};
|
||||||
|
|
||||||
|
environment = {
|
||||||
|
sessionVariables.LD_LIBRARY_PATH = lib.mkDefault libPath;
|
||||||
|
variables.LD_LIBRARY_PATH = lib.mkDefault libPath;
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,8 +4,7 @@
|
|||||||
console.keyMap = lib.mkDefault "us";
|
console.keyMap = lib.mkDefault "us";
|
||||||
|
|
||||||
i18n = {
|
i18n = {
|
||||||
defaultLocale = lib.mkDefault "en_US.utf8";
|
defaultLocale = lib.mkDefault "en_US.UTF-8";
|
||||||
supportedLocales = lib.mkDefault [ "en_US.UTF-8/UTF-8" ];
|
|
||||||
extraLocaleSettings = lib.mkDefault {
|
extraLocaleSettings = lib.mkDefault {
|
||||||
LC_ADDRESS = "en_US.UTF-8";
|
LC_ADDRESS = "en_US.UTF-8";
|
||||||
LC_IDENTIFICATION = "en_US.UTF-8";
|
LC_IDENTIFICATION = "en_US.UTF-8";
|
||||||
|
|||||||
@@ -2,9 +2,11 @@
|
|||||||
inputs,
|
inputs,
|
||||||
lib,
|
lib,
|
||||||
...
|
...
|
||||||
}: let
|
}:
|
||||||
|
let
|
||||||
flakeInputs = lib.filterAttrs (_: lib.isType "flake") inputs;
|
flakeInputs = lib.filterAttrs (_: lib.isType "flake") inputs;
|
||||||
in {
|
in
|
||||||
|
{
|
||||||
nix = {
|
nix = {
|
||||||
settings = {
|
settings = {
|
||||||
trusted-users = [
|
trusted-users = [
|
||||||
@@ -13,17 +15,14 @@ in {
|
|||||||
];
|
];
|
||||||
trusted-substituters = [
|
trusted-substituters = [
|
||||||
"https://cache.nixos.org"
|
"https://cache.nixos.org"
|
||||||
"https://cache.tmmworkshop.com"
|
|
||||||
"https://nix-community.cachix.org"
|
"https://nix-community.cachix.org"
|
||||||
];
|
];
|
||||||
substituters = [
|
substituters = [
|
||||||
"https://cache.nixos.org/?priority=2&want-mass-query=true"
|
"https://cache.nixos.org/?priority=2&want-mass-query=true"
|
||||||
"https://cache.tmmworkshop.com/?priority=2&want-mass-query=true"
|
|
||||||
"https://nix-community.cachix.org/?priority=10&want-mass-query=true"
|
"https://nix-community.cachix.org/?priority=10&want-mass-query=true"
|
||||||
];
|
];
|
||||||
trusted-public-keys = [
|
trusted-public-keys = [
|
||||||
"cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY="
|
"cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY="
|
||||||
"cache.tmmworkshop.com:jHffkpgbmEdstQPoihJPYW9TQe6jnQbWR2LqkNGV3iA="
|
|
||||||
"nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs="
|
"nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs="
|
||||||
];
|
];
|
||||||
auto-optimise-store = lib.mkDefault true;
|
auto-optimise-store = lib.mkDefault true;
|
||||||
@@ -34,10 +33,12 @@ in {
|
|||||||
];
|
];
|
||||||
warn-dirty = false;
|
warn-dirty = false;
|
||||||
flake-registry = ""; # disable global flake registries
|
flake-registry = ""; # disable global flake registries
|
||||||
|
connect-timeout = 10;
|
||||||
|
fallback = true;
|
||||||
};
|
};
|
||||||
|
|
||||||
# Add each flake input as a registry and nix_path
|
# Add each flake input as a registry and nix_path
|
||||||
registry = lib.mapAttrs (_: flake: {inherit flake;}) flakeInputs;
|
registry = lib.mapAttrs (_: flake: { inherit flake; }) flakeInputs;
|
||||||
nixPath = lib.mapAttrsToList (n: _: "${n}=flake:${n}") flakeInputs;
|
nixPath = lib.mapAttrsToList (n: _: "${n}=flake:${n}") flakeInputs;
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,6 +2,6 @@
|
|||||||
{
|
{
|
||||||
environment.systemPackages = with pkgs; [
|
environment.systemPackages = with pkgs; [
|
||||||
git
|
git
|
||||||
python313
|
my_python
|
||||||
];
|
];
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,37 +1,57 @@
|
|||||||
{ inputs, pkgs, lib, config, ... }:
|
{
|
||||||
|
pkgs,
|
||||||
|
lib,
|
||||||
|
config,
|
||||||
|
...
|
||||||
|
}:
|
||||||
let
|
let
|
||||||
cfg = config.services.snapshot_manager;
|
cfg = config.services.snapshot_manager;
|
||||||
in
|
in
|
||||||
{
|
{
|
||||||
options = {
|
options = {
|
||||||
services.snapshot_manager = {
|
services.snapshot_manager = {
|
||||||
enable = lib.mkOption {
|
enable = lib.mkEnableOption "ZFS snapshot manager";
|
||||||
default = true;
|
|
||||||
example = true;
|
|
||||||
description = "Whether to enable k3s-net.";
|
|
||||||
type = lib.types.bool;
|
|
||||||
};
|
|
||||||
path = lib.mkOption {
|
path = lib.mkOption {
|
||||||
type = lib.types.path;
|
type = lib.types.path;
|
||||||
description = "Path that needs to be updated via git pull";
|
|
||||||
default = ./snapshot_config.toml;
|
default = ./snapshot_config.toml;
|
||||||
|
description = "Path to the snapshot_manager TOML config.";
|
||||||
|
};
|
||||||
|
PYTHONPATH = lib.mkOption {
|
||||||
|
type = lib.types.str;
|
||||||
|
description = ''
|
||||||
|
the PYTHONPATH to use for the snapshot_manager service.
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
EnvironmentFile = lib.mkOption {
|
||||||
|
type = lib.types.nullOr (lib.types.coercedTo lib.types.path toString lib.types.str);
|
||||||
|
default = null;
|
||||||
|
description = ''
|
||||||
|
Single environment file for the service (e.g. /etc/snapshot-manager/env).
|
||||||
|
Use a leading "-" to ignore if missing (systemd feature).
|
||||||
|
'';
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
config = lib.mkIf cfg.enable {
|
config = lib.mkIf cfg.enable {
|
||||||
systemd = {
|
systemd = {
|
||||||
services."snapshot_manager" = {
|
services.snapshot_manager = {
|
||||||
description = "ZFS Snapshot Manager";
|
description = "ZFS Snapshot Manager";
|
||||||
requires = [ "zfs-import.target" ];
|
requires = [ "zfs-import.target" ];
|
||||||
after = [ "zfs-import.target" ];
|
after = [ "zfs-import.target" ];
|
||||||
path = [ pkgs.zfs ];
|
path = [ pkgs.zfs ];
|
||||||
|
environment = {
|
||||||
|
PYTHONPATH = cfg.PYTHONPATH;
|
||||||
|
};
|
||||||
serviceConfig = {
|
serviceConfig = {
|
||||||
Type = "oneshot";
|
Type = "oneshot";
|
||||||
ExecStart = "${inputs.system_tools.packages.x86_64-linux.default}/bin/snapshot_manager --config-file='${cfg.path}'";
|
ExecStart = "${pkgs.my_python}/bin/python -m python.tools.snapshot_manager ${lib.escapeShellArg cfg.path}";
|
||||||
|
}
|
||||||
|
// lib.optionalAttrs (cfg.EnvironmentFile != null) {
|
||||||
|
EnvironmentFile = cfg.EnvironmentFile;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
timers."snapshot_manager" = {
|
timers.snapshot_manager = {
|
||||||
wantedBy = [ "timers.target" ];
|
wantedBy = [ "timers.target" ];
|
||||||
timerConfig = {
|
timerConfig = {
|
||||||
OnBootSec = "15m";
|
OnBootSec = "15m";
|
||||||
@@ -41,4 +61,4 @@ in
|
|||||||
};
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -37,6 +37,8 @@
|
|||||||
TcpKeepAlive = "no";
|
TcpKeepAlive = "no";
|
||||||
X11Forwarding = lib.mkDefault false;
|
X11Forwarding = lib.mkDefault false;
|
||||||
KexAlgorithms = [
|
KexAlgorithms = [
|
||||||
|
"sntrup761x25519-sha512@openssh.com"
|
||||||
|
"mlkem768x25519-sha256"
|
||||||
"curve25519-sha256@libssh.org"
|
"curve25519-sha256@libssh.org"
|
||||||
"diffie-hellman-group-exchange-sha256"
|
"diffie-hellman-group-exchange-sha256"
|
||||||
];
|
];
|
||||||
@@ -67,5 +69,5 @@
|
|||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
networking.firewall.allowedTCPPorts = [ 22 ];
|
networking.firewall.allowedTCPPorts = [ 22 ];
|
||||||
}
|
}
|
||||||
|
|||||||
6
common/optional/brain_substituter.nix
Normal file
6
common/optional/brain_substituter.nix
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
{
|
||||||
|
nix.settings = {
|
||||||
|
trusted-substituters = [ "http://192.168.95.35:5000" ];
|
||||||
|
substituters = [ "http://192.168.95.35:5000/?priority=1&want-mass-query=true" ];
|
||||||
|
};
|
||||||
|
}
|
||||||
@@ -1,10 +1,18 @@
|
|||||||
{ pkgs, ... }:
|
{ pkgs, ... }:
|
||||||
{
|
{
|
||||||
boot = {
|
boot = {
|
||||||
kernelPackages = pkgs.linuxPackages_6_12;
|
kernelPackages = pkgs.linuxPackages_6_18;
|
||||||
zfs.package = pkgs.zfs;
|
zfs.package = pkgs.zfs_2_4;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
hardware.bluetooth = {
|
||||||
|
enable = true;
|
||||||
|
powerOnBoot = true;
|
||||||
|
};
|
||||||
|
|
||||||
|
# rtkit is optional but recommended for pipewire
|
||||||
|
security.rtkit.enable = true;
|
||||||
|
|
||||||
services = {
|
services = {
|
||||||
displayManager.sddm = {
|
displayManager.sddm = {
|
||||||
enable = true;
|
enable = true;
|
||||||
@@ -21,6 +29,8 @@
|
|||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
pulseaudio.enable = false;
|
||||||
|
|
||||||
pipewire = {
|
pipewire = {
|
||||||
enable = true;
|
enable = true;
|
||||||
alsa.enable = true;
|
alsa.enable = true;
|
||||||
|
|||||||
@@ -1,5 +1,7 @@
|
|||||||
{ config, ... }:
|
{ config, ... }:
|
||||||
{
|
{
|
||||||
|
nixpkgs.config.cudaSupport = true;
|
||||||
|
|
||||||
services.xserver.videoDrivers = [ "nvidia" ];
|
services.xserver.videoDrivers = [ "nvidia" ];
|
||||||
hardware = {
|
hardware = {
|
||||||
nvidia = {
|
nvidia = {
|
||||||
10
common/optional/printing.nix
Normal file
10
common/optional/printing.nix
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{ pkgs, ... }:
|
||||||
|
{
|
||||||
|
services.printing = {
|
||||||
|
enable = true;
|
||||||
|
drivers = with pkgs; [
|
||||||
|
gutenprint
|
||||||
|
hplip
|
||||||
|
];
|
||||||
|
};
|
||||||
|
}
|
||||||
@@ -10,6 +10,9 @@
|
|||||||
authorizedKeys = config.users.users.richie.openssh.authorizedKeys.keys;
|
authorizedKeys = config.users.users.richie.openssh.authorizedKeys.keys;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
availableKernelModules = [ "igb" ];
|
availableKernelModules = [
|
||||||
|
"igb"
|
||||||
|
"r8152"
|
||||||
|
];
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,9 @@
|
|||||||
{ pkgs, ... }:
|
{ pkgs, ... }:
|
||||||
{
|
{
|
||||||
environment.systemPackages = with pkgs; [mangohud steam-run];
|
environment.systemPackages = with pkgs; [
|
||||||
|
mangohud
|
||||||
|
steam-run
|
||||||
|
];
|
||||||
hardware.steam-hardware.enable = true;
|
hardware.steam-hardware.enable = true;
|
||||||
|
|
||||||
programs = {
|
programs = {
|
||||||
@@ -11,7 +14,7 @@
|
|||||||
remotePlay.openFirewall = true;
|
remotePlay.openFirewall = true;
|
||||||
localNetworkGameTransfers.openFirewall = true;
|
localNetworkGameTransfers.openFirewall = true;
|
||||||
protontricks.enable = true;
|
protontricks.enable = true;
|
||||||
extraCompatPackages = with pkgs; [proton-ge-bin];
|
extraCompatPackages = with pkgs; [ proton-ge-bin ];
|
||||||
extest.enable = true;
|
extest.enable = true;
|
||||||
};
|
};
|
||||||
gamescope = {
|
gamescope = {
|
||||||
|
|||||||
@@ -1,19 +1,19 @@
|
|||||||
|
{ lib, ... }:
|
||||||
{
|
{
|
||||||
services.syncthing = {
|
services.syncthing = {
|
||||||
enable = true;
|
enable = true;
|
||||||
user = "richie";
|
user = "richie";
|
||||||
overrideDevices = true;
|
overrideDevices = true;
|
||||||
overrideFolders = true;
|
overrideFolders = lib.mkDefault true;
|
||||||
dataDir = "/home/richie/Syncthing";
|
dataDir = "/home/richie/Syncthing";
|
||||||
configDir = "/home/richie/.config/syncthing";
|
configDir = "/home/richie/.config/syncthing";
|
||||||
settings = {
|
settings.devices = {
|
||||||
devices = {
|
bob.id = "CJIAPEJ-VO74RR4-F75VU6M-QNZAMYG-FYUJG7Y-6AT62HJ-355PRPL-PJFETAZ"; # cspell:disable-line
|
||||||
phone.id = "LTGPLAE-M4ZDJTM-TZ3DJGY-SLLAVWF-CQDVEVS-RGCS75T-GAPZYK3-KUM6LA5"; # cspell:disable-line
|
brain.id = "SSCGIPI-IV3VYKB-TRNIJE3-COV4T2H-CDBER7F-I2CGHYA-NWOEUDU-3T5QAAN"; # cspell:disable-line
|
||||||
jeeves.id = "ICRHXZW-ECYJCUZ-I4CZ64R-3XRK7CG-LL2HAAK-FGOHD22-BQA4AI6-5OAL6AG"; # cspell:disable-line
|
ipad.id = "KI76T3X-SFUGV2L-VSNYTKR-TSIUV5L-SHWD3HE-GQRGRCN-GY4UFMD-CW6Z6AX"; # cspell:disable-line
|
||||||
ipad.id = "KI76T3X-SFUGV2L-VSNYTKR-TSIUV5L-SHWD3HE-GQRGRCN-GY4UFMD-CW6Z6AX"; # cspell:disable-line
|
jeeves.id = "ICRHXZW-ECYJCUZ-I4CZ64R-3XRK7CG-LL2HAAK-FGOHD22-BQA4AI6-5OAL6AG"; # cspell:disable-line
|
||||||
bob.id = "CJIAPEJ-VO74RR4-F75VU6M-QNZAMYG-FYUJG7Y-6AT62HJ-355PRPL-PJFETAZ"; # cspell:disable-line
|
phone.id = "TBRULKD-7DZPGGZ-F6LLB7J-MSO54AY-7KLPBIN-QOFK6PX-W2HBEWI-PHM2CQI"; # cspell:disable-line
|
||||||
rhapsody-in-green.id = "ASL3KC4-3XEN6PA-7BQBRKE-A7JXLI6-DJT43BY-Q4WPOER-7UALUAZ-VTPQ6Q4"; # cspell:disable-line
|
rhapsody-in-green.id = "ASL3KC4-3XEN6PA-7BQBRKE-A7JXLI6-DJT43BY-Q4WPOER-7UALUAZ-VTPQ6Q4"; # cspell:disable-line
|
||||||
};
|
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
7
common/optional/tmmworkshop_cache.nix
Normal file
7
common/optional/tmmworkshop_cache.nix
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
{
|
||||||
|
nix.settings = {
|
||||||
|
trusted-substituters = [ "http://cache.tmmworkshop.com" ];
|
||||||
|
substituters = [ "http://cache.tmmworkshop.com/?priority=1&want-mass-query=true" ];
|
||||||
|
trusted-public-keys = [ "cache.tmmworkshop.com:jHffkpgbmEdstQPoihJPYW9TQe6jnQbWR2LqkNGV3iA=" ];
|
||||||
|
};
|
||||||
|
}
|
||||||
@@ -1,32 +1,11 @@
|
|||||||
{ lib, pkgs, ... }:
|
|
||||||
{
|
{
|
||||||
systemd = {
|
|
||||||
services."autopull@dotfiles" = {
|
|
||||||
requires = [ "multi-user.target" ];
|
|
||||||
after = [ "multi-user.target" ];
|
|
||||||
description = "Pull the latest data for dotfiles";
|
|
||||||
serviceConfig = {
|
|
||||||
Type = "oneshot";
|
|
||||||
User = "root";
|
|
||||||
WorkingDirectory = /root/dotfiles;
|
|
||||||
ExecStart = "${pkgs.git}/bin/git pull --all --prune";
|
|
||||||
};
|
|
||||||
};
|
|
||||||
timers."autopull@dotfiles" = {
|
|
||||||
wantedBy = [ "timers.target" ];
|
|
||||||
timerConfig = {
|
|
||||||
OnBootSec = "1h";
|
|
||||||
OnUnitActiveSec = "1h";
|
|
||||||
Unit = "autopull@dotfiles.service";
|
|
||||||
};
|
|
||||||
};
|
|
||||||
};
|
|
||||||
|
|
||||||
system.autoUpgrade = {
|
system.autoUpgrade = {
|
||||||
enable = lib.mkDefault true;
|
enable = true;
|
||||||
flags = [ "--accept-flake-config" ];
|
flags = [ "--accept-flake-config" ];
|
||||||
randomizedDelaySec = "1h";
|
randomizedDelaySec = "1h";
|
||||||
persistent = true;
|
persistent = true;
|
||||||
flake = "github:RAD-Development/nix-dotfiles";
|
flake = "github:RichieCahill/dotfiles";
|
||||||
|
allowReboot = true;
|
||||||
|
dates = "Sat *-*-* 06:00:00";
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
4
docs/Gemfile
Normal file
4
docs/Gemfile
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
source "https://rubygems.org"
|
||||||
|
|
||||||
|
# The github-pages gem pins all compatible versions of Jekyll and its plugins
|
||||||
|
gem "github-pages", group: :jekyll_plugins
|
||||||
23
docs/_config.yml
Normal file
23
docs/_config.yml
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
title: "Richie Cahill"
|
||||||
|
description: "ALL THE CHAOS THAT I CANT DO AT WORK"
|
||||||
|
baseurl: "/dotfiles"
|
||||||
|
url: "https://richiecahill.github.io"
|
||||||
|
|
||||||
|
remote_theme: pages-themes/hacker@v0.2.0
|
||||||
|
plugins:
|
||||||
|
- jekyll-feed
|
||||||
|
- jekyll-remote-theme
|
||||||
|
- jekyll-seo-tag
|
||||||
|
- jekyll-sitemap
|
||||||
|
- jekyll-paginate
|
||||||
|
|
||||||
|
paginate: 5
|
||||||
|
paginate_path: "/page:num"
|
||||||
|
|
||||||
|
author:
|
||||||
|
name: "Richie Cahill"
|
||||||
|
email: "richie@tmmworkshop.com"
|
||||||
|
|
||||||
|
social_links:
|
||||||
|
github: "RichieCahill"
|
||||||
|
website: "https://tmmworkshop.com"
|
||||||
13
docs/_posts/2025-10-31-MONOREPO.md
Normal file
13
docs/_posts/2025-10-31-MONOREPO.md
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
# The MONOREPO experiment
|
||||||
|
|
||||||
|
Im testing a [MONOREPO](https://en.wikipedia.org/wiki/Monorepo) because Phil said this was a bad idea. To that i say hold my beer.
|
||||||
|
|
||||||
|
In all seriousness, I Think that for a small dev team/solo dev. The simplicity is worth higher barer to entry. One of my most annoying processes was updating my system tools. I had to build my update in a feature branch and then merge it into my main branch. then go to my dotfiles create a feature branch update the system tools merge it into main.
|
||||||
|
|
||||||
|
It will be starting with my Nix Dotfiles Python tools and now my blog.
|
||||||
|
|
||||||
|
I will be reaching ot to phil on 2030-10-31 and 2035-10-31 to give him updates on the progress.
|
||||||
|
|
||||||
|
Known Issues:
|
||||||
|
|
||||||
|
- the python tests are running on the current derivation not the one the derivation im updating to.
|
||||||
17
docs/index.md
Normal file
17
docs/index.md
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
---
|
||||||
|
layout: default
|
||||||
|
title: "Welcome"
|
||||||
|
---
|
||||||
|
|
||||||
|
Welcome to my build logs, notes, and experiments.
|
||||||
|
|
||||||
|
You can read my latest posts below
|
||||||
|
|
||||||
|
<ul>
|
||||||
|
{% for post in site.posts %}
|
||||||
|
<li>
|
||||||
|
<a href="{{ post.url | relative_url }}">{{ post.title }}</a>
|
||||||
|
<small>— {{ post.date | date: "%Y-%m-%d" }}</small>
|
||||||
|
</li>
|
||||||
|
{% endfor %}
|
||||||
|
</ul>
|
||||||
3
esphome/.gitignore
vendored
Normal file
3
esphome/.gitignore
vendored
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
# esphome
|
||||||
|
/.esphome/
|
||||||
|
/secrets.yaml
|
||||||
132
esphome/battery0.yml
Normal file
132
esphome/battery0.yml
Normal file
@@ -0,0 +1,132 @@
|
|||||||
|
esphome:
|
||||||
|
name: batteries
|
||||||
|
friendly_name: batteries
|
||||||
|
|
||||||
|
esp32:
|
||||||
|
board: esp32dev
|
||||||
|
framework:
|
||||||
|
type: arduino
|
||||||
|
|
||||||
|
logger:
|
||||||
|
|
||||||
|
api:
|
||||||
|
encryption:
|
||||||
|
key: !secret api_key
|
||||||
|
|
||||||
|
external_components:
|
||||||
|
- source: github://syssi/esphome-jk-bms@main
|
||||||
|
|
||||||
|
ota:
|
||||||
|
- platform: esphome
|
||||||
|
password: !secret ota_password
|
||||||
|
|
||||||
|
wifi:
|
||||||
|
ssid: !secret wifi_ssid
|
||||||
|
password: !secret wifi_password
|
||||||
|
fast_connect: on
|
||||||
|
|
||||||
|
captive_portal:
|
||||||
|
|
||||||
|
esp32_ble_tracker:
|
||||||
|
scan_parameters:
|
||||||
|
interval: 1100ms
|
||||||
|
window: 1100ms
|
||||||
|
active: true
|
||||||
|
|
||||||
|
ble_client:
|
||||||
|
- mac_address: "C8:47:80:29:0F:DB"
|
||||||
|
id: jk_ble0
|
||||||
|
|
||||||
|
jk_bms_ble:
|
||||||
|
- ble_client_id: jk_ble0
|
||||||
|
protocol_version: JK02_32S
|
||||||
|
throttle: 1s
|
||||||
|
id: jk_bms0
|
||||||
|
|
||||||
|
button:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
retrieve_settings:
|
||||||
|
name: "JK0 retrieve settings"
|
||||||
|
retrieve_device_info:
|
||||||
|
name: "JK0 retrieve device info"
|
||||||
|
|
||||||
|
sensor:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms0
|
||||||
|
total_voltage:
|
||||||
|
name: "JK0 Total Voltage"
|
||||||
|
state_of_charge:
|
||||||
|
name: "JK0 SoC"
|
||||||
|
charging_power:
|
||||||
|
name: "JK0 charging power"
|
||||||
|
discharging_power:
|
||||||
|
name: "JK0 discharging power"
|
||||||
|
temperature_sensor_1:
|
||||||
|
name: "JK0 Temp 1"
|
||||||
|
temperature_sensor_2:
|
||||||
|
name: "JK0 Temp 2"
|
||||||
|
balancing:
|
||||||
|
name: "JK0 balancing"
|
||||||
|
total_runtime:
|
||||||
|
name: "JK0 total runtime"
|
||||||
|
balancing_current:
|
||||||
|
name: "JK0 balancing current"
|
||||||
|
delta_cell_voltage:
|
||||||
|
name: "JK0 cell delta voltage"
|
||||||
|
average_cell_voltage:
|
||||||
|
name: "JK0 cell average voltage"
|
||||||
|
cell_voltage_1:
|
||||||
|
name: "JK0 cell voltage 1"
|
||||||
|
cell_voltage_2:
|
||||||
|
name: "JK0 cell voltage 2"
|
||||||
|
cell_voltage_3:
|
||||||
|
name: "JK0 cell voltage 3"
|
||||||
|
cell_voltage_4:
|
||||||
|
name: "JK0 cell voltage 4"
|
||||||
|
cell_voltage_5:
|
||||||
|
name: "JK0 cell voltage 5"
|
||||||
|
cell_voltage_6:
|
||||||
|
name: "JK0 cell voltage 6"
|
||||||
|
cell_voltage_7:
|
||||||
|
name: "JK0 cell voltage 7"
|
||||||
|
cell_voltage_8:
|
||||||
|
name: "JK0 cell voltage 8"
|
||||||
|
cell_resistance_1:
|
||||||
|
name: "JK0 cell resistance 1"
|
||||||
|
cell_resistance_2:
|
||||||
|
name: "JK0 cell resistance 2"
|
||||||
|
cell_resistance_3:
|
||||||
|
name: "JK0 cell resistance 3"
|
||||||
|
cell_resistance_4:
|
||||||
|
name: "JK0 cell resistance 4"
|
||||||
|
cell_resistance_5:
|
||||||
|
name: "JK0 cell resistance 5"
|
||||||
|
cell_resistance_6:
|
||||||
|
name: "JK0 cell resistance 6"
|
||||||
|
cell_resistance_7:
|
||||||
|
name: "JK0 cell resistance 7"
|
||||||
|
cell_resistance_8:
|
||||||
|
name: "JK0 cell resistance 8"
|
||||||
|
total_charging_cycle_capacity:
|
||||||
|
name: "JK0 total charging cycle capacity"
|
||||||
|
|
||||||
|
text_sensor:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms0
|
||||||
|
errors:
|
||||||
|
name: "JK0 Errors"
|
||||||
|
|
||||||
|
switch:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms0
|
||||||
|
charging:
|
||||||
|
name: "JK0 Charging"
|
||||||
|
discharging:
|
||||||
|
name: "JK0 Discharging"
|
||||||
|
balancer:
|
||||||
|
name: "JK0 Balancing"
|
||||||
|
|
||||||
|
- platform: ble_client
|
||||||
|
ble_client_id: jk_ble0
|
||||||
|
name: "JK0 enable bluetooth connection"
|
||||||
|
id: ble_client_switch0
|
||||||
132
esphome/battery1.yml
Normal file
132
esphome/battery1.yml
Normal file
@@ -0,0 +1,132 @@
|
|||||||
|
esphome:
|
||||||
|
name: battery1
|
||||||
|
friendly_name: battery1
|
||||||
|
|
||||||
|
esp32:
|
||||||
|
board: esp32dev
|
||||||
|
framework:
|
||||||
|
type: arduino
|
||||||
|
|
||||||
|
logger:
|
||||||
|
|
||||||
|
api:
|
||||||
|
encryption:
|
||||||
|
key: !secret api_key
|
||||||
|
|
||||||
|
external_components:
|
||||||
|
- source: github://syssi/esphome-jk-bms@main
|
||||||
|
|
||||||
|
ota:
|
||||||
|
- platform: esphome
|
||||||
|
password: !secret ota_password
|
||||||
|
|
||||||
|
wifi:
|
||||||
|
ssid: !secret wifi_ssid
|
||||||
|
password: !secret wifi_password
|
||||||
|
fast_connect: on
|
||||||
|
|
||||||
|
captive_portal:
|
||||||
|
|
||||||
|
esp32_ble_tracker:
|
||||||
|
scan_parameters:
|
||||||
|
interval: 1100ms
|
||||||
|
window: 1100ms
|
||||||
|
active: true
|
||||||
|
|
||||||
|
ble_client:
|
||||||
|
- mac_address: "C8:47:80:37:9D:DD"
|
||||||
|
id: jk_ble1
|
||||||
|
|
||||||
|
jk_bms_ble:
|
||||||
|
- ble_client_id: jk_ble1
|
||||||
|
protocol_version: JK02_32S
|
||||||
|
throttle: 1s
|
||||||
|
id: jk_bms1
|
||||||
|
|
||||||
|
button:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
retrieve_settings:
|
||||||
|
name: "JK1 retrieve settings"
|
||||||
|
retrieve_device_info:
|
||||||
|
name: "JK1 retrieve device info"
|
||||||
|
|
||||||
|
sensor:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms1
|
||||||
|
total_voltage:
|
||||||
|
name: "JK1 Total Voltage"
|
||||||
|
state_of_charge:
|
||||||
|
name: "JK1 SoC"
|
||||||
|
charging_power:
|
||||||
|
name: "JK1 charging power"
|
||||||
|
discharging_power:
|
||||||
|
name: "JK1 discharging power"
|
||||||
|
temperature_sensor_1:
|
||||||
|
name: "JK1 Temp 1"
|
||||||
|
temperature_sensor_2:
|
||||||
|
name: "JK1 Temp 2"
|
||||||
|
balancing:
|
||||||
|
name: "JK1 balancing"
|
||||||
|
total_runtime:
|
||||||
|
name: "JK1 total runtime"
|
||||||
|
balancing_current:
|
||||||
|
name: "JK1 balancing current"
|
||||||
|
delta_cell_voltage:
|
||||||
|
name: "JK1 cell delta voltage"
|
||||||
|
average_cell_voltage:
|
||||||
|
name: "JK1 cell average voltage"
|
||||||
|
cell_voltage_1:
|
||||||
|
name: "JK1 cell voltage 1"
|
||||||
|
cell_voltage_2:
|
||||||
|
name: "JK1 cell voltage 2"
|
||||||
|
cell_voltage_3:
|
||||||
|
name: "JK1 cell voltage 3"
|
||||||
|
cell_voltage_4:
|
||||||
|
name: "JK1 cell voltage 4"
|
||||||
|
cell_voltage_5:
|
||||||
|
name: "JK1 cell voltage 5"
|
||||||
|
cell_voltage_6:
|
||||||
|
name: "JK1 cell voltage 6"
|
||||||
|
cell_voltage_7:
|
||||||
|
name: "JK1 cell voltage 7"
|
||||||
|
cell_voltage_8:
|
||||||
|
name: "JK1 cell voltage 8"
|
||||||
|
cell_resistance_1:
|
||||||
|
name: "JK1 cell resistance 1"
|
||||||
|
cell_resistance_2:
|
||||||
|
name: "JK1 cell resistance 2"
|
||||||
|
cell_resistance_3:
|
||||||
|
name: "JK1 cell resistance 3"
|
||||||
|
cell_resistance_4:
|
||||||
|
name: "JK1 cell resistance 4"
|
||||||
|
cell_resistance_5:
|
||||||
|
name: "JK1 cell resistance 5"
|
||||||
|
cell_resistance_6:
|
||||||
|
name: "JK1 cell resistance 6"
|
||||||
|
cell_resistance_7:
|
||||||
|
name: "JK1 cell resistance 7"
|
||||||
|
cell_resistance_8:
|
||||||
|
name: "JK1 cell resistance 8"
|
||||||
|
total_charging_cycle_capacity:
|
||||||
|
name: "JK1 total charging cycle capacity"
|
||||||
|
|
||||||
|
text_sensor:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms1
|
||||||
|
errors:
|
||||||
|
name: "JK1 Errors"
|
||||||
|
|
||||||
|
switch:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms1
|
||||||
|
charging:
|
||||||
|
name: "JK1 Charging"
|
||||||
|
discharging:
|
||||||
|
name: "JK1 Discharging"
|
||||||
|
balancer:
|
||||||
|
name: "JK1 Balancing"
|
||||||
|
|
||||||
|
- platform: ble_client
|
||||||
|
ble_client_id: jk_ble1
|
||||||
|
name: "JK1 enable bluetooth connection"
|
||||||
|
id: ble_client_switch0
|
||||||
48
esphome/environment.yml
Normal file
48
esphome/environment.yml
Normal file
@@ -0,0 +1,48 @@
|
|||||||
|
esphome:
|
||||||
|
name: "environment"
|
||||||
|
friendly_name: "environment"
|
||||||
|
|
||||||
|
esp32:
|
||||||
|
board: esp32dev
|
||||||
|
framework:
|
||||||
|
type: arduino
|
||||||
|
|
||||||
|
i2c:
|
||||||
|
sda: GPIO21
|
||||||
|
scl: GPIO22
|
||||||
|
scan: True
|
||||||
|
id: bus_a
|
||||||
|
|
||||||
|
sensor:
|
||||||
|
- platform: aht10
|
||||||
|
i2c_id: bus_a
|
||||||
|
address: 0x38
|
||||||
|
variant: AHT20
|
||||||
|
temperature:
|
||||||
|
name: "environment Temperature"
|
||||||
|
id: aht10_temperature
|
||||||
|
humidity:
|
||||||
|
name: "environment Humidity"
|
||||||
|
id: aht10_humidity
|
||||||
|
update_interval: 5s
|
||||||
|
|
||||||
|
web_server:
|
||||||
|
port: 80
|
||||||
|
|
||||||
|
logger:
|
||||||
|
level: DEBUG
|
||||||
|
|
||||||
|
api:
|
||||||
|
encryption:
|
||||||
|
key: !secret api_key
|
||||||
|
|
||||||
|
ota:
|
||||||
|
- platform: esphome
|
||||||
|
password: !secret ota_password
|
||||||
|
|
||||||
|
wifi:
|
||||||
|
ssid: !secret wifi_ssid
|
||||||
|
password: !secret wifi_password
|
||||||
|
fast_connect: on
|
||||||
|
|
||||||
|
captive_portal:
|
||||||
1
file_sizes.txt.new
Normal file
1
file_sizes.txt.new
Normal file
File diff suppressed because one or more lines are too long
272
flake.lock
generated
272
flake.lock
generated
@@ -2,18 +2,17 @@
|
|||||||
"nodes": {
|
"nodes": {
|
||||||
"firefox-addons": {
|
"firefox-addons": {
|
||||||
"inputs": {
|
"inputs": {
|
||||||
"flake-utils": "flake-utils",
|
|
||||||
"nixpkgs": [
|
"nixpkgs": [
|
||||||
"nixpkgs"
|
"nixpkgs"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"dir": "pkgs/firefox-addons",
|
"dir": "pkgs/firefox-addons",
|
||||||
"lastModified": 1735099403,
|
"lastModified": 1773979456,
|
||||||
"narHash": "sha256-22NF70bxkMY7/IY2NG0pC7WzgJ0bI67FGSUP37d2mQ8=",
|
"narHash": "sha256-9kBMJ5IvxqNlkkj/swmE8uK1Sc7TL/LIRUI958m7uBM=",
|
||||||
"owner": "rycee",
|
"owner": "rycee",
|
||||||
"repo": "nur-expressions",
|
"repo": "nur-expressions",
|
||||||
"rev": "5b2c380332cf5a3022fde931d0346e2b868b544e",
|
"rev": "81e28f47ac18d9e89513929c77e711e657b64851",
|
||||||
"type": "gitlab"
|
"type": "gitlab"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -23,55 +22,6 @@
|
|||||||
"type": "gitlab"
|
"type": "gitlab"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"flake-compat": {
|
|
||||||
"flake": false,
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1717312683,
|
|
||||||
"narHash": "sha256-FrlieJH50AuvagamEvWMIE6D2OAnERuDboFDYAED/dE=",
|
|
||||||
"owner": "nix-community",
|
|
||||||
"repo": "flake-compat",
|
|
||||||
"rev": "38fd3954cf65ce6faf3d0d45cd26059e059f07ea",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "nix-community",
|
|
||||||
"repo": "flake-compat",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"flake-utils": {
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1629284811,
|
|
||||||
"narHash": "sha256-JHgasjPR0/J1J3DRm4KxM4zTyAj4IOJY8vIl75v/kPI=",
|
|
||||||
"owner": "numtide",
|
|
||||||
"repo": "flake-utils",
|
|
||||||
"rev": "c5d161cc0af116a2e17f54316f0bf43f0819785c",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "numtide",
|
|
||||||
"repo": "flake-utils",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"flake-utils_2": {
|
|
||||||
"inputs": {
|
|
||||||
"systems": "systems"
|
|
||||||
},
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1731533236,
|
|
||||||
"narHash": "sha256-l0KFg5HjrsfsO/JpG+r7fRrqm12kzFHyUHqHCVpMMbI=",
|
|
||||||
"owner": "numtide",
|
|
||||||
"repo": "flake-utils",
|
|
||||||
"rev": "11707dc2f618dd54ca8739b309ec4fc024de578b",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "numtide",
|
|
||||||
"repo": "flake-utils",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"home-manager": {
|
"home-manager": {
|
||||||
"inputs": {
|
"inputs": {
|
||||||
"nixpkgs": [
|
"nixpkgs": [
|
||||||
@@ -79,11 +29,11 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1735053786,
|
"lastModified": 1774007980,
|
||||||
"narHash": "sha256-Gm+0DcbUS338vvkwyYWms5jsWlx8z8MeQBzcnIDuIkw=",
|
"narHash": "sha256-FOnZjElEI8pqqCvB6K/1JRHTE8o4rer8driivTpq2uo=",
|
||||||
"owner": "nix-community",
|
"owner": "nix-community",
|
||||||
"repo": "home-manager",
|
"repo": "home-manager",
|
||||||
"rev": "35b98d20ca8f4ca1f6a2c30b8a2c8bb305a36d84",
|
"rev": "9670de2921812bc4e0452f6e3efd8c859696c183",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -92,58 +42,13 @@
|
|||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"nix-github-actions": {
|
|
||||||
"inputs": {
|
|
||||||
"nixpkgs": [
|
|
||||||
"system_tools",
|
|
||||||
"poetry2nix",
|
|
||||||
"nixpkgs"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1729742964,
|
|
||||||
"narHash": "sha256-B4mzTcQ0FZHdpeWcpDYPERtyjJd/NIuaQ9+BV1h+MpA=",
|
|
||||||
"owner": "nix-community",
|
|
||||||
"repo": "nix-github-actions",
|
|
||||||
"rev": "e04df33f62cdcf93d73e9a04142464753a16db67",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "nix-community",
|
|
||||||
"repo": "nix-github-actions",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"nixos-cosmic": {
|
|
||||||
"inputs": {
|
|
||||||
"flake-compat": "flake-compat",
|
|
||||||
"nixpkgs": [
|
|
||||||
"nixpkgs"
|
|
||||||
],
|
|
||||||
"nixpkgs-stable": "nixpkgs-stable",
|
|
||||||
"rust-overlay": "rust-overlay"
|
|
||||||
},
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1735090583,
|
|
||||||
"narHash": "sha256-Tm+BsKXJS/EdJd9DvLxDbw+chPI1o7A9RHKIFxho36I=",
|
|
||||||
"owner": "lilyinstarlight",
|
|
||||||
"repo": "nixos-cosmic",
|
|
||||||
"rev": "847b93e3b63bcea9a477dd86bb4b56ce7e051f0e",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "lilyinstarlight",
|
|
||||||
"repo": "nixos-cosmic",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"nixos-hardware": {
|
"nixos-hardware": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1734954597,
|
"lastModified": 1774018263,
|
||||||
"narHash": "sha256-QIhd8/0x30gEv8XEE1iAnrdMlKuQ0EzthfDR7Hwl+fk=",
|
"narHash": "sha256-HHYEwK1A22aSaxv2ibhMMkKvrDGKGlA/qObG4smrSqc=",
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixos-hardware",
|
"repo": "nixos-hardware",
|
||||||
"rev": "def1d472c832d77885f174089b0d34854b007198",
|
"rev": "2d4b4717b2534fad5c715968c1cece04a172b365",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -155,11 +60,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs": {
|
"nixpkgs": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1734649271,
|
"lastModified": 1773821835,
|
||||||
"narHash": "sha256-4EVBRhOjMDuGtMaofAIqzJbg4Ql7Ai0PSeuVZTHjyKQ=",
|
"narHash": "sha256-TJ3lSQtW0E2JrznGVm8hOQGVpXjJyXY2guAxku2O9A4=",
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "d70bd19e0a38ad4790d3913bf08fcbfc9eeca507",
|
"rev": "b40629efe5d6ec48dd1efba650c797ddbd39ace0",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -171,11 +76,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs-master": {
|
"nixpkgs-master": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1735170897,
|
"lastModified": 1774051532,
|
||||||
"narHash": "sha256-O1OXa12dHaObowGsfiM8pgphWfSXJj5v8sTxFoPXR0A=",
|
"narHash": "sha256-d3CGMweyYIcPuTj5BKq+1Lx4zwlgL31nVtN647tOZKo=",
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "138326d6e9841ab9e6161730d40e09cedd670c40",
|
"rev": "8620c0b5cc8fbe76502442181be1d0514bc3a1b7",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -187,27 +92,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs-stable": {
|
"nixpkgs-stable": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1734875076,
|
"lastModified": 1735563628,
|
||||||
"narHash": "sha256-Pzyb+YNG5u3zP79zoi8HXYMs15Q5dfjDgwCdUI5B0nY=",
|
"narHash": "sha256-OnSAY7XDSx7CtDoqNh8jwVwh4xNL/2HaJxGjryLWzX8=",
|
||||||
"owner": "NixOS",
|
|
||||||
"repo": "nixpkgs",
|
|
||||||
"rev": "1807c2b91223227ad5599d7067a61665c52d1295",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "NixOS",
|
|
||||||
"ref": "nixos-24.11",
|
|
||||||
"repo": "nixpkgs",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"nixpkgs-stable_2": {
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1734835170,
|
|
||||||
"narHash": "sha256-JG6n9tQET7ZLjvwjH9BdYE6pES/Alcduxv2FDXliuBM=",
|
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "6df924734e1b626efd70925545de844b1c25246f",
|
"rev": "b134951a4c9f3c995fd7be05f3243f8ecd65d798",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -217,120 +106,39 @@
|
|||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"poetry2nix": {
|
|
||||||
"inputs": {
|
|
||||||
"flake-utils": [
|
|
||||||
"system_tools",
|
|
||||||
"flake-utils"
|
|
||||||
],
|
|
||||||
"nix-github-actions": "nix-github-actions",
|
|
||||||
"nixpkgs": [
|
|
||||||
"system_tools",
|
|
||||||
"nixpkgs"
|
|
||||||
],
|
|
||||||
"systems": "systems_2",
|
|
||||||
"treefmt-nix": "treefmt-nix"
|
|
||||||
},
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1731205797,
|
|
||||||
"narHash": "sha256-F7N1mxH1VrkVNHR3JGNMRvp9+98KYO4b832KS8Gl2xI=",
|
|
||||||
"owner": "nix-community",
|
|
||||||
"repo": "poetry2nix",
|
|
||||||
"rev": "f554d27c1544d9c56e5f1f8e2b8aff399803674e",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "nix-community",
|
|
||||||
"repo": "poetry2nix",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"root": {
|
"root": {
|
||||||
"inputs": {
|
"inputs": {
|
||||||
"firefox-addons": "firefox-addons",
|
"firefox-addons": "firefox-addons",
|
||||||
"home-manager": "home-manager",
|
"home-manager": "home-manager",
|
||||||
"nixos-cosmic": "nixos-cosmic",
|
|
||||||
"nixos-hardware": "nixos-hardware",
|
"nixos-hardware": "nixos-hardware",
|
||||||
"nixpkgs": "nixpkgs",
|
"nixpkgs": "nixpkgs",
|
||||||
"nixpkgs-master": "nixpkgs-master",
|
"nixpkgs-master": "nixpkgs-master",
|
||||||
"nixpkgs-stable": "nixpkgs-stable_2",
|
"nixpkgs-stable": "nixpkgs-stable",
|
||||||
"system_tools": "system_tools",
|
"sops-nix": "sops-nix",
|
||||||
"systems": "systems_3"
|
"systems": "systems"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"rust-overlay": {
|
"sops-nix": {
|
||||||
"inputs": {
|
"inputs": {
|
||||||
"nixpkgs": [
|
"nixpkgs": [
|
||||||
"nixos-cosmic",
|
|
||||||
"nixpkgs"
|
"nixpkgs"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1735007320,
|
"lastModified": 1773889674,
|
||||||
"narHash": "sha256-NdhUgB9BkLGW9I+Q1GyUUCc3CbDgsg7HLWjG7WZBR5Q=",
|
"narHash": "sha256-+ycaiVAk3MEshJTg35cBTUa0MizGiS+bgpYw/f8ohkg=",
|
||||||
"owner": "oxalica",
|
"owner": "Mic92",
|
||||||
"repo": "rust-overlay",
|
"repo": "sops-nix",
|
||||||
"rev": "fb5fdba697ee9a2391ca9ceea3b853b4e3ce37a5",
|
"rev": "29b6519f3e0780452bca0ac0be4584f04ac16cc5",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
"owner": "oxalica",
|
"owner": "Mic92",
|
||||||
"repo": "rust-overlay",
|
"repo": "sops-nix",
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"system_tools": {
|
|
||||||
"inputs": {
|
|
||||||
"flake-utils": "flake-utils_2",
|
|
||||||
"nixpkgs": [
|
|
||||||
"nixpkgs"
|
|
||||||
],
|
|
||||||
"poetry2nix": "poetry2nix"
|
|
||||||
},
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1733083480,
|
|
||||||
"narHash": "sha256-B13faNyBbA3MeI7Jp6pFVbp58rI2Rx5Uvd83csW2p48=",
|
|
||||||
"owner": "RichieCahill",
|
|
||||||
"repo": "system_tools",
|
|
||||||
"rev": "bc357d8fabd83c1423611829091e5b1d86dd913c",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "RichieCahill",
|
|
||||||
"repo": "system_tools",
|
|
||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"systems": {
|
"systems": {
|
||||||
"locked": {
|
|
||||||
"lastModified": 1681028828,
|
|
||||||
"narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=",
|
|
||||||
"owner": "nix-systems",
|
|
||||||
"repo": "default",
|
|
||||||
"rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "nix-systems",
|
|
||||||
"repo": "default",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"systems_2": {
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1681028828,
|
|
||||||
"narHash": "sha256-Vy1rq5AaRuLzOxct8nz4T6wlgyUR7zLU309k9mBC768=",
|
|
||||||
"owner": "nix-systems",
|
|
||||||
"repo": "default",
|
|
||||||
"rev": "da67096a3b9bf56a91d16901293e51ba5b49a27e",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"id": "systems",
|
|
||||||
"type": "indirect"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"systems_3": {
|
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1689347949,
|
"lastModified": 1689347949,
|
||||||
"narHash": "sha256-12tWmuL2zgBgZkdoB6qXZsgJEH9LR3oUgpaQq2RbI80=",
|
"narHash": "sha256-12tWmuL2zgBgZkdoB6qXZsgJEH9LR3oUgpaQq2RbI80=",
|
||||||
@@ -344,28 +152,6 @@
|
|||||||
"repo": "default-linux",
|
"repo": "default-linux",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
}
|
}
|
||||||
},
|
|
||||||
"treefmt-nix": {
|
|
||||||
"inputs": {
|
|
||||||
"nixpkgs": [
|
|
||||||
"system_tools",
|
|
||||||
"poetry2nix",
|
|
||||||
"nixpkgs"
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"locked": {
|
|
||||||
"lastModified": 1730120726,
|
|
||||||
"narHash": "sha256-LqHYIxMrl/1p3/kvm2ir925tZ8DkI0KA10djk8wecSk=",
|
|
||||||
"owner": "numtide",
|
|
||||||
"repo": "treefmt-nix",
|
|
||||||
"rev": "9ef337e492a5555d8e17a51c911ff1f02635be15",
|
|
||||||
"type": "github"
|
|
||||||
},
|
|
||||||
"original": {
|
|
||||||
"owner": "numtide",
|
|
||||||
"repo": "treefmt-nix",
|
|
||||||
"type": "github"
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"root": "root",
|
"root": "root",
|
||||||
|
|||||||
107
flake.nix
107
flake.nix
@@ -4,14 +4,12 @@
|
|||||||
nixConfig = {
|
nixConfig = {
|
||||||
extra-substituters = [
|
extra-substituters = [
|
||||||
"https://cache.nixos.org/?priority=2&want-mass-query=true"
|
"https://cache.nixos.org/?priority=2&want-mass-query=true"
|
||||||
"https://cache.tmmworkshop.com/?priority=2&want-mass-query=true"
|
|
||||||
"https://nix-community.cachix.org/?priority=10&want-mass-query=true"
|
"https://nix-community.cachix.org/?priority=10&want-mass-query=true"
|
||||||
];
|
];
|
||||||
extra-trusted-public-keys = [
|
extra-trusted-public-keys = [
|
||||||
"cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY="
|
"cache.nixos.org-1:6NCHdD59X431o0gWypbMrAURkbJ16ZPMQFGspcDShjY=" # cspell:disable-line
|
||||||
"cache.tmmworkshop.com:jHffkpgbmEdstQPoihJPYW9TQe6jnQbWR2LqkNGV3iA="
|
"nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs=" # cspell:disable-line
|
||||||
"nix-community.cachix.org-1:mB9FSh9qf2dCimDSUo8Zy7bkq5CX+/rkCWyvRCYg3Fs="
|
"cache-nix-dot:Od9KN34LXc6Lu7y1ozzV1kIXZa8coClozgth/SYE7dU=" # cspell:disable-line
|
||||||
"cache-nix-dot:Od9KN34LXc6Lu7y1ozzV1kIXZa8coClozgth/SYE7dU="
|
|
||||||
];
|
];
|
||||||
};
|
};
|
||||||
|
|
||||||
@@ -33,59 +31,72 @@
|
|||||||
inputs.nixpkgs.follows = "nixpkgs";
|
inputs.nixpkgs.follows = "nixpkgs";
|
||||||
};
|
};
|
||||||
|
|
||||||
system_tools = {
|
sops-nix = {
|
||||||
url = "github:RichieCahill/system_tools";
|
url = "github:Mic92/sops-nix";
|
||||||
inputs.nixpkgs.follows = "nixpkgs";
|
|
||||||
};
|
|
||||||
|
|
||||||
nixos-cosmic = {
|
|
||||||
url = "github:lilyinstarlight/nixos-cosmic";
|
|
||||||
inputs.nixpkgs.follows = "nixpkgs";
|
inputs.nixpkgs.follows = "nixpkgs";
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
outputs = {
|
outputs =
|
||||||
self,
|
{
|
||||||
nixpkgs,
|
self,
|
||||||
home-manager,
|
nixpkgs,
|
||||||
systems,
|
home-manager,
|
||||||
nixos-cosmic,
|
systems,
|
||||||
...
|
sops-nix,
|
||||||
} @ inputs: let
|
...
|
||||||
inherit (self) outputs;
|
}@inputs:
|
||||||
lib = nixpkgs.lib // home-manager.lib;
|
let
|
||||||
forEachSystem = f: lib.genAttrs (import systems) (system: f pkgsFor.${system});
|
inherit (self) outputs;
|
||||||
pkgsFor = lib.genAttrs (import systems) (
|
lib = nixpkgs.lib // home-manager.lib;
|
||||||
system:
|
forEachSystem = f: lib.genAttrs (import systems) (system: f pkgsFor.${system});
|
||||||
|
pkgsFor = lib.genAttrs (import systems) (
|
||||||
|
system:
|
||||||
import nixpkgs {
|
import nixpkgs {
|
||||||
inherit system;
|
inherit system;
|
||||||
|
overlays = builtins.attrValues outputs.overlays;
|
||||||
config.allowUnfree = true;
|
config.allowUnfree = true;
|
||||||
}
|
}
|
||||||
);
|
);
|
||||||
in {
|
in
|
||||||
inherit lib;
|
{
|
||||||
overlays = import ./overlays {inherit inputs outputs;};
|
inherit lib;
|
||||||
|
overlays = import ./overlays { inherit inputs outputs; };
|
||||||
|
|
||||||
devShells = forEachSystem (pkgs: import ./shell.nix {inherit pkgs;});
|
devShells = forEachSystem (pkgs: import ./shell.nix { inherit pkgs; });
|
||||||
formatter = forEachSystem (pkgs: pkgs.alejandra);
|
formatter = forEachSystem (pkgs: pkgs.treefmt);
|
||||||
|
|
||||||
nixosConfigurations = {
|
nixosConfigurations = {
|
||||||
bob = lib.nixosSystem {
|
bob = lib.nixosSystem {
|
||||||
modules = [./systems/bob];
|
modules = [
|
||||||
specialArgs = {inherit inputs outputs;};
|
./systems/bob
|
||||||
};
|
];
|
||||||
jeeves = lib.nixosSystem {
|
specialArgs = { inherit inputs outputs; };
|
||||||
modules = [./systems/jeeves];
|
};
|
||||||
specialArgs = {inherit inputs outputs;};
|
brain = lib.nixosSystem {
|
||||||
};
|
modules = [
|
||||||
rhapsody-in-green = lib.nixosSystem {
|
./systems/brain
|
||||||
modules = [./systems/rhapsody-in-green];
|
];
|
||||||
specialArgs = {inherit inputs outputs;};
|
specialArgs = { inherit inputs outputs; };
|
||||||
};
|
};
|
||||||
muninn = lib.nixosSystem {
|
jeeves = lib.nixosSystem {
|
||||||
modules = [./systems/muninn];
|
modules = [
|
||||||
specialArgs = {inherit inputs outputs;};
|
./systems/jeeves
|
||||||
|
];
|
||||||
|
specialArgs = { inherit inputs outputs; };
|
||||||
|
};
|
||||||
|
rhapsody-in-green = lib.nixosSystem {
|
||||||
|
modules = [
|
||||||
|
./systems/rhapsody-in-green
|
||||||
|
];
|
||||||
|
specialArgs = { inherit inputs outputs; };
|
||||||
|
};
|
||||||
|
leviathan = lib.nixosSystem {
|
||||||
|
modules = [
|
||||||
|
./systems/leviathan
|
||||||
|
];
|
||||||
|
specialArgs = { inherit inputs outputs; };
|
||||||
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
};
|
|
||||||
}
|
}
|
||||||
|
|||||||
24
frontend/.gitignore
vendored
Normal file
24
frontend/.gitignore
vendored
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
# Logs
|
||||||
|
logs
|
||||||
|
*.log
|
||||||
|
npm-debug.log*
|
||||||
|
yarn-debug.log*
|
||||||
|
yarn-error.log*
|
||||||
|
pnpm-debug.log*
|
||||||
|
lerna-debug.log*
|
||||||
|
|
||||||
|
node_modules
|
||||||
|
dist
|
||||||
|
dist-ssr
|
||||||
|
*.local
|
||||||
|
|
||||||
|
# Editor directories and files
|
||||||
|
.vscode/*
|
||||||
|
!.vscode/extensions.json
|
||||||
|
.idea
|
||||||
|
.DS_Store
|
||||||
|
*.suo
|
||||||
|
*.ntvs*
|
||||||
|
*.njsproj
|
||||||
|
*.sln
|
||||||
|
*.sw?
|
||||||
@@ -1,16 +1,49 @@
|
|||||||
{inputs, ...}: {
|
{ inputs, ... }:
|
||||||
|
{
|
||||||
# When applied, the stable nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.stable'
|
# When applied, the stable nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.stable'
|
||||||
stable = final: _prev: {
|
stable = final: _prev: {
|
||||||
stable = import inputs.nixpkgs-stable {
|
stable = import inputs.nixpkgs-stable {
|
||||||
system = final.system;
|
system = final.stdenv.hostPlatform.system;
|
||||||
config.allowUnfree = true;
|
config.allowUnfree = true;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
# When applied, the master nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.master'
|
# When applied, the master nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.master'
|
||||||
master = final: _prev: {
|
master = final: _prev: {
|
||||||
master = import inputs.nixpkgs-master {
|
master = import inputs.nixpkgs-master {
|
||||||
system = final.system;
|
system = final.stdenv.hostPlatform.system;
|
||||||
config.allowUnfree = true;
|
config.allowUnfree = true;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
|
python-env = final: _prev: {
|
||||||
|
my_python = final.python314.withPackages (
|
||||||
|
ps: with ps; [
|
||||||
|
alembic
|
||||||
|
apprise
|
||||||
|
apscheduler
|
||||||
|
fastapi
|
||||||
|
fastapi-cli
|
||||||
|
httpx
|
||||||
|
mypy
|
||||||
|
polars
|
||||||
|
psycopg
|
||||||
|
pydantic
|
||||||
|
pyfakefs
|
||||||
|
pytest
|
||||||
|
pytest-cov
|
||||||
|
pytest-mock
|
||||||
|
pytest-xdist
|
||||||
|
python-multipart
|
||||||
|
ruff
|
||||||
|
scalene
|
||||||
|
sqlalchemy
|
||||||
|
sqlalchemy
|
||||||
|
tenacity
|
||||||
|
textual
|
||||||
|
tinytuya
|
||||||
|
typer
|
||||||
|
websockets
|
||||||
|
]
|
||||||
|
);
|
||||||
|
};
|
||||||
}
|
}
|
||||||
|
|||||||
114
pyproject.toml
Normal file
114
pyproject.toml
Normal file
@@ -0,0 +1,114 @@
|
|||||||
|
[project]
|
||||||
|
name = "system_tools"
|
||||||
|
version = "0.1.0"
|
||||||
|
description = ""
|
||||||
|
authors = [{ name = "Richie Cahill", email = "richie@tmmworkshop.com" }]
|
||||||
|
requires-python = "~=3.13.0"
|
||||||
|
readme = "README.md"
|
||||||
|
license = "MIT"
|
||||||
|
# these dependencies are a best effort and aren't guaranteed to work
|
||||||
|
# for up-to-date dependencies, see overlays/default.nix
|
||||||
|
dependencies = [
|
||||||
|
"alembic",
|
||||||
|
"apprise",
|
||||||
|
"apscheduler",
|
||||||
|
"httpx",
|
||||||
|
"python-multipart",
|
||||||
|
"polars",
|
||||||
|
"psycopg[binary]",
|
||||||
|
"pydantic",
|
||||||
|
"pyyaml",
|
||||||
|
"sqlalchemy",
|
||||||
|
"typer",
|
||||||
|
"websockets",
|
||||||
|
]
|
||||||
|
|
||||||
|
[project.scripts]
|
||||||
|
database = "python.database_cli:app"
|
||||||
|
van-inventory = "python.van_inventory.main:serve"
|
||||||
|
|
||||||
|
[dependency-groups]
|
||||||
|
dev = [
|
||||||
|
"mypy",
|
||||||
|
"pyfakefs",
|
||||||
|
"pytest-cov",
|
||||||
|
"pytest-mock",
|
||||||
|
"pytest-xdist",
|
||||||
|
"pytest",
|
||||||
|
"ruff",
|
||||||
|
]
|
||||||
|
|
||||||
|
[tool.ruff]
|
||||||
|
|
||||||
|
target-version = "py313"
|
||||||
|
|
||||||
|
line-length = 120
|
||||||
|
|
||||||
|
lint.select = ["ALL"]
|
||||||
|
lint.ignore = [
|
||||||
|
"G004", # (PERM) This is a performers nit
|
||||||
|
"COM812", # (TEMP) conflicts when used with the formatter
|
||||||
|
"ISC001", # (TEMP) conflicts when used with the formatter
|
||||||
|
"S603", # (PERM) This is known to cause a false positive
|
||||||
|
]
|
||||||
|
|
||||||
|
[tool.ruff.lint.per-file-ignores]
|
||||||
|
|
||||||
|
"tests/**" = [
|
||||||
|
"ANN", # (perm) type annotations not needed in tests
|
||||||
|
"D", # (perm) docstrings not needed in tests
|
||||||
|
"PLR2004", # (perm) magic values are fine in test assertions
|
||||||
|
"S101", # (perm) pytest needs asserts
|
||||||
|
]
|
||||||
|
"python/stuff/**" = [
|
||||||
|
"T201", # (perm) I don't care about print statements dir
|
||||||
|
]
|
||||||
|
"python/testing/**" = [
|
||||||
|
"T201", # (perm) I don't care about print statements dir
|
||||||
|
"ERA001", # (perm) I don't care about print statements dir
|
||||||
|
]
|
||||||
|
"python/splendor/**" = [
|
||||||
|
"S311", # (perm) there is no security issue here
|
||||||
|
"T201", # (perm) I don't care about print statements dir
|
||||||
|
"PLR2004", # (temps) need to think about this
|
||||||
|
]
|
||||||
|
"python/orm/**" = [
|
||||||
|
"TC003", # (perm) this creates issues because sqlalchemy uses these at runtime
|
||||||
|
]
|
||||||
|
"python/congress_tracker/**" = [
|
||||||
|
"TC003", # (perm) this creates issues because sqlalchemy uses these at runtime
|
||||||
|
]
|
||||||
|
"python/eval_warnings/**" = [
|
||||||
|
"S607", # (perm) gh and git are expected on PATH in the runner environment
|
||||||
|
]
|
||||||
|
"python/alembic/**" = [
|
||||||
|
"INP001", # (perm) this creates LSP issues for alembic
|
||||||
|
]
|
||||||
|
"python/signal_bot/**" = [
|
||||||
|
"D107", # (perm) class docstrings cover __init__
|
||||||
|
]
|
||||||
|
|
||||||
|
[tool.ruff.lint.pydocstyle]
|
||||||
|
convention = "google"
|
||||||
|
|
||||||
|
[tool.ruff.lint.flake8-builtins]
|
||||||
|
builtins-ignorelist = ["id"]
|
||||||
|
|
||||||
|
[tool.ruff.lint.pylint]
|
||||||
|
max-args = 9
|
||||||
|
|
||||||
|
[tool.coverage.run]
|
||||||
|
source = ["system_tools"]
|
||||||
|
|
||||||
|
[tool.coverage.report]
|
||||||
|
exclude_lines = [
|
||||||
|
"pragma: no cover",
|
||||||
|
"if TYPE_CHECKING:",
|
||||||
|
"raise NotImplementedError",
|
||||||
|
"if __name__ == \"__main__\":",
|
||||||
|
]
|
||||||
|
|
||||||
|
[tool.pytest.ini_options]
|
||||||
|
addopts = "-n auto -ra"
|
||||||
|
testpaths = ["tests"]
|
||||||
|
# --cov=system_tools --cov-report=term-missing --cov-report=xml --cov-report=html --cov-branch
|
||||||
1
python/__init__.py
Normal file
1
python/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Server Tools."""
|
||||||
121
python/alembic/env.py
Normal file
121
python/alembic/env.py
Normal file
@@ -0,0 +1,121 @@
|
|||||||
|
"""Alembic."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import logging
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
from typing import TYPE_CHECKING, Any, Literal
|
||||||
|
|
||||||
|
from alembic import context
|
||||||
|
from alembic.script import write_hooks
|
||||||
|
from sqlalchemy.schema import CreateSchema
|
||||||
|
|
||||||
|
from python.common import bash_wrapper
|
||||||
|
from python.orm.common import get_postgres_engine
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import MutableMapping
|
||||||
|
|
||||||
|
from sqlalchemy.orm import DeclarativeBase
|
||||||
|
|
||||||
|
config = context.config
|
||||||
|
|
||||||
|
base_class: type[DeclarativeBase] = config.attributes.get("base")
|
||||||
|
if base_class is None:
|
||||||
|
error = "No base class provided. Use the database CLI to run alembic commands."
|
||||||
|
raise RuntimeError(error)
|
||||||
|
|
||||||
|
target_metadata = base_class.metadata
|
||||||
|
logging.basicConfig(
|
||||||
|
level="DEBUG",
|
||||||
|
datefmt="%Y-%m-%dT%H:%M:%S%z",
|
||||||
|
format="%(asctime)s %(levelname)s %(filename)s:%(lineno)d - %(message)s",
|
||||||
|
handlers=[logging.StreamHandler(sys.stdout)],
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@write_hooks.register("dynamic_schema")
|
||||||
|
def dynamic_schema(filename: str, _options: dict[Any, Any]) -> None:
|
||||||
|
"""Dynamic schema."""
|
||||||
|
original_file = Path(filename).read_text()
|
||||||
|
schema_name = base_class.schema_name
|
||||||
|
dynamic_schema_file_part1 = original_file.replace(f"schema='{schema_name}'", "schema=schema")
|
||||||
|
dynamic_schema_file = dynamic_schema_file_part1.replace(f"'{schema_name}.", "f'{schema}.")
|
||||||
|
Path(filename).write_text(dynamic_schema_file)
|
||||||
|
|
||||||
|
|
||||||
|
@write_hooks.register("import_postgresql")
|
||||||
|
def import_postgresql(filename: str, _options: dict[Any, Any]) -> None:
|
||||||
|
"""Add postgresql dialect import when postgresql types are used."""
|
||||||
|
content = Path(filename).read_text()
|
||||||
|
if "postgresql." in content and "from sqlalchemy.dialects import postgresql" not in content:
|
||||||
|
content = content.replace(
|
||||||
|
"import sqlalchemy as sa\n",
|
||||||
|
"import sqlalchemy as sa\nfrom sqlalchemy.dialects import postgresql\n",
|
||||||
|
)
|
||||||
|
Path(filename).write_text(content)
|
||||||
|
|
||||||
|
|
||||||
|
@write_hooks.register("ruff")
|
||||||
|
def ruff_check_and_format(filename: str, _options: dict[Any, Any]) -> None:
|
||||||
|
"""Docstring for ruff_check_and_format."""
|
||||||
|
bash_wrapper(f"ruff check --fix {filename}")
|
||||||
|
bash_wrapper(f"ruff format {filename}")
|
||||||
|
|
||||||
|
|
||||||
|
def include_name(
|
||||||
|
name: str | None,
|
||||||
|
type_: Literal["schema", "table", "column", "index", "unique_constraint", "foreign_key_constraint"],
|
||||||
|
_parent_names: MutableMapping[Literal["schema_name", "table_name", "schema_qualified_table_name"], str | None],
|
||||||
|
) -> bool:
|
||||||
|
"""Filter tables to be included in the migration.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
name (str): The name of the table.
|
||||||
|
type_ (str): The type of the table.
|
||||||
|
_parent_names (MutableMapping): The names of the parent tables.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
bool: True if the table should be included, False otherwise.
|
||||||
|
|
||||||
|
"""
|
||||||
|
if type_ == "schema":
|
||||||
|
return name == target_metadata.schema
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
def run_migrations_online() -> None:
|
||||||
|
"""Run migrations in 'online' mode.
|
||||||
|
|
||||||
|
In this scenario we need to create an Engine
|
||||||
|
and associate a connection with the context.
|
||||||
|
|
||||||
|
"""
|
||||||
|
env_prefix = config.attributes.get("env_prefix", "POSTGRES")
|
||||||
|
connectable = get_postgres_engine(name=env_prefix)
|
||||||
|
|
||||||
|
with connectable.connect() as connection:
|
||||||
|
schema = base_class.schema_name
|
||||||
|
if not connectable.dialect.has_schema(connection, schema):
|
||||||
|
answer = input(f"Schema {schema!r} does not exist. Create it? [y/N] ")
|
||||||
|
if answer.lower() != "y":
|
||||||
|
error = f"Schema {schema!r} does not exist. Exiting."
|
||||||
|
raise SystemExit(error)
|
||||||
|
connection.execute(CreateSchema(schema))
|
||||||
|
connection.commit()
|
||||||
|
|
||||||
|
context.configure(
|
||||||
|
connection=connection,
|
||||||
|
target_metadata=target_metadata,
|
||||||
|
include_schemas=True,
|
||||||
|
version_table_schema=schema,
|
||||||
|
include_name=include_name,
|
||||||
|
)
|
||||||
|
|
||||||
|
with context.begin_transaction():
|
||||||
|
context.run_migrations()
|
||||||
|
connection.commit()
|
||||||
|
|
||||||
|
|
||||||
|
run_migrations_online()
|
||||||
@@ -0,0 +1,113 @@
|
|||||||
|
"""created contact api.
|
||||||
|
|
||||||
|
Revision ID: edd7dd61a3d2
|
||||||
|
Revises:
|
||||||
|
Create Date: 2026-01-11 15:45:59.909266
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "edd7dd61a3d2"
|
||||||
|
down_revision: str | None = None
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"contact",
|
||||||
|
sa.Column("name", sa.String(), nullable=False),
|
||||||
|
sa.Column("age", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("bio", sa.String(), nullable=True),
|
||||||
|
sa.Column("current_job", sa.String(), nullable=True),
|
||||||
|
sa.Column("gender", sa.String(), nullable=True),
|
||||||
|
sa.Column("goals", sa.String(), nullable=True),
|
||||||
|
sa.Column("legal_name", sa.String(), nullable=True),
|
||||||
|
sa.Column("profile_pic", sa.String(), nullable=True),
|
||||||
|
sa.Column("safe_conversation_starters", sa.String(), nullable=True),
|
||||||
|
sa.Column("self_sufficiency_score", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("social_structure_style", sa.String(), nullable=True),
|
||||||
|
sa.Column("ssn", sa.String(), nullable=True),
|
||||||
|
sa.Column("suffix", sa.String(), nullable=True),
|
||||||
|
sa.Column("timezone", sa.String(), nullable=True),
|
||||||
|
sa.Column("topics_to_avoid", sa.String(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_contact")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"need",
|
||||||
|
sa.Column("name", sa.String(), nullable=False),
|
||||||
|
sa.Column("description", sa.String(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_need")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"contact_need",
|
||||||
|
sa.Column("contact_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("need_id", sa.Integer(), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["contact_id"],
|
||||||
|
[f"{schema}.contact.id"],
|
||||||
|
name=op.f("fk_contact_need_contact_id_contact"),
|
||||||
|
ondelete="CASCADE",
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["need_id"], [f"{schema}.need.id"], name=op.f("fk_contact_need_need_id_need"), ondelete="CASCADE"
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("contact_id", "need_id", name=op.f("pk_contact_need")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"contact_relationship",
|
||||||
|
sa.Column("contact_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("related_contact_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("relationship_type", sa.String(length=100), nullable=False),
|
||||||
|
sa.Column("closeness_weight", sa.Integer(), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["contact_id"],
|
||||||
|
[f"{schema}.contact.id"],
|
||||||
|
name=op.f("fk_contact_relationship_contact_id_contact"),
|
||||||
|
ondelete="CASCADE",
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["related_contact_id"],
|
||||||
|
[f"{schema}.contact.id"],
|
||||||
|
name=op.f("fk_contact_relationship_related_contact_id_contact"),
|
||||||
|
ondelete="CASCADE",
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("contact_id", "related_contact_id", name=op.f("pk_contact_relationship")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("contact_relationship", schema=schema)
|
||||||
|
op.drop_table("contact_need", schema=schema)
|
||||||
|
op.drop_table("need", schema=schema)
|
||||||
|
op.drop_table("contact", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,135 @@
|
|||||||
|
"""add congress tracker tables.
|
||||||
|
|
||||||
|
Revision ID: 3f71565e38de
|
||||||
|
Revises: edd7dd61a3d2
|
||||||
|
Create Date: 2026-02-12 16:36:09.457303
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "3f71565e38de"
|
||||||
|
down_revision: str | None = "edd7dd61a3d2"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"bill",
|
||||||
|
sa.Column("congress", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("bill_type", sa.String(), nullable=False),
|
||||||
|
sa.Column("number", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("title", sa.String(), nullable=True),
|
||||||
|
sa.Column("title_short", sa.String(), nullable=True),
|
||||||
|
sa.Column("official_title", sa.String(), nullable=True),
|
||||||
|
sa.Column("status", sa.String(), nullable=True),
|
||||||
|
sa.Column("status_at", sa.Date(), nullable=True),
|
||||||
|
sa.Column("sponsor_bioguide_id", sa.String(), nullable=True),
|
||||||
|
sa.Column("subjects_top_term", sa.String(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_bill")),
|
||||||
|
sa.UniqueConstraint("congress", "bill_type", "number", name="uq_bill_congress_type_number"),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_index("ix_bill_congress", "bill", ["congress"], unique=False, schema=schema)
|
||||||
|
op.create_table(
|
||||||
|
"legislator",
|
||||||
|
sa.Column("bioguide_id", sa.Text(), nullable=False),
|
||||||
|
sa.Column("thomas_id", sa.String(), nullable=True),
|
||||||
|
sa.Column("lis_id", sa.String(), nullable=True),
|
||||||
|
sa.Column("govtrack_id", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("opensecrets_id", sa.String(), nullable=True),
|
||||||
|
sa.Column("fec_ids", sa.String(), nullable=True),
|
||||||
|
sa.Column("first_name", sa.String(), nullable=False),
|
||||||
|
sa.Column("last_name", sa.String(), nullable=False),
|
||||||
|
sa.Column("official_full_name", sa.String(), nullable=True),
|
||||||
|
sa.Column("nickname", sa.String(), nullable=True),
|
||||||
|
sa.Column("birthday", sa.Date(), nullable=True),
|
||||||
|
sa.Column("gender", sa.String(), nullable=True),
|
||||||
|
sa.Column("current_party", sa.String(), nullable=True),
|
||||||
|
sa.Column("current_state", sa.String(), nullable=True),
|
||||||
|
sa.Column("current_district", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("current_chamber", sa.String(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_legislator")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_index(op.f("ix_legislator_bioguide_id"), "legislator", ["bioguide_id"], unique=True, schema=schema)
|
||||||
|
op.create_table(
|
||||||
|
"vote",
|
||||||
|
sa.Column("congress", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("chamber", sa.String(), nullable=False),
|
||||||
|
sa.Column("session", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("number", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("vote_type", sa.String(), nullable=True),
|
||||||
|
sa.Column("question", sa.String(), nullable=True),
|
||||||
|
sa.Column("result", sa.String(), nullable=True),
|
||||||
|
sa.Column("result_text", sa.String(), nullable=True),
|
||||||
|
sa.Column("vote_date", sa.Date(), nullable=False),
|
||||||
|
sa.Column("yea_count", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("nay_count", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("not_voting_count", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("present_count", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("bill_id", sa.Integer(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(["bill_id"], [f"{schema}.bill.id"], name=op.f("fk_vote_bill_id_bill")),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_vote")),
|
||||||
|
sa.UniqueConstraint("congress", "chamber", "session", "number", name="uq_vote_congress_chamber_session_number"),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_index("ix_vote_congress_chamber", "vote", ["congress", "chamber"], unique=False, schema=schema)
|
||||||
|
op.create_index("ix_vote_date", "vote", ["vote_date"], unique=False, schema=schema)
|
||||||
|
op.create_table(
|
||||||
|
"vote_record",
|
||||||
|
sa.Column("vote_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("legislator_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("position", sa.String(), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["legislator_id"],
|
||||||
|
[f"{schema}.legislator.id"],
|
||||||
|
name=op.f("fk_vote_record_legislator_id_legislator"),
|
||||||
|
ondelete="CASCADE",
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["vote_id"], [f"{schema}.vote.id"], name=op.f("fk_vote_record_vote_id_vote"), ondelete="CASCADE"
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("vote_id", "legislator_id", name=op.f("pk_vote_record")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("vote_record", schema=schema)
|
||||||
|
op.drop_index("ix_vote_date", table_name="vote", schema=schema)
|
||||||
|
op.drop_index("ix_vote_congress_chamber", table_name="vote", schema=schema)
|
||||||
|
op.drop_table("vote", schema=schema)
|
||||||
|
op.drop_index(op.f("ix_legislator_bioguide_id"), table_name="legislator", schema=schema)
|
||||||
|
op.drop_table("legislator", schema=schema)
|
||||||
|
op.drop_index("ix_bill_congress", table_name="bill", schema=schema)
|
||||||
|
op.drop_table("bill", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,58 @@
|
|||||||
|
"""adding SignalDevice for DeviceRegistry for signal bot.
|
||||||
|
|
||||||
|
Revision ID: 4c410c16e39c
|
||||||
|
Revises: 3f71565e38de
|
||||||
|
Create Date: 2026-03-09 14:51:24.228976
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
from sqlalchemy.dialects import postgresql
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "4c410c16e39c"
|
||||||
|
down_revision: str | None = "3f71565e38de"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"signal_device",
|
||||||
|
sa.Column("phone_number", sa.String(length=50), nullable=False),
|
||||||
|
sa.Column("safety_number", sa.String(), nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"trust_level",
|
||||||
|
postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column("last_seen", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_signal_device")),
|
||||||
|
sa.UniqueConstraint("phone_number", name=op.f("uq_signal_device_phone_number")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("signal_device", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,41 @@
|
|||||||
|
"""fixed safety number logic.
|
||||||
|
|
||||||
|
Revision ID: 99fec682516c
|
||||||
|
Revises: 4c410c16e39c
|
||||||
|
Create Date: 2026-03-09 16:25:25.085806
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "99fec682516c"
|
||||||
|
down_revision: str | None = "4c410c16e39c"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.alter_column("signal_device", "safety_number", existing_type=sa.VARCHAR(), nullable=True, schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.alter_column("signal_device", "safety_number", existing_type=sa.VARCHAR(), nullable=False, schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,54 @@
|
|||||||
|
"""add dead_letter_message table.
|
||||||
|
|
||||||
|
Revision ID: a1b2c3d4e5f6
|
||||||
|
Revises: 99fec682516c
|
||||||
|
Create Date: 2026-03-10 12:00:00.000000
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
from sqlalchemy.dialects import postgresql
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "a1b2c3d4e5f6"
|
||||||
|
down_revision: str | None = "99fec682516c"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
op.create_table(
|
||||||
|
"dead_letter_message",
|
||||||
|
sa.Column("source", sa.String(), nullable=False),
|
||||||
|
sa.Column("message", sa.Text(), nullable=False),
|
||||||
|
sa.Column("received_at", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"status",
|
||||||
|
postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_dead_letter_message")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
op.drop_table("dead_letter_message", schema=schema)
|
||||||
|
op.execute(sa.text(f"DROP TYPE IF EXISTS {schema}.message_status"))
|
||||||
@@ -0,0 +1,66 @@
|
|||||||
|
"""adding roles to signal devices.
|
||||||
|
|
||||||
|
Revision ID: 2ef7ba690159
|
||||||
|
Revises: a1b2c3d4e5f6
|
||||||
|
Create Date: 2026-03-16 19:22:38.020350
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "2ef7ba690159"
|
||||||
|
down_revision: str | None = "a1b2c3d4e5f6"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"role",
|
||||||
|
sa.Column("name", sa.String(length=50), nullable=False),
|
||||||
|
sa.Column("id", sa.SmallInteger(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_role")),
|
||||||
|
sa.UniqueConstraint("name", name=op.f("uq_role_name")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"device_role",
|
||||||
|
sa.Column("device_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("role_id", sa.SmallInteger(), nullable=False),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["device_id"], [f"{schema}.signal_device.id"], name=op.f("fk_device_role_device_id_signal_device")
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(["role_id"], [f"{schema}.role.id"], name=op.f("fk_device_role_role_id_role")),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_device_role")),
|
||||||
|
sa.UniqueConstraint("device_id", "role_id", name="uq_device_role_device_role"),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("device_role", schema=schema)
|
||||||
|
op.drop_table("role", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,171 @@
|
|||||||
|
"""seprating signal_bot database.
|
||||||
|
|
||||||
|
Revision ID: 6b275323f435
|
||||||
|
Revises: 2ef7ba690159
|
||||||
|
Create Date: 2026-03-18 08:34:28.785885
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
from sqlalchemy.dialects import postgresql
|
||||||
|
|
||||||
|
from python.orm import RichieBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "6b275323f435"
|
||||||
|
down_revision: str | None = "2ef7ba690159"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = RichieBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("device_role", schema=schema)
|
||||||
|
op.drop_table("signal_device", schema=schema)
|
||||||
|
op.drop_table("role", schema=schema)
|
||||||
|
op.drop_table("dead_letter_message", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"dead_letter_message",
|
||||||
|
sa.Column("source", sa.VARCHAR(), autoincrement=False, nullable=False),
|
||||||
|
sa.Column("message", sa.TEXT(), autoincrement=False, nullable=False),
|
||||||
|
sa.Column("received_at", postgresql.TIMESTAMP(timezone=True), autoincrement=False, nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"status",
|
||||||
|
postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"created",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column(
|
||||||
|
"updated",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_dead_letter_message")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"role",
|
||||||
|
sa.Column("name", sa.VARCHAR(length=50), autoincrement=False, nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"id",
|
||||||
|
sa.SMALLINT(),
|
||||||
|
server_default=sa.text(f"nextval('{schema}.role_id_seq'::regclass)"),
|
||||||
|
autoincrement=True,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column(
|
||||||
|
"created",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column(
|
||||||
|
"updated",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_role")),
|
||||||
|
sa.UniqueConstraint(
|
||||||
|
"name", name=op.f("uq_role_name"), postgresql_include=[], postgresql_nulls_not_distinct=False
|
||||||
|
),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"signal_device",
|
||||||
|
sa.Column("phone_number", sa.VARCHAR(length=50), autoincrement=False, nullable=False),
|
||||||
|
sa.Column("safety_number", sa.VARCHAR(), autoincrement=False, nullable=True),
|
||||||
|
sa.Column(
|
||||||
|
"trust_level",
|
||||||
|
postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column("last_seen", postgresql.TIMESTAMP(timezone=True), autoincrement=False, nullable=False),
|
||||||
|
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"created",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column(
|
||||||
|
"updated",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_signal_device")),
|
||||||
|
sa.UniqueConstraint(
|
||||||
|
"phone_number",
|
||||||
|
name=op.f("uq_signal_device_phone_number"),
|
||||||
|
postgresql_include=[],
|
||||||
|
postgresql_nulls_not_distinct=False,
|
||||||
|
),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"device_role",
|
||||||
|
sa.Column("device_id", sa.INTEGER(), autoincrement=False, nullable=False),
|
||||||
|
sa.Column("role_id", sa.SMALLINT(), autoincrement=False, nullable=False),
|
||||||
|
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"created",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column(
|
||||||
|
"updated",
|
||||||
|
postgresql.TIMESTAMP(timezone=True),
|
||||||
|
server_default=sa.text("now()"),
|
||||||
|
autoincrement=False,
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["device_id"], [f"{schema}.signal_device.id"], name=op.f("fk_device_role_device_id_signal_device")
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(["role_id"], [f"{schema}.role.id"], name=op.f("fk_device_role_role_id_role")),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_device_role")),
|
||||||
|
sa.UniqueConstraint(
|
||||||
|
"device_id",
|
||||||
|
"role_id",
|
||||||
|
name=op.f("uq_device_role_device_role"),
|
||||||
|
postgresql_include=[],
|
||||||
|
postgresql_nulls_not_distinct=False,
|
||||||
|
),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
36
python/alembic/script.py.mako
Normal file
36
python/alembic/script.py.mako
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
"""${message}.
|
||||||
|
|
||||||
|
Revision ID: ${up_revision}
|
||||||
|
Revises: ${down_revision | comma,n}
|
||||||
|
Create Date: ${create_date}
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
from python.orm import ${config.attributes["base"].__name__}
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = ${repr(up_revision)}
|
||||||
|
down_revision: str | None = ${repr(down_revision)}
|
||||||
|
branch_labels: str | Sequence[str] | None = ${repr(branch_labels)}
|
||||||
|
depends_on: str | Sequence[str] | None = ${repr(depends_on)}
|
||||||
|
|
||||||
|
schema=${config.attributes["base"].__name__}.schema_name
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
${upgrades if upgrades else "pass"}
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
${downgrades if downgrades else "pass"}
|
||||||
@@ -0,0 +1,100 @@
|
|||||||
|
"""seprating signal_bot database.
|
||||||
|
|
||||||
|
Revision ID: 6eaf696e07a5
|
||||||
|
Revises:
|
||||||
|
Create Date: 2026-03-17 21:35:37.612672
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
from sqlalchemy.dialects import postgresql
|
||||||
|
|
||||||
|
from python.orm import SignalBotBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "6eaf696e07a5"
|
||||||
|
down_revision: str | None = None
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = SignalBotBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"dead_letter_message",
|
||||||
|
sa.Column("source", sa.String(), nullable=False),
|
||||||
|
sa.Column("message", sa.Text(), nullable=False),
|
||||||
|
sa.Column("received_at", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"status", postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema), nullable=False
|
||||||
|
),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_dead_letter_message")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"role",
|
||||||
|
sa.Column("name", sa.String(length=50), nullable=False),
|
||||||
|
sa.Column("id", sa.SmallInteger(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_role")),
|
||||||
|
sa.UniqueConstraint("name", name=op.f("uq_role_name")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"signal_device",
|
||||||
|
sa.Column("phone_number", sa.String(length=50), nullable=False),
|
||||||
|
sa.Column("safety_number", sa.String(), nullable=True),
|
||||||
|
sa.Column(
|
||||||
|
"trust_level",
|
||||||
|
postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
|
sa.Column("last_seen", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_signal_device")),
|
||||||
|
sa.UniqueConstraint("phone_number", name=op.f("uq_signal_device_phone_number")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"device_role",
|
||||||
|
sa.Column("device_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("role_id", sa.SmallInteger(), nullable=False),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["device_id"], [f"{schema}.signal_device.id"], name=op.f("fk_device_role_device_id_signal_device")
|
||||||
|
),
|
||||||
|
sa.ForeignKeyConstraint(["role_id"], [f"{schema}.role.id"], name=op.f("fk_device_role_role_id_role")),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_device_role")),
|
||||||
|
sa.UniqueConstraint("device_id", "role_id", name="uq_device_role_device_role"),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("device_role", schema=schema)
|
||||||
|
op.drop_table("signal_device", schema=schema)
|
||||||
|
op.drop_table("role", schema=schema)
|
||||||
|
op.drop_table("dead_letter_message", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,72 @@
|
|||||||
|
"""test.
|
||||||
|
|
||||||
|
Revision ID: 66bdd532bcab
|
||||||
|
Revises: 6eaf696e07a5
|
||||||
|
Create Date: 2026-03-18 19:21:14.561568
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
from sqlalchemy.dialects import postgresql
|
||||||
|
|
||||||
|
from python.orm import SignalBotBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "66bdd532bcab"
|
||||||
|
down_revision: str | None = "6eaf696e07a5"
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = SignalBotBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.alter_column(
|
||||||
|
"dead_letter_message",
|
||||||
|
"status",
|
||||||
|
existing_type=postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
||||||
|
type_=sa.Enum("UNPROCESSED", "PROCESSED", name="message_status", native_enum=False),
|
||||||
|
existing_nullable=False,
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.alter_column(
|
||||||
|
"signal_device",
|
||||||
|
"trust_level",
|
||||||
|
existing_type=postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
||||||
|
type_=sa.Enum("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", native_enum=False),
|
||||||
|
existing_nullable=False,
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.alter_column(
|
||||||
|
"signal_device",
|
||||||
|
"trust_level",
|
||||||
|
existing_type=sa.Enum("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", native_enum=False),
|
||||||
|
type_=postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
||||||
|
existing_nullable=False,
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.alter_column(
|
||||||
|
"dead_letter_message",
|
||||||
|
"status",
|
||||||
|
existing_type=sa.Enum("UNPROCESSED", "PROCESSED", name="message_status", native_enum=False),
|
||||||
|
type_=postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
||||||
|
existing_nullable=False,
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,80 @@
|
|||||||
|
"""starting van invintory.
|
||||||
|
|
||||||
|
Revision ID: 15e733499804
|
||||||
|
Revises:
|
||||||
|
Create Date: 2026-03-08 00:18:20.759720
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
from python.orm import VanInventoryBase
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "15e733499804"
|
||||||
|
down_revision: str | None = None
|
||||||
|
branch_labels: str | Sequence[str] | None = None
|
||||||
|
depends_on: str | Sequence[str] | None = None
|
||||||
|
|
||||||
|
schema = VanInventoryBase.schema_name
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Upgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"items",
|
||||||
|
sa.Column("name", sa.String(), nullable=False),
|
||||||
|
sa.Column("quantity", sa.Float(), nullable=False),
|
||||||
|
sa.Column("unit", sa.String(), nullable=False),
|
||||||
|
sa.Column("category", sa.String(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_items")),
|
||||||
|
sa.UniqueConstraint("name", name=op.f("uq_items_name")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"meals",
|
||||||
|
sa.Column("name", sa.String(), nullable=False),
|
||||||
|
sa.Column("instructions", sa.String(), nullable=True),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_meals")),
|
||||||
|
sa.UniqueConstraint("name", name=op.f("uq_meals_name")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
op.create_table(
|
||||||
|
"meal_ingredients",
|
||||||
|
sa.Column("meal_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("item_id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("quantity_needed", sa.Float(), nullable=False),
|
||||||
|
sa.Column("id", sa.Integer(), nullable=False),
|
||||||
|
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(["item_id"], [f"{schema}.items.id"], name=op.f("fk_meal_ingredients_item_id_items")),
|
||||||
|
sa.ForeignKeyConstraint(["meal_id"], [f"{schema}.meals.id"], name=op.f("fk_meal_ingredients_meal_id_meals")),
|
||||||
|
sa.PrimaryKeyConstraint("id", name=op.f("pk_meal_ingredients")),
|
||||||
|
sa.UniqueConstraint("meal_id", "item_id", name=op.f("uq_meal_ingredients_meal_id")),
|
||||||
|
schema=schema,
|
||||||
|
)
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Downgrade."""
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.drop_table("meal_ingredients", schema=schema)
|
||||||
|
op.drop_table("meals", schema=schema)
|
||||||
|
op.drop_table("items", schema=schema)
|
||||||
|
# ### end Alembic commands ###
|
||||||
1
python/api/__init__.py
Normal file
1
python/api/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""FastAPI applications."""
|
||||||
16
python/api/dependencies.py
Normal file
16
python/api/dependencies.py
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
"""FastAPI dependencies."""
|
||||||
|
|
||||||
|
from collections.abc import Iterator
|
||||||
|
from typing import Annotated
|
||||||
|
|
||||||
|
from fastapi import Depends, Request
|
||||||
|
from sqlalchemy.orm import Session
|
||||||
|
|
||||||
|
|
||||||
|
def get_db(request: Request) -> Iterator[Session]:
|
||||||
|
"""Get database session from app state."""
|
||||||
|
with Session(request.app.state.engine) as session:
|
||||||
|
yield session
|
||||||
|
|
||||||
|
|
||||||
|
DbSession = Annotated[Session, Depends(get_db)]
|
||||||
52
python/api/main.py
Normal file
52
python/api/main.py
Normal file
@@ -0,0 +1,52 @@
|
|||||||
|
"""FastAPI interface for Contact database."""
|
||||||
|
|
||||||
|
import logging
|
||||||
|
from collections.abc import AsyncIterator
|
||||||
|
from contextlib import asynccontextmanager
|
||||||
|
from typing import Annotated
|
||||||
|
|
||||||
|
import typer
|
||||||
|
import uvicorn
|
||||||
|
from fastapi import FastAPI
|
||||||
|
|
||||||
|
from python.api.middleware import ZstdMiddleware
|
||||||
|
from python.api.routers import contact_router, views_router
|
||||||
|
from python.common import configure_logger
|
||||||
|
from python.orm.common import get_postgres_engine
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def create_app() -> FastAPI:
|
||||||
|
"""Create and configure the FastAPI application."""
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def lifespan(app: FastAPI) -> AsyncIterator[None]:
|
||||||
|
"""Manage application lifespan."""
|
||||||
|
app.state.engine = get_postgres_engine()
|
||||||
|
yield
|
||||||
|
app.state.engine.dispose()
|
||||||
|
|
||||||
|
app = FastAPI(title="Contact Database API", lifespan=lifespan)
|
||||||
|
app.add_middleware(ZstdMiddleware)
|
||||||
|
|
||||||
|
app.include_router(contact_router)
|
||||||
|
app.include_router(views_router)
|
||||||
|
|
||||||
|
return app
|
||||||
|
|
||||||
|
|
||||||
|
def serve(
|
||||||
|
host: Annotated[str, typer.Option("--host", "-h", help="Host to bind to")],
|
||||||
|
port: Annotated[int, typer.Option("--port", "-p", help="Port to bind to")] = 8000,
|
||||||
|
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
||||||
|
) -> None:
|
||||||
|
"""Start the Contact API server."""
|
||||||
|
configure_logger(log_level)
|
||||||
|
|
||||||
|
app = create_app()
|
||||||
|
uvicorn.run(app, host=host, port=port)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
typer.run(serve)
|
||||||
49
python/api/middleware.py
Normal file
49
python/api/middleware.py
Normal file
@@ -0,0 +1,49 @@
|
|||||||
|
"""Middleware for the FastAPI application."""
|
||||||
|
|
||||||
|
from compression import zstd
|
||||||
|
from starlette.middleware.base import BaseHTTPMiddleware, RequestResponseEndpoint
|
||||||
|
from starlette.requests import Request
|
||||||
|
from starlette.responses import Response
|
||||||
|
|
||||||
|
MINIMUM_RESPONSE_SIZE = 500
|
||||||
|
|
||||||
|
|
||||||
|
class ZstdMiddleware(BaseHTTPMiddleware):
|
||||||
|
"""Middleware that compresses responses with zstd when the client supports it."""
|
||||||
|
|
||||||
|
async def dispatch(self, request: Request, call_next: RequestResponseEndpoint) -> Response:
|
||||||
|
"""Compress the response with zstd if the client accepts it."""
|
||||||
|
accepted_encodings = request.headers.get("accept-encoding", "")
|
||||||
|
if "zstd" not in accepted_encodings:
|
||||||
|
return await call_next(request)
|
||||||
|
|
||||||
|
response = await call_next(request)
|
||||||
|
|
||||||
|
if response.headers.get("content-encoding") or "text/event-stream" in response.headers.get("content-type", ""):
|
||||||
|
return response
|
||||||
|
|
||||||
|
body = b""
|
||||||
|
async for chunk in response.body_iterator:
|
||||||
|
body += chunk if isinstance(chunk, bytes) else chunk.encode()
|
||||||
|
|
||||||
|
if len(body) < MINIMUM_RESPONSE_SIZE:
|
||||||
|
return Response(
|
||||||
|
content=body,
|
||||||
|
status_code=response.status_code,
|
||||||
|
headers=dict(response.headers),
|
||||||
|
media_type=response.media_type,
|
||||||
|
)
|
||||||
|
|
||||||
|
compressed = zstd.compress(body)
|
||||||
|
|
||||||
|
headers = dict(response.headers)
|
||||||
|
headers["content-encoding"] = "zstd"
|
||||||
|
headers["content-length"] = str(len(compressed))
|
||||||
|
headers.pop("transfer-encoding", None)
|
||||||
|
|
||||||
|
return Response(
|
||||||
|
content=compressed,
|
||||||
|
status_code=response.status_code,
|
||||||
|
headers=headers,
|
||||||
|
media_type=response.media_type,
|
||||||
|
)
|
||||||
6
python/api/routers/__init__.py
Normal file
6
python/api/routers/__init__.py
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
"""API routers."""
|
||||||
|
|
||||||
|
from python.api.routers.contact import router as contact_router
|
||||||
|
from python.api.routers.views import router as views_router
|
||||||
|
|
||||||
|
__all__ = ["contact_router", "views_router"]
|
||||||
481
python/api/routers/contact.py
Normal file
481
python/api/routers/contact.py
Normal file
@@ -0,0 +1,481 @@
|
|||||||
|
"""Contact API router."""
|
||||||
|
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
from fastapi import APIRouter, HTTPException, Request
|
||||||
|
from fastapi.responses import HTMLResponse
|
||||||
|
from fastapi.templating import Jinja2Templates
|
||||||
|
from pydantic import BaseModel
|
||||||
|
from sqlalchemy import select
|
||||||
|
from sqlalchemy.orm import selectinload
|
||||||
|
|
||||||
|
from python.api.dependencies import DbSession
|
||||||
|
from python.orm.richie.contact import Contact, ContactRelationship, Need, RelationshipType
|
||||||
|
|
||||||
|
TEMPLATES_DIR = Path(__file__).parent.parent / "templates"
|
||||||
|
templates = Jinja2Templates(directory=TEMPLATES_DIR)
|
||||||
|
|
||||||
|
|
||||||
|
def _is_htmx(request: Request) -> bool:
|
||||||
|
"""Check if the request is from HTMX."""
|
||||||
|
return request.headers.get("HX-Request") == "true"
|
||||||
|
|
||||||
|
|
||||||
|
class NeedBase(BaseModel):
|
||||||
|
"""Base schema for Need."""
|
||||||
|
|
||||||
|
name: str
|
||||||
|
description: str | None = None
|
||||||
|
|
||||||
|
|
||||||
|
class NeedCreate(NeedBase):
|
||||||
|
"""Schema for creating a Need."""
|
||||||
|
|
||||||
|
|
||||||
|
class NeedResponse(NeedBase):
|
||||||
|
"""Schema for Need response."""
|
||||||
|
|
||||||
|
id: int
|
||||||
|
|
||||||
|
model_config = {"from_attributes": True}
|
||||||
|
|
||||||
|
|
||||||
|
class ContactRelationshipCreate(BaseModel):
|
||||||
|
"""Schema for creating a contact relationship."""
|
||||||
|
|
||||||
|
related_contact_id: int
|
||||||
|
relationship_type: RelationshipType
|
||||||
|
closeness_weight: int | None = None
|
||||||
|
|
||||||
|
|
||||||
|
class ContactRelationshipUpdate(BaseModel):
|
||||||
|
"""Schema for updating a contact relationship."""
|
||||||
|
|
||||||
|
relationship_type: RelationshipType | None = None
|
||||||
|
closeness_weight: int | None = None
|
||||||
|
|
||||||
|
|
||||||
|
class ContactRelationshipResponse(BaseModel):
|
||||||
|
"""Schema for contact relationship response."""
|
||||||
|
|
||||||
|
contact_id: int
|
||||||
|
related_contact_id: int
|
||||||
|
relationship_type: str
|
||||||
|
closeness_weight: int
|
||||||
|
|
||||||
|
model_config = {"from_attributes": True}
|
||||||
|
|
||||||
|
|
||||||
|
class RelationshipTypeInfo(BaseModel):
|
||||||
|
"""Information about a relationship type."""
|
||||||
|
|
||||||
|
value: str
|
||||||
|
display_name: str
|
||||||
|
default_weight: int
|
||||||
|
|
||||||
|
|
||||||
|
class GraphNode(BaseModel):
|
||||||
|
"""Node in the relationship graph."""
|
||||||
|
|
||||||
|
id: int
|
||||||
|
name: str
|
||||||
|
current_job: str | None = None
|
||||||
|
|
||||||
|
|
||||||
|
class GraphEdge(BaseModel):
|
||||||
|
"""Edge in the relationship graph."""
|
||||||
|
|
||||||
|
source: int
|
||||||
|
target: int
|
||||||
|
relationship_type: str
|
||||||
|
closeness_weight: int
|
||||||
|
|
||||||
|
|
||||||
|
class GraphData(BaseModel):
|
||||||
|
"""Complete graph data for visualization."""
|
||||||
|
|
||||||
|
nodes: list[GraphNode]
|
||||||
|
edges: list[GraphEdge]
|
||||||
|
|
||||||
|
|
||||||
|
class ContactBase(BaseModel):
|
||||||
|
"""Base schema for Contact."""
|
||||||
|
|
||||||
|
name: str
|
||||||
|
age: int | None = None
|
||||||
|
bio: str | None = None
|
||||||
|
current_job: str | None = None
|
||||||
|
gender: str | None = None
|
||||||
|
goals: str | None = None
|
||||||
|
legal_name: str | None = None
|
||||||
|
profile_pic: str | None = None
|
||||||
|
safe_conversation_starters: str | None = None
|
||||||
|
self_sufficiency_score: int | None = None
|
||||||
|
social_structure_style: str | None = None
|
||||||
|
ssn: str | None = None
|
||||||
|
suffix: str | None = None
|
||||||
|
timezone: str | None = None
|
||||||
|
topics_to_avoid: str | None = None
|
||||||
|
|
||||||
|
|
||||||
|
class ContactCreate(ContactBase):
|
||||||
|
"""Schema for creating a Contact."""
|
||||||
|
|
||||||
|
need_ids: list[int] = []
|
||||||
|
|
||||||
|
|
||||||
|
class ContactUpdate(BaseModel):
|
||||||
|
"""Schema for updating a Contact."""
|
||||||
|
|
||||||
|
name: str | None = None
|
||||||
|
age: int | None = None
|
||||||
|
bio: str | None = None
|
||||||
|
current_job: str | None = None
|
||||||
|
gender: str | None = None
|
||||||
|
goals: str | None = None
|
||||||
|
legal_name: str | None = None
|
||||||
|
profile_pic: str | None = None
|
||||||
|
safe_conversation_starters: str | None = None
|
||||||
|
self_sufficiency_score: int | None = None
|
||||||
|
social_structure_style: str | None = None
|
||||||
|
ssn: str | None = None
|
||||||
|
suffix: str | None = None
|
||||||
|
timezone: str | None = None
|
||||||
|
topics_to_avoid: str | None = None
|
||||||
|
need_ids: list[int] | None = None
|
||||||
|
|
||||||
|
|
||||||
|
class ContactResponse(ContactBase):
|
||||||
|
"""Schema for Contact response with relationships."""
|
||||||
|
|
||||||
|
id: int
|
||||||
|
needs: list[NeedResponse] = []
|
||||||
|
related_to: list[ContactRelationshipResponse] = []
|
||||||
|
related_from: list[ContactRelationshipResponse] = []
|
||||||
|
|
||||||
|
model_config = {"from_attributes": True}
|
||||||
|
|
||||||
|
|
||||||
|
class ContactListResponse(ContactBase):
|
||||||
|
"""Schema for Contact list response."""
|
||||||
|
|
||||||
|
id: int
|
||||||
|
|
||||||
|
model_config = {"from_attributes": True}
|
||||||
|
|
||||||
|
|
||||||
|
router = APIRouter(prefix="/api", tags=["contacts"])
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/needs", response_model=NeedResponse)
|
||||||
|
def create_need(need: NeedCreate, db: DbSession) -> Need:
|
||||||
|
"""Create a new need."""
|
||||||
|
db_need = Need(name=need.name, description=need.description)
|
||||||
|
db.add(db_need)
|
||||||
|
db.commit()
|
||||||
|
db.refresh(db_need)
|
||||||
|
return db_need
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/needs", response_model=list[NeedResponse])
|
||||||
|
def list_needs(db: DbSession) -> list[Need]:
|
||||||
|
"""List all needs."""
|
||||||
|
return list(db.scalars(select(Need)).all())
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/needs/{need_id}", response_model=NeedResponse)
|
||||||
|
def get_need(need_id: int, db: DbSession) -> Need:
|
||||||
|
"""Get a need by ID."""
|
||||||
|
need = db.get(Need, need_id)
|
||||||
|
if not need:
|
||||||
|
raise HTTPException(status_code=404, detail="Need not found")
|
||||||
|
return need
|
||||||
|
|
||||||
|
|
||||||
|
@router.delete("/needs/{need_id}", response_model=None)
|
||||||
|
def delete_need(need_id: int, request: Request, db: DbSession) -> dict[str, bool] | HTMLResponse:
|
||||||
|
"""Delete a need by ID."""
|
||||||
|
need = db.get(Need, need_id)
|
||||||
|
if not need:
|
||||||
|
raise HTTPException(status_code=404, detail="Need not found")
|
||||||
|
db.delete(need)
|
||||||
|
db.commit()
|
||||||
|
if _is_htmx(request):
|
||||||
|
return HTMLResponse("")
|
||||||
|
return {"deleted": True}
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/contacts", response_model=ContactResponse)
|
||||||
|
def create_contact(contact: ContactCreate, db: DbSession) -> Contact:
|
||||||
|
"""Create a new contact."""
|
||||||
|
need_ids = contact.need_ids
|
||||||
|
contact_data = contact.model_dump(exclude={"need_ids"})
|
||||||
|
db_contact = Contact(**contact_data)
|
||||||
|
|
||||||
|
if need_ids:
|
||||||
|
needs = list(db.scalars(select(Need).where(Need.id.in_(need_ids))).all())
|
||||||
|
db_contact.needs = needs
|
||||||
|
|
||||||
|
db.add(db_contact)
|
||||||
|
db.commit()
|
||||||
|
db.refresh(db_contact)
|
||||||
|
return db_contact
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/contacts", response_model=list[ContactListResponse])
|
||||||
|
def list_contacts(
|
||||||
|
db: DbSession,
|
||||||
|
skip: int = 0,
|
||||||
|
limit: int = 100,
|
||||||
|
) -> list[Contact]:
|
||||||
|
"""List all contacts with pagination."""
|
||||||
|
return list(db.scalars(select(Contact).offset(skip).limit(limit)).all())
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/contacts/{contact_id}", response_model=ContactResponse)
|
||||||
|
def get_contact(contact_id: int, db: DbSession) -> Contact:
|
||||||
|
"""Get a contact by ID with all relationships."""
|
||||||
|
contact = db.scalar(
|
||||||
|
select(Contact)
|
||||||
|
.where(Contact.id == contact_id)
|
||||||
|
.options(
|
||||||
|
selectinload(Contact.needs),
|
||||||
|
selectinload(Contact.related_to),
|
||||||
|
selectinload(Contact.related_from),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
return contact
|
||||||
|
|
||||||
|
|
||||||
|
@router.patch("/contacts/{contact_id}", response_model=ContactResponse)
|
||||||
|
def update_contact(
|
||||||
|
contact_id: int,
|
||||||
|
contact: ContactUpdate,
|
||||||
|
db: DbSession,
|
||||||
|
) -> Contact:
|
||||||
|
"""Update a contact by ID."""
|
||||||
|
db_contact = db.get(Contact, contact_id)
|
||||||
|
if not db_contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
update_data = contact.model_dump(exclude_unset=True)
|
||||||
|
need_ids = update_data.pop("need_ids", None)
|
||||||
|
|
||||||
|
for key, value in update_data.items():
|
||||||
|
setattr(db_contact, key, value)
|
||||||
|
|
||||||
|
if need_ids is not None:
|
||||||
|
needs = list(db.scalars(select(Need).where(Need.id.in_(need_ids))).all())
|
||||||
|
db_contact.needs = needs
|
||||||
|
|
||||||
|
db.commit()
|
||||||
|
db.refresh(db_contact)
|
||||||
|
return db_contact
|
||||||
|
|
||||||
|
|
||||||
|
@router.delete("/contacts/{contact_id}", response_model=None)
|
||||||
|
def delete_contact(contact_id: int, request: Request, db: DbSession) -> dict[str, bool] | HTMLResponse:
|
||||||
|
"""Delete a contact by ID."""
|
||||||
|
contact = db.get(Contact, contact_id)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
db.delete(contact)
|
||||||
|
db.commit()
|
||||||
|
if _is_htmx(request):
|
||||||
|
return HTMLResponse("")
|
||||||
|
return {"deleted": True}
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/contacts/{contact_id}/needs/{need_id}")
|
||||||
|
def add_need_to_contact(
|
||||||
|
contact_id: int,
|
||||||
|
need_id: int,
|
||||||
|
db: DbSession,
|
||||||
|
) -> dict[str, bool]:
|
||||||
|
"""Add a need to a contact."""
|
||||||
|
contact = db.get(Contact, contact_id)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
need = db.get(Need, need_id)
|
||||||
|
if not need:
|
||||||
|
raise HTTPException(status_code=404, detail="Need not found")
|
||||||
|
|
||||||
|
if need not in contact.needs:
|
||||||
|
contact.needs.append(need)
|
||||||
|
db.commit()
|
||||||
|
|
||||||
|
return {"added": True}
|
||||||
|
|
||||||
|
|
||||||
|
@router.delete("/contacts/{contact_id}/needs/{need_id}", response_model=None)
|
||||||
|
def remove_need_from_contact(
|
||||||
|
contact_id: int,
|
||||||
|
need_id: int,
|
||||||
|
request: Request,
|
||||||
|
db: DbSession,
|
||||||
|
) -> dict[str, bool] | HTMLResponse:
|
||||||
|
"""Remove a need from a contact."""
|
||||||
|
contact = db.get(Contact, contact_id)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
need = db.get(Need, need_id)
|
||||||
|
if not need:
|
||||||
|
raise HTTPException(status_code=404, detail="Need not found")
|
||||||
|
|
||||||
|
if need in contact.needs:
|
||||||
|
contact.needs.remove(need)
|
||||||
|
db.commit()
|
||||||
|
|
||||||
|
if _is_htmx(request):
|
||||||
|
return HTMLResponse("")
|
||||||
|
return {"removed": True}
|
||||||
|
|
||||||
|
|
||||||
|
@router.post(
|
||||||
|
"/contacts/{contact_id}/relationships",
|
||||||
|
response_model=ContactRelationshipResponse,
|
||||||
|
)
|
||||||
|
def add_contact_relationship(
|
||||||
|
contact_id: int,
|
||||||
|
relationship: ContactRelationshipCreate,
|
||||||
|
db: DbSession,
|
||||||
|
) -> ContactRelationship:
|
||||||
|
"""Add a relationship between two contacts."""
|
||||||
|
contact = db.get(Contact, contact_id)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
related_contact = db.get(Contact, relationship.related_contact_id)
|
||||||
|
if not related_contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Related contact not found")
|
||||||
|
|
||||||
|
if contact_id == relationship.related_contact_id:
|
||||||
|
raise HTTPException(status_code=400, detail="Cannot relate contact to itself")
|
||||||
|
|
||||||
|
# Use provided weight or default from relationship type
|
||||||
|
weight = relationship.closeness_weight
|
||||||
|
if weight is None:
|
||||||
|
weight = relationship.relationship_type.default_weight
|
||||||
|
|
||||||
|
db_relationship = ContactRelationship(
|
||||||
|
contact_id=contact_id,
|
||||||
|
related_contact_id=relationship.related_contact_id,
|
||||||
|
relationship_type=relationship.relationship_type.value,
|
||||||
|
closeness_weight=weight,
|
||||||
|
)
|
||||||
|
db.add(db_relationship)
|
||||||
|
db.commit()
|
||||||
|
db.refresh(db_relationship)
|
||||||
|
return db_relationship
|
||||||
|
|
||||||
|
|
||||||
|
@router.get(
|
||||||
|
"/contacts/{contact_id}/relationships",
|
||||||
|
response_model=list[ContactRelationshipResponse],
|
||||||
|
)
|
||||||
|
def get_contact_relationships(
|
||||||
|
contact_id: int,
|
||||||
|
db: DbSession,
|
||||||
|
) -> list[ContactRelationship]:
|
||||||
|
"""Get all relationships for a contact."""
|
||||||
|
contact = db.get(Contact, contact_id)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
outgoing = list(db.scalars(select(ContactRelationship).where(ContactRelationship.contact_id == contact_id)).all())
|
||||||
|
incoming = list(
|
||||||
|
db.scalars(select(ContactRelationship).where(ContactRelationship.related_contact_id == contact_id)).all()
|
||||||
|
)
|
||||||
|
return outgoing + incoming
|
||||||
|
|
||||||
|
|
||||||
|
@router.patch(
|
||||||
|
"/contacts/{contact_id}/relationships/{related_contact_id}",
|
||||||
|
response_model=ContactRelationshipResponse,
|
||||||
|
)
|
||||||
|
def update_contact_relationship(
|
||||||
|
contact_id: int,
|
||||||
|
related_contact_id: int,
|
||||||
|
update: ContactRelationshipUpdate,
|
||||||
|
db: DbSession,
|
||||||
|
) -> ContactRelationship:
|
||||||
|
"""Update a relationship between two contacts."""
|
||||||
|
relationship = db.scalar(
|
||||||
|
select(ContactRelationship).where(
|
||||||
|
ContactRelationship.contact_id == contact_id,
|
||||||
|
ContactRelationship.related_contact_id == related_contact_id,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if not relationship:
|
||||||
|
raise HTTPException(status_code=404, detail="Relationship not found")
|
||||||
|
|
||||||
|
if update.relationship_type is not None:
|
||||||
|
relationship.relationship_type = update.relationship_type.value
|
||||||
|
if update.closeness_weight is not None:
|
||||||
|
relationship.closeness_weight = update.closeness_weight
|
||||||
|
|
||||||
|
db.commit()
|
||||||
|
db.refresh(relationship)
|
||||||
|
return relationship
|
||||||
|
|
||||||
|
|
||||||
|
@router.delete("/contacts/{contact_id}/relationships/{related_contact_id}", response_model=None)
|
||||||
|
def remove_contact_relationship(
|
||||||
|
contact_id: int,
|
||||||
|
related_contact_id: int,
|
||||||
|
request: Request,
|
||||||
|
db: DbSession,
|
||||||
|
) -> dict[str, bool] | HTMLResponse:
|
||||||
|
"""Remove a relationship between two contacts."""
|
||||||
|
relationship = db.scalar(
|
||||||
|
select(ContactRelationship).where(
|
||||||
|
ContactRelationship.contact_id == contact_id,
|
||||||
|
ContactRelationship.related_contact_id == related_contact_id,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if not relationship:
|
||||||
|
raise HTTPException(status_code=404, detail="Relationship not found")
|
||||||
|
|
||||||
|
db.delete(relationship)
|
||||||
|
db.commit()
|
||||||
|
if _is_htmx(request):
|
||||||
|
return HTMLResponse("")
|
||||||
|
return {"deleted": True}
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/relationship-types")
|
||||||
|
def list_relationship_types() -> list[RelationshipTypeInfo]:
|
||||||
|
"""List all available relationship types with their default weights."""
|
||||||
|
return [
|
||||||
|
RelationshipTypeInfo(
|
||||||
|
value=rt.value,
|
||||||
|
display_name=rt.display_name,
|
||||||
|
default_weight=rt.default_weight,
|
||||||
|
)
|
||||||
|
for rt in RelationshipType
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/graph")
|
||||||
|
def get_relationship_graph(db: DbSession) -> GraphData:
|
||||||
|
"""Get all contacts and relationships as graph data for visualization."""
|
||||||
|
contacts = list(db.scalars(select(Contact)).all())
|
||||||
|
relationships = list(db.scalars(select(ContactRelationship)).all())
|
||||||
|
|
||||||
|
nodes = [GraphNode(id=c.id, name=c.name, current_job=c.current_job) for c in contacts]
|
||||||
|
|
||||||
|
edges = [
|
||||||
|
GraphEdge(
|
||||||
|
source=rel.contact_id,
|
||||||
|
target=rel.related_contact_id,
|
||||||
|
relationship_type=rel.relationship_type,
|
||||||
|
closeness_weight=rel.closeness_weight,
|
||||||
|
)
|
||||||
|
for rel in relationships
|
||||||
|
]
|
||||||
|
|
||||||
|
return GraphData(nodes=nodes, edges=edges)
|
||||||
345
python/api/routers/views.py
Normal file
345
python/api/routers/views.py
Normal file
@@ -0,0 +1,345 @@
|
|||||||
|
"""HTMX server-rendered view router."""
|
||||||
|
|
||||||
|
from pathlib import Path
|
||||||
|
from typing import Annotated, Any
|
||||||
|
|
||||||
|
from fastapi import APIRouter, Form, HTTPException, Request
|
||||||
|
from fastapi.responses import HTMLResponse, RedirectResponse
|
||||||
|
from fastapi.templating import Jinja2Templates
|
||||||
|
from sqlalchemy import select
|
||||||
|
from sqlalchemy.orm import Session, selectinload
|
||||||
|
|
||||||
|
from python.api.dependencies import DbSession
|
||||||
|
from python.orm.richie.contact import Contact, ContactRelationship, Need, RelationshipType
|
||||||
|
|
||||||
|
TEMPLATES_DIR = Path(__file__).parent.parent / "templates"
|
||||||
|
templates = Jinja2Templates(directory=TEMPLATES_DIR)
|
||||||
|
|
||||||
|
router = APIRouter(tags=["views"])
|
||||||
|
|
||||||
|
FAMILIAL_TYPES = {
|
||||||
|
"parent",
|
||||||
|
"child",
|
||||||
|
"sibling",
|
||||||
|
"grandparent",
|
||||||
|
"grandchild",
|
||||||
|
"aunt_uncle",
|
||||||
|
"niece_nephew",
|
||||||
|
"cousin",
|
||||||
|
"in_law",
|
||||||
|
}
|
||||||
|
FRIEND_TYPES = {"best_friend", "close_friend", "friend", "acquaintance", "neighbor"}
|
||||||
|
PARTNER_TYPES = {"spouse", "partner"}
|
||||||
|
PROFESSIONAL_TYPES = {"mentor", "mentee", "business_partner", "colleague", "manager", "direct_report", "client"}
|
||||||
|
|
||||||
|
CONTACT_STRING_FIELDS = (
|
||||||
|
"name",
|
||||||
|
"legal_name",
|
||||||
|
"suffix",
|
||||||
|
"gender",
|
||||||
|
"current_job",
|
||||||
|
"timezone",
|
||||||
|
"profile_pic",
|
||||||
|
"bio",
|
||||||
|
"goals",
|
||||||
|
"social_structure_style",
|
||||||
|
"safe_conversation_starters",
|
||||||
|
"topics_to_avoid",
|
||||||
|
"ssn",
|
||||||
|
)
|
||||||
|
|
||||||
|
CONTACT_INT_FIELDS = ("age", "self_sufficiency_score")
|
||||||
|
|
||||||
|
|
||||||
|
def _group_relationships(relationships: list[ContactRelationship]) -> dict[str, list[ContactRelationship]]:
|
||||||
|
"""Group relationships by category."""
|
||||||
|
groups: dict[str, list[ContactRelationship]] = {
|
||||||
|
"familial": [],
|
||||||
|
"partners": [],
|
||||||
|
"friends": [],
|
||||||
|
"professional": [],
|
||||||
|
"other": [],
|
||||||
|
}
|
||||||
|
for rel in relationships:
|
||||||
|
if rel.relationship_type in FAMILIAL_TYPES:
|
||||||
|
groups["familial"].append(rel)
|
||||||
|
elif rel.relationship_type in PARTNER_TYPES:
|
||||||
|
groups["partners"].append(rel)
|
||||||
|
elif rel.relationship_type in FRIEND_TYPES:
|
||||||
|
groups["friends"].append(rel)
|
||||||
|
elif rel.relationship_type in PROFESSIONAL_TYPES:
|
||||||
|
groups["professional"].append(rel)
|
||||||
|
else:
|
||||||
|
groups["other"].append(rel)
|
||||||
|
return groups
|
||||||
|
|
||||||
|
|
||||||
|
def _build_contact_name_map(database: Session, contact: Contact) -> dict[int, str]:
|
||||||
|
"""Build a mapping of contact IDs to names for relationship display."""
|
||||||
|
related_ids = {rel.related_contact_id for rel in contact.related_to}
|
||||||
|
related_ids |= {rel.contact_id for rel in contact.related_from}
|
||||||
|
related_ids.discard(contact.id)
|
||||||
|
|
||||||
|
if not related_ids:
|
||||||
|
return {}
|
||||||
|
|
||||||
|
related_contacts = list(database.scalars(select(Contact).where(Contact.id.in_(related_ids))).all())
|
||||||
|
return {related.id: related.name for related in related_contacts}
|
||||||
|
|
||||||
|
|
||||||
|
def _get_relationship_type_display() -> dict[str, str]:
|
||||||
|
"""Build a mapping of relationship type values to display names."""
|
||||||
|
return {rel_type.value: rel_type.display_name for rel_type in RelationshipType}
|
||||||
|
|
||||||
|
|
||||||
|
async def _parse_contact_form(request: Request) -> dict[str, Any]:
|
||||||
|
"""Parse contact form data from a multipart/form request."""
|
||||||
|
form_data = await request.form()
|
||||||
|
result: dict[str, Any] = {}
|
||||||
|
|
||||||
|
for field in CONTACT_STRING_FIELDS:
|
||||||
|
value = form_data.get(field, "")
|
||||||
|
result[field] = str(value) if value else None
|
||||||
|
|
||||||
|
for field in CONTACT_INT_FIELDS:
|
||||||
|
value = form_data.get(field, "")
|
||||||
|
result[field] = int(value) if value else None
|
||||||
|
|
||||||
|
result["need_ids"] = [int(value) for value in form_data.getlist("need_ids")]
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
|
def _save_contact_from_form(database: Session, contact: Contact, form_result: dict[str, Any]) -> None:
|
||||||
|
"""Apply parsed form data to a Contact and save associated needs."""
|
||||||
|
need_ids = form_result.pop("need_ids")
|
||||||
|
|
||||||
|
for key, value in form_result.items():
|
||||||
|
setattr(contact, key, value)
|
||||||
|
|
||||||
|
if need_ids:
|
||||||
|
contact.needs = list(database.scalars(select(Need).where(Need.id.in_(need_ids))).all())
|
||||||
|
else:
|
||||||
|
contact.needs = []
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/", response_class=HTMLResponse)
|
||||||
|
@router.get("/contacts", response_class=HTMLResponse)
|
||||||
|
def contact_list_page(request: Request, database: DbSession) -> HTMLResponse:
|
||||||
|
"""Render the contacts list page."""
|
||||||
|
contacts = list(database.scalars(select(Contact)).all())
|
||||||
|
return templates.TemplateResponse(request, "contact_list.html", {"contacts": contacts})
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/contacts/new", response_class=HTMLResponse)
|
||||||
|
def new_contact_page(request: Request, database: DbSession) -> HTMLResponse:
|
||||||
|
"""Render the new contact form page."""
|
||||||
|
all_needs = list(database.scalars(select(Need)).all())
|
||||||
|
return templates.TemplateResponse(request, "contact_form.html", {"contact": None, "all_needs": all_needs})
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/htmx/contacts/new")
|
||||||
|
async def create_contact_form(request: Request, database: DbSession) -> RedirectResponse:
|
||||||
|
"""Handle the create contact form submission."""
|
||||||
|
form_result = await _parse_contact_form(request)
|
||||||
|
contact = Contact()
|
||||||
|
_save_contact_from_form(database, contact, form_result)
|
||||||
|
|
||||||
|
database.add(contact)
|
||||||
|
database.commit()
|
||||||
|
database.refresh(contact)
|
||||||
|
return RedirectResponse(url=f"/contacts/{contact.id}", status_code=303)
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/contacts/{contact_id}", response_class=HTMLResponse)
|
||||||
|
def contact_detail_page(contact_id: int, request: Request, database: DbSession) -> HTMLResponse:
|
||||||
|
"""Render the contact detail page."""
|
||||||
|
contact = database.scalar(
|
||||||
|
select(Contact)
|
||||||
|
.where(Contact.id == contact_id)
|
||||||
|
.options(
|
||||||
|
selectinload(Contact.needs),
|
||||||
|
selectinload(Contact.related_to),
|
||||||
|
selectinload(Contact.related_from),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
contact_names = _build_contact_name_map(database, contact)
|
||||||
|
grouped_relationships = _group_relationships(contact.related_to)
|
||||||
|
all_contacts = list(database.scalars(select(Contact)).all())
|
||||||
|
all_needs = list(database.scalars(select(Need)).all())
|
||||||
|
available_needs = [need for need in all_needs if need not in contact.needs]
|
||||||
|
|
||||||
|
return templates.TemplateResponse(
|
||||||
|
request,
|
||||||
|
"contact_detail.html",
|
||||||
|
{
|
||||||
|
"contact": contact,
|
||||||
|
"contact_names": contact_names,
|
||||||
|
"grouped_relationships": grouped_relationships,
|
||||||
|
"all_contacts": all_contacts,
|
||||||
|
"available_needs": available_needs,
|
||||||
|
"relationship_types": list(RelationshipType),
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/contacts/{contact_id}/edit", response_class=HTMLResponse)
|
||||||
|
def edit_contact_page(contact_id: int, request: Request, database: DbSession) -> HTMLResponse:
|
||||||
|
"""Render the edit contact form page."""
|
||||||
|
contact = database.scalar(select(Contact).where(Contact.id == contact_id).options(selectinload(Contact.needs)))
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
all_needs = list(database.scalars(select(Need)).all())
|
||||||
|
return templates.TemplateResponse(request, "contact_form.html", {"contact": contact, "all_needs": all_needs})
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/htmx/contacts/{contact_id}/edit")
|
||||||
|
async def update_contact_form(contact_id: int, request: Request, database: DbSession) -> RedirectResponse:
|
||||||
|
"""Handle the edit contact form submission."""
|
||||||
|
contact = database.get(Contact, contact_id)
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
form_result = await _parse_contact_form(request)
|
||||||
|
_save_contact_from_form(database, contact, form_result)
|
||||||
|
|
||||||
|
database.commit()
|
||||||
|
return RedirectResponse(url=f"/contacts/{contact_id}", status_code=303)
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/htmx/contacts/{contact_id}/add-need", response_class=HTMLResponse)
|
||||||
|
def add_need_to_contact_htmx(
|
||||||
|
contact_id: int,
|
||||||
|
request: Request,
|
||||||
|
database: DbSession,
|
||||||
|
need_id: Annotated[int, Form()],
|
||||||
|
) -> HTMLResponse:
|
||||||
|
"""Add a need to a contact and return updated manage-needs partial."""
|
||||||
|
contact = database.scalar(select(Contact).where(Contact.id == contact_id).options(selectinload(Contact.needs)))
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
need = database.get(Need, need_id)
|
||||||
|
if not need:
|
||||||
|
raise HTTPException(status_code=404, detail="Need not found")
|
||||||
|
|
||||||
|
if need not in contact.needs:
|
||||||
|
contact.needs.append(need)
|
||||||
|
database.commit()
|
||||||
|
database.refresh(contact)
|
||||||
|
|
||||||
|
return templates.TemplateResponse(request, "partials/manage_needs.html", {"contact": contact})
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/htmx/contacts/{contact_id}/add-relationship", response_class=HTMLResponse)
|
||||||
|
def add_relationship_htmx(
|
||||||
|
contact_id: int,
|
||||||
|
request: Request,
|
||||||
|
database: DbSession,
|
||||||
|
related_contact_id: Annotated[int, Form()],
|
||||||
|
relationship_type: Annotated[str, Form()],
|
||||||
|
) -> HTMLResponse:
|
||||||
|
"""Add a relationship and return updated manage-relationships partial."""
|
||||||
|
contact = database.scalar(select(Contact).where(Contact.id == contact_id).options(selectinload(Contact.related_to)))
|
||||||
|
if not contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Contact not found")
|
||||||
|
|
||||||
|
related_contact = database.get(Contact, related_contact_id)
|
||||||
|
if not related_contact:
|
||||||
|
raise HTTPException(status_code=404, detail="Related contact not found")
|
||||||
|
|
||||||
|
rel_type = RelationshipType(relationship_type)
|
||||||
|
weight = rel_type.default_weight
|
||||||
|
|
||||||
|
relationship = ContactRelationship(
|
||||||
|
contact_id=contact_id,
|
||||||
|
related_contact_id=related_contact_id,
|
||||||
|
relationship_type=relationship_type,
|
||||||
|
closeness_weight=weight,
|
||||||
|
)
|
||||||
|
database.add(relationship)
|
||||||
|
database.commit()
|
||||||
|
database.refresh(contact)
|
||||||
|
|
||||||
|
contact_names = _build_contact_name_map(database, contact)
|
||||||
|
return templates.TemplateResponse(
|
||||||
|
request,
|
||||||
|
"partials/manage_relationships.html",
|
||||||
|
{"contact": contact, "contact_names": contact_names},
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/htmx/contacts/{contact_id}/relationships/{related_contact_id}/weight")
|
||||||
|
def update_relationship_weight_htmx(
|
||||||
|
contact_id: int,
|
||||||
|
related_contact_id: int,
|
||||||
|
database: DbSession,
|
||||||
|
closeness_weight: Annotated[int, Form()],
|
||||||
|
) -> HTMLResponse:
|
||||||
|
"""Update a relationship's closeness weight from HTMX range input."""
|
||||||
|
relationship = database.scalar(
|
||||||
|
select(ContactRelationship).where(
|
||||||
|
ContactRelationship.contact_id == contact_id,
|
||||||
|
ContactRelationship.related_contact_id == related_contact_id,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
if not relationship:
|
||||||
|
raise HTTPException(status_code=404, detail="Relationship not found")
|
||||||
|
|
||||||
|
relationship.closeness_weight = closeness_weight
|
||||||
|
database.commit()
|
||||||
|
return HTMLResponse("")
|
||||||
|
|
||||||
|
|
||||||
|
@router.post("/htmx/needs", response_class=HTMLResponse)
|
||||||
|
def create_need_htmx(
|
||||||
|
request: Request,
|
||||||
|
database: DbSession,
|
||||||
|
name: Annotated[str, Form()],
|
||||||
|
description: Annotated[str, Form()] = "",
|
||||||
|
) -> HTMLResponse:
|
||||||
|
"""Create a need via form data and return updated needs list."""
|
||||||
|
need = Need(name=name, description=description or None)
|
||||||
|
database.add(need)
|
||||||
|
database.commit()
|
||||||
|
needs = list(database.scalars(select(Need)).all())
|
||||||
|
return templates.TemplateResponse(request, "partials/need_items.html", {"needs": needs})
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/needs", response_class=HTMLResponse)
|
||||||
|
def needs_page(request: Request, database: DbSession) -> HTMLResponse:
|
||||||
|
"""Render the needs list page."""
|
||||||
|
needs = list(database.scalars(select(Need)).all())
|
||||||
|
return templates.TemplateResponse(request, "need_list.html", {"needs": needs})
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/graph", response_class=HTMLResponse)
|
||||||
|
def graph_page(request: Request, database: DbSession) -> HTMLResponse:
|
||||||
|
"""Render the relationship graph page."""
|
||||||
|
contacts = list(database.scalars(select(Contact)).all())
|
||||||
|
relationships = list(database.scalars(select(ContactRelationship)).all())
|
||||||
|
|
||||||
|
graph_data = {
|
||||||
|
"nodes": [{"id": contact.id, "name": contact.name, "current_job": contact.current_job} for contact in contacts],
|
||||||
|
"edges": [
|
||||||
|
{
|
||||||
|
"source": rel.contact_id,
|
||||||
|
"target": rel.related_contact_id,
|
||||||
|
"relationship_type": rel.relationship_type,
|
||||||
|
"closeness_weight": rel.closeness_weight,
|
||||||
|
}
|
||||||
|
for rel in relationships
|
||||||
|
],
|
||||||
|
}
|
||||||
|
|
||||||
|
return templates.TemplateResponse(
|
||||||
|
request,
|
||||||
|
"graph.html",
|
||||||
|
{
|
||||||
|
"graph_data": graph_data,
|
||||||
|
"relationship_type_display": _get_relationship_type_display(),
|
||||||
|
},
|
||||||
|
)
|
||||||
198
python/api/templates/base.html
Normal file
198
python/api/templates/base.html
Normal file
@@ -0,0 +1,198 @@
|
|||||||
|
<!DOCTYPE html>
|
||||||
|
<html lang="en" data-theme="light">
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8">
|
||||||
|
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||||
|
<title>{% block title %}Contact Database{% endblock %}</title>
|
||||||
|
<script src="https://unpkg.com/htmx.org@2.0.4"></script>
|
||||||
|
<style>
|
||||||
|
:root {
|
||||||
|
--color-bg: #f5f5f5;
|
||||||
|
--color-bg-card: #ffffff;
|
||||||
|
--color-bg-hover: #f0f0f0;
|
||||||
|
--color-bg-muted: #f9f9f9;
|
||||||
|
--color-bg-error: #ffe0e0;
|
||||||
|
--color-text: #333333;
|
||||||
|
--color-text-muted: #666666;
|
||||||
|
--color-text-error: #cc0000;
|
||||||
|
--color-border: #dddddd;
|
||||||
|
--color-border-light: #eeeeee;
|
||||||
|
--color-border-lighter: #f0f0f0;
|
||||||
|
--color-primary: #0066cc;
|
||||||
|
--color-primary-hover: #0055aa;
|
||||||
|
--color-danger: #cc3333;
|
||||||
|
--color-danger-hover: #aa2222;
|
||||||
|
--color-tag-bg: #e0e0e0;
|
||||||
|
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
|
||||||
|
font-family: -apple-system, BlinkMacSystemFont, "Segoe UI", Roboto, sans-serif;
|
||||||
|
line-height: 1.5;
|
||||||
|
color: var(--color-text);
|
||||||
|
background-color: var(--color-bg);
|
||||||
|
}
|
||||||
|
[data-theme="dark"] {
|
||||||
|
--color-bg: #1a1a1a;
|
||||||
|
--color-bg-card: #2d2d2d;
|
||||||
|
--color-bg-hover: #3d3d3d;
|
||||||
|
--color-bg-muted: #252525;
|
||||||
|
--color-bg-error: #4a2020;
|
||||||
|
--color-text: #e0e0e0;
|
||||||
|
--color-text-muted: #a0a0a0;
|
||||||
|
--color-text-error: #ff6b6b;
|
||||||
|
--color-border: #404040;
|
||||||
|
--color-border-light: #353535;
|
||||||
|
--color-border-lighter: #303030;
|
||||||
|
--color-primary: #4da6ff;
|
||||||
|
--color-primary-hover: #7dbfff;
|
||||||
|
--color-danger: #ff6b6b;
|
||||||
|
--color-danger-hover: #ff8a8a;
|
||||||
|
--color-tag-bg: #404040;
|
||||||
|
--shadow: 0 1px 3px rgba(0, 0, 0, 0.3);
|
||||||
|
}
|
||||||
|
* { box-sizing: border-box; }
|
||||||
|
body { margin: 0; background: var(--color-bg); color: var(--color-text); }
|
||||||
|
.app { max-width: 1000px; margin: 0 auto; padding: 20px; }
|
||||||
|
nav { display: flex; align-items: center; gap: 20px; padding: 15px 0; border-bottom: 1px solid var(--color-border); margin-bottom: 20px; }
|
||||||
|
nav a { color: var(--color-primary); text-decoration: none; font-weight: 500; }
|
||||||
|
nav a:hover { text-decoration: underline; }
|
||||||
|
.theme-toggle { margin-left: auto; }
|
||||||
|
main { background: var(--color-bg-card); padding: 20px; border-radius: 8px; box-shadow: var(--shadow); }
|
||||||
|
.header { display: flex; justify-content: space-between; align-items: center; margin-bottom: 20px; }
|
||||||
|
.header h1 { margin: 0; }
|
||||||
|
a { color: var(--color-primary); }
|
||||||
|
a:hover { text-decoration: underline; }
|
||||||
|
|
||||||
|
.btn { display: inline-block; padding: 8px 16px; border: 1px solid var(--color-border); border-radius: 4px; background: var(--color-bg-card); color: var(--color-text); text-decoration: none; cursor: pointer; font-size: 14px; margin-left: 8px; }
|
||||||
|
.btn:hover { background: var(--color-bg-hover); }
|
||||||
|
.btn-primary { background: var(--color-primary); border-color: var(--color-primary); color: white; }
|
||||||
|
.btn-primary:hover { background: var(--color-primary-hover); }
|
||||||
|
.btn-danger { background: var(--color-danger); border-color: var(--color-danger); color: white; }
|
||||||
|
.btn-danger:hover { background: var(--color-danger-hover); }
|
||||||
|
.btn-small { padding: 4px 8px; font-size: 12px; }
|
||||||
|
.btn:disabled { opacity: 0.6; cursor: not-allowed; }
|
||||||
|
|
||||||
|
table { width: 100%; border-collapse: collapse; }
|
||||||
|
th, td { padding: 12px; text-align: left; border-bottom: 1px solid var(--color-border-light); }
|
||||||
|
th { font-weight: 600; background: var(--color-bg-muted); }
|
||||||
|
tr:hover { background: var(--color-bg-muted); }
|
||||||
|
|
||||||
|
.error { background: var(--color-bg-error); color: var(--color-text-error); padding: 10px; border-radius: 4px; margin-bottom: 20px; }
|
||||||
|
.tag { display: inline-block; background: var(--color-tag-bg); padding: 2px 8px; border-radius: 12px; font-size: 12px; color: var(--color-text-muted); }
|
||||||
|
|
||||||
|
.add-form { display: flex; gap: 10px; margin-top: 15px; flex-wrap: wrap; }
|
||||||
|
.add-form select, .add-form input { padding: 8px; border: 1px solid var(--color-border); border-radius: 4px; min-width: 200px; background: var(--color-bg-card); color: var(--color-text); }
|
||||||
|
|
||||||
|
.form-group { margin-bottom: 20px; }
|
||||||
|
.form-group label { display: block; font-weight: 500; margin-bottom: 5px; }
|
||||||
|
.form-group input, .form-group textarea, .form-group select { width: 100%; padding: 10px; border: 1px solid var(--color-border); border-radius: 4px; font-size: 14px; background: var(--color-bg-card); color: var(--color-text); }
|
||||||
|
.form-group textarea { resize: vertical; }
|
||||||
|
.form-row { display: grid; grid-template-columns: 1fr 1fr; gap: 20px; }
|
||||||
|
.checkbox-group { display: flex; flex-wrap: wrap; gap: 15px; }
|
||||||
|
.checkbox-label { display: flex; align-items: center; gap: 5px; cursor: pointer; }
|
||||||
|
.form-actions { display: flex; gap: 10px; margin-top: 30px; padding-top: 20px; border-top: 1px solid var(--color-border-light); }
|
||||||
|
|
||||||
|
.need-form { background: var(--color-bg-muted); padding: 20px; border-radius: 4px; margin-bottom: 20px; }
|
||||||
|
.need-items { list-style: none; padding: 0; }
|
||||||
|
.need-items li { display: flex; justify-content: space-between; align-items: flex-start; padding: 15px; border: 1px solid var(--color-border-light); border-radius: 4px; margin-bottom: 10px; }
|
||||||
|
.need-info p { margin: 5px 0 0; color: var(--color-text-muted); font-size: 14px; }
|
||||||
|
|
||||||
|
.graph-container { width: 100%; }
|
||||||
|
.graph-hint { color: var(--color-text-muted); font-size: 14px; margin-bottom: 15px; }
|
||||||
|
.selected-info { margin-top: 15px; padding: 15px; background: var(--color-bg-muted); border-radius: 8px; }
|
||||||
|
.selected-info h3 { margin: 0 0 10px; }
|
||||||
|
.selected-info p { margin: 5px 0; color: var(--color-text-muted); }
|
||||||
|
.legend { margin-top: 20px; padding: 15px; background: var(--color-bg-muted); border-radius: 8px; }
|
||||||
|
.legend h4 { margin: 0 0 10px; font-size: 14px; }
|
||||||
|
.legend-items { display: flex; flex-wrap: wrap; gap: 15px; }
|
||||||
|
.legend-item { display: flex; align-items: center; gap: 8px; font-size: 12px; color: var(--color-text-muted); }
|
||||||
|
.legend-line { width: 30px; border-radius: 2px; }
|
||||||
|
|
||||||
|
.id-card { width: 100%; }
|
||||||
|
.id-card-inner { background: linear-gradient(135deg, #0a0a0f 0%, #1a1a2e 50%, #0a0a0f 100%); background-image: radial-gradient(white 1px, transparent 1px), linear-gradient(135deg, #0a0a0f 0%, #1a1a2e 50%, #0a0a0f 100%); background-size: 50px 50px, 100% 100%; color: #fff; border-radius: 12px; padding: 25px; min-height: 500px; position: relative; overflow: hidden; }
|
||||||
|
.id-card-header { display: flex; justify-content: space-between; align-items: flex-start; margin-bottom: 15px; }
|
||||||
|
.id-card-header-left { flex: 1; }
|
||||||
|
.id-card-header-right { display: flex; flex-direction: column; align-items: flex-end; gap: 10px; }
|
||||||
|
.id-card-title { font-size: 2.5rem; font-weight: 700; margin: 0; color: #fff; text-shadow: 2px 2px 4px rgba(0,0,0,0.5); }
|
||||||
|
.id-profile-pic { width: 80px; height: 80px; border-radius: 8px; object-fit: cover; border: 2px solid rgba(255,255,255,0.3); }
|
||||||
|
.id-profile-placeholder { width: 80px; height: 80px; border-radius: 8px; background: linear-gradient(135deg, #4ecdc4 0%, #44a8a0 100%); display: flex; align-items: center; justify-content: center; border: 2px solid rgba(255,255,255,0.3); }
|
||||||
|
.id-profile-placeholder span { font-size: 2rem; font-weight: 700; color: #fff; text-shadow: 1px 1px 2px rgba(0,0,0,0.3); }
|
||||||
|
.id-card-actions { display: flex; gap: 8px; }
|
||||||
|
.id-card-actions .btn { background: rgba(255,255,255,0.1); border-color: rgba(255,255,255,0.3); color: #fff; }
|
||||||
|
.id-card-actions .btn:hover { background: rgba(255,255,255,0.2); }
|
||||||
|
.id-card-body { display: grid; grid-template-columns: 1fr 1.5fr; gap: 30px; }
|
||||||
|
.id-card-left { display: flex; flex-direction: column; gap: 8px; }
|
||||||
|
.id-field { font-size: 1rem; line-height: 1.4; }
|
||||||
|
.id-field-block { margin-top: 15px; font-size: 0.95rem; line-height: 1.5; }
|
||||||
|
.id-label { color: #4ecdc4; font-weight: 500; }
|
||||||
|
.id-card-right { display: flex; flex-direction: column; gap: 20px; }
|
||||||
|
.id-bio { font-size: 0.9rem; line-height: 1.6; color: #e0e0e0; }
|
||||||
|
.id-relationships { margin-top: 10px; }
|
||||||
|
.id-section-title { font-size: 1.5rem; margin: 0 0 15px; color: #fff; border-bottom: 1px solid rgba(255,255,255,0.2); padding-bottom: 8px; }
|
||||||
|
.id-rel-group { margin-bottom: 12px; font-size: 0.9rem; line-height: 1.6; }
|
||||||
|
.id-rel-label { color: #a0a0a0; }
|
||||||
|
.id-rel-group a { color: #4ecdc4; text-decoration: none; }
|
||||||
|
.id-rel-group a:hover { text-decoration: underline; }
|
||||||
|
.id-rel-type { color: #888; font-size: 0.85em; }
|
||||||
|
.id-card-warnings { margin-top: 30px; padding-top: 20px; border-top: 1px solid rgba(255,255,255,0.2); display: flex; flex-wrap: wrap; gap: 20px; }
|
||||||
|
.id-warning { display: flex; align-items: center; gap: 8px; font-size: 0.9rem; color: #ff6b6b; }
|
||||||
|
.warning-dot { width: 8px; height: 8px; background: #ff6b6b; border-radius: 50%; flex-shrink: 0; }
|
||||||
|
.warning-desc { color: #ccc; }
|
||||||
|
|
||||||
|
.id-card-manage { margin-top: 20px; background: var(--color-bg-muted); border-radius: 8px; padding: 15px; }
|
||||||
|
.id-card-manage summary { cursor: pointer; font-weight: 600; font-size: 1.1rem; padding: 5px 0; }
|
||||||
|
.id-card-manage[open] summary { margin-bottom: 15px; border-bottom: 1px solid var(--color-border-light); padding-bottom: 10px; }
|
||||||
|
.manage-section { margin-bottom: 25px; }
|
||||||
|
.manage-section h3 { margin: 0 0 15px; font-size: 1rem; }
|
||||||
|
.manage-relationships { display: flex; flex-direction: column; gap: 10px; margin-bottom: 15px; }
|
||||||
|
.manage-rel-item { display: flex; align-items: center; gap: 12px; padding: 10px; background: var(--color-bg-card); border-radius: 6px; flex-wrap: wrap; }
|
||||||
|
.manage-rel-item a { font-weight: 500; min-width: 120px; }
|
||||||
|
.weight-control { display: flex; align-items: center; gap: 8px; font-size: 12px; color: var(--color-text-muted); }
|
||||||
|
.weight-control input[type="range"] { width: 80px; cursor: pointer; }
|
||||||
|
.weight-value { min-width: 20px; text-align: center; font-weight: 600; }
|
||||||
|
.manage-needs-list { list-style: none; padding: 0; margin: 0 0 15px; }
|
||||||
|
.manage-needs-list li { display: flex; align-items: center; gap: 12px; padding: 10px; background: var(--color-bg-card); border-radius: 6px; margin-bottom: 8px; }
|
||||||
|
.manage-needs-list li .btn { margin-left: auto; }
|
||||||
|
|
||||||
|
.htmx-indicator { display: none; }
|
||||||
|
.htmx-request .htmx-indicator { display: inline; }
|
||||||
|
.htmx-request.htmx-indicator { display: inline; }
|
||||||
|
|
||||||
|
@media (max-width: 768px) {
|
||||||
|
.id-card-body { grid-template-columns: 1fr; }
|
||||||
|
.id-card-title { font-size: 1.8rem; }
|
||||||
|
.id-card-header { flex-direction: column; gap: 15px; }
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<div class="app">
|
||||||
|
<nav>
|
||||||
|
<a href="/contacts">Contacts</a>
|
||||||
|
<a href="/graph">Graph</a>
|
||||||
|
<a href="/needs">Needs</a>
|
||||||
|
<button class="btn btn-small theme-toggle" onclick="toggleTheme()">
|
||||||
|
<span id="theme-label">Dark</span>
|
||||||
|
</button>
|
||||||
|
</nav>
|
||||||
|
|
||||||
|
<main id="main-content">
|
||||||
|
{% block content %}{% endblock %}
|
||||||
|
</main>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<script>
|
||||||
|
function toggleTheme() {
|
||||||
|
const html = document.documentElement;
|
||||||
|
const current = html.getAttribute('data-theme');
|
||||||
|
const next = current === 'light' ? 'dark' : 'light';
|
||||||
|
html.setAttribute('data-theme', next);
|
||||||
|
localStorage.setItem('theme', next);
|
||||||
|
document.getElementById('theme-label').textContent = next === 'light' ? 'Dark' : 'Light';
|
||||||
|
}
|
||||||
|
(function() {
|
||||||
|
const saved = localStorage.getItem('theme') || 'light';
|
||||||
|
document.documentElement.setAttribute('data-theme', saved);
|
||||||
|
document.getElementById('theme-label').textContent = saved === 'light' ? 'Dark' : 'Light';
|
||||||
|
})();
|
||||||
|
</script>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
204
python/api/templates/contact_detail.html
Normal file
204
python/api/templates/contact_detail.html
Normal file
@@ -0,0 +1,204 @@
|
|||||||
|
{% extends "base.html" %}
|
||||||
|
{% block title %}{{ contact.name }}{% endblock %}
|
||||||
|
{% block content %}
|
||||||
|
<div class="id-card">
|
||||||
|
<div class="id-card-inner">
|
||||||
|
<div class="id-card-header">
|
||||||
|
<div class="id-card-header-left">
|
||||||
|
<h1 class="id-card-title">I.D.: {{ contact.name }}</h1>
|
||||||
|
</div>
|
||||||
|
<div class="id-card-header-right">
|
||||||
|
{% if contact.profile_pic %}
|
||||||
|
<img src="{{ contact.profile_pic }}" alt="{{ contact.name }}'s profile" class="id-profile-pic">
|
||||||
|
{% else %}
|
||||||
|
<div class="id-profile-placeholder">
|
||||||
|
<span>{{ contact.name[0]|upper }}</span>
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
<div class="id-card-actions">
|
||||||
|
<a href="/contacts/{{ contact.id }}/edit" class="btn btn-small">Edit</a>
|
||||||
|
<a href="/contacts" class="btn btn-small">Back</a>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="id-card-body">
|
||||||
|
<div class="id-card-left">
|
||||||
|
{% if contact.legal_name %}
|
||||||
|
<div class="id-field">Legal name: {{ contact.legal_name }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.suffix %}
|
||||||
|
<div class="id-field">Suffix: {{ contact.suffix }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.gender %}
|
||||||
|
<div class="id-field">Gender: {{ contact.gender }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.age %}
|
||||||
|
<div class="id-field">Age: {{ contact.age }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.current_job %}
|
||||||
|
<div class="id-field">Job: {{ contact.current_job }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.social_structure_style %}
|
||||||
|
<div class="id-field">Social style: {{ contact.social_structure_style }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.self_sufficiency_score is not none %}
|
||||||
|
<div class="id-field">Self-Sufficiency: {{ contact.self_sufficiency_score }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.timezone %}
|
||||||
|
<div class="id-field">Timezone: {{ contact.timezone }}</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.safe_conversation_starters %}
|
||||||
|
<div class="id-field-block">
|
||||||
|
<span class="id-label">Safe con starters:</span> {{ contact.safe_conversation_starters }}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.topics_to_avoid %}
|
||||||
|
<div class="id-field-block">
|
||||||
|
<span class="id-label">Topics to avoid:</span> {{ contact.topics_to_avoid }}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
{% if contact.goals %}
|
||||||
|
<div class="id-field-block">
|
||||||
|
<span class="id-label">Goals:</span> {{ contact.goals }}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="id-card-right">
|
||||||
|
{% if contact.bio %}
|
||||||
|
<div class="id-bio">
|
||||||
|
<span class="id-label">Bio:</span> {{ contact.bio }}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
<div class="id-relationships">
|
||||||
|
<h2 class="id-section-title">Relationships</h2>
|
||||||
|
|
||||||
|
{% if grouped_relationships.familial %}
|
||||||
|
<div class="id-rel-group">
|
||||||
|
<span class="id-rel-label">Familial:</span>
|
||||||
|
{% for rel in grouped_relationships.familial %}
|
||||||
|
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a><span class="id-rel-type">({{ rel.relationship_type|replace("_", " ")|title }})</span>{% if not loop.last %}, {% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
{% if grouped_relationships.partners %}
|
||||||
|
<div class="id-rel-group">
|
||||||
|
<span class="id-rel-label">Partners:</span>
|
||||||
|
{% for rel in grouped_relationships.partners %}
|
||||||
|
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a>{% if not loop.last %}, {% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
{% if grouped_relationships.friends %}
|
||||||
|
<div class="id-rel-group">
|
||||||
|
<span class="id-rel-label">Friends:</span>
|
||||||
|
{% for rel in grouped_relationships.friends %}
|
||||||
|
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a>{% if not loop.last %}, {% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
{% if grouped_relationships.professional %}
|
||||||
|
<div class="id-rel-group">
|
||||||
|
<span class="id-rel-label">Professional:</span>
|
||||||
|
{% for rel in grouped_relationships.professional %}
|
||||||
|
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a><span class="id-rel-type">({{ rel.relationship_type|replace("_", " ")|title }})</span>{% if not loop.last %}, {% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
{% if grouped_relationships.other %}
|
||||||
|
<div class="id-rel-group">
|
||||||
|
<span class="id-rel-label">Other:</span>
|
||||||
|
{% for rel in grouped_relationships.other %}
|
||||||
|
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a><span class="id-rel-type">({{ rel.relationship_type|replace("_", " ")|title }})</span>{% if not loop.last %}, {% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
{% if contact.related_from %}
|
||||||
|
<div class="id-rel-group">
|
||||||
|
<span class="id-rel-label">Known by:</span>
|
||||||
|
{% for rel in contact.related_from %}
|
||||||
|
<a href="/contacts/{{ rel.contact_id }}">{{ contact_names[rel.contact_id] }}</a>{% if not loop.last %}, {% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
{% if contact.needs %}
|
||||||
|
<div class="id-card-warnings">
|
||||||
|
{% for need in contact.needs %}
|
||||||
|
<div class="id-warning">
|
||||||
|
<span class="warning-dot"></span>
|
||||||
|
Warning: {{ need.name }}
|
||||||
|
{% if need.description %}<span class="warning-desc"> - {{ need.description }}</span>{% endif %}
|
||||||
|
</div>
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<details class="id-card-manage">
|
||||||
|
<summary>Manage Contact</summary>
|
||||||
|
|
||||||
|
<div class="manage-section">
|
||||||
|
<h3>Manage Relationships</h3>
|
||||||
|
<div id="manage-relationships" class="manage-relationships">
|
||||||
|
{% include "partials/manage_relationships.html" %}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
{% if all_contacts %}
|
||||||
|
<form hx-post="/htmx/contacts/{{ contact.id }}/add-relationship"
|
||||||
|
hx-target="#manage-relationships"
|
||||||
|
hx-swap="innerHTML"
|
||||||
|
class="add-form">
|
||||||
|
<select name="related_contact_id" required>
|
||||||
|
<option value="">Select contact...</option>
|
||||||
|
{% for other in all_contacts %}
|
||||||
|
{% if other.id != contact.id %}
|
||||||
|
<option value="{{ other.id }}">{{ other.name }}</option>
|
||||||
|
{% endif %}
|
||||||
|
{% endfor %}
|
||||||
|
</select>
|
||||||
|
<select name="relationship_type" required>
|
||||||
|
<option value="">Select relationship type...</option>
|
||||||
|
{% for rel_type in relationship_types %}
|
||||||
|
<option value="{{ rel_type.value }}">{{ rel_type.display_name }}</option>
|
||||||
|
{% endfor %}
|
||||||
|
</select>
|
||||||
|
<button type="submit" class="btn btn-primary">Add Relationship</button>
|
||||||
|
</form>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="manage-section">
|
||||||
|
<h3>Manage Needs/Warnings</h3>
|
||||||
|
<div id="manage-needs">
|
||||||
|
{% include "partials/manage_needs.html" %}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
{% if available_needs %}
|
||||||
|
<form hx-post="/htmx/contacts/{{ contact.id }}/add-need"
|
||||||
|
hx-target="#manage-needs"
|
||||||
|
hx-swap="innerHTML"
|
||||||
|
class="add-form">
|
||||||
|
<select name="need_id" required>
|
||||||
|
<option value="">Select a need...</option>
|
||||||
|
{% for need in available_needs %}
|
||||||
|
<option value="{{ need.id }}">{{ need.name }}</option>
|
||||||
|
{% endfor %}
|
||||||
|
</select>
|
||||||
|
<button type="submit" class="btn btn-primary">Add Need</button>
|
||||||
|
</form>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
</details>
|
||||||
|
</div>
|
||||||
|
{% endblock %}
|
||||||
115
python/api/templates/contact_form.html
Normal file
115
python/api/templates/contact_form.html
Normal file
@@ -0,0 +1,115 @@
|
|||||||
|
{% extends "base.html" %}
|
||||||
|
{% block title %}{{ "Edit " + contact.name if contact else "New Contact" }}{% endblock %}
|
||||||
|
{% block content %}
|
||||||
|
<div class="contact-form">
|
||||||
|
<h1>{{ "Edit Contact" if contact else "New Contact" }}</h1>
|
||||||
|
|
||||||
|
{% if contact %}
|
||||||
|
<form method="post" action="/htmx/contacts/{{ contact.id }}/edit">
|
||||||
|
{% else %}
|
||||||
|
<form method="post" action="/htmx/contacts/new">
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="name">Name *</label>
|
||||||
|
<input id="name" name="name" type="text" value="{{ contact.name if contact else '' }}" required>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-row">
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="legal_name">Legal Name</label>
|
||||||
|
<input id="legal_name" name="legal_name" type="text" value="{{ contact.legal_name or '' }}">
|
||||||
|
</div>
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="suffix">Suffix</label>
|
||||||
|
<input id="suffix" name="suffix" type="text" value="{{ contact.suffix or '' }}">
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-row">
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="age">Age</label>
|
||||||
|
<input id="age" name="age" type="number" value="{{ contact.age if contact and contact.age is not none else '' }}">
|
||||||
|
</div>
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="gender">Gender</label>
|
||||||
|
<input id="gender" name="gender" type="text" value="{{ contact.gender or '' }}">
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="current_job">Current Job</label>
|
||||||
|
<input id="current_job" name="current_job" type="text" value="{{ contact.current_job or '' }}">
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="timezone">Timezone</label>
|
||||||
|
<input id="timezone" name="timezone" type="text" value="{{ contact.timezone or '' }}">
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="profile_pic">Profile Picture URL</label>
|
||||||
|
<input id="profile_pic" name="profile_pic" type="url" placeholder="https://example.com/photo.jpg" value="{{ contact.profile_pic or '' }}">
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="bio">Bio</label>
|
||||||
|
<textarea id="bio" name="bio" rows="3">{{ contact.bio or '' }}</textarea>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="goals">Goals</label>
|
||||||
|
<textarea id="goals" name="goals" rows="3">{{ contact.goals or '' }}</textarea>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="social_structure_style">Social Structure Style</label>
|
||||||
|
<input id="social_structure_style" name="social_structure_style" type="text" value="{{ contact.social_structure_style or '' }}">
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="self_sufficiency_score">Self-Sufficiency Score (1-10)</label>
|
||||||
|
<input id="self_sufficiency_score" name="self_sufficiency_score" type="number" min="1" max="10" value="{{ contact.self_sufficiency_score if contact and contact.self_sufficiency_score is not none else '' }}">
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="safe_conversation_starters">Safe Conversation Starters</label>
|
||||||
|
<textarea id="safe_conversation_starters" name="safe_conversation_starters" rows="2">{{ contact.safe_conversation_starters or '' }}</textarea>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="topics_to_avoid">Topics to Avoid</label>
|
||||||
|
<textarea id="topics_to_avoid" name="topics_to_avoid" rows="2">{{ contact.topics_to_avoid or '' }}</textarea>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="ssn">SSN</label>
|
||||||
|
<input id="ssn" name="ssn" type="text" value="{{ contact.ssn or '' }}">
|
||||||
|
</div>
|
||||||
|
|
||||||
|
{% if all_needs %}
|
||||||
|
<div class="form-group">
|
||||||
|
<label>Needs/Accommodations</label>
|
||||||
|
<div class="checkbox-group">
|
||||||
|
{% for need in all_needs %}
|
||||||
|
<label class="checkbox-label">
|
||||||
|
<input type="checkbox" name="need_ids" value="{{ need.id }}"
|
||||||
|
{% if contact and need in contact.needs %}checked{% endif %}>
|
||||||
|
{{ need.name }}
|
||||||
|
</label>
|
||||||
|
{% endfor %}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
|
<div class="form-actions">
|
||||||
|
<button type="submit" class="btn btn-primary">Save</button>
|
||||||
|
{% if contact %}
|
||||||
|
<a href="/contacts/{{ contact.id }}" class="btn">Cancel</a>
|
||||||
|
{% else %}
|
||||||
|
<a href="/contacts" class="btn">Cancel</a>
|
||||||
|
{% endif %}
|
||||||
|
</div>
|
||||||
|
</form>
|
||||||
|
</div>
|
||||||
|
{% endblock %}
|
||||||
14
python/api/templates/contact_list.html
Normal file
14
python/api/templates/contact_list.html
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
{% extends "base.html" %}
|
||||||
|
{% block title %}Contacts{% endblock %}
|
||||||
|
{% block content %}
|
||||||
|
<div class="contact-list">
|
||||||
|
<div class="header">
|
||||||
|
<h1>Contacts</h1>
|
||||||
|
<a href="/contacts/new" class="btn btn-primary">Add Contact</a>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<div id="contact-table">
|
||||||
|
{% include "partials/contact_table.html" %}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
{% endblock %}
|
||||||
198
python/api/templates/graph.html
Normal file
198
python/api/templates/graph.html
Normal file
@@ -0,0 +1,198 @@
|
|||||||
|
{% extends "base.html" %}
|
||||||
|
{% block title %}Relationship Graph{% endblock %}
|
||||||
|
{% block content %}
|
||||||
|
<div class="graph-container">
|
||||||
|
<div class="header">
|
||||||
|
<h1>Relationship Graph</h1>
|
||||||
|
</div>
|
||||||
|
<p class="graph-hint">Drag nodes to reposition. Closer relationships have shorter, darker edges.</p>
|
||||||
|
<canvas id="graph-canvas" width="900" height="600"
|
||||||
|
style="border: 1px solid var(--color-border); border-radius: 8px; background: var(--color-bg); cursor: grab;">
|
||||||
|
</canvas>
|
||||||
|
<div id="selected-info"></div>
|
||||||
|
<div class="legend">
|
||||||
|
<h4>Relationship Closeness (1-10)</h4>
|
||||||
|
<div class="legend-items">
|
||||||
|
<div class="legend-item">
|
||||||
|
<span class="legend-line" style="background: hsl(220, 70%, 40%); height: 4px; display: inline-block;"></span>
|
||||||
|
<span>10 - Very Close (Spouse, Partner)</span>
|
||||||
|
</div>
|
||||||
|
<div class="legend-item">
|
||||||
|
<span class="legend-line" style="background: hsl(220, 70%, 52%); height: 3px; display: inline-block;"></span>
|
||||||
|
<span>7 - Close (Family, Best Friend)</span>
|
||||||
|
</div>
|
||||||
|
<div class="legend-item">
|
||||||
|
<span class="legend-line" style="background: hsl(220, 70%, 64%); height: 2px; display: inline-block;"></span>
|
||||||
|
<span>4 - Moderate (Friend, Colleague)</span>
|
||||||
|
</div>
|
||||||
|
<div class="legend-item">
|
||||||
|
<span class="legend-line" style="background: hsl(220, 70%, 72%); height: 1px; display: inline-block;"></span>
|
||||||
|
<span>2 - Distant (Acquaintance)</span>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<script>
|
||||||
|
(function() {
|
||||||
|
const RELATIONSHIP_DISPLAY = {{ relationship_type_display|tojson }};
|
||||||
|
const graphData = {{ graph_data|tojson }};
|
||||||
|
|
||||||
|
const canvas = document.getElementById('graph-canvas');
|
||||||
|
const ctx = canvas.getContext('2d');
|
||||||
|
const width = canvas.width;
|
||||||
|
const height = canvas.height;
|
||||||
|
const centerX = width / 2;
|
||||||
|
const centerY = height / 2;
|
||||||
|
|
||||||
|
const nodes = graphData.nodes.map(function(node) {
|
||||||
|
return Object.assign({}, node, {
|
||||||
|
x: centerX + (Math.random() - 0.5) * 300,
|
||||||
|
y: centerY + (Math.random() - 0.5) * 300,
|
||||||
|
vx: 0,
|
||||||
|
vy: 0
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
const nodeMap = new Map(nodes.map(function(node) { return [node.id, node]; }));
|
||||||
|
|
||||||
|
const edges = graphData.edges.map(function(edge) {
|
||||||
|
const sourceNode = nodeMap.get(edge.source);
|
||||||
|
const targetNode = nodeMap.get(edge.target);
|
||||||
|
if (!sourceNode || !targetNode) return null;
|
||||||
|
return Object.assign({}, edge, { sourceNode: sourceNode, targetNode: targetNode });
|
||||||
|
}).filter(function(edge) { return edge !== null; });
|
||||||
|
|
||||||
|
let dragNode = null;
|
||||||
|
let selectedNode = null;
|
||||||
|
|
||||||
|
const repulsion = 5000;
|
||||||
|
const springStrength = 0.05;
|
||||||
|
const baseSpringLength = 150;
|
||||||
|
const damping = 0.9;
|
||||||
|
const centerPull = 0.01;
|
||||||
|
|
||||||
|
function simulate() {
|
||||||
|
for (const node of nodes) { node.vx = 0; node.vy = 0; }
|
||||||
|
for (let i = 0; i < nodes.length; i++) {
|
||||||
|
for (let j = i + 1; j < nodes.length; j++) {
|
||||||
|
const dx = nodes[j].x - nodes[i].x;
|
||||||
|
const dy = nodes[j].y - nodes[i].y;
|
||||||
|
const dist = Math.sqrt(dx * dx + dy * dy) || 1;
|
||||||
|
const force = repulsion / (dist * dist);
|
||||||
|
const fx = (dx / dist) * force;
|
||||||
|
const fy = (dy / dist) * force;
|
||||||
|
nodes[i].vx -= fx; nodes[i].vy -= fy;
|
||||||
|
nodes[j].vx += fx; nodes[j].vy += fy;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for (const edge of edges) {
|
||||||
|
const dx = edge.targetNode.x - edge.sourceNode.x;
|
||||||
|
const dy = edge.targetNode.y - edge.sourceNode.y;
|
||||||
|
const dist = Math.sqrt(dx * dx + dy * dy) || 1;
|
||||||
|
const normalizedWeight = edge.closeness_weight / 10;
|
||||||
|
const idealLength = baseSpringLength * (1.5 - normalizedWeight);
|
||||||
|
const displacement = dist - idealLength;
|
||||||
|
const force = springStrength * displacement;
|
||||||
|
const fx = (dx / dist) * force;
|
||||||
|
const fy = (dy / dist) * force;
|
||||||
|
edge.sourceNode.vx += fx; edge.sourceNode.vy += fy;
|
||||||
|
edge.targetNode.vx -= fx; edge.targetNode.vy -= fy;
|
||||||
|
}
|
||||||
|
for (const node of nodes) {
|
||||||
|
node.vx += (centerX - node.x) * centerPull;
|
||||||
|
node.vy += (centerY - node.y) * centerPull;
|
||||||
|
}
|
||||||
|
for (const node of nodes) {
|
||||||
|
if (node === dragNode) continue;
|
||||||
|
node.x += node.vx * damping;
|
||||||
|
node.y += node.vy * damping;
|
||||||
|
node.x = Math.max(30, Math.min(width - 30, node.x));
|
||||||
|
node.y = Math.max(30, Math.min(height - 30, node.y));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function getEdgeColor(weight) {
|
||||||
|
const normalized = weight / 10;
|
||||||
|
return 'hsl(220, 70%, ' + (80 - normalized * 40) + '%)';
|
||||||
|
}
|
||||||
|
|
||||||
|
function draw() {
|
||||||
|
ctx.clearRect(0, 0, width, height);
|
||||||
|
for (const edge of edges) {
|
||||||
|
const lineWidth = 1 + (edge.closeness_weight / 10) * 3;
|
||||||
|
ctx.strokeStyle = getEdgeColor(edge.closeness_weight);
|
||||||
|
ctx.lineWidth = lineWidth;
|
||||||
|
ctx.beginPath();
|
||||||
|
ctx.moveTo(edge.sourceNode.x, edge.sourceNode.y);
|
||||||
|
ctx.lineTo(edge.targetNode.x, edge.targetNode.y);
|
||||||
|
ctx.stroke();
|
||||||
|
const midX = (edge.sourceNode.x + edge.targetNode.x) / 2;
|
||||||
|
const midY = (edge.sourceNode.y + edge.targetNode.y) / 2;
|
||||||
|
ctx.fillStyle = '#666';
|
||||||
|
ctx.font = '10px sans-serif';
|
||||||
|
ctx.textAlign = 'center';
|
||||||
|
const label = RELATIONSHIP_DISPLAY[edge.relationship_type] || edge.relationship_type;
|
||||||
|
ctx.fillText(label, midX, midY - 5);
|
||||||
|
}
|
||||||
|
for (const node of nodes) {
|
||||||
|
const isSelected = node === selectedNode;
|
||||||
|
const radius = isSelected ? 25 : 20;
|
||||||
|
ctx.beginPath();
|
||||||
|
ctx.arc(node.x, node.y, radius, 0, Math.PI * 2);
|
||||||
|
ctx.fillStyle = isSelected ? '#0066cc' : '#fff';
|
||||||
|
ctx.fill();
|
||||||
|
ctx.strokeStyle = '#0066cc';
|
||||||
|
ctx.lineWidth = 2;
|
||||||
|
ctx.stroke();
|
||||||
|
ctx.fillStyle = isSelected ? '#fff' : '#333';
|
||||||
|
ctx.font = '12px sans-serif';
|
||||||
|
ctx.textAlign = 'center';
|
||||||
|
ctx.textBaseline = 'middle';
|
||||||
|
const name = node.name.length > 10 ? node.name.slice(0, 9) + '\u2026' : node.name;
|
||||||
|
ctx.fillText(name, node.x, node.y);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
function animate() {
|
||||||
|
simulate();
|
||||||
|
draw();
|
||||||
|
requestAnimationFrame(animate);
|
||||||
|
}
|
||||||
|
animate();
|
||||||
|
|
||||||
|
function getNodeAt(x, y) {
|
||||||
|
for (const node of nodes) {
|
||||||
|
const dx = x - node.x;
|
||||||
|
const dy = y - node.y;
|
||||||
|
if (dx * dx + dy * dy < 400) return node;
|
||||||
|
}
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
canvas.addEventListener('mousedown', function(event) {
|
||||||
|
const rect = canvas.getBoundingClientRect();
|
||||||
|
const node = getNodeAt(event.clientX - rect.left, event.clientY - rect.top);
|
||||||
|
if (node) {
|
||||||
|
dragNode = node;
|
||||||
|
selectedNode = node;
|
||||||
|
const infoDiv = document.getElementById('selected-info');
|
||||||
|
let html = '<div class="selected-info"><h3>' + node.name + '</h3>';
|
||||||
|
if (node.current_job) html += '<p>Job: ' + node.current_job + '</p>';
|
||||||
|
html += '<a href="/contacts/' + node.id + '">View details</a></div>';
|
||||||
|
infoDiv.innerHTML = html;
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
canvas.addEventListener('mousemove', function(event) {
|
||||||
|
if (!dragNode) return;
|
||||||
|
const rect = canvas.getBoundingClientRect();
|
||||||
|
dragNode.x = event.clientX - rect.left;
|
||||||
|
dragNode.y = event.clientY - rect.top;
|
||||||
|
});
|
||||||
|
|
||||||
|
canvas.addEventListener('mouseup', function() { dragNode = null; });
|
||||||
|
canvas.addEventListener('mouseleave', function() { dragNode = null; });
|
||||||
|
})();
|
||||||
|
</script>
|
||||||
|
{% endblock %}
|
||||||
31
python/api/templates/need_list.html
Normal file
31
python/api/templates/need_list.html
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
{% extends "base.html" %}
|
||||||
|
{% block title %}Needs{% endblock %}
|
||||||
|
{% block content %}
|
||||||
|
<div class="need-list">
|
||||||
|
<div class="header">
|
||||||
|
<h1>Needs / Accommodations</h1>
|
||||||
|
<button class="btn btn-primary" onclick="document.getElementById('need-form').toggleAttribute('hidden')">Add Need</button>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<form id="need-form" hidden
|
||||||
|
hx-post="/htmx/needs"
|
||||||
|
hx-target="#need-items"
|
||||||
|
hx-swap="innerHTML"
|
||||||
|
hx-on::after-request="if(event.detail.successful) this.reset()"
|
||||||
|
class="need-form">
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="name">Name *</label>
|
||||||
|
<input id="name" name="name" type="text" placeholder="e.g., Light Sensitive, ADHD" required>
|
||||||
|
</div>
|
||||||
|
<div class="form-group">
|
||||||
|
<label for="description">Description</label>
|
||||||
|
<textarea id="description" name="description" placeholder="Optional description..." rows="2"></textarea>
|
||||||
|
</div>
|
||||||
|
<button type="submit" class="btn btn-primary">Create</button>
|
||||||
|
</form>
|
||||||
|
|
||||||
|
<div id="need-items">
|
||||||
|
{% include "partials/need_items.html" %}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
{% endblock %}
|
||||||
33
python/api/templates/partials/contact_table.html
Normal file
33
python/api/templates/partials/contact_table.html
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
{% if contacts %}
|
||||||
|
<table>
|
||||||
|
<thead>
|
||||||
|
<tr>
|
||||||
|
<th>Name</th>
|
||||||
|
<th>Job</th>
|
||||||
|
<th>Timezone</th>
|
||||||
|
<th>Actions</th>
|
||||||
|
</tr>
|
||||||
|
</thead>
|
||||||
|
<tbody>
|
||||||
|
{% for contact in contacts %}
|
||||||
|
<tr id="contact-row-{{ contact.id }}">
|
||||||
|
<td><a href="/contacts/{{ contact.id }}">{{ contact.name }}</a></td>
|
||||||
|
<td>{{ contact.current_job or "-" }}</td>
|
||||||
|
<td>{{ contact.timezone or "-" }}</td>
|
||||||
|
<td>
|
||||||
|
<a href="/contacts/{{ contact.id }}/edit" class="btn">Edit</a>
|
||||||
|
<button class="btn btn-danger"
|
||||||
|
hx-delete="/api/contacts/{{ contact.id }}"
|
||||||
|
hx-target="#contact-row-{{ contact.id }}"
|
||||||
|
hx-swap="outerHTML"
|
||||||
|
hx-confirm="Delete this contact?">
|
||||||
|
Delete
|
||||||
|
</button>
|
||||||
|
</td>
|
||||||
|
</tr>
|
||||||
|
{% endfor %}
|
||||||
|
</tbody>
|
||||||
|
</table>
|
||||||
|
{% else %}
|
||||||
|
<p>No contacts yet.</p>
|
||||||
|
{% endif %}
|
||||||
14
python/api/templates/partials/manage_needs.html
Normal file
14
python/api/templates/partials/manage_needs.html
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
<ul class="manage-needs-list">
|
||||||
|
{% for need in contact.needs %}
|
||||||
|
<li id="contact-need-{{ need.id }}">
|
||||||
|
<strong>{{ need.name }}</strong>
|
||||||
|
{% if need.description %}<span> - {{ need.description }}</span>{% endif %}
|
||||||
|
<button class="btn btn-small btn-danger"
|
||||||
|
hx-delete="/api/contacts/{{ contact.id }}/needs/{{ need.id }}"
|
||||||
|
hx-target="#contact-need-{{ need.id }}"
|
||||||
|
hx-swap="outerHTML">
|
||||||
|
Remove
|
||||||
|
</button>
|
||||||
|
</li>
|
||||||
|
{% endfor %}
|
||||||
|
</ul>
|
||||||
23
python/api/templates/partials/manage_relationships.html
Normal file
23
python/api/templates/partials/manage_relationships.html
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
{% for rel in contact.related_to %}
|
||||||
|
<div class="manage-rel-item" id="rel-{{ contact.id }}-{{ rel.related_contact_id }}">
|
||||||
|
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a>
|
||||||
|
<span class="tag">{{ rel.relationship_type|replace("_", " ")|title }}</span>
|
||||||
|
<label class="weight-control">
|
||||||
|
<span>Closeness:</span>
|
||||||
|
<input type="range" min="1" max="10" value="{{ rel.closeness_weight }}"
|
||||||
|
hx-post="/htmx/contacts/{{ contact.id }}/relationships/{{ rel.related_contact_id }}/weight"
|
||||||
|
hx-trigger="change"
|
||||||
|
hx-include="this"
|
||||||
|
name="closeness_weight"
|
||||||
|
hx-swap="none"
|
||||||
|
oninput="this.nextElementSibling.textContent = this.value">
|
||||||
|
<span class="weight-value">{{ rel.closeness_weight }}</span>
|
||||||
|
</label>
|
||||||
|
<button class="btn btn-small btn-danger"
|
||||||
|
hx-delete="/api/contacts/{{ contact.id }}/relationships/{{ rel.related_contact_id }}"
|
||||||
|
hx-target="#rel-{{ contact.id }}-{{ rel.related_contact_id }}"
|
||||||
|
hx-swap="outerHTML">
|
||||||
|
Remove
|
||||||
|
</button>
|
||||||
|
</div>
|
||||||
|
{% endfor %}
|
||||||
21
python/api/templates/partials/need_items.html
Normal file
21
python/api/templates/partials/need_items.html
Normal file
@@ -0,0 +1,21 @@
|
|||||||
|
{% if needs %}
|
||||||
|
<ul class="need-items">
|
||||||
|
{% for need in needs %}
|
||||||
|
<li id="need-item-{{ need.id }}">
|
||||||
|
<div class="need-info">
|
||||||
|
<strong>{{ need.name }}</strong>
|
||||||
|
{% if need.description %}<p>{{ need.description }}</p>{% endif %}
|
||||||
|
</div>
|
||||||
|
<button class="btn btn-danger"
|
||||||
|
hx-delete="/api/needs/{{ need.id }}"
|
||||||
|
hx-target="#need-item-{{ need.id }}"
|
||||||
|
hx-swap="outerHTML"
|
||||||
|
hx-confirm="Delete this need?">
|
||||||
|
Delete
|
||||||
|
</button>
|
||||||
|
</li>
|
||||||
|
{% endfor %}
|
||||||
|
</ul>
|
||||||
|
{% else %}
|
||||||
|
<p>No needs defined yet.</p>
|
||||||
|
{% endif %}
|
||||||
72
python/common.py
Normal file
72
python/common.py
Normal file
@@ -0,0 +1,72 @@
|
|||||||
|
"""common."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import logging
|
||||||
|
import sys
|
||||||
|
from datetime import UTC, datetime
|
||||||
|
from os import getenv
|
||||||
|
from subprocess import PIPE, Popen
|
||||||
|
|
||||||
|
from apprise import Apprise
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def configure_logger(level: str = "INFO") -> None:
|
||||||
|
"""Configure the logger.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
level (str, optional): The logging level. Defaults to "INFO".
|
||||||
|
"""
|
||||||
|
logging.basicConfig(
|
||||||
|
level=level,
|
||||||
|
datefmt="%Y-%m-%dT%H:%M:%S%z",
|
||||||
|
format="%(asctime)s %(levelname)s %(filename)s:%(lineno)d - %(message)s",
|
||||||
|
handlers=[logging.StreamHandler(sys.stdout)],
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def bash_wrapper(command: str) -> tuple[str, int]:
|
||||||
|
"""Execute a bash command and capture the output.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
command (str): The bash command to be executed.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tuple[str, int]: A tuple containing the output of the command (stdout) as a string,
|
||||||
|
the error output (stderr) as a string (optional), and the return code as an integer.
|
||||||
|
"""
|
||||||
|
# This is a acceptable risk
|
||||||
|
process = Popen(command.split(), stdout=PIPE, stderr=PIPE)
|
||||||
|
output, error = process.communicate()
|
||||||
|
if error:
|
||||||
|
logger.error(f"{error=}")
|
||||||
|
return error.decode(), process.returncode
|
||||||
|
|
||||||
|
return output.decode(), process.returncode
|
||||||
|
|
||||||
|
|
||||||
|
def signal_alert(body: str, title: str = "") -> None:
|
||||||
|
"""Send a signal alert.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
body (str): The body of the alert.
|
||||||
|
title (str, optional): The title of the alert. Defaults to "".
|
||||||
|
"""
|
||||||
|
apprise_client = Apprise()
|
||||||
|
|
||||||
|
from_phone = getenv("SIGNAL_ALERT_FROM_PHONE")
|
||||||
|
to_phone = getenv("SIGNAL_ALERT_TO_PHONE")
|
||||||
|
if not from_phone or not to_phone:
|
||||||
|
logger.info("SIGNAL_ALERT_FROM_PHONE or SIGNAL_ALERT_TO_PHONE not set")
|
||||||
|
return
|
||||||
|
|
||||||
|
apprise_client.add(f"signal://localhost:8989/{from_phone}/{to_phone}")
|
||||||
|
|
||||||
|
apprise_client.notify(title=title, body=body)
|
||||||
|
|
||||||
|
|
||||||
|
def utcnow() -> datetime:
|
||||||
|
"""Get the current UTC time."""
|
||||||
|
return datetime.now(tz=UTC)
|
||||||
59
python/database.py
Normal file
59
python/database.py
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
"""database."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import logging
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
from sqlalchemy import inspect
|
||||||
|
from sqlalchemy.exc import NoInspectionAvailable
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
from sqlalchemy.orm import Session
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def safe_insert(orm_objects: Sequence[object], session: Session) -> list[tuple[Exception, object]]:
|
||||||
|
"""Safer insert at allows for partial rollbacks.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
orm_objects (Sequence[object]): Tables to insert.
|
||||||
|
session (Session): Database session.
|
||||||
|
"""
|
||||||
|
if unmapped := [orm_object for orm_object in orm_objects if not _is_mapped_instance(orm_object)]:
|
||||||
|
error = f"safe_insert expects ORM-mapped instances {unmapped}"
|
||||||
|
raise TypeError(error)
|
||||||
|
return _safe_insert(orm_objects, session)
|
||||||
|
|
||||||
|
|
||||||
|
def _safe_insert(objects: Sequence[object], session: Session) -> list[tuple[Exception, object]]:
|
||||||
|
exceptions: list[tuple[Exception, object]] = []
|
||||||
|
try:
|
||||||
|
session.add_all(objects)
|
||||||
|
session.commit()
|
||||||
|
|
||||||
|
except Exception as error:
|
||||||
|
session.rollback()
|
||||||
|
|
||||||
|
objects_len = len(objects)
|
||||||
|
if objects_len == 1:
|
||||||
|
logger.exception(objects)
|
||||||
|
return [(error, objects[0])]
|
||||||
|
|
||||||
|
middle = objects_len // 2
|
||||||
|
exceptions.extend(_safe_insert(objects=objects[:middle], session=session))
|
||||||
|
exceptions.extend(_safe_insert(objects=objects[middle:], session=session))
|
||||||
|
return exceptions
|
||||||
|
|
||||||
|
|
||||||
|
def _is_mapped_instance(obj: object) -> bool:
|
||||||
|
"""Return True if `obj` is a SQLAlchemy ORM-mapped instance."""
|
||||||
|
try:
|
||||||
|
inspect(obj) # raises NoInspectionAvailable if not mapped
|
||||||
|
except NoInspectionAvailable:
|
||||||
|
return False
|
||||||
|
else:
|
||||||
|
return True
|
||||||
122
python/database_cli.py
Normal file
122
python/database_cli.py
Normal file
@@ -0,0 +1,122 @@
|
|||||||
|
"""CLI wrapper around alembic for multi-database support.
|
||||||
|
|
||||||
|
Usage:
|
||||||
|
database <db_name> <command> [args...]
|
||||||
|
|
||||||
|
Examples:
|
||||||
|
database van_inventory upgrade head
|
||||||
|
database van_inventory downgrade head-1
|
||||||
|
database van_inventory revision --autogenerate -m "add meals table"
|
||||||
|
database van_inventory check
|
||||||
|
database richie check
|
||||||
|
database richie upgrade head
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from dataclasses import dataclass
|
||||||
|
from importlib import import_module
|
||||||
|
from typing import TYPE_CHECKING, Annotated
|
||||||
|
|
||||||
|
import typer
|
||||||
|
from alembic.config import CommandLine, Config
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from sqlalchemy.orm import DeclarativeBase
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass(frozen=True)
|
||||||
|
class DatabaseConfig:
|
||||||
|
"""Configuration for a database."""
|
||||||
|
|
||||||
|
env_prefix: str
|
||||||
|
version_location: str
|
||||||
|
base_module: str
|
||||||
|
base_class_name: str
|
||||||
|
models_module: str
|
||||||
|
script_location: str = "python/alembic"
|
||||||
|
file_template: str = "%%(year)d_%%(month).2d_%%(day).2d-%%(slug)s_%%(rev)s"
|
||||||
|
|
||||||
|
def get_base(self) -> type[DeclarativeBase]:
|
||||||
|
"""Import and return the Base class."""
|
||||||
|
module = import_module(self.base_module)
|
||||||
|
return getattr(module, self.base_class_name)
|
||||||
|
|
||||||
|
def import_models(self) -> None:
|
||||||
|
"""Import ORM models so alembic autogenerate can detect them."""
|
||||||
|
import_module(self.models_module)
|
||||||
|
|
||||||
|
def alembic_config(self) -> Config:
|
||||||
|
"""Build an alembic Config for this database."""
|
||||||
|
# Runtime import needed — Config is in TYPE_CHECKING for the return type annotation
|
||||||
|
from alembic.config import Config as AlembicConfig # noqa: PLC0415
|
||||||
|
|
||||||
|
cfg = AlembicConfig()
|
||||||
|
cfg.set_main_option("script_location", self.script_location)
|
||||||
|
cfg.set_main_option("file_template", self.file_template)
|
||||||
|
cfg.set_main_option("prepend_sys_path", ".")
|
||||||
|
cfg.set_main_option("version_path_separator", "os")
|
||||||
|
cfg.set_main_option("version_locations", self.version_location)
|
||||||
|
cfg.set_main_option("revision_environment", "true")
|
||||||
|
cfg.set_section_option("post_write_hooks", "hooks", "dynamic_schema,import_postgresql,ruff")
|
||||||
|
cfg.set_section_option("post_write_hooks", "dynamic_schema.type", "dynamic_schema")
|
||||||
|
cfg.set_section_option("post_write_hooks", "import_postgresql.type", "import_postgresql")
|
||||||
|
cfg.set_section_option("post_write_hooks", "ruff.type", "ruff")
|
||||||
|
cfg.attributes["base"] = self.get_base()
|
||||||
|
cfg.attributes["env_prefix"] = self.env_prefix
|
||||||
|
self.import_models()
|
||||||
|
return cfg
|
||||||
|
|
||||||
|
|
||||||
|
DATABASES: dict[str, DatabaseConfig] = {
|
||||||
|
"richie": DatabaseConfig(
|
||||||
|
env_prefix="RICHIE",
|
||||||
|
version_location="python/alembic/richie/versions",
|
||||||
|
base_module="python.orm.richie.base",
|
||||||
|
base_class_name="RichieBase",
|
||||||
|
models_module="python.orm.richie",
|
||||||
|
),
|
||||||
|
"van_inventory": DatabaseConfig(
|
||||||
|
env_prefix="VAN_INVENTORY",
|
||||||
|
version_location="python/alembic/van_inventory/versions",
|
||||||
|
base_module="python.orm.van_inventory.base",
|
||||||
|
base_class_name="VanInventoryBase",
|
||||||
|
models_module="python.orm.van_inventory.models",
|
||||||
|
),
|
||||||
|
"signal_bot": DatabaseConfig(
|
||||||
|
env_prefix="SIGNALBOT",
|
||||||
|
version_location="python/alembic/signal_bot/versions",
|
||||||
|
base_module="python.orm.signal_bot.base",
|
||||||
|
base_class_name="SignalBotBase",
|
||||||
|
models_module="python.orm.signal_bot.models",
|
||||||
|
),
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
app = typer.Typer(help="Multi-database alembic wrapper.")
|
||||||
|
|
||||||
|
|
||||||
|
@app.command(
|
||||||
|
context_settings={"allow_extra_args": True, "ignore_unknown_options": True},
|
||||||
|
)
|
||||||
|
def main(
|
||||||
|
ctx: typer.Context,
|
||||||
|
db_name: Annotated[str, typer.Argument(help=f"Database name. Options: {', '.join(DATABASES)}")],
|
||||||
|
command: Annotated[str, typer.Argument(help="Alembic command (upgrade, downgrade, revision, check, etc.)")],
|
||||||
|
) -> None:
|
||||||
|
"""Run an alembic command against the specified database."""
|
||||||
|
db_config = DATABASES.get(db_name)
|
||||||
|
if not db_config:
|
||||||
|
typer.echo(f"Unknown database: {db_name!r}. Available: {', '.join(DATABASES)}", err=True)
|
||||||
|
raise typer.Exit(code=1)
|
||||||
|
|
||||||
|
alembic_cfg = db_config.alembic_config()
|
||||||
|
|
||||||
|
cmd_line = CommandLine()
|
||||||
|
options = cmd_line.parser.parse_args([command, *ctx.args])
|
||||||
|
cmd_line.run_cmd(alembic_cfg, options)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
app()
|
||||||
|
|
||||||
1
python/eval_warnings/__init__.py
Normal file
1
python/eval_warnings/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Detect Nix evaluation warnings from build logs and create PRs with LLM-suggested fixes."""
|
||||||
449
python/eval_warnings/main.py
Normal file
449
python/eval_warnings/main.py
Normal file
@@ -0,0 +1,449 @@
|
|||||||
|
"""Detect Nix evaluation warnings and create PRs with LLM-suggested fixes."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import hashlib
|
||||||
|
import logging
|
||||||
|
import re
|
||||||
|
import subprocess
|
||||||
|
from dataclasses import dataclass
|
||||||
|
from io import BytesIO
|
||||||
|
from pathlib import Path
|
||||||
|
from typing import Annotated
|
||||||
|
from zipfile import ZipFile
|
||||||
|
|
||||||
|
import typer
|
||||||
|
from httpx import HTTPError, post
|
||||||
|
|
||||||
|
from python.common import configure_logger
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass(frozen=True)
|
||||||
|
class EvalWarning:
|
||||||
|
"""A single Nix evaluation warning."""
|
||||||
|
|
||||||
|
system: str
|
||||||
|
message: str
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class FileChange:
|
||||||
|
"""A file change suggested by the LLM."""
|
||||||
|
|
||||||
|
file_path: str
|
||||||
|
original: str
|
||||||
|
fixed: str
|
||||||
|
|
||||||
|
|
||||||
|
def run_cmd(cmd: list[str], *, check: bool = True) -> subprocess.CompletedProcess[str]:
|
||||||
|
"""Run a subprocess command and return the result.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
cmd: Command and arguments.
|
||||||
|
check: Whether to raise on non-zero exit.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
CompletedProcess with captured stdout/stderr.
|
||||||
|
"""
|
||||||
|
logger.debug("Running: %s", " ".join(cmd))
|
||||||
|
return subprocess.run(cmd, capture_output=True, text=True, check=check)
|
||||||
|
|
||||||
|
|
||||||
|
def download_logs(run_id: str, repo: str) -> dict[str, str]:
|
||||||
|
"""Download build logs for a GitHub Actions run.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
run_id: The workflow run ID.
|
||||||
|
repo: The GitHub repository (owner/repo).
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dict mapping zip entry names to their text content, filtered to build log files.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
RuntimeError: If log download fails.
|
||||||
|
"""
|
||||||
|
result = subprocess.run(
|
||||||
|
["gh", "api", f"repos/{repo}/actions/runs/{run_id}/logs"],
|
||||||
|
capture_output=True,
|
||||||
|
check=False,
|
||||||
|
)
|
||||||
|
if result.returncode != 0:
|
||||||
|
msg = f"Failed to download logs: {result.stderr.decode(errors='replace')}"
|
||||||
|
raise RuntimeError(msg)
|
||||||
|
|
||||||
|
logs: dict[str, str] = {}
|
||||||
|
with ZipFile(BytesIO(result.stdout)) as zip_file:
|
||||||
|
for name in zip_file.namelist():
|
||||||
|
if name.startswith("build-") and name.endswith(".txt"):
|
||||||
|
logs[name] = zip_file.read(name).decode(errors="replace")
|
||||||
|
|
||||||
|
return logs
|
||||||
|
|
||||||
|
|
||||||
|
def parse_warnings(logs: dict[str, str]) -> set[EvalWarning]:
|
||||||
|
"""Parse Nix evaluation warnings from build log contents.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
logs: Dict mapping zip entry names (e.g. "build-bob/2_Build.txt") to their text.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Deduplicated set of warnings.
|
||||||
|
"""
|
||||||
|
warnings: set[EvalWarning] = set()
|
||||||
|
warning_pattern = re.compile(r"(?:^[\d\-T:.Z]+ )?(warning:|trace: warning:)")
|
||||||
|
timestamp_prefix = re.compile(r"^[\d\-T:.Z]+ ")
|
||||||
|
|
||||||
|
for name, content in sorted(logs.items()):
|
||||||
|
system = name.split("/")[0].removeprefix("build-")
|
||||||
|
for line in content.splitlines():
|
||||||
|
if warning_pattern.search(line):
|
||||||
|
message = timestamp_prefix.sub("", line).strip()
|
||||||
|
if message.startswith("warning: ignoring untrusted flake configuration setting"):
|
||||||
|
continue
|
||||||
|
logger.debug(f"Found warning: {line}")
|
||||||
|
warnings.add(EvalWarning(system=system, message=message))
|
||||||
|
|
||||||
|
logger.info("Found %d unique warnings", len(warnings))
|
||||||
|
return warnings
|
||||||
|
|
||||||
|
|
||||||
|
def extract_referenced_files(warnings: set[EvalWarning]) -> dict[str, str]:
|
||||||
|
"""Extract file paths referenced in warnings and read their contents.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
warnings: List of parsed warnings.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Dict mapping repo-relative file paths to their contents.
|
||||||
|
"""
|
||||||
|
paths: set[str] = set()
|
||||||
|
warning_text = "\n".join(w.message for w in warnings)
|
||||||
|
|
||||||
|
nix_store_path = re.compile(r"/nix/store/[^/]+-source/([^:]+\.nix)")
|
||||||
|
for match in nix_store_path.finditer(warning_text):
|
||||||
|
paths.add(match.group(1))
|
||||||
|
|
||||||
|
repo_relative_path = re.compile(r"(?<![/\w])(systems|common|users|overlays)/[^:\s]+\.nix")
|
||||||
|
for match in repo_relative_path.finditer(warning_text):
|
||||||
|
paths.add(match.group(0))
|
||||||
|
|
||||||
|
files: dict[str, str] = {}
|
||||||
|
for path_str in sorted(paths):
|
||||||
|
path = Path(path_str)
|
||||||
|
if path.is_file():
|
||||||
|
files[path_str] = path.read_text()
|
||||||
|
|
||||||
|
if not files and Path("flake.nix").is_file():
|
||||||
|
files["flake.nix"] = Path("flake.nix").read_text()
|
||||||
|
|
||||||
|
logger.info("Extracted %d referenced files", len(files))
|
||||||
|
return files
|
||||||
|
|
||||||
|
|
||||||
|
def compute_warning_hash(warnings: set[EvalWarning]) -> str:
|
||||||
|
"""Compute a short hash of the warning set for deduplication.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
warnings: List of warnings.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
8-character hex hash.
|
||||||
|
"""
|
||||||
|
text = "\n".join(sorted(f"[{w.system}] {w.message}" for w in warnings))
|
||||||
|
return hashlib.sha256(text.encode()).hexdigest()[:8]
|
||||||
|
|
||||||
|
|
||||||
|
def check_duplicate_pr(warning_hash: str) -> bool:
|
||||||
|
"""Check if an open PR already exists for this warning hash.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
warning_hash: The hash to check.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
True if a duplicate PR exists.
|
||||||
|
|
||||||
|
Raises:
|
||||||
|
RuntimeError: If the gh CLI call fails.
|
||||||
|
"""
|
||||||
|
result = run_cmd(
|
||||||
|
[
|
||||||
|
"gh",
|
||||||
|
"pr",
|
||||||
|
"list",
|
||||||
|
"--state",
|
||||||
|
"open",
|
||||||
|
"--label",
|
||||||
|
"eval-warning-fix",
|
||||||
|
"--json",
|
||||||
|
"title",
|
||||||
|
"--jq",
|
||||||
|
".[].title",
|
||||||
|
],
|
||||||
|
check=False,
|
||||||
|
)
|
||||||
|
if result.returncode != 0:
|
||||||
|
msg = f"Failed to check for duplicate PRs: {result.stderr}"
|
||||||
|
raise RuntimeError(msg)
|
||||||
|
|
||||||
|
for title in result.stdout.splitlines():
|
||||||
|
if warning_hash in title:
|
||||||
|
logger.info("Duplicate PR found for hash %s", warning_hash)
|
||||||
|
return True
|
||||||
|
return False
|
||||||
|
|
||||||
|
|
||||||
|
def query_ollama(
|
||||||
|
warnings: set[EvalWarning],
|
||||||
|
files: dict[str, str],
|
||||||
|
ollama_url: str,
|
||||||
|
) -> str | None:
|
||||||
|
"""Query Ollama for a fix suggestion.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
warnings: List of warnings.
|
||||||
|
files: Referenced file contents.
|
||||||
|
ollama_url: Ollama API base URL.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
LLM response text, or None on failure.
|
||||||
|
"""
|
||||||
|
warning_text = "\n".join(f"[{w.system}] {w.message}" for w in warnings)
|
||||||
|
file_context = "\n".join(f"--- FILE: {path} ---\n{content}\n--- END FILE ---" for path, content in files.items())
|
||||||
|
|
||||||
|
prompt = f"""You are a NixOS configuration expert. \
|
||||||
|
Analyze the following Nix evaluation warnings and suggest fixes.
|
||||||
|
|
||||||
|
## Warnings
|
||||||
|
{warning_text}
|
||||||
|
|
||||||
|
## Referenced Files
|
||||||
|
{file_context}
|
||||||
|
|
||||||
|
## Instructions
|
||||||
|
- Identify the root cause of each warning
|
||||||
|
- Provide the exact file changes needed to fix the warnings
|
||||||
|
- Output your response in two clearly separated sections:
|
||||||
|
1. **REASONING**: Brief explanation of what causes each warning and how to fix it
|
||||||
|
2. **CHANGES**: For each file that needs changes, output a block like:
|
||||||
|
FILE: path/to/file.nix
|
||||||
|
<<<<<<< ORIGINAL
|
||||||
|
the original lines to replace
|
||||||
|
=======
|
||||||
|
the replacement lines
|
||||||
|
>>>>>>> FIXED
|
||||||
|
- Only suggest changes for files that exist in the repository
|
||||||
|
- Do not add unnecessary complexity
|
||||||
|
- Preserve the existing code style
|
||||||
|
- If a warning comes from upstream nixpkgs and cannot be fixed in this repo, \
|
||||||
|
say so in REASONING and do not suggest changes"""
|
||||||
|
|
||||||
|
try:
|
||||||
|
response = post(
|
||||||
|
f"{ollama_url}/api/generate",
|
||||||
|
json={
|
||||||
|
"model": "qwen3-coder:30b",
|
||||||
|
"prompt": prompt,
|
||||||
|
"stream": False,
|
||||||
|
"options": {"num_predict": 4096},
|
||||||
|
},
|
||||||
|
timeout=300,
|
||||||
|
)
|
||||||
|
response.raise_for_status()
|
||||||
|
except HTTPError:
|
||||||
|
logger.exception("Ollama request failed")
|
||||||
|
return None
|
||||||
|
|
||||||
|
return response.json().get("response")
|
||||||
|
|
||||||
|
|
||||||
|
def parse_changes(response: str) -> list[FileChange]:
|
||||||
|
"""Parse file changes from the **CHANGES** section of the LLM response.
|
||||||
|
|
||||||
|
Expects blocks in the format:
|
||||||
|
FILE: path/to/file.nix
|
||||||
|
<<<<<<< ORIGINAL
|
||||||
|
...
|
||||||
|
=======
|
||||||
|
...
|
||||||
|
>>>>>>> FIXED
|
||||||
|
|
||||||
|
Args:
|
||||||
|
response: Raw LLM response text.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
List of parsed file changes.
|
||||||
|
"""
|
||||||
|
if "**CHANGES**" not in response:
|
||||||
|
logger.warning("LLM response missing **CHANGES** section")
|
||||||
|
return []
|
||||||
|
|
||||||
|
changes_section = response.split("**CHANGES**", 1)[1]
|
||||||
|
|
||||||
|
changes: list[FileChange] = []
|
||||||
|
current_file = ""
|
||||||
|
section: str | None = None
|
||||||
|
original_lines: list[str] = []
|
||||||
|
fixed_lines: list[str] = []
|
||||||
|
|
||||||
|
for line in changes_section.splitlines():
|
||||||
|
stripped = line.strip()
|
||||||
|
if stripped.startswith("FILE:"):
|
||||||
|
current_file = stripped.removeprefix("FILE:").strip()
|
||||||
|
elif stripped == "<<<<<<< ORIGINAL":
|
||||||
|
section = "original"
|
||||||
|
original_lines = []
|
||||||
|
elif stripped == "=======" and section == "original":
|
||||||
|
section = "fixed"
|
||||||
|
fixed_lines = []
|
||||||
|
elif stripped == ">>>>>>> FIXED" and section == "fixed":
|
||||||
|
section = None
|
||||||
|
if current_file:
|
||||||
|
changes.append(FileChange(current_file, "\n".join(original_lines), "\n".join(fixed_lines)))
|
||||||
|
elif section == "original":
|
||||||
|
original_lines.append(line)
|
||||||
|
elif section == "fixed":
|
||||||
|
fixed_lines.append(line)
|
||||||
|
|
||||||
|
logger.info("Parsed %d file changes", len(changes))
|
||||||
|
return changes
|
||||||
|
|
||||||
|
|
||||||
|
def apply_changes(changes: list[FileChange]) -> int:
|
||||||
|
"""Apply file changes to the working directory.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
changes: List of changes to apply.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Number of changes successfully applied.
|
||||||
|
"""
|
||||||
|
applied = 0
|
||||||
|
cwd = Path.cwd().resolve()
|
||||||
|
for change in changes:
|
||||||
|
path = Path(change.file_path).resolve()
|
||||||
|
if not path.is_relative_to(cwd):
|
||||||
|
logger.warning("Path traversal blocked: %s", change.file_path)
|
||||||
|
continue
|
||||||
|
if not path.is_file():
|
||||||
|
logger.warning("File not found: %s", change.file_path)
|
||||||
|
continue
|
||||||
|
|
||||||
|
content = path.read_text()
|
||||||
|
if change.original not in content:
|
||||||
|
logger.warning("Original text not found in %s", change.file_path)
|
||||||
|
continue
|
||||||
|
|
||||||
|
path.write_text(content.replace(change.original, change.fixed, 1))
|
||||||
|
logger.info("Applied fix to %s", change.file_path)
|
||||||
|
applied += 1
|
||||||
|
|
||||||
|
return applied
|
||||||
|
|
||||||
|
|
||||||
|
def create_pr(
|
||||||
|
warning_hash: str,
|
||||||
|
warnings: set[EvalWarning],
|
||||||
|
llm_response: str,
|
||||||
|
run_url: str,
|
||||||
|
) -> None:
|
||||||
|
"""Create a git branch and PR with the applied fixes.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
warning_hash: Short hash for branch naming and deduplication.
|
||||||
|
warnings: Original warnings for the PR body.
|
||||||
|
llm_response: Full LLM response for extracting reasoning.
|
||||||
|
run_url: URL to the triggering build run.
|
||||||
|
"""
|
||||||
|
branch = f"fix/eval-warning-{warning_hash}"
|
||||||
|
warning_text = "\n".join(f"[{w.system}] {w.message}" for w in warnings)
|
||||||
|
|
||||||
|
if "**REASONING**" not in llm_response:
|
||||||
|
logger.warning("LLM response missing **REASONING** section")
|
||||||
|
reasoning = ""
|
||||||
|
else:
|
||||||
|
_, after = llm_response.split("**REASONING**", 1)
|
||||||
|
reasoning = "\n".join(after.split("**CHANGES**", 1)[0].strip().splitlines()[:50])
|
||||||
|
|
||||||
|
run_cmd(["git", "config", "user.name", "github-actions[bot]"])
|
||||||
|
run_cmd(["git", "config", "user.email", "github-actions[bot]@users.noreply.github.com"])
|
||||||
|
run_cmd(["git", "checkout", "-b", branch])
|
||||||
|
run_cmd(["git", "add", "-A"])
|
||||||
|
|
||||||
|
diff_result = run_cmd(["git", "diff", "--cached", "--quiet"], check=False)
|
||||||
|
if diff_result.returncode == 0:
|
||||||
|
logger.info("No file changes to commit")
|
||||||
|
return
|
||||||
|
|
||||||
|
run_cmd(["git", "commit", "-m", f"fix: resolve nix evaluation warnings ({warning_hash})"])
|
||||||
|
run_cmd(["git", "push", "origin", branch, "--force"])
|
||||||
|
|
||||||
|
body = f"""## Nix Evaluation Warnings
|
||||||
|
|
||||||
|
Detected in [build_systems run]({run_url}):
|
||||||
|
|
||||||
|
```
|
||||||
|
{warning_text}
|
||||||
|
```
|
||||||
|
|
||||||
|
## LLM Analysis (qwen3-coder:30b)
|
||||||
|
|
||||||
|
{reasoning}
|
||||||
|
|
||||||
|
---
|
||||||
|
*Auto-generated by fix_eval_warnings. Review carefully before merging.*"""
|
||||||
|
|
||||||
|
run_cmd(
|
||||||
|
[
|
||||||
|
"gh",
|
||||||
|
"pr",
|
||||||
|
"create",
|
||||||
|
"--title",
|
||||||
|
f"fix: resolve nix eval warnings ({warning_hash})",
|
||||||
|
"--label",
|
||||||
|
"automated",
|
||||||
|
"--label",
|
||||||
|
"eval-warning-fix",
|
||||||
|
"--body",
|
||||||
|
body,
|
||||||
|
]
|
||||||
|
)
|
||||||
|
logger.info("PR created on branch %s", branch)
|
||||||
|
|
||||||
|
|
||||||
|
def main(
|
||||||
|
run_id: Annotated[str, typer.Option("--run-id", help="GitHub Actions run ID")],
|
||||||
|
repo: Annotated[str, typer.Option("--repo", help="GitHub repository (owner/repo)")],
|
||||||
|
ollama_url: Annotated[str, typer.Option("--ollama-url", help="Ollama API base URL")],
|
||||||
|
run_url: Annotated[str, typer.Option("--run-url", help="URL to the triggering build run")],
|
||||||
|
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
||||||
|
) -> None:
|
||||||
|
"""Detect Nix evaluation warnings and create PRs with LLM-suggested fixes."""
|
||||||
|
configure_logger(log_level)
|
||||||
|
|
||||||
|
logs = download_logs(run_id, repo)
|
||||||
|
warnings = parse_warnings(logs)
|
||||||
|
if not warnings:
|
||||||
|
return
|
||||||
|
|
||||||
|
warning_hash = compute_warning_hash(warnings)
|
||||||
|
if check_duplicate_pr(warning_hash):
|
||||||
|
return
|
||||||
|
|
||||||
|
files = extract_referenced_files(warnings)
|
||||||
|
llm_response = query_ollama(warnings, files, ollama_url)
|
||||||
|
if not llm_response:
|
||||||
|
return
|
||||||
|
|
||||||
|
changes = parse_changes(llm_response)
|
||||||
|
applied = apply_changes(changes)
|
||||||
|
if applied == 0:
|
||||||
|
logger.info("No changes could be applied")
|
||||||
|
return
|
||||||
|
|
||||||
|
create_pr(warning_hash, warnings, llm_response, run_url)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
typer.run(main)
|
||||||
1
python/heater/__init__.py
Normal file
1
python/heater/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Tuya heater control service."""
|
||||||
69
python/heater/controller.py
Normal file
69
python/heater/controller.py
Normal file
@@ -0,0 +1,69 @@
|
|||||||
|
"""TinyTuya device controller for heater."""
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
import tinytuya
|
||||||
|
|
||||||
|
from python.heater.models import ActionResult, DeviceConfig, HeaterStatus
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
# DPS mapping for heater
|
||||||
|
DPS_POWER = "1" # bool: on/off
|
||||||
|
DPS_SETPOINT = "101" # int: target temp (read-only)
|
||||||
|
DPS_STATE = "102" # str: "Stop", "Heat", etc.
|
||||||
|
DPS_UNKNOWN = "104" # int: unknown
|
||||||
|
DPS_ERROR = "108" # int: last error code
|
||||||
|
|
||||||
|
|
||||||
|
class HeaterController:
|
||||||
|
"""Controls a Tuya heater device via local network."""
|
||||||
|
|
||||||
|
def __init__(self, config: DeviceConfig) -> None:
|
||||||
|
"""Initialize the controller."""
|
||||||
|
self.device = tinytuya.Device(config.device_id, config.ip, config.local_key)
|
||||||
|
self.device.set_version(config.version)
|
||||||
|
self.device.set_socketTimeout(0.5)
|
||||||
|
self.device.set_socketRetryLimit(1)
|
||||||
|
|
||||||
|
def status(self) -> HeaterStatus:
|
||||||
|
"""Get current heater status."""
|
||||||
|
data = self.device.status()
|
||||||
|
|
||||||
|
if "Error" in data:
|
||||||
|
logger.error("Device error: %s", data)
|
||||||
|
return HeaterStatus(power=False, raw_dps={"error": data["Error"]})
|
||||||
|
|
||||||
|
dps = data.get("dps", {})
|
||||||
|
return HeaterStatus(
|
||||||
|
power=bool(dps.get(DPS_POWER, False)),
|
||||||
|
setpoint=dps.get(DPS_SETPOINT),
|
||||||
|
state=dps.get(DPS_STATE),
|
||||||
|
error_code=dps.get(DPS_ERROR),
|
||||||
|
raw_dps=dps,
|
||||||
|
)
|
||||||
|
|
||||||
|
def turn_on(self) -> ActionResult:
|
||||||
|
"""Turn heater on."""
|
||||||
|
try:
|
||||||
|
self.device.set_value(index=DPS_POWER, value=True)
|
||||||
|
return ActionResult(success=True, action="on", power=True)
|
||||||
|
except Exception as error:
|
||||||
|
logger.exception("Failed to turn on")
|
||||||
|
return ActionResult(success=False, action="on", error=str(error))
|
||||||
|
|
||||||
|
def turn_off(self) -> ActionResult:
|
||||||
|
"""Turn heater off."""
|
||||||
|
try:
|
||||||
|
self.device.set_value(index=DPS_POWER, value=False)
|
||||||
|
return ActionResult(success=True, action="off", power=False)
|
||||||
|
except Exception as error:
|
||||||
|
logger.exception("Failed to turn off")
|
||||||
|
return ActionResult(success=False, action="off", error=str(error))
|
||||||
|
|
||||||
|
def toggle(self) -> ActionResult:
|
||||||
|
"""Toggle heater power state."""
|
||||||
|
status = self.status()
|
||||||
|
if status.power:
|
||||||
|
return self.turn_off()
|
||||||
|
return self.turn_on()
|
||||||
85
python/heater/main.py
Normal file
85
python/heater/main.py
Normal file
@@ -0,0 +1,85 @@
|
|||||||
|
"""FastAPI heater control service."""
|
||||||
|
|
||||||
|
import logging
|
||||||
|
from collections.abc import AsyncIterator
|
||||||
|
from contextlib import asynccontextmanager
|
||||||
|
from typing import Annotated
|
||||||
|
|
||||||
|
import typer
|
||||||
|
import uvicorn
|
||||||
|
from fastapi import FastAPI, HTTPException
|
||||||
|
|
||||||
|
from python.common import configure_logger
|
||||||
|
from python.heater.controller import HeaterController
|
||||||
|
from python.heater.models import ActionResult, DeviceConfig, HeaterStatus
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def create_app(config: DeviceConfig) -> FastAPI:
|
||||||
|
"""Create FastAPI application."""
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def lifespan(app: FastAPI) -> AsyncIterator[None]:
|
||||||
|
app.state.controller = HeaterController(config)
|
||||||
|
yield
|
||||||
|
|
||||||
|
app = FastAPI(
|
||||||
|
title="Heater Control API",
|
||||||
|
description="Fast local control for Tuya heater",
|
||||||
|
lifespan=lifespan,
|
||||||
|
)
|
||||||
|
|
||||||
|
@app.get("/status")
|
||||||
|
def get_status() -> HeaterStatus:
|
||||||
|
return app.state.controller.status()
|
||||||
|
|
||||||
|
@app.post("/on")
|
||||||
|
def heater_on() -> ActionResult:
|
||||||
|
result = app.state.controller.turn_on()
|
||||||
|
if not result.success:
|
||||||
|
raise HTTPException(status_code=500, detail=result.error)
|
||||||
|
return result
|
||||||
|
|
||||||
|
@app.post("/off")
|
||||||
|
def heater_off() -> ActionResult:
|
||||||
|
result = app.state.controller.turn_off()
|
||||||
|
if not result.success:
|
||||||
|
raise HTTPException(status_code=500, detail=result.error)
|
||||||
|
return result
|
||||||
|
|
||||||
|
@app.post("/toggle")
|
||||||
|
def heater_toggle() -> ActionResult:
|
||||||
|
result = app.state.controller.toggle()
|
||||||
|
if not result.success:
|
||||||
|
raise HTTPException(status_code=500, detail=result.error)
|
||||||
|
return result
|
||||||
|
|
||||||
|
return app
|
||||||
|
|
||||||
|
|
||||||
|
def serve(
|
||||||
|
host: Annotated[str, typer.Option("--host", "-h", help="Host to bind to")],
|
||||||
|
port: Annotated[int, typer.Option("--port", "-p", help="Port to bind to")] = 8124,
|
||||||
|
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
||||||
|
device_id: Annotated[str | None, typer.Option("--device-id", envvar="TUYA_DEVICE_ID")] = None,
|
||||||
|
device_ip: Annotated[str | None, typer.Option("--device-ip", envvar="TUYA_DEVICE_IP")] = None,
|
||||||
|
local_key: Annotated[str | None, typer.Option("--local-key", envvar="TUYA_LOCAL_KEY")] = None,
|
||||||
|
) -> None:
|
||||||
|
"""Start the heater control API server."""
|
||||||
|
configure_logger(log_level)
|
||||||
|
|
||||||
|
logger.info("Starting heater control API server")
|
||||||
|
|
||||||
|
if not device_id or not device_ip or not local_key:
|
||||||
|
error = "Must provide device ID, IP, and local key"
|
||||||
|
raise typer.Exit(error)
|
||||||
|
|
||||||
|
config = DeviceConfig(device_id=device_id, ip=device_ip, local_key=local_key)
|
||||||
|
|
||||||
|
app = create_app(config)
|
||||||
|
uvicorn.run(app, host=host, port=port)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
typer.run(serve)
|
||||||
31
python/heater/models.py
Normal file
31
python/heater/models.py
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
"""Pydantic models for heater API."""
|
||||||
|
|
||||||
|
from pydantic import BaseModel, Field
|
||||||
|
|
||||||
|
|
||||||
|
class DeviceConfig(BaseModel):
|
||||||
|
"""Tuya device configuration."""
|
||||||
|
|
||||||
|
device_id: str
|
||||||
|
ip: str
|
||||||
|
local_key: str
|
||||||
|
version: float = 3.5
|
||||||
|
|
||||||
|
|
||||||
|
class HeaterStatus(BaseModel):
|
||||||
|
"""Current heater status."""
|
||||||
|
|
||||||
|
power: bool
|
||||||
|
setpoint: int | None = None
|
||||||
|
state: str | None = None # "Stop", "Heat", etc.
|
||||||
|
error_code: int | None = None
|
||||||
|
raw_dps: dict[str, object] = Field(default_factory=dict)
|
||||||
|
|
||||||
|
|
||||||
|
class ActionResult(BaseModel):
|
||||||
|
"""Result of a heater action."""
|
||||||
|
|
||||||
|
success: bool
|
||||||
|
action: str
|
||||||
|
power: bool | None = None
|
||||||
|
error: str | None = None
|
||||||
1
python/installer/__init__.py
Normal file
1
python/installer/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""installer."""
|
||||||
308
python/installer/__main__.py
Normal file
308
python/installer/__main__.py
Normal file
@@ -0,0 +1,308 @@
|
|||||||
|
"""Install NixOS on a ZFS pool."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import curses
|
||||||
|
import logging
|
||||||
|
import sys
|
||||||
|
from os import getenv
|
||||||
|
from pathlib import Path
|
||||||
|
from random import getrandbits
|
||||||
|
from subprocess import PIPE, Popen, run
|
||||||
|
from time import sleep
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
from python.common import configure_logger
|
||||||
|
from python.installer.tui import draw_menu
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def bash_wrapper(command: str) -> str:
|
||||||
|
"""Execute a bash command and capture the output.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
command (str): The bash command to be executed.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tuple[str, int]: A tuple containing the output of the command (stdout) as a string,
|
||||||
|
the error output (stderr) as a string (optional), and the return code as an integer.
|
||||||
|
"""
|
||||||
|
logger.debug(f"running {command=}")
|
||||||
|
# This is a acceptable risk
|
||||||
|
process = Popen(command.split(), stdout=PIPE, stderr=PIPE)
|
||||||
|
output, _ = process.communicate()
|
||||||
|
if process.returncode != 0:
|
||||||
|
error = f"Failed to run command {command=} return code {process.returncode=}"
|
||||||
|
raise RuntimeError(error)
|
||||||
|
|
||||||
|
return output.decode()
|
||||||
|
|
||||||
|
|
||||||
|
def partition_disk(disk: str, swap_size: int, reserve: int = 0) -> None:
|
||||||
|
"""Partition a disk.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
disk (str): The disk to partition.
|
||||||
|
swap_size (int): The size of the swap partition in GB.
|
||||||
|
minimum value is 1.
|
||||||
|
reserve (int, optional): The size of the reserve partition in GB. Defaults to 0.
|
||||||
|
minimum value is 0.
|
||||||
|
"""
|
||||||
|
logger.info(f"partitioning {disk=}")
|
||||||
|
swap_size = max(swap_size, 1)
|
||||||
|
reserve = max(reserve, 0)
|
||||||
|
|
||||||
|
bash_wrapper(f"blkdiscard -f {disk}")
|
||||||
|
|
||||||
|
if reserve > 0:
|
||||||
|
msg = f"Creating swap partition on {disk=} with size {swap_size=}GiB and reserve {reserve=}GiB"
|
||||||
|
logger.info(msg)
|
||||||
|
|
||||||
|
swap_start = swap_size + reserve
|
||||||
|
swap_partition = f"mkpart swap -{swap_start}GiB -{reserve}GiB "
|
||||||
|
else:
|
||||||
|
logger.info(f"Creating swap partition on {disk=} with size {swap_size=}GiB")
|
||||||
|
swap_start = swap_size
|
||||||
|
swap_partition = f"mkpart swap -{swap_start}GiB 100% "
|
||||||
|
|
||||||
|
logger.debug(f"{swap_partition=}")
|
||||||
|
|
||||||
|
create_partitions = (
|
||||||
|
f"parted --script --align=optimal {disk} -- "
|
||||||
|
"mklabel gpt "
|
||||||
|
"mkpart EFI 1MiB 4GiB "
|
||||||
|
f"mkpart root_pool 4GiB -{swap_start}GiB "
|
||||||
|
f"{swap_partition}"
|
||||||
|
"set 1 esp on"
|
||||||
|
)
|
||||||
|
bash_wrapper(create_partitions)
|
||||||
|
|
||||||
|
logger.info(f"{disk=} successfully partitioned")
|
||||||
|
|
||||||
|
|
||||||
|
def create_zfs_pool(pool_disks: Sequence[str], mnt_dir: str) -> None:
|
||||||
|
"""Create a ZFS pool.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
pool_disks (Sequence[str]): A tuple of disks to use for the pool.
|
||||||
|
mnt_dir (str): The mount directory.
|
||||||
|
"""
|
||||||
|
if len(pool_disks) <= 0:
|
||||||
|
error = "disks must be a tuple of at least length 1"
|
||||||
|
raise ValueError(error)
|
||||||
|
|
||||||
|
zpool_create = (
|
||||||
|
"zpool create "
|
||||||
|
"-o ashift=12 "
|
||||||
|
"-o autotrim=on "
|
||||||
|
f"-R {mnt_dir} "
|
||||||
|
"-O acltype=posixacl "
|
||||||
|
"-O canmount=off "
|
||||||
|
"-O dnodesize=auto "
|
||||||
|
"-O normalization=formD "
|
||||||
|
"-O relatime=on "
|
||||||
|
"-O xattr=sa "
|
||||||
|
"-O mountpoint=legacy "
|
||||||
|
"-O compression=zstd "
|
||||||
|
"-O atime=off "
|
||||||
|
"root_pool "
|
||||||
|
)
|
||||||
|
if len(pool_disks) == 1:
|
||||||
|
zpool_create += pool_disks[0]
|
||||||
|
else:
|
||||||
|
zpool_create += "mirror "
|
||||||
|
zpool_create += " ".join(pool_disks)
|
||||||
|
|
||||||
|
bash_wrapper(zpool_create)
|
||||||
|
zpools = bash_wrapper("zpool list -o name")
|
||||||
|
if "root_pool" not in zpools.splitlines():
|
||||||
|
logger.critical("Failed to create root_pool")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
def create_zfs_datasets() -> None:
|
||||||
|
"""Create ZFS datasets."""
|
||||||
|
bash_wrapper("zfs create -o canmount=noauto -o reservation=10G root_pool/root")
|
||||||
|
bash_wrapper("zfs create root_pool/home")
|
||||||
|
bash_wrapper("zfs create root_pool/var -o reservation=1G")
|
||||||
|
bash_wrapper("zfs create -o compression=zstd-9 -o reservation=10G root_pool/nix")
|
||||||
|
datasets = bash_wrapper("zfs list -o name")
|
||||||
|
|
||||||
|
expected_datasets = {
|
||||||
|
"root_pool/root",
|
||||||
|
"root_pool/home",
|
||||||
|
"root_pool/var",
|
||||||
|
"root_pool/nix",
|
||||||
|
}
|
||||||
|
missing_datasets = expected_datasets.difference(datasets.splitlines())
|
||||||
|
if missing_datasets:
|
||||||
|
logger.critical(f"Failed to create pools {missing_datasets}")
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
|
||||||
|
def get_cpu_manufacturer() -> str:
|
||||||
|
"""Get the CPU manufacturer."""
|
||||||
|
output = bash_wrapper("cat /proc/cpuinfo")
|
||||||
|
|
||||||
|
id_vendor = {"AuthenticAMD": "amd", "GenuineIntel": "intel"}
|
||||||
|
|
||||||
|
for line in output.splitlines():
|
||||||
|
if "vendor_id" in line:
|
||||||
|
return id_vendor[line.split(": ")[1].strip()]
|
||||||
|
|
||||||
|
error = "Failed to get CPU manufacturer"
|
||||||
|
raise RuntimeError(error)
|
||||||
|
|
||||||
|
|
||||||
|
def get_boot_drive_id(disk: str) -> str:
|
||||||
|
"""Get the boot drive ID."""
|
||||||
|
output = bash_wrapper(f"lsblk -o UUID {disk}-part1")
|
||||||
|
return output.splitlines()[1]
|
||||||
|
|
||||||
|
|
||||||
|
def create_nix_hardware_file(mnt_dir: str, disks: Sequence[str], encrypt: str | None) -> None:
|
||||||
|
"""Create a NixOS hardware file."""
|
||||||
|
cpu_manufacturer = get_cpu_manufacturer()
|
||||||
|
|
||||||
|
devices = ""
|
||||||
|
if encrypt:
|
||||||
|
disk = disks[0]
|
||||||
|
|
||||||
|
devices = (
|
||||||
|
f' luks.devices."luks-root-pool-{disk.split("/")[-1]}-part2"'
|
||||||
|
"= {\n"
|
||||||
|
f' device = "{disk}-part2";\n'
|
||||||
|
" bypassWorkqueues = true;\n"
|
||||||
|
" allowDiscards = true;\n"
|
||||||
|
" };\n"
|
||||||
|
)
|
||||||
|
|
||||||
|
host_id = format(getrandbits(32), "08x")
|
||||||
|
|
||||||
|
nix_hardware = (
|
||||||
|
"{ config, lib, modulesPath, ... }:\n"
|
||||||
|
"{\n"
|
||||||
|
' imports = [ (modulesPath + "/installer/scan/not-detected.nix") ];\n\n'
|
||||||
|
" boot = {\n"
|
||||||
|
" initrd = {\n"
|
||||||
|
' availableKernelModules = [ \n "ahci"\n "ehci_pci"\n "nvme"\n "sd_mod"\n'
|
||||||
|
' "usb_storage"\n "usbhid"\n "xhci_pci"\n ];\n'
|
||||||
|
" kernelModules = [ ];\n"
|
||||||
|
f" {devices}"
|
||||||
|
" };\n"
|
||||||
|
f' kernelModules = [ "kvm-{cpu_manufacturer}" ];\n'
|
||||||
|
" extraModulePackages = [ ];\n"
|
||||||
|
" };\n\n"
|
||||||
|
" fileSystems = {\n"
|
||||||
|
' "/" = lib.mkDefault {\n device = "root_pool/root";\n fsType = "zfs";\n };\n\n'
|
||||||
|
' "/home" = {\n device = "root_pool/home";\n fsType = "zfs";\n };\n\n'
|
||||||
|
' "/var" = {\n device = "root_pool/var";\n fsType = "zfs";\n };\n\n'
|
||||||
|
' "/nix" = {\n device = "root_pool/nix";\n fsType = "zfs";\n };\n\n'
|
||||||
|
' "/boot" = {\n'
|
||||||
|
f' device = "/dev/disk/by-uuid/{get_boot_drive_id(disks[0])}";\n'
|
||||||
|
' fsType = "vfat";\n options = [\n "fmask=0077"\n'
|
||||||
|
' "dmask=0077"\n ];\n };\n };\n\n'
|
||||||
|
" swapDevices = [ ];\n\n"
|
||||||
|
" networking.useDHCP = lib.mkDefault true;\n\n"
|
||||||
|
' nixpkgs.hostPlatform = lib.mkDefault "x86_64-linux";\n'
|
||||||
|
f" hardware.cpu.{cpu_manufacturer}.updateMicrocode = "
|
||||||
|
"lib.mkDefault config.hardware.enableRedistributableFirmware;\n"
|
||||||
|
f' networking.hostId = "{host_id}";\n'
|
||||||
|
"}\n"
|
||||||
|
)
|
||||||
|
|
||||||
|
Path(f"{mnt_dir}/etc/nixos/hardware-configuration.nix").write_text(nix_hardware)
|
||||||
|
|
||||||
|
|
||||||
|
def install_nixos(mnt_dir: str, disks: Sequence[str], encrypt: str | None) -> None:
|
||||||
|
"""Install NixOS."""
|
||||||
|
bash_wrapper(f"mount -o X-mount.mkdir -t zfs root_pool/root {mnt_dir}")
|
||||||
|
bash_wrapper(f"mount -o X-mount.mkdir -t zfs root_pool/home {mnt_dir}/home")
|
||||||
|
bash_wrapper(f"mount -o X-mount.mkdir -t zfs root_pool/var {mnt_dir}/var")
|
||||||
|
bash_wrapper(f"mount -o X-mount.mkdir -t zfs root_pool/nix {mnt_dir}/nix")
|
||||||
|
|
||||||
|
for disk in disks:
|
||||||
|
bash_wrapper(f"mkfs.vfat -n EFI {disk}-part1")
|
||||||
|
|
||||||
|
# set up mirroring afterwards if more than one disk
|
||||||
|
boot_partition = (
|
||||||
|
f"mount -t vfat -o fmask=0077,dmask=0077,iocharset=iso8859-1,X-mount.mkdir {disks[0]}-part1 {mnt_dir}/boot"
|
||||||
|
)
|
||||||
|
bash_wrapper(boot_partition)
|
||||||
|
|
||||||
|
bash_wrapper(f"nixos-generate-config --root {mnt_dir}")
|
||||||
|
|
||||||
|
create_nix_hardware_file(mnt_dir, disks, encrypt)
|
||||||
|
|
||||||
|
run(("nixos-install", "--root", mnt_dir), check=True)
|
||||||
|
|
||||||
|
|
||||||
|
def installer(
|
||||||
|
disks: Sequence[str],
|
||||||
|
swap_size: int,
|
||||||
|
reserve: int,
|
||||||
|
encrypt_key: str | None,
|
||||||
|
) -> None:
|
||||||
|
"""Main."""
|
||||||
|
logger.info("Starting installation")
|
||||||
|
|
||||||
|
for disk in disks:
|
||||||
|
partition_disk(disk, swap_size, reserve)
|
||||||
|
|
||||||
|
test = Popen(("printf", f"'{encrypt_key}'"), stdout=PIPE)
|
||||||
|
if encrypt_key:
|
||||||
|
sleep(1)
|
||||||
|
for command in (
|
||||||
|
f"cryptsetup luksFormat --type luks2 {disk}-part2 -",
|
||||||
|
f"cryptsetup luksOpen {disk}-part2 luks-root-pool-{disk.split('/')[-1]}-part2 -",
|
||||||
|
):
|
||||||
|
run(command, check=True, stdin=test.stdout)
|
||||||
|
|
||||||
|
mnt_dir = "/tmp/nix_install" # noqa: S108
|
||||||
|
|
||||||
|
Path(mnt_dir).mkdir(parents=True, exist_ok=True)
|
||||||
|
|
||||||
|
if encrypt_key:
|
||||||
|
pool_disks = [f"/dev/mapper/luks-root-pool-{disk.split('/')[-1]}-part2" for disk in disks]
|
||||||
|
else:
|
||||||
|
pool_disks = [f"{disk}-part2" for disk in disks]
|
||||||
|
|
||||||
|
create_zfs_pool(pool_disks, mnt_dir)
|
||||||
|
|
||||||
|
create_zfs_datasets()
|
||||||
|
|
||||||
|
install_nixos(mnt_dir, disks, encrypt_key)
|
||||||
|
|
||||||
|
logger.info("Installation complete")
|
||||||
|
|
||||||
|
|
||||||
|
def main() -> None:
|
||||||
|
"""Main."""
|
||||||
|
configure_logger("DEBUG")
|
||||||
|
|
||||||
|
state = curses.wrapper(draw_menu)
|
||||||
|
|
||||||
|
encrypt_key = getenv("ENCRYPT_KEY")
|
||||||
|
|
||||||
|
logger.info("installing_nixos")
|
||||||
|
logger.info(f"disks: {state.selected_device_ids}")
|
||||||
|
logger.info(f"swap_size: {state.swap_size}")
|
||||||
|
logger.info(f"reserve: {state.reserve_size}")
|
||||||
|
logger.info(f"encrypted: {bool(encrypt_key)}")
|
||||||
|
|
||||||
|
sleep(3)
|
||||||
|
|
||||||
|
installer(
|
||||||
|
disks=state.get_selected_devices(),
|
||||||
|
swap_size=state.swap_size,
|
||||||
|
reserve=state.reserve_size,
|
||||||
|
encrypt_key=encrypt_key,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
||||||
498
python/installer/tui.py
Normal file
498
python/installer/tui.py
Normal file
@@ -0,0 +1,498 @@
|
|||||||
|
"""TUI module."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import curses
|
||||||
|
import logging
|
||||||
|
from collections import defaultdict
|
||||||
|
from subprocess import PIPE, Popen
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def bash_wrapper(command: str) -> str:
|
||||||
|
"""Execute a bash command and capture the output.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
command (str): The bash command to be executed.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tuple[str, int]: A tuple containing the output of the command (stdout) as a string,
|
||||||
|
the error output (stderr) as a string (optional), and the return code as an integer.
|
||||||
|
"""
|
||||||
|
logger.debug(f"running {command=}")
|
||||||
|
# This is a acceptable risk
|
||||||
|
process = Popen(command.split(), stdout=PIPE, stderr=PIPE)
|
||||||
|
output, _ = process.communicate()
|
||||||
|
if process.returncode != 0:
|
||||||
|
error = f"Failed to run command {command=} return code {process.returncode=}"
|
||||||
|
raise RuntimeError(error)
|
||||||
|
|
||||||
|
return output.decode()
|
||||||
|
|
||||||
|
|
||||||
|
class Cursor:
|
||||||
|
"""Cursor class."""
|
||||||
|
|
||||||
|
def __init__(self) -> None:
|
||||||
|
"""Initialize the Cursor class."""
|
||||||
|
self.x_position = 0
|
||||||
|
self.y_position = 0
|
||||||
|
self.height = 0
|
||||||
|
self.width = 0
|
||||||
|
|
||||||
|
def set_height(self, height: int) -> None:
|
||||||
|
"""Set height."""
|
||||||
|
self.height = height
|
||||||
|
|
||||||
|
def set_width(self, width: int) -> None:
|
||||||
|
"""Set width."""
|
||||||
|
self.width = width
|
||||||
|
|
||||||
|
def x_bounce_check(self, cursor: int) -> int:
|
||||||
|
"""X bounce check."""
|
||||||
|
cursor = max(0, cursor)
|
||||||
|
return min(self.width - 1, cursor)
|
||||||
|
|
||||||
|
def y_bounce_check(self, cursor: int) -> int:
|
||||||
|
"""Y bounce check."""
|
||||||
|
cursor = max(0, cursor)
|
||||||
|
return min(self.height - 1, cursor)
|
||||||
|
|
||||||
|
def set_x(self, x: int) -> None:
|
||||||
|
"""Set x."""
|
||||||
|
self.x_position = self.x_bounce_check(x)
|
||||||
|
|
||||||
|
def set_y(self, y: int) -> None:
|
||||||
|
"""Set y."""
|
||||||
|
self.y_position = self.y_bounce_check(y)
|
||||||
|
|
||||||
|
def get_x(self) -> int:
|
||||||
|
"""Get x."""
|
||||||
|
return self.x_position
|
||||||
|
|
||||||
|
def get_y(self) -> int:
|
||||||
|
"""Get y."""
|
||||||
|
return self.y_position
|
||||||
|
|
||||||
|
def move_up(self) -> None:
|
||||||
|
"""Move up."""
|
||||||
|
self.set_y(self.y_position - 1)
|
||||||
|
|
||||||
|
def move_down(self) -> None:
|
||||||
|
"""Move down."""
|
||||||
|
self.set_y(self.y_position + 1)
|
||||||
|
|
||||||
|
def move_left(self) -> None:
|
||||||
|
"""Move left."""
|
||||||
|
self.set_x(self.x_position - 1)
|
||||||
|
|
||||||
|
def move_right(self) -> None:
|
||||||
|
"""Move right."""
|
||||||
|
self.set_x(self.x_position + 1)
|
||||||
|
|
||||||
|
def navigation(self, key: int) -> None:
|
||||||
|
"""Navigation.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
key (int): The key.
|
||||||
|
"""
|
||||||
|
action = {
|
||||||
|
curses.KEY_DOWN: self.move_down,
|
||||||
|
curses.KEY_UP: self.move_up,
|
||||||
|
curses.KEY_RIGHT: self.move_right,
|
||||||
|
curses.KEY_LEFT: self.move_left,
|
||||||
|
}
|
||||||
|
|
||||||
|
action.get(key, lambda: None)()
|
||||||
|
|
||||||
|
|
||||||
|
class State:
|
||||||
|
"""State class to store the state of the program."""
|
||||||
|
|
||||||
|
def __init__(self) -> None:
|
||||||
|
"""Initialize the State class."""
|
||||||
|
self.key = 0
|
||||||
|
self.cursor = Cursor()
|
||||||
|
|
||||||
|
self.swap_size = 0
|
||||||
|
self.show_swap_input = False
|
||||||
|
|
||||||
|
self.reserve_size = 0
|
||||||
|
self.show_reserve_input = False
|
||||||
|
|
||||||
|
self.selected_device_ids: set[str] = set()
|
||||||
|
|
||||||
|
def get_selected_devices(self) -> tuple[str, ...]:
|
||||||
|
"""Get selected devices."""
|
||||||
|
return tuple(self.selected_device_ids)
|
||||||
|
|
||||||
|
|
||||||
|
def get_device(raw_device: str) -> dict[str, str]:
|
||||||
|
"""Get a device.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
raw_device (str): The raw device.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
dict[str, str]: The device.
|
||||||
|
"""
|
||||||
|
raw_device_components = raw_device.split(" ")
|
||||||
|
return {thing.split("=")[0].lower(): thing.split("=")[1].strip('"') for thing in raw_device_components}
|
||||||
|
|
||||||
|
|
||||||
|
def get_devices() -> list[dict[str, str]]:
|
||||||
|
"""Get a list of devices."""
|
||||||
|
# --bytes
|
||||||
|
raw_devices = bash_wrapper("lsblk --paths --pairs").splitlines()
|
||||||
|
return [get_device(raw_device) for raw_device in raw_devices]
|
||||||
|
|
||||||
|
|
||||||
|
def set_color() -> None:
|
||||||
|
"""Set the color."""
|
||||||
|
curses.start_color()
|
||||||
|
curses.use_default_colors()
|
||||||
|
for i in range(curses.COLORS):
|
||||||
|
curses.init_pair(i + 1, i, -1)
|
||||||
|
|
||||||
|
|
||||||
|
def debug_menu(std_screen: curses.window, key: int) -> None:
|
||||||
|
"""Debug menu.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): The curses window.
|
||||||
|
key (int): The key.
|
||||||
|
"""
|
||||||
|
height, width = std_screen.getmaxyx()
|
||||||
|
std_screen.addstr(height - 4, 0, f"Width: {width}, Height: {height}", curses.color_pair(5))
|
||||||
|
|
||||||
|
key_pressed = f"Last key pressed: {key}"[: width - 1]
|
||||||
|
if key == 0:
|
||||||
|
key_pressed = "No key press detected..."[: width - 1]
|
||||||
|
std_screen.addstr(height - 3, 0, key_pressed)
|
||||||
|
|
||||||
|
for i in range(8):
|
||||||
|
std_screen.addstr(height - 2, i * 3, f"{i}██", curses.color_pair(i))
|
||||||
|
|
||||||
|
|
||||||
|
def get_text_input(std_screen: curses.window, prompt: str, y: int, x: int) -> str:
|
||||||
|
"""Get text input.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): The curses window.
|
||||||
|
prompt (str): The prompt.
|
||||||
|
y (int): The y position.
|
||||||
|
x (int): The x position.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
str: The input string.
|
||||||
|
"""
|
||||||
|
esc_key = 27
|
||||||
|
curses.echo()
|
||||||
|
std_screen.addstr(y, x, prompt)
|
||||||
|
input_str = ""
|
||||||
|
while True:
|
||||||
|
key = std_screen.getch()
|
||||||
|
if key == ord("\n"):
|
||||||
|
break
|
||||||
|
if key == esc_key:
|
||||||
|
input_str = ""
|
||||||
|
break
|
||||||
|
if key in (curses.KEY_BACKSPACE, ord("\b"), 127):
|
||||||
|
input_str = input_str[:-1]
|
||||||
|
std_screen.addstr(y, x + len(prompt), input_str + " ")
|
||||||
|
else:
|
||||||
|
input_str += chr(key)
|
||||||
|
std_screen.refresh()
|
||||||
|
curses.noecho()
|
||||||
|
return input_str
|
||||||
|
|
||||||
|
|
||||||
|
def swap_size_input(
|
||||||
|
std_screen: curses.window,
|
||||||
|
state: State,
|
||||||
|
swap_offset: int,
|
||||||
|
) -> State:
|
||||||
|
"""Reserve size input.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): The curses window.
|
||||||
|
state (State): The state object.
|
||||||
|
swap_offset (int): The swap offset.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
State: The updated state object.
|
||||||
|
"""
|
||||||
|
swap_size_text = "Swap size (GB): "
|
||||||
|
std_screen.addstr(swap_offset, 0, f"{swap_size_text}{state.swap_size}")
|
||||||
|
if state.key == ord("\n") and state.cursor.get_y() == swap_offset:
|
||||||
|
state.show_swap_input = True
|
||||||
|
|
||||||
|
if state.show_swap_input:
|
||||||
|
swap_size_str = get_text_input(std_screen, swap_size_text, swap_offset, 0)
|
||||||
|
try:
|
||||||
|
state.swap_size = int(swap_size_str)
|
||||||
|
state.show_swap_input = False
|
||||||
|
except ValueError:
|
||||||
|
std_screen.addstr(swap_offset, 0, "Invalid input. Press any key to continue.")
|
||||||
|
std_screen.getch()
|
||||||
|
state.show_swap_input = False
|
||||||
|
|
||||||
|
return state
|
||||||
|
|
||||||
|
|
||||||
|
def reserve_size_input(
|
||||||
|
std_screen: curses.window,
|
||||||
|
state: State,
|
||||||
|
reserve_offset: int,
|
||||||
|
) -> State:
|
||||||
|
"""Reserve size input.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): The curses window.
|
||||||
|
state (State): The state object.
|
||||||
|
reserve_offset (int): The reserve offset.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
State: The updated state object.
|
||||||
|
"""
|
||||||
|
reserve_size_text = "reserve size (GB): "
|
||||||
|
std_screen.addstr(reserve_offset, 0, f"{reserve_size_text}{state.reserve_size}")
|
||||||
|
if state.key == ord("\n") and state.cursor.get_y() == reserve_offset:
|
||||||
|
state.show_reserve_input = True
|
||||||
|
|
||||||
|
if state.show_reserve_input:
|
||||||
|
reserve_size_str = get_text_input(std_screen, reserve_size_text, reserve_offset, 0)
|
||||||
|
try:
|
||||||
|
state.reserve_size = int(reserve_size_str)
|
||||||
|
state.show_reserve_input = False
|
||||||
|
except ValueError:
|
||||||
|
std_screen.addstr(reserve_offset, 0, "Invalid input. Press any key to continue.")
|
||||||
|
std_screen.getch()
|
||||||
|
state.show_reserve_input = False
|
||||||
|
|
||||||
|
return state
|
||||||
|
|
||||||
|
|
||||||
|
def status_bar(
|
||||||
|
std_screen: curses.window,
|
||||||
|
cursor: Cursor,
|
||||||
|
width: int,
|
||||||
|
height: int,
|
||||||
|
) -> None:
|
||||||
|
"""Draw the status bar.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): The curses window.
|
||||||
|
cursor (Cursor): The cursor.
|
||||||
|
width (int): The width.
|
||||||
|
height (int): The height.
|
||||||
|
"""
|
||||||
|
std_screen.attron(curses.A_REVERSE)
|
||||||
|
std_screen.attron(curses.color_pair(3))
|
||||||
|
|
||||||
|
status_bar = f"Press 'q' to exit | STATUS BAR | Pos: {cursor.get_x()}, {cursor.get_y()}"
|
||||||
|
std_screen.addstr(height - 1, 0, status_bar)
|
||||||
|
std_screen.addstr(height - 1, len(status_bar), " " * (width - len(status_bar) - 1))
|
||||||
|
|
||||||
|
std_screen.attroff(curses.color_pair(3))
|
||||||
|
std_screen.attroff(curses.A_REVERSE)
|
||||||
|
|
||||||
|
|
||||||
|
def get_device_id_mapping() -> dict[str, set[str]]:
|
||||||
|
"""Get a list of device ids.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
list[str]: the list of device ids
|
||||||
|
"""
|
||||||
|
device_ids = bash_wrapper("find /dev/disk/by-id -type l").splitlines()
|
||||||
|
|
||||||
|
device_id_mapping: dict[str, set[str]] = defaultdict(set)
|
||||||
|
|
||||||
|
for device_id in device_ids:
|
||||||
|
device = bash_wrapper(f"readlink -f {device_id}").strip()
|
||||||
|
device_id_mapping[device].add(device_id)
|
||||||
|
|
||||||
|
return device_id_mapping
|
||||||
|
|
||||||
|
|
||||||
|
def calculate_device_menu_padding(devices: list[dict[str, str]], column: str, padding: int = 0) -> int:
|
||||||
|
"""Calculate the device menu padding.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
devices (list[dict[str, str]]): The devices.
|
||||||
|
column (str): The column.
|
||||||
|
padding (int, optional): The padding. Defaults to 0.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
int: The calculated padding.
|
||||||
|
"""
|
||||||
|
return max(len(device[column]) for device in devices) + padding
|
||||||
|
|
||||||
|
|
||||||
|
def draw_device_ids(
|
||||||
|
state: State,
|
||||||
|
row_number: int,
|
||||||
|
menu_start_x: int,
|
||||||
|
std_screen: curses.window,
|
||||||
|
menu_width: list[int],
|
||||||
|
device_ids: set[str],
|
||||||
|
) -> tuple[State, int]:
|
||||||
|
"""Draw device IDs.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
state (State): The state object.
|
||||||
|
row_number (int): The row number.
|
||||||
|
menu_start_x (int): The menu start x.
|
||||||
|
std_screen (curses.window): The curses window.
|
||||||
|
menu_width (list[int]): The menu width.
|
||||||
|
device_ids (set[str]): The device IDs.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
tuple[State, int]: The updated state object and the row number.
|
||||||
|
"""
|
||||||
|
for device_id in sorted(device_ids):
|
||||||
|
row_number = row_number + 1
|
||||||
|
if row_number == state.cursor.get_y() and state.cursor.get_x() in menu_width:
|
||||||
|
std_screen.attron(curses.A_BOLD)
|
||||||
|
if state.key == ord(" "):
|
||||||
|
if device_id not in state.selected_device_ids:
|
||||||
|
state.selected_device_ids.add(device_id)
|
||||||
|
else:
|
||||||
|
state.selected_device_ids.remove(device_id)
|
||||||
|
|
||||||
|
if device_id in state.selected_device_ids:
|
||||||
|
std_screen.attron(curses.color_pair(7))
|
||||||
|
|
||||||
|
std_screen.addstr(row_number, menu_start_x, f" {device_id}")
|
||||||
|
|
||||||
|
std_screen.attroff(curses.color_pair(7))
|
||||||
|
std_screen.attroff(curses.A_BOLD)
|
||||||
|
|
||||||
|
return state, row_number
|
||||||
|
|
||||||
|
|
||||||
|
def draw_device_menu(
|
||||||
|
std_screen: curses.window,
|
||||||
|
devices: list[dict[str, str]],
|
||||||
|
device_id_mapping: dict[str, set[str]],
|
||||||
|
state: State,
|
||||||
|
menu_start_y: int = 0,
|
||||||
|
menu_start_x: int = 0,
|
||||||
|
) -> tuple[State, int]:
|
||||||
|
"""Draw the device menu and handle user input.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): the curses window to draw on
|
||||||
|
devices (list[dict[str, str]]): the list of devices to draw
|
||||||
|
device_id_mapping (dict[str, set[str]]): the list of device ids to draw
|
||||||
|
state (State): the state object to update
|
||||||
|
menu_start_y (int, optional): the y position to start drawing the menu. Defaults to 0.
|
||||||
|
menu_start_x (int, optional): the x position to start drawing the menu. Defaults to 0.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
State: the updated state object
|
||||||
|
"""
|
||||||
|
padding = 2
|
||||||
|
|
||||||
|
name_padding = calculate_device_menu_padding(devices, "name", padding)
|
||||||
|
size_padding = calculate_device_menu_padding(devices, "size", padding)
|
||||||
|
type_padding = calculate_device_menu_padding(devices, "type", padding)
|
||||||
|
mountpoints_padding = calculate_device_menu_padding(devices, "mountpoints", padding)
|
||||||
|
|
||||||
|
device_header = (
|
||||||
|
f"{'Name':{name_padding}}{'Size':{size_padding}}{'Type':{type_padding}}{'Mountpoints':{mountpoints_padding}}"
|
||||||
|
)
|
||||||
|
|
||||||
|
menu_width = list(range(menu_start_x, len(device_header) + menu_start_x))
|
||||||
|
|
||||||
|
std_screen.addstr(menu_start_y, menu_start_x, device_header, curses.color_pair(5))
|
||||||
|
devises_list_start = menu_start_y + 1
|
||||||
|
|
||||||
|
row_number = devises_list_start
|
||||||
|
|
||||||
|
for device in devices:
|
||||||
|
row_number = row_number + 1
|
||||||
|
device_name = device["name"]
|
||||||
|
device_row = (
|
||||||
|
f"{device_name:{name_padding}}"
|
||||||
|
f"{device['size']:{size_padding}}"
|
||||||
|
f"{device['type']:{type_padding}}"
|
||||||
|
f"{device['mountpoints']:{mountpoints_padding}}"
|
||||||
|
)
|
||||||
|
std_screen.addstr(row_number, menu_start_x, device_row)
|
||||||
|
|
||||||
|
state, row_number = draw_device_ids(
|
||||||
|
state=state,
|
||||||
|
row_number=row_number,
|
||||||
|
menu_start_x=menu_start_x,
|
||||||
|
std_screen=std_screen,
|
||||||
|
menu_width=menu_width,
|
||||||
|
device_ids=device_id_mapping[device_name],
|
||||||
|
)
|
||||||
|
|
||||||
|
return state, row_number
|
||||||
|
|
||||||
|
|
||||||
|
def draw_menu(std_screen: curses.window) -> State:
|
||||||
|
"""Draw the menu and handle user input.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
std_screen (curses.window): the curses window to draw on
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
State: the state object
|
||||||
|
"""
|
||||||
|
# Clear and refresh the screen for a blank canvas
|
||||||
|
std_screen.clear()
|
||||||
|
std_screen.refresh()
|
||||||
|
|
||||||
|
set_color()
|
||||||
|
|
||||||
|
state = State()
|
||||||
|
|
||||||
|
devices = get_devices()
|
||||||
|
|
||||||
|
device_id_mapping = get_device_id_mapping()
|
||||||
|
|
||||||
|
# Loop where k is the last character pressed
|
||||||
|
while state.key != ord("q"):
|
||||||
|
std_screen.clear()
|
||||||
|
height, width = std_screen.getmaxyx()
|
||||||
|
|
||||||
|
state.cursor.set_height(height)
|
||||||
|
state.cursor.set_width(width)
|
||||||
|
|
||||||
|
state.cursor.navigation(state.key)
|
||||||
|
|
||||||
|
state, device_menu_size = draw_device_menu(
|
||||||
|
std_screen=std_screen,
|
||||||
|
state=state,
|
||||||
|
devices=devices,
|
||||||
|
device_id_mapping=device_id_mapping,
|
||||||
|
)
|
||||||
|
|
||||||
|
swap_offset = device_menu_size + 2
|
||||||
|
|
||||||
|
swap_size_input(
|
||||||
|
std_screen=std_screen,
|
||||||
|
state=state,
|
||||||
|
swap_offset=swap_offset,
|
||||||
|
)
|
||||||
|
reserve_size_input(
|
||||||
|
std_screen=std_screen,
|
||||||
|
state=state,
|
||||||
|
reserve_offset=swap_offset + 1,
|
||||||
|
)
|
||||||
|
|
||||||
|
status_bar(std_screen, state.cursor, width, height)
|
||||||
|
|
||||||
|
debug_menu(std_screen, state.key)
|
||||||
|
|
||||||
|
std_screen.move(state.cursor.get_y(), state.cursor.get_x())
|
||||||
|
|
||||||
|
std_screen.refresh()
|
||||||
|
|
||||||
|
state.key = std_screen.getch()
|
||||||
|
|
||||||
|
return state
|
||||||
11
python/orm/__init__.py
Normal file
11
python/orm/__init__.py
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
"""ORM package exports."""
|
||||||
|
|
||||||
|
from python.orm.richie.base import RichieBase
|
||||||
|
from python.orm.signal_bot.base import SignalBotBase
|
||||||
|
from python.orm.van_inventory.base import VanInventoryBase
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
"RichieBase",
|
||||||
|
"SignalBotBase",
|
||||||
|
"VanInventoryBase",
|
||||||
|
]
|
||||||
51
python/orm/common.py
Normal file
51
python/orm/common.py
Normal file
@@ -0,0 +1,51 @@
|
|||||||
|
"""Shared ORM definitions."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from os import getenv
|
||||||
|
from typing import cast
|
||||||
|
|
||||||
|
from sqlalchemy import create_engine
|
||||||
|
from sqlalchemy.engine import URL, Engine
|
||||||
|
|
||||||
|
NAMING_CONVENTION = {
|
||||||
|
"ix": "ix_%(table_name)s_%(column_0_name)s",
|
||||||
|
"uq": "uq_%(table_name)s_%(column_0_name)s",
|
||||||
|
"ck": "ck_%(table_name)s_%(constraint_name)s",
|
||||||
|
"fk": "fk_%(table_name)s_%(column_0_name)s_%(referred_table_name)s",
|
||||||
|
"pk": "pk_%(table_name)s",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def get_connection_info(name: str) -> tuple[str, str, str, str, str | None]:
|
||||||
|
"""Get connection info from environment variables."""
|
||||||
|
database = getenv(f"{name}_DB")
|
||||||
|
host = getenv(f"{name}_HOST")
|
||||||
|
port = getenv(f"{name}_PORT")
|
||||||
|
username = getenv(f"{name}_USER")
|
||||||
|
password = getenv(f"{name}_PASSWORD")
|
||||||
|
|
||||||
|
if None in (database, host, port, username):
|
||||||
|
error = f"Missing environment variables for Postgres connection.\n{database=}\n{host=}\n{port=}\n{username=}\n"
|
||||||
|
raise ValueError(error)
|
||||||
|
return cast("tuple[str, str, str, str, str | None]", (database, host, port, username, password))
|
||||||
|
|
||||||
|
|
||||||
|
def get_postgres_engine(*, name: str = "POSTGRES", pool_pre_ping: bool = True) -> Engine:
|
||||||
|
"""Create a SQLAlchemy engine from environment variables."""
|
||||||
|
database, host, port, username, password = get_connection_info(name)
|
||||||
|
|
||||||
|
url = URL.create(
|
||||||
|
drivername="postgresql+psycopg",
|
||||||
|
username=username,
|
||||||
|
password=password,
|
||||||
|
host=host,
|
||||||
|
port=int(port),
|
||||||
|
database=database,
|
||||||
|
)
|
||||||
|
|
||||||
|
return create_engine(
|
||||||
|
url=url,
|
||||||
|
pool_pre_ping=pool_pre_ping,
|
||||||
|
pool_recycle=1800,
|
||||||
|
)
|
||||||
29
python/orm/richie/__init__.py
Normal file
29
python/orm/richie/__init__.py
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
"""Richie database ORM exports."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from python.orm.richie.base import RichieBase, TableBase, TableBaseBig, TableBaseSmall
|
||||||
|
from python.orm.richie.congress import Bill, Legislator, Vote, VoteRecord
|
||||||
|
from python.orm.richie.contact import (
|
||||||
|
Contact,
|
||||||
|
ContactNeed,
|
||||||
|
ContactRelationship,
|
||||||
|
Need,
|
||||||
|
RelationshipType,
|
||||||
|
)
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
"Bill",
|
||||||
|
"Contact",
|
||||||
|
"ContactNeed",
|
||||||
|
"ContactRelationship",
|
||||||
|
"Legislator",
|
||||||
|
"Need",
|
||||||
|
"RelationshipType",
|
||||||
|
"RichieBase",
|
||||||
|
"TableBase",
|
||||||
|
"TableBaseBig",
|
||||||
|
"TableBaseSmall",
|
||||||
|
"Vote",
|
||||||
|
"VoteRecord",
|
||||||
|
]
|
||||||
60
python/orm/richie/base.py
Normal file
60
python/orm/richie/base.py
Normal file
@@ -0,0 +1,60 @@
|
|||||||
|
"""Richie database ORM base."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
from sqlalchemy import BigInteger, DateTime, MetaData, SmallInteger, func
|
||||||
|
from sqlalchemy.ext.declarative import AbstractConcreteBase
|
||||||
|
from sqlalchemy.orm import DeclarativeBase, Mapped, mapped_column
|
||||||
|
|
||||||
|
from python.orm.common import NAMING_CONVENTION
|
||||||
|
|
||||||
|
|
||||||
|
class RichieBase(DeclarativeBase):
|
||||||
|
"""Base class for richie database ORM models."""
|
||||||
|
|
||||||
|
schema_name = "main"
|
||||||
|
|
||||||
|
metadata = MetaData(
|
||||||
|
schema=schema_name,
|
||||||
|
naming_convention=NAMING_CONVENTION,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class _TableMixin:
|
||||||
|
"""Shared timestamp columns for all table bases."""
|
||||||
|
|
||||||
|
created: Mapped[datetime] = mapped_column(
|
||||||
|
DateTime(timezone=True),
|
||||||
|
server_default=func.now(),
|
||||||
|
)
|
||||||
|
updated: Mapped[datetime] = mapped_column(
|
||||||
|
DateTime(timezone=True),
|
||||||
|
server_default=func.now(),
|
||||||
|
onupdate=func.now(),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class TableBaseSmall(_TableMixin, AbstractConcreteBase, RichieBase):
|
||||||
|
"""Table with SmallInteger primary key."""
|
||||||
|
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
id: Mapped[int] = mapped_column(SmallInteger, primary_key=True)
|
||||||
|
|
||||||
|
|
||||||
|
class TableBase(_TableMixin, AbstractConcreteBase, RichieBase):
|
||||||
|
"""Table with Integer primary key."""
|
||||||
|
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
id: Mapped[int] = mapped_column(primary_key=True)
|
||||||
|
|
||||||
|
|
||||||
|
class TableBaseBig(_TableMixin, AbstractConcreteBase, RichieBase):
|
||||||
|
"""Table with BigInteger primary key."""
|
||||||
|
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
id: Mapped[int] = mapped_column(BigInteger, primary_key=True)
|
||||||
150
python/orm/richie/congress.py
Normal file
150
python/orm/richie/congress.py
Normal file
@@ -0,0 +1,150 @@
|
|||||||
|
"""Congress Tracker database models."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from datetime import date
|
||||||
|
|
||||||
|
from sqlalchemy import ForeignKey, Index, Text, UniqueConstraint
|
||||||
|
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||||
|
|
||||||
|
from python.orm.richie.base import RichieBase, TableBase
|
||||||
|
|
||||||
|
|
||||||
|
class Legislator(TableBase):
|
||||||
|
"""Legislator model - members of Congress."""
|
||||||
|
|
||||||
|
__tablename__ = "legislator"
|
||||||
|
|
||||||
|
# Natural key - bioguide ID is the authoritative identifier
|
||||||
|
bioguide_id: Mapped[str] = mapped_column(Text, unique=True, index=True)
|
||||||
|
|
||||||
|
# Other IDs for cross-referencing
|
||||||
|
thomas_id: Mapped[str | None]
|
||||||
|
lis_id: Mapped[str | None]
|
||||||
|
govtrack_id: Mapped[int | None]
|
||||||
|
opensecrets_id: Mapped[str | None]
|
||||||
|
fec_ids: Mapped[str | None] # JSON array stored as string
|
||||||
|
|
||||||
|
# Name info
|
||||||
|
first_name: Mapped[str]
|
||||||
|
last_name: Mapped[str]
|
||||||
|
official_full_name: Mapped[str | None]
|
||||||
|
nickname: Mapped[str | None]
|
||||||
|
|
||||||
|
# Bio
|
||||||
|
birthday: Mapped[date | None]
|
||||||
|
gender: Mapped[str | None] # M/F
|
||||||
|
|
||||||
|
# Current term info (denormalized for query efficiency)
|
||||||
|
current_party: Mapped[str | None]
|
||||||
|
current_state: Mapped[str | None]
|
||||||
|
current_district: Mapped[int | None] # House only
|
||||||
|
current_chamber: Mapped[str | None] # rep/sen
|
||||||
|
|
||||||
|
# Relationships
|
||||||
|
vote_records: Mapped[list[VoteRecord]] = relationship(
|
||||||
|
"VoteRecord",
|
||||||
|
back_populates="legislator",
|
||||||
|
cascade="all, delete-orphan",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class Bill(TableBase):
|
||||||
|
"""Bill model - legislation introduced in Congress."""
|
||||||
|
|
||||||
|
__tablename__ = "bill"
|
||||||
|
|
||||||
|
# Composite natural key: congress + bill_type + number
|
||||||
|
congress: Mapped[int]
|
||||||
|
bill_type: Mapped[str] # hr, s, hres, sres, hjres, sjres
|
||||||
|
number: Mapped[int]
|
||||||
|
|
||||||
|
# Bill info
|
||||||
|
title: Mapped[str | None]
|
||||||
|
title_short: Mapped[str | None]
|
||||||
|
official_title: Mapped[str | None]
|
||||||
|
|
||||||
|
# Status
|
||||||
|
status: Mapped[str | None]
|
||||||
|
status_at: Mapped[date | None]
|
||||||
|
|
||||||
|
# Sponsor
|
||||||
|
sponsor_bioguide_id: Mapped[str | None]
|
||||||
|
|
||||||
|
# Subjects
|
||||||
|
subjects_top_term: Mapped[str | None]
|
||||||
|
|
||||||
|
# Relationships
|
||||||
|
votes: Mapped[list[Vote]] = relationship(
|
||||||
|
"Vote",
|
||||||
|
back_populates="bill",
|
||||||
|
)
|
||||||
|
|
||||||
|
__table_args__ = (
|
||||||
|
UniqueConstraint("congress", "bill_type", "number", name="uq_bill_congress_type_number"),
|
||||||
|
Index("ix_bill_congress", "congress"),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class Vote(TableBase):
|
||||||
|
"""Vote model - roll call votes in Congress."""
|
||||||
|
|
||||||
|
__tablename__ = "vote"
|
||||||
|
|
||||||
|
# Composite natural key: congress + chamber + session + number
|
||||||
|
congress: Mapped[int]
|
||||||
|
chamber: Mapped[str] # house/senate
|
||||||
|
session: Mapped[int]
|
||||||
|
number: Mapped[int]
|
||||||
|
|
||||||
|
# Vote details
|
||||||
|
vote_type: Mapped[str | None]
|
||||||
|
question: Mapped[str | None]
|
||||||
|
result: Mapped[str | None]
|
||||||
|
result_text: Mapped[str | None]
|
||||||
|
|
||||||
|
# Timing
|
||||||
|
vote_date: Mapped[date]
|
||||||
|
|
||||||
|
# Vote counts (denormalized for efficiency)
|
||||||
|
yea_count: Mapped[int | None]
|
||||||
|
nay_count: Mapped[int | None]
|
||||||
|
not_voting_count: Mapped[int | None]
|
||||||
|
present_count: Mapped[int | None]
|
||||||
|
|
||||||
|
# Related bill (optional - not all votes are on bills)
|
||||||
|
bill_id: Mapped[int | None] = mapped_column(ForeignKey("main.bill.id"))
|
||||||
|
|
||||||
|
# Relationships
|
||||||
|
bill: Mapped[Bill | None] = relationship("Bill", back_populates="votes")
|
||||||
|
vote_records: Mapped[list[VoteRecord]] = relationship(
|
||||||
|
"VoteRecord",
|
||||||
|
back_populates="vote",
|
||||||
|
cascade="all, delete-orphan",
|
||||||
|
)
|
||||||
|
|
||||||
|
__table_args__ = (
|
||||||
|
UniqueConstraint("congress", "chamber", "session", "number", name="uq_vote_congress_chamber_session_number"),
|
||||||
|
Index("ix_vote_date", "vote_date"),
|
||||||
|
Index("ix_vote_congress_chamber", "congress", "chamber"),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class VoteRecord(RichieBase):
|
||||||
|
"""Association table: Vote <-> Legislator with position."""
|
||||||
|
|
||||||
|
__tablename__ = "vote_record"
|
||||||
|
|
||||||
|
vote_id: Mapped[int] = mapped_column(
|
||||||
|
ForeignKey("main.vote.id", ondelete="CASCADE"),
|
||||||
|
primary_key=True,
|
||||||
|
)
|
||||||
|
legislator_id: Mapped[int] = mapped_column(
|
||||||
|
ForeignKey("main.legislator.id", ondelete="CASCADE"),
|
||||||
|
primary_key=True,
|
||||||
|
)
|
||||||
|
position: Mapped[str] # Yea, Nay, Not Voting, Present
|
||||||
|
|
||||||
|
# Relationships
|
||||||
|
vote: Mapped[Vote] = relationship("Vote", back_populates="vote_records")
|
||||||
|
legislator: Mapped[Legislator] = relationship("Legislator", back_populates="vote_records")
|
||||||
168
python/orm/richie/contact.py
Normal file
168
python/orm/richie/contact.py
Normal file
@@ -0,0 +1,168 @@
|
|||||||
|
"""Contact database models."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from enum import StrEnum
|
||||||
|
|
||||||
|
from sqlalchemy import ForeignKey, String
|
||||||
|
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||||
|
|
||||||
|
from python.orm.richie.base import RichieBase, TableBase
|
||||||
|
|
||||||
|
|
||||||
|
class RelationshipType(StrEnum):
|
||||||
|
"""Relationship types with default closeness weights.
|
||||||
|
|
||||||
|
Default weight is an integer 1-10 where 10 = closest relationship.
|
||||||
|
Users can override this per-relationship in the UI.
|
||||||
|
"""
|
||||||
|
|
||||||
|
SPOUSE = "spouse"
|
||||||
|
PARTNER = "partner"
|
||||||
|
PARENT = "parent"
|
||||||
|
CHILD = "child"
|
||||||
|
SIBLING = "sibling"
|
||||||
|
BEST_FRIEND = "best_friend"
|
||||||
|
GRANDPARENT = "grandparent"
|
||||||
|
GRANDCHILD = "grandchild"
|
||||||
|
AUNT_UNCLE = "aunt_uncle"
|
||||||
|
NIECE_NEPHEW = "niece_nephew"
|
||||||
|
COUSIN = "cousin"
|
||||||
|
IN_LAW = "in_law"
|
||||||
|
CLOSE_FRIEND = "close_friend"
|
||||||
|
FRIEND = "friend"
|
||||||
|
MENTOR = "mentor"
|
||||||
|
MENTEE = "mentee"
|
||||||
|
BUSINESS_PARTNER = "business_partner"
|
||||||
|
COLLEAGUE = "colleague"
|
||||||
|
MANAGER = "manager"
|
||||||
|
DIRECT_REPORT = "direct_report"
|
||||||
|
CLIENT = "client"
|
||||||
|
ACQUAINTANCE = "acquaintance"
|
||||||
|
NEIGHBOR = "neighbor"
|
||||||
|
EX = "ex"
|
||||||
|
OTHER = "other"
|
||||||
|
|
||||||
|
@property
|
||||||
|
def default_weight(self) -> int:
|
||||||
|
"""Return the default closeness weight (1-10) for this relationship type."""
|
||||||
|
weights = {
|
||||||
|
RelationshipType.SPOUSE: 10,
|
||||||
|
RelationshipType.PARTNER: 10,
|
||||||
|
RelationshipType.PARENT: 9,
|
||||||
|
RelationshipType.CHILD: 9,
|
||||||
|
RelationshipType.SIBLING: 9,
|
||||||
|
RelationshipType.BEST_FRIEND: 8,
|
||||||
|
RelationshipType.GRANDPARENT: 7,
|
||||||
|
RelationshipType.GRANDCHILD: 7,
|
||||||
|
RelationshipType.AUNT_UNCLE: 7,
|
||||||
|
RelationshipType.NIECE_NEPHEW: 7,
|
||||||
|
RelationshipType.COUSIN: 7,
|
||||||
|
RelationshipType.IN_LAW: 7,
|
||||||
|
RelationshipType.CLOSE_FRIEND: 6,
|
||||||
|
RelationshipType.FRIEND: 6,
|
||||||
|
RelationshipType.MENTOR: 5,
|
||||||
|
RelationshipType.MENTEE: 5,
|
||||||
|
RelationshipType.BUSINESS_PARTNER: 5,
|
||||||
|
RelationshipType.COLLEAGUE: 4,
|
||||||
|
RelationshipType.MANAGER: 4,
|
||||||
|
RelationshipType.DIRECT_REPORT: 4,
|
||||||
|
RelationshipType.CLIENT: 4,
|
||||||
|
RelationshipType.ACQUAINTANCE: 3,
|
||||||
|
RelationshipType.NEIGHBOR: 3,
|
||||||
|
RelationshipType.EX: 2,
|
||||||
|
RelationshipType.OTHER: 2,
|
||||||
|
}
|
||||||
|
return weights.get(self, 5)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def display_name(self) -> str:
|
||||||
|
"""Return a human-readable display name."""
|
||||||
|
return self.value.replace("_", " ").title()
|
||||||
|
|
||||||
|
|
||||||
|
class ContactNeed(RichieBase):
|
||||||
|
"""Association table: Contact <-> Need."""
|
||||||
|
|
||||||
|
__tablename__ = "contact_need"
|
||||||
|
|
||||||
|
contact_id: Mapped[int] = mapped_column(
|
||||||
|
ForeignKey("main.contact.id", ondelete="CASCADE"),
|
||||||
|
primary_key=True,
|
||||||
|
)
|
||||||
|
need_id: Mapped[int] = mapped_column(
|
||||||
|
ForeignKey("main.need.id", ondelete="CASCADE"),
|
||||||
|
primary_key=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class ContactRelationship(RichieBase):
|
||||||
|
"""Association table: Contact <-> Contact with relationship type and weight."""
|
||||||
|
|
||||||
|
__tablename__ = "contact_relationship"
|
||||||
|
|
||||||
|
contact_id: Mapped[int] = mapped_column(
|
||||||
|
ForeignKey("main.contact.id", ondelete="CASCADE"),
|
||||||
|
primary_key=True,
|
||||||
|
)
|
||||||
|
related_contact_id: Mapped[int] = mapped_column(
|
||||||
|
ForeignKey("main.contact.id", ondelete="CASCADE"),
|
||||||
|
primary_key=True,
|
||||||
|
)
|
||||||
|
relationship_type: Mapped[str] = mapped_column(String(100))
|
||||||
|
closeness_weight: Mapped[int] = mapped_column(default=5)
|
||||||
|
|
||||||
|
|
||||||
|
class Contact(TableBase):
|
||||||
|
"""Contact model."""
|
||||||
|
|
||||||
|
__tablename__ = "contact"
|
||||||
|
|
||||||
|
name: Mapped[str]
|
||||||
|
|
||||||
|
age: Mapped[int | None]
|
||||||
|
bio: Mapped[str | None]
|
||||||
|
current_job: Mapped[str | None]
|
||||||
|
gender: Mapped[str | None]
|
||||||
|
goals: Mapped[str | None]
|
||||||
|
legal_name: Mapped[str | None]
|
||||||
|
profile_pic: Mapped[str | None]
|
||||||
|
safe_conversation_starters: Mapped[str | None]
|
||||||
|
self_sufficiency_score: Mapped[int | None]
|
||||||
|
social_structure_style: Mapped[str | None]
|
||||||
|
ssn: Mapped[str | None]
|
||||||
|
suffix: Mapped[str | None]
|
||||||
|
timezone: Mapped[str | None]
|
||||||
|
topics_to_avoid: Mapped[str | None]
|
||||||
|
|
||||||
|
needs: Mapped[list[Need]] = relationship(
|
||||||
|
"Need",
|
||||||
|
secondary=ContactNeed.__table__,
|
||||||
|
back_populates="contacts",
|
||||||
|
)
|
||||||
|
|
||||||
|
related_to: Mapped[list[ContactRelationship]] = relationship(
|
||||||
|
"ContactRelationship",
|
||||||
|
foreign_keys=[ContactRelationship.contact_id],
|
||||||
|
cascade="all, delete-orphan",
|
||||||
|
)
|
||||||
|
related_from: Mapped[list[ContactRelationship]] = relationship(
|
||||||
|
"ContactRelationship",
|
||||||
|
foreign_keys=[ContactRelationship.related_contact_id],
|
||||||
|
cascade="all, delete-orphan",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class Need(TableBase):
|
||||||
|
"""Need/accommodation model (e.g., light sensitive, ADHD)."""
|
||||||
|
|
||||||
|
__tablename__ = "need"
|
||||||
|
|
||||||
|
name: Mapped[str]
|
||||||
|
description: Mapped[str | None]
|
||||||
|
|
||||||
|
contacts: Mapped[list[Contact]] = relationship(
|
||||||
|
"Contact",
|
||||||
|
secondary=ContactNeed.__table__,
|
||||||
|
back_populates="needs",
|
||||||
|
)
|
||||||
16
python/orm/signal_bot/__init__.py
Normal file
16
python/orm/signal_bot/__init__.py
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
"""Signal bot database ORM exports."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from python.orm.signal_bot.base import SignalBotBase, SignalBotTableBase, SignalBotTableBaseSmall
|
||||||
|
from python.orm.signal_bot.models import DeadLetterMessage, DeviceRole, RoleRecord, SignalDevice
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
"DeadLetterMessage",
|
||||||
|
"DeviceRole",
|
||||||
|
"RoleRecord",
|
||||||
|
"SignalBotBase",
|
||||||
|
"SignalBotTableBase",
|
||||||
|
"SignalBotTableBaseSmall",
|
||||||
|
"SignalDevice",
|
||||||
|
]
|
||||||
52
python/orm/signal_bot/base.py
Normal file
52
python/orm/signal_bot/base.py
Normal file
@@ -0,0 +1,52 @@
|
|||||||
|
"""Signal bot database ORM base."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
from sqlalchemy import DateTime, MetaData, SmallInteger, func
|
||||||
|
from sqlalchemy.ext.declarative import AbstractConcreteBase
|
||||||
|
from sqlalchemy.orm import DeclarativeBase, Mapped, mapped_column
|
||||||
|
|
||||||
|
from python.orm.common import NAMING_CONVENTION
|
||||||
|
|
||||||
|
|
||||||
|
class SignalBotBase(DeclarativeBase):
|
||||||
|
"""Base class for signal_bot database ORM models."""
|
||||||
|
|
||||||
|
schema_name = "main"
|
||||||
|
|
||||||
|
metadata = MetaData(
|
||||||
|
schema=schema_name,
|
||||||
|
naming_convention=NAMING_CONVENTION,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class _TableMixin:
|
||||||
|
"""Shared timestamp columns for all table bases."""
|
||||||
|
|
||||||
|
created: Mapped[datetime] = mapped_column(
|
||||||
|
DateTime(timezone=True),
|
||||||
|
server_default=func.now(),
|
||||||
|
)
|
||||||
|
updated: Mapped[datetime] = mapped_column(
|
||||||
|
DateTime(timezone=True),
|
||||||
|
server_default=func.now(),
|
||||||
|
onupdate=func.now(),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class SignalBotTableBaseSmall(_TableMixin, AbstractConcreteBase, SignalBotBase):
|
||||||
|
"""Table with SmallInteger primary key."""
|
||||||
|
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
id: Mapped[int] = mapped_column(SmallInteger, primary_key=True)
|
||||||
|
|
||||||
|
|
||||||
|
class SignalBotTableBase(_TableMixin, AbstractConcreteBase, SignalBotBase):
|
||||||
|
"""Table with Integer primary key."""
|
||||||
|
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
id: Mapped[int] = mapped_column(primary_key=True)
|
||||||
62
python/orm/signal_bot/models.py
Normal file
62
python/orm/signal_bot/models.py
Normal file
@@ -0,0 +1,62 @@
|
|||||||
|
"""Signal bot device, role, and dead letter ORM models."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
from sqlalchemy import DateTime, Enum, ForeignKey, SmallInteger, String, Text, UniqueConstraint
|
||||||
|
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||||
|
|
||||||
|
from python.orm.signal_bot.base import SignalBotTableBase, SignalBotTableBaseSmall
|
||||||
|
from python.signal_bot.models import MessageStatus, TrustLevel
|
||||||
|
|
||||||
|
|
||||||
|
class RoleRecord(SignalBotTableBaseSmall):
|
||||||
|
"""Lookup table for RBAC roles, keyed by smallint."""
|
||||||
|
|
||||||
|
__tablename__ = "role"
|
||||||
|
|
||||||
|
name: Mapped[str] = mapped_column(String(50), unique=True)
|
||||||
|
|
||||||
|
|
||||||
|
class DeviceRole(SignalBotTableBase):
|
||||||
|
"""Association between a device and a role."""
|
||||||
|
|
||||||
|
__tablename__ = "device_role"
|
||||||
|
__table_args__ = (
|
||||||
|
UniqueConstraint("device_id", "role_id", name="uq_device_role_device_role"),
|
||||||
|
{"schema": "main"},
|
||||||
|
)
|
||||||
|
|
||||||
|
device_id: Mapped[int] = mapped_column(ForeignKey("main.signal_device.id"))
|
||||||
|
role_id: Mapped[int] = mapped_column(SmallInteger, ForeignKey("main.role.id"))
|
||||||
|
|
||||||
|
|
||||||
|
class SignalDevice(SignalBotTableBase):
|
||||||
|
"""A Signal device tracked by phone number and safety number."""
|
||||||
|
|
||||||
|
__tablename__ = "signal_device"
|
||||||
|
|
||||||
|
phone_number: Mapped[str] = mapped_column(String(50), unique=True)
|
||||||
|
safety_number: Mapped[str | None]
|
||||||
|
trust_level: Mapped[TrustLevel] = mapped_column(
|
||||||
|
Enum(TrustLevel, name="trust_level", create_constraint=False, native_enum=False),
|
||||||
|
default=TrustLevel.UNVERIFIED,
|
||||||
|
)
|
||||||
|
last_seen: Mapped[datetime] = mapped_column(DateTime(timezone=True))
|
||||||
|
|
||||||
|
roles: Mapped[list[RoleRecord]] = relationship(secondary=DeviceRole.__table__)
|
||||||
|
|
||||||
|
|
||||||
|
class DeadLetterMessage(SignalBotTableBase):
|
||||||
|
"""A Signal message that failed processing and was sent to the dead letter queue."""
|
||||||
|
|
||||||
|
__tablename__ = "dead_letter_message"
|
||||||
|
|
||||||
|
source: Mapped[str]
|
||||||
|
message: Mapped[str] = mapped_column(Text)
|
||||||
|
received_at: Mapped[datetime] = mapped_column(DateTime(timezone=True))
|
||||||
|
status: Mapped[MessageStatus] = mapped_column(
|
||||||
|
Enum(MessageStatus, name="message_status", create_constraint=False, native_enum=False),
|
||||||
|
default=MessageStatus.UNPROCESSED,
|
||||||
|
)
|
||||||
1
python/orm/van_inventory/__init__.py
Normal file
1
python/orm/van_inventory/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Van inventory database ORM exports."""
|
||||||
39
python/orm/van_inventory/base.py
Normal file
39
python/orm/van_inventory/base.py
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
"""Van inventory database ORM base."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from datetime import datetime
|
||||||
|
|
||||||
|
from sqlalchemy import DateTime, MetaData, func
|
||||||
|
from sqlalchemy.ext.declarative import AbstractConcreteBase
|
||||||
|
from sqlalchemy.orm import DeclarativeBase, Mapped, mapped_column
|
||||||
|
|
||||||
|
from python.orm.common import NAMING_CONVENTION
|
||||||
|
|
||||||
|
|
||||||
|
class VanInventoryBase(DeclarativeBase):
|
||||||
|
"""Base class for van_inventory database ORM models."""
|
||||||
|
|
||||||
|
schema_name = "main"
|
||||||
|
|
||||||
|
metadata = MetaData(
|
||||||
|
schema=schema_name,
|
||||||
|
naming_convention=NAMING_CONVENTION,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class VanTableBase(AbstractConcreteBase, VanInventoryBase):
|
||||||
|
"""Abstract concrete base for van_inventory tables with IDs and timestamps."""
|
||||||
|
|
||||||
|
__abstract__ = True
|
||||||
|
|
||||||
|
id: Mapped[int] = mapped_column(primary_key=True)
|
||||||
|
created: Mapped[datetime] = mapped_column(
|
||||||
|
DateTime(timezone=True),
|
||||||
|
server_default=func.now(),
|
||||||
|
)
|
||||||
|
updated: Mapped[datetime] = mapped_column(
|
||||||
|
DateTime(timezone=True),
|
||||||
|
server_default=func.now(),
|
||||||
|
onupdate=func.now(),
|
||||||
|
)
|
||||||
46
python/orm/van_inventory/models.py
Normal file
46
python/orm/van_inventory/models.py
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
"""Van inventory ORM models."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
from sqlalchemy import ForeignKey, UniqueConstraint
|
||||||
|
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
||||||
|
|
||||||
|
from python.orm.van_inventory.base import VanTableBase
|
||||||
|
|
||||||
|
|
||||||
|
class Item(VanTableBase):
|
||||||
|
"""A food item in the van."""
|
||||||
|
|
||||||
|
__tablename__ = "items"
|
||||||
|
|
||||||
|
name: Mapped[str] = mapped_column(unique=True)
|
||||||
|
quantity: Mapped[float] = mapped_column(default=0)
|
||||||
|
unit: Mapped[str]
|
||||||
|
category: Mapped[str | None]
|
||||||
|
|
||||||
|
meal_ingredients: Mapped[list[MealIngredient]] = relationship(back_populates="item")
|
||||||
|
|
||||||
|
|
||||||
|
class Meal(VanTableBase):
|
||||||
|
"""A meal that can be made from items in the van."""
|
||||||
|
|
||||||
|
__tablename__ = "meals"
|
||||||
|
|
||||||
|
name: Mapped[str] = mapped_column(unique=True)
|
||||||
|
instructions: Mapped[str | None]
|
||||||
|
|
||||||
|
ingredients: Mapped[list[MealIngredient]] = relationship(back_populates="meal")
|
||||||
|
|
||||||
|
|
||||||
|
class MealIngredient(VanTableBase):
|
||||||
|
"""Links a meal to the items it requires, with quantities."""
|
||||||
|
|
||||||
|
__tablename__ = "meal_ingredients"
|
||||||
|
__table_args__ = (UniqueConstraint("meal_id", "item_id"),)
|
||||||
|
|
||||||
|
meal_id: Mapped[int] = mapped_column(ForeignKey("meals.id"))
|
||||||
|
item_id: Mapped[int] = mapped_column(ForeignKey("items.id"))
|
||||||
|
quantity_needed: Mapped[float]
|
||||||
|
|
||||||
|
meal: Mapped[Meal] = relationship(back_populates="ingredients")
|
||||||
|
item: Mapped[Item] = relationship(back_populates="meal_ingredients")
|
||||||
155
python/parallelize.py
Normal file
155
python/parallelize.py
Normal file
@@ -0,0 +1,155 @@
|
|||||||
|
"""Thing."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import logging
|
||||||
|
from concurrent.futures import ProcessPoolExecutor, ThreadPoolExecutor
|
||||||
|
from dataclasses import dataclass
|
||||||
|
from multiprocessing import cpu_count
|
||||||
|
from typing import TYPE_CHECKING, Any, Literal, TypeVar
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Callable, Mapping, Sequence
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
R = TypeVar("R")
|
||||||
|
|
||||||
|
modes = Literal["normal", "early_error"]
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass
|
||||||
|
class ExecutorResults[R]:
|
||||||
|
"""Dataclass to store the results and exceptions of the parallel execution."""
|
||||||
|
|
||||||
|
results: list[R]
|
||||||
|
exceptions: list[BaseException]
|
||||||
|
|
||||||
|
def __repr__(self) -> str:
|
||||||
|
"""Return a string representation of the object."""
|
||||||
|
return f"results={self.results} exceptions={self.exceptions}"
|
||||||
|
|
||||||
|
|
||||||
|
def _parallelize_base[R](
|
||||||
|
executor_type: type[ThreadPoolExecutor | ProcessPoolExecutor],
|
||||||
|
func: Callable[..., R],
|
||||||
|
kwargs_list: Sequence[Mapping[str, Any]],
|
||||||
|
max_workers: int | None,
|
||||||
|
progress_tracker: int | None,
|
||||||
|
mode: modes,
|
||||||
|
) -> ExecutorResults:
|
||||||
|
total_work = len(kwargs_list)
|
||||||
|
|
||||||
|
with executor_type(max_workers=max_workers) as executor:
|
||||||
|
futures = [executor.submit(func, **kwarg) for kwarg in kwargs_list]
|
||||||
|
|
||||||
|
results = []
|
||||||
|
exceptions = []
|
||||||
|
for index, future in enumerate(futures, 1):
|
||||||
|
if exception := future.exception():
|
||||||
|
logger.error(f"{future} raised {exception.__class__.__name__}")
|
||||||
|
exceptions.append(exception)
|
||||||
|
if mode == "early_error":
|
||||||
|
executor.shutdown(wait=False)
|
||||||
|
raise exception
|
||||||
|
continue
|
||||||
|
|
||||||
|
results.append(future.result())
|
||||||
|
|
||||||
|
if progress_tracker and index % progress_tracker == 0:
|
||||||
|
logger.info(f"Progress: {index}/{total_work}")
|
||||||
|
|
||||||
|
return ExecutorResults(results, exceptions)
|
||||||
|
|
||||||
|
|
||||||
|
def parallelize_thread[R](
|
||||||
|
func: Callable[..., R],
|
||||||
|
kwargs_list: Sequence[Mapping[str, Any]],
|
||||||
|
max_workers: int | None = None,
|
||||||
|
progress_tracker: int | None = None,
|
||||||
|
mode: modes = "normal",
|
||||||
|
) -> ExecutorResults:
|
||||||
|
"""Generic function to run a function with multiple arguments in threads.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
func (Callable[..., R]): Function to run in threads.
|
||||||
|
kwargs_list (Sequence[Mapping[str, Any]]): List of dictionaries with the arguments for the function.
|
||||||
|
max_workers (int, optional): Number of workers to use. Defaults to 8.
|
||||||
|
progress_tracker (int, optional): Number of tasks to complete before logging progress.
|
||||||
|
mode (modes, optional): Mode to use. Defaults to "normal".
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
tuple[list[R], list[Exception]]: List with the results and a list with the exceptions.
|
||||||
|
"""
|
||||||
|
return _parallelize_base(
|
||||||
|
executor_type=ThreadPoolExecutor,
|
||||||
|
func=func,
|
||||||
|
kwargs_list=kwargs_list,
|
||||||
|
max_workers=max_workers,
|
||||||
|
progress_tracker=progress_tracker,
|
||||||
|
mode=mode,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def parallelize_process[R](
|
||||||
|
func: Callable[..., R],
|
||||||
|
kwargs_list: Sequence[Mapping[str, Any]],
|
||||||
|
max_workers: int | None = None,
|
||||||
|
progress_tracker: int | None = None,
|
||||||
|
mode: modes = "normal",
|
||||||
|
) -> ExecutorResults:
|
||||||
|
"""Generic function to run a function with multiple arguments in process.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
func (Callable[..., R]): Function to run in process.
|
||||||
|
kwargs_list (Sequence[Mapping[str, Any]]): List of dictionaries with the arguments for the function.
|
||||||
|
max_workers (int, optional): Number of workers to use. Defaults to 4.
|
||||||
|
progress_tracker (int, optional): Number of tasks to complete before logging progress.
|
||||||
|
mode (modes, optional): Mode to use. Defaults to "normal".
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
tuple[list[R], list[Exception]]: List with the results and a list with the exceptions.
|
||||||
|
"""
|
||||||
|
if max_workers and max_workers > cpu_count():
|
||||||
|
error = f"max_workers must be less than or equal to {cpu_count()}"
|
||||||
|
raise RuntimeError(error)
|
||||||
|
|
||||||
|
return process_executor_unchecked(
|
||||||
|
func=func,
|
||||||
|
kwargs_list=kwargs_list,
|
||||||
|
max_workers=max_workers,
|
||||||
|
progress_tracker=progress_tracker,
|
||||||
|
mode=mode,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def process_executor_unchecked[R](
|
||||||
|
func: Callable[..., R],
|
||||||
|
kwargs_list: Sequence[Mapping[str, Any]],
|
||||||
|
max_workers: int | None,
|
||||||
|
progress_tracker: int | None,
|
||||||
|
mode: modes = "normal",
|
||||||
|
) -> ExecutorResults:
|
||||||
|
"""Generic function to run a function with multiple arguments in parallel.
|
||||||
|
|
||||||
|
Note: this function does not check if the number of workers is greater than the number of CPUs.
|
||||||
|
This can cause the system to become unresponsive.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
func (Callable[..., R]): Function to run in parallel.
|
||||||
|
kwargs_list (Sequence[Mapping[str, Any]]): List of dictionaries with the arguments for the function.
|
||||||
|
max_workers (int, optional): Number of workers to use. Defaults to 8.
|
||||||
|
progress_tracker (int, optional): Number of tasks to complete before logging progress.
|
||||||
|
mode (modes, optional): Mode to use. Defaults to "normal".
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
tuple[list[R], list[Exception]]: List with the results and a list with the exceptions.
|
||||||
|
"""
|
||||||
|
return _parallelize_base(
|
||||||
|
executor_type=ProcessPoolExecutor,
|
||||||
|
func=func,
|
||||||
|
kwargs_list=kwargs_list,
|
||||||
|
max_workers=max_workers,
|
||||||
|
progress_tracker=progress_tracker,
|
||||||
|
mode=mode,
|
||||||
|
)
|
||||||
1
python/signal_bot/__init__.py
Normal file
1
python/signal_bot/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Signal command and control bot."""
|
||||||
1
python/signal_bot/commands/__init__.py
Normal file
1
python/signal_bot/commands/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Signal bot commands."""
|
||||||
137
python/signal_bot/commands/inventory.py
Normal file
137
python/signal_bot/commands/inventory.py
Normal file
@@ -0,0 +1,137 @@
|
|||||||
|
"""Van inventory command — parse receipts and item lists via LLM, push to API."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import json
|
||||||
|
import logging
|
||||||
|
from typing import TYPE_CHECKING, Any
|
||||||
|
|
||||||
|
import httpx
|
||||||
|
|
||||||
|
from python.signal_bot.models import InventoryItem, InventoryUpdate
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from python.signal_bot.llm_client import LLMClient
|
||||||
|
from python.signal_bot.models import SignalMessage
|
||||||
|
from python.signal_bot.signal_client import SignalClient
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
SYSTEM_PROMPT = """\
|
||||||
|
You are an inventory assistant. Extract items from the input and return ONLY
|
||||||
|
a JSON array. Each element must have these fields:
|
||||||
|
- "name": item name (string)
|
||||||
|
- "quantity": numeric count or amount (default 1)
|
||||||
|
- "unit": unit of measure (e.g. "each", "lb", "oz", "gallon", "bag", "box")
|
||||||
|
- "category": category like "food", "tools", "supplies", etc.
|
||||||
|
- "notes": any extra detail (empty string if none)
|
||||||
|
|
||||||
|
Example output:
|
||||||
|
[{"name": "water bottles", "quantity": 6, "unit": "gallon", "category": "supplies", "notes": "1 gallon each"}]
|
||||||
|
|
||||||
|
Return ONLY the JSON array, no other text.\
|
||||||
|
"""
|
||||||
|
|
||||||
|
IMAGE_PROMPT = "Extract all items from this receipt or inventory photo."
|
||||||
|
TEXT_PROMPT = "Extract all items from this inventory list."
|
||||||
|
|
||||||
|
|
||||||
|
def parse_llm_response(raw: str) -> list[InventoryItem]:
|
||||||
|
"""Parse the LLM JSON response into InventoryItem list."""
|
||||||
|
text = raw.strip()
|
||||||
|
# Strip markdown code fences if present
|
||||||
|
if text.startswith("```"):
|
||||||
|
lines = text.split("\n")
|
||||||
|
lines = [line for line in lines if not line.startswith("```")]
|
||||||
|
text = "\n".join(lines)
|
||||||
|
|
||||||
|
items_data: list[dict[str, Any]] = json.loads(text)
|
||||||
|
return [InventoryItem.model_validate(item) for item in items_data]
|
||||||
|
|
||||||
|
|
||||||
|
def _upsert_item(api_url: str, item: InventoryItem) -> None:
|
||||||
|
"""Create or update an item via the van_inventory API.
|
||||||
|
|
||||||
|
Fetches existing items, and if one with the same name exists,
|
||||||
|
patches its quantity (summing). Otherwise creates a new item.
|
||||||
|
"""
|
||||||
|
base = api_url.rstrip("/")
|
||||||
|
response = httpx.get(f"{base}/api/items", timeout=10)
|
||||||
|
response.raise_for_status()
|
||||||
|
existing: list[dict[str, Any]] = response.json()
|
||||||
|
|
||||||
|
match = next((e for e in existing if e["name"].lower() == item.name.lower()), None)
|
||||||
|
|
||||||
|
if match:
|
||||||
|
new_qty = match["quantity"] + item.quantity
|
||||||
|
patch = {"quantity": new_qty}
|
||||||
|
if item.category:
|
||||||
|
patch["category"] = item.category
|
||||||
|
response = httpx.patch(f"{base}/api/items/{match['id']}", json=patch, timeout=10)
|
||||||
|
response.raise_for_status()
|
||||||
|
return
|
||||||
|
payload = {
|
||||||
|
"name": item.name,
|
||||||
|
"quantity": item.quantity,
|
||||||
|
"unit": item.unit,
|
||||||
|
"category": item.category or None,
|
||||||
|
}
|
||||||
|
response = httpx.post(f"{base}/api/items", json=payload, timeout=10)
|
||||||
|
response.raise_for_status()
|
||||||
|
|
||||||
|
|
||||||
|
def handle_inventory_update(
|
||||||
|
message: SignalMessage,
|
||||||
|
signal: SignalClient,
|
||||||
|
llm: LLMClient,
|
||||||
|
api_url: str,
|
||||||
|
) -> InventoryUpdate:
|
||||||
|
"""Process an inventory update from a Signal message.
|
||||||
|
|
||||||
|
Accepts either an image (receipt photo) or text list.
|
||||||
|
Uses the LLM to extract structured items, then pushes to the van_inventory API.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
logger.info(f"Processing inventory update from {message.source}")
|
||||||
|
if message.attachments:
|
||||||
|
image_data = signal.get_attachment(message.attachments[0])
|
||||||
|
raw_response = llm.chat(
|
||||||
|
IMAGE_PROMPT,
|
||||||
|
image_data=image_data,
|
||||||
|
system=SYSTEM_PROMPT,
|
||||||
|
)
|
||||||
|
source_type = "receipt_photo"
|
||||||
|
elif message.message.strip():
|
||||||
|
raw_response = llm.chat(
|
||||||
|
f"{TEXT_PROMPT}\n\n{message.message}",
|
||||||
|
system=SYSTEM_PROMPT,
|
||||||
|
)
|
||||||
|
source_type = "text_list"
|
||||||
|
else:
|
||||||
|
signal.reply(message, "Send a photo of a receipt or a text list of items to update inventory.")
|
||||||
|
return InventoryUpdate()
|
||||||
|
|
||||||
|
logger.info(f"{raw_response=}")
|
||||||
|
|
||||||
|
new_items = parse_llm_response(raw_response)
|
||||||
|
|
||||||
|
logger.info(f"{new_items=}")
|
||||||
|
|
||||||
|
for item in new_items:
|
||||||
|
_upsert_item(api_url, item)
|
||||||
|
|
||||||
|
summary = _format_summary(new_items)
|
||||||
|
signal.reply(message, f"Inventory updated with {len(new_items)} item(s):\n{summary}")
|
||||||
|
|
||||||
|
return InventoryUpdate(items=new_items, raw_response=raw_response, source_type=source_type)
|
||||||
|
|
||||||
|
except Exception:
|
||||||
|
logger.exception("Failed to process inventory update")
|
||||||
|
signal.reply(message, "Failed to process inventory update. Check logs for details.")
|
||||||
|
return InventoryUpdate()
|
||||||
|
|
||||||
|
|
||||||
|
def _format_summary(items: list[InventoryItem]) -> str:
|
||||||
|
"""Format items into a readable summary."""
|
||||||
|
lines = [f" - {item.name} x{item.quantity} {item.unit} [{item.category}]" for item in items]
|
||||||
|
return "\n".join(lines)
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user