mirror of
https://github.com/RichieCahill/dotfiles.git
synced 2026-04-21 14:49:10 -04:00
Compare commits
15 Commits
claude/add
...
feature/bu
| Author | SHA1 | Date | |
|---|---|---|---|
| 2b2d9135c5 | |||
| 81c0ce0928 | |||
| a16daf0e5e | |||
| 55e45d5e84 | |||
| dbd8d5eee3 | |||
| c89bc227f1 | |||
| 6ec355f567 | |||
| 78e51d7c12 | |||
| 96fc29e23a | |||
| 251089990e | |||
| 7364707277 | |||
| 6cefc275a8 | |||
| 74bb89f863 | |||
| ee581e6904 | |||
| c2182c7feb |
2
.github/workflows/build_systems.yml
vendored
2
.github/workflows/build_systems.yml
vendored
@@ -25,4 +25,4 @@ jobs:
|
|||||||
- name: Build default package
|
- name: Build default package
|
||||||
run: "nixos-rebuild build --flake ./#${{ matrix.system }}"
|
run: "nixos-rebuild build --flake ./#${{ matrix.system }}"
|
||||||
- name: copy to nix-cache
|
- name: copy to nix-cache
|
||||||
run: nix copy --accept-flake-config --to unix:///host-nix/var/nix/daemon-socket/socket .#nixosConfigurations.${{ matrix.system }}.config.system.build.toplevel
|
run: nix copy --to ssh://jeeves .#nixosConfigurations.${{ matrix.system }}.config.system.build.toplevel
|
||||||
|
|||||||
30
.github/workflows/fix_eval_warnings.yml
vendored
30
.github/workflows/fix_eval_warnings.yml
vendored
@@ -1,30 +0,0 @@
|
|||||||
name: fix_eval_warnings
|
|
||||||
on:
|
|
||||||
workflow_run:
|
|
||||||
workflows: ["build_systems"]
|
|
||||||
types: [completed]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
check-warnings:
|
|
||||||
if: >-
|
|
||||||
github.event.workflow_run.conclusion != 'cancelled' &&
|
|
||||||
github.event.workflow_run.head_branch == 'main' &&
|
|
||||||
(github.event.workflow_run.event == 'push' || github.event.workflow_run.event == 'schedule')
|
|
||||||
runs-on: self-hosted
|
|
||||||
permissions:
|
|
||||||
contents: write
|
|
||||||
pull-requests: write
|
|
||||||
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v4
|
|
||||||
|
|
||||||
- name: Fix eval warnings
|
|
||||||
env:
|
|
||||||
GH_TOKEN: ${{ secrets.GH_TOKEN_FOR_UPDATES }}
|
|
||||||
run: >-
|
|
||||||
nix develop .#devShells.x86_64-linux.default -c
|
|
||||||
python -m python.eval_warnings.main
|
|
||||||
--run-id "${{ github.event.workflow_run.id }}"
|
|
||||||
--repo "${{ github.repository }}"
|
|
||||||
--ollama-url "${{ secrets.OLLAMA_URL }}"
|
|
||||||
--run-url "${{ github.event.workflow_run.html_url }}"
|
|
||||||
4
.gitignore
vendored
4
.gitignore
vendored
@@ -165,7 +165,3 @@ test.*
|
|||||||
|
|
||||||
# syncthing
|
# syncthing
|
||||||
.stfolder
|
.stfolder
|
||||||
|
|
||||||
# Frontend build output
|
|
||||||
frontend/dist/
|
|
||||||
frontend/node_modules/
|
|
||||||
|
|||||||
19
.vscode/settings.json
vendored
19
.vscode/settings.json
vendored
@@ -77,11 +77,11 @@
|
|||||||
"esphome",
|
"esphome",
|
||||||
"extest",
|
"extest",
|
||||||
"fadvise",
|
"fadvise",
|
||||||
"fastfetch",
|
|
||||||
"fastforwardteam",
|
"fastforwardteam",
|
||||||
"FASTFOX",
|
"FASTFOX",
|
||||||
"ffmpegthumbnailer",
|
"ffmpegthumbnailer",
|
||||||
"filebot",
|
"filebot",
|
||||||
|
"filebrowser",
|
||||||
"fileroller",
|
"fileroller",
|
||||||
"findbar",
|
"findbar",
|
||||||
"Fira",
|
"Fira",
|
||||||
@@ -98,7 +98,6 @@
|
|||||||
"getch",
|
"getch",
|
||||||
"getmaxyx",
|
"getmaxyx",
|
||||||
"ghdeploy",
|
"ghdeploy",
|
||||||
"gitea",
|
|
||||||
"globalprivacycontrol",
|
"globalprivacycontrol",
|
||||||
"gparted",
|
"gparted",
|
||||||
"gtts",
|
"gtts",
|
||||||
@@ -117,9 +116,7 @@
|
|||||||
"httpchk",
|
"httpchk",
|
||||||
"hurlenko",
|
"hurlenko",
|
||||||
"hwloc",
|
"hwloc",
|
||||||
"ical",
|
|
||||||
"ignorelist",
|
"ignorelist",
|
||||||
"improv",
|
|
||||||
"INITDB",
|
"INITDB",
|
||||||
"iocharset",
|
"iocharset",
|
||||||
"ioit",
|
"ioit",
|
||||||
@@ -129,8 +126,6 @@
|
|||||||
"jnoortheen",
|
"jnoortheen",
|
||||||
"jsbc",
|
"jsbc",
|
||||||
"kagi",
|
"kagi",
|
||||||
"keyformat",
|
|
||||||
"keylocation",
|
|
||||||
"kuma",
|
"kuma",
|
||||||
"lazer",
|
"lazer",
|
||||||
"levelname",
|
"levelname",
|
||||||
@@ -167,6 +162,7 @@
|
|||||||
"mypy",
|
"mypy",
|
||||||
"ncdu",
|
"ncdu",
|
||||||
"nemo",
|
"nemo",
|
||||||
|
"neofetch",
|
||||||
"nerdfonts",
|
"nerdfonts",
|
||||||
"netdev",
|
"netdev",
|
||||||
"netdevs",
|
"netdevs",
|
||||||
@@ -229,10 +225,12 @@
|
|||||||
"pylint",
|
"pylint",
|
||||||
"pymetno",
|
"pymetno",
|
||||||
"pymodbus",
|
"pymodbus",
|
||||||
"pyopenweathermap",
|
|
||||||
"pyownet",
|
"pyownet",
|
||||||
"pytest",
|
"pytest",
|
||||||
"qalculate",
|
"qbit",
|
||||||
|
"qbittorrent",
|
||||||
|
"qbittorrentvpn",
|
||||||
|
"qbitvpn",
|
||||||
"quicksuggest",
|
"quicksuggest",
|
||||||
"radarr",
|
"radarr",
|
||||||
"readahead",
|
"readahead",
|
||||||
@@ -257,7 +255,6 @@
|
|||||||
"sessionmaker",
|
"sessionmaker",
|
||||||
"sessionstore",
|
"sessionstore",
|
||||||
"shellcheck",
|
"shellcheck",
|
||||||
"signalbot",
|
|
||||||
"signon",
|
"signon",
|
||||||
"Signons",
|
"Signons",
|
||||||
"skia",
|
"skia",
|
||||||
@@ -289,14 +286,11 @@
|
|||||||
"topstories",
|
"topstories",
|
||||||
"treefmt",
|
"treefmt",
|
||||||
"twimg",
|
"twimg",
|
||||||
"typedmonarchmoney",
|
|
||||||
"typer",
|
"typer",
|
||||||
"uaccess",
|
"uaccess",
|
||||||
"ubiquiti",
|
|
||||||
"ublock",
|
"ublock",
|
||||||
"uiprotect",
|
"uiprotect",
|
||||||
"uitour",
|
"uitour",
|
||||||
"unifi",
|
|
||||||
"unrar",
|
"unrar",
|
||||||
"unsubmitted",
|
"unsubmitted",
|
||||||
"uptimekuma",
|
"uptimekuma",
|
||||||
@@ -307,7 +301,6 @@
|
|||||||
"useragent",
|
"useragent",
|
||||||
"usernamehw",
|
"usernamehw",
|
||||||
"userprefs",
|
"userprefs",
|
||||||
"vaninventory",
|
|
||||||
"vfat",
|
"vfat",
|
||||||
"victron",
|
"victron",
|
||||||
"virt",
|
"virt",
|
||||||
|
|||||||
12
AGENTS.md
12
AGENTS.md
@@ -1,12 +0,0 @@
|
|||||||
## Dev environment tips
|
|
||||||
|
|
||||||
- use treefmt to format all files
|
|
||||||
- make python code ruff compliant
|
|
||||||
- use pytest to test python code
|
|
||||||
- always use the minimum amount of complexity
|
|
||||||
- if judgment calls are easy to reverse make them. if not ask me first
|
|
||||||
- Match existing code style.
|
|
||||||
- Use builtin helpers getenv() over os.environ.get.
|
|
||||||
- Prefer single-purpose functions over “do everything” helpers.
|
|
||||||
- Avoid compatibility branches like PG_USER and POSTGRESQL_URL unless requested.
|
|
||||||
- Keep helpers only if reused or they simplify the code otherwise inline.
|
|
||||||
25
README.md
25
README.md
@@ -1,26 +1 @@
|
|||||||
# dotfiles
|
# dotfiles
|
||||||
|
|
||||||
<!-- LINE-COUNT-START -->
|
|
||||||
This repo has **20,055** lines of technical debt.
|
|
||||||
|
|
||||||
| File Type | Lines | Percentage |
|
|
||||||
|-----------|------:|-----------:|
|
|
||||||
| .py | 11,441 | 57.0% |
|
|
||||||
| .nix | 4,471 | 22.3% |
|
|
||||||
| .yaml | 1,121 | 5.6% |
|
|
||||||
| .html | 1,009 | 5.0% |
|
|
||||||
| .json | 555 | 2.8% |
|
|
||||||
| .yml | 479 | 2.4% |
|
|
||||||
| .toml | 290 | 1.4% |
|
|
||||||
| .css | 212 | 1.1% |
|
|
||||||
| .gitignore | 199 | 1.0% |
|
|
||||||
| .md | 75 | 0.4% |
|
|
||||||
| .cfg | 73 | 0.4% |
|
|
||||||
| .sh | 48 | 0.2% |
|
|
||||||
| .mako | 36 | 0.2% |
|
|
||||||
| .LICENSE | 21 | 0.1% |
|
|
||||||
| .conf | 17 | 0.1% |
|
|
||||||
| .Gemfile | 4 | 0.0% |
|
|
||||||
| .svg | 3 | 0.0% |
|
|
||||||
| .new | 1 | 0.0% |
|
|
||||||
<!-- LINE-COUNT-END -->
|
|
||||||
|
|||||||
@@ -16,6 +16,7 @@
|
|||||||
./nh.nix
|
./nh.nix
|
||||||
./nix.nix
|
./nix.nix
|
||||||
./programs.nix
|
./programs.nix
|
||||||
|
./safe_reboot.nix
|
||||||
./ssh.nix
|
./ssh.nix
|
||||||
./snapshot_manager.nix
|
./snapshot_manager.nix
|
||||||
];
|
];
|
||||||
@@ -23,7 +24,7 @@
|
|||||||
boot = {
|
boot = {
|
||||||
tmp.useTmpfs = true;
|
tmp.useTmpfs = true;
|
||||||
kernelPackages = lib.mkDefault pkgs.linuxPackages_6_12;
|
kernelPackages = lib.mkDefault pkgs.linuxPackages_6_12;
|
||||||
zfs.package = lib.mkDefault pkgs.zfs_2_4;
|
zfs.package = lib.mkDefault pkgs.zfs_2_3;
|
||||||
};
|
};
|
||||||
|
|
||||||
hardware.enableRedistributableFirmware = true;
|
hardware.enableRedistributableFirmware = true;
|
||||||
@@ -49,6 +50,11 @@
|
|||||||
PYTHONPATH = "${inputs.self}/";
|
PYTHONPATH = "${inputs.self}/";
|
||||||
};
|
};
|
||||||
|
|
||||||
|
safe_reboot = {
|
||||||
|
enable = lib.mkDefault true;
|
||||||
|
datasetPrefix = "root_pool/";
|
||||||
|
};
|
||||||
|
|
||||||
zfs = {
|
zfs = {
|
||||||
trim.enable = lib.mkDefault true;
|
trim.enable = lib.mkDefault true;
|
||||||
autoScrub.enable = lib.mkDefault true;
|
autoScrub.enable = lib.mkDefault true;
|
||||||
|
|||||||
@@ -33,8 +33,6 @@ in
|
|||||||
];
|
];
|
||||||
warn-dirty = false;
|
warn-dirty = false;
|
||||||
flake-registry = ""; # disable global flake registries
|
flake-registry = ""; # disable global flake registries
|
||||||
connect-timeout = 10;
|
|
||||||
fallback = true;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
# Add each flake input as a registry and nix_path
|
# Add each flake input as a registry and nix_path
|
||||||
|
|||||||
56
common/global/safe_reboot.nix
Normal file
56
common/global/safe_reboot.nix
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
{
|
||||||
|
config,
|
||||||
|
inputs,
|
||||||
|
lib,
|
||||||
|
pkgs,
|
||||||
|
...
|
||||||
|
}:
|
||||||
|
let
|
||||||
|
cfg = config.services.safe_reboot;
|
||||||
|
python_command =
|
||||||
|
lib.escapeShellArgs (
|
||||||
|
[
|
||||||
|
"${pkgs.my_python}/bin/python"
|
||||||
|
"-m"
|
||||||
|
"python.tools.safe_reboot"
|
||||||
|
]
|
||||||
|
++ lib.optionals (cfg.drivePath != null) [ cfg.drivePath ]
|
||||||
|
++ [
|
||||||
|
"--dataset-prefix"
|
||||||
|
cfg.datasetPrefix
|
||||||
|
"--check-only"
|
||||||
|
]
|
||||||
|
);
|
||||||
|
in
|
||||||
|
{
|
||||||
|
options.services.safe_reboot = {
|
||||||
|
enable = lib.mkEnableOption "Safe reboot dataset/drive validation";
|
||||||
|
datasetPrefix = lib.mkOption {
|
||||||
|
type = lib.types.str;
|
||||||
|
default = "root_pool/";
|
||||||
|
description = "Dataset prefix that must have exec enabled before rebooting.";
|
||||||
|
};
|
||||||
|
drivePath = lib.mkOption {
|
||||||
|
type = lib.types.nullOr lib.types.str;
|
||||||
|
default = null;
|
||||||
|
description = "Drive path that must exist before rebooting. Set to null to skip.";
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
config = lib.mkIf cfg.enable {
|
||||||
|
systemd.services.safe-reboot-check = {
|
||||||
|
description = "Safe reboot validation";
|
||||||
|
before = [ "systemd-reboot.service" ];
|
||||||
|
wantedBy = [ "reboot.target" ];
|
||||||
|
partOf = [ "reboot.target" ];
|
||||||
|
path = [ pkgs.zfs ];
|
||||||
|
environment = {
|
||||||
|
PYTHONPATH = "${inputs.self}/";
|
||||||
|
};
|
||||||
|
serviceConfig = {
|
||||||
|
Type = "oneshot";
|
||||||
|
ExecStart = python_command;
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
}
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
{
|
|
||||||
nix.settings = {
|
|
||||||
trusted-substituters = [ "http://192.168.95.35:5000" ];
|
|
||||||
substituters = [ "http://192.168.95.35:5000/?priority=1&want-mass-query=true" ];
|
|
||||||
};
|
|
||||||
}
|
|
||||||
@@ -1,8 +1,8 @@
|
|||||||
{ pkgs, ... }:
|
{ pkgs, ... }:
|
||||||
{
|
{
|
||||||
boot = {
|
boot = {
|
||||||
kernelPackages = pkgs.linuxPackages_6_18;
|
kernelPackages = pkgs.linuxPackages_6_17;
|
||||||
zfs.package = pkgs.zfs_2_4;
|
zfs.package = pkgs.zfs_unstable;
|
||||||
};
|
};
|
||||||
|
|
||||||
hardware.bluetooth = {
|
hardware.bluetooth = {
|
||||||
|
|||||||
129
esphome/batteries.yml
Normal file
129
esphome/batteries.yml
Normal file
@@ -0,0 +1,129 @@
|
|||||||
|
esphome:
|
||||||
|
name: batteries
|
||||||
|
friendly_name: batteries
|
||||||
|
|
||||||
|
esp32:
|
||||||
|
board: esp32dev
|
||||||
|
framework:
|
||||||
|
type: arduino
|
||||||
|
|
||||||
|
logger:
|
||||||
|
|
||||||
|
api:
|
||||||
|
encryption:
|
||||||
|
key: !secret api_key
|
||||||
|
|
||||||
|
external_components:
|
||||||
|
- source: github://syssi/esphome-jk-bms@main
|
||||||
|
|
||||||
|
ota:
|
||||||
|
- platform: esphome
|
||||||
|
password: !secret ota_password
|
||||||
|
|
||||||
|
wifi:
|
||||||
|
ssid: !secret wifi_ssid
|
||||||
|
password: !secret wifi_password
|
||||||
|
|
||||||
|
captive_portal:
|
||||||
|
|
||||||
|
esp32_ble_tracker:
|
||||||
|
scan_parameters:
|
||||||
|
interval: 1100ms
|
||||||
|
window: 1100ms
|
||||||
|
active: true
|
||||||
|
|
||||||
|
ble_client:
|
||||||
|
- mac_address: "C8:47:80:29:0F:DB"
|
||||||
|
id: jk_ble0
|
||||||
|
- mac_address: "C8:47:80:37:9D:DD"
|
||||||
|
id: jk_ble1
|
||||||
|
|
||||||
|
jk_bms_ble:
|
||||||
|
- ble_client_id: jk_ble0
|
||||||
|
protocol_version: JK02_32S
|
||||||
|
throttle: 1s
|
||||||
|
id: jk_bms0
|
||||||
|
|
||||||
|
- ble_client_id: jk_ble1
|
||||||
|
protocol_version: JK02_32S
|
||||||
|
throttle: 1s
|
||||||
|
id: jk_bms1
|
||||||
|
|
||||||
|
sensor:
|
||||||
|
# BMS1 sensors
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms0
|
||||||
|
total_voltage:
|
||||||
|
name: "JK0 Total Voltage"
|
||||||
|
current:
|
||||||
|
name: "JK0 Current"
|
||||||
|
state_of_charge:
|
||||||
|
name: "JK0 SoC"
|
||||||
|
power:
|
||||||
|
name: "JK0 Power"
|
||||||
|
temperature_sensor_1:
|
||||||
|
name: "JK0 Temp 1"
|
||||||
|
temperature_sensor_2:
|
||||||
|
name: "JK0 Temp 2"
|
||||||
|
balancing:
|
||||||
|
name: "JK0 balancing"
|
||||||
|
charging_cycles:
|
||||||
|
name: "JK0 charging cycles"
|
||||||
|
total_runtime:
|
||||||
|
name: "JK0 total runtime"
|
||||||
|
balancing_current:
|
||||||
|
name: "JK0 balancing current"
|
||||||
|
|
||||||
|
# BMS2 sensors
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms1
|
||||||
|
total_voltage:
|
||||||
|
name: "JK1 Total Voltage"
|
||||||
|
current:
|
||||||
|
name: "JK1 Current"
|
||||||
|
state_of_charge:
|
||||||
|
name: "JK1 SoC"
|
||||||
|
power:
|
||||||
|
name: "Jk1 Power"
|
||||||
|
temperature_sensor_1:
|
||||||
|
name: "JK1 Temp 1"
|
||||||
|
temperature_sensor_2:
|
||||||
|
name: "Jk1 Temp 2"
|
||||||
|
balancing:
|
||||||
|
name: "JK1 balancing"
|
||||||
|
charging_cycles:
|
||||||
|
name: "JK1 charging cycles"
|
||||||
|
total_runtime:
|
||||||
|
name: "JK1 total runtime"
|
||||||
|
balancing_current:
|
||||||
|
name: "JK1 balancing current"
|
||||||
|
|
||||||
|
text_sensor:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms0
|
||||||
|
errors:
|
||||||
|
name: "JK0 Errors"
|
||||||
|
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms1
|
||||||
|
errors:
|
||||||
|
name: "JK1 Errors"
|
||||||
|
|
||||||
|
switch:
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms0
|
||||||
|
charging:
|
||||||
|
name: "JK0 Charging"
|
||||||
|
discharging:
|
||||||
|
name: "JK0 Discharging"
|
||||||
|
balancer:
|
||||||
|
name: "JK0 Balancing"
|
||||||
|
|
||||||
|
- platform: jk_bms_ble
|
||||||
|
jk_bms_ble_id: jk_bms1
|
||||||
|
charging:
|
||||||
|
name: "JK1 Charging"
|
||||||
|
discharging:
|
||||||
|
name: "JK1 Discharging"
|
||||||
|
balancer:
|
||||||
|
name: "JK1 Balancing"
|
||||||
@@ -1,132 +0,0 @@
|
|||||||
esphome:
|
|
||||||
name: batteries
|
|
||||||
friendly_name: batteries
|
|
||||||
|
|
||||||
esp32:
|
|
||||||
board: esp32dev
|
|
||||||
framework:
|
|
||||||
type: arduino
|
|
||||||
|
|
||||||
logger:
|
|
||||||
|
|
||||||
api:
|
|
||||||
encryption:
|
|
||||||
key: !secret api_key
|
|
||||||
|
|
||||||
external_components:
|
|
||||||
- source: github://syssi/esphome-jk-bms@main
|
|
||||||
|
|
||||||
ota:
|
|
||||||
- platform: esphome
|
|
||||||
password: !secret ota_password
|
|
||||||
|
|
||||||
wifi:
|
|
||||||
ssid: !secret wifi_ssid
|
|
||||||
password: !secret wifi_password
|
|
||||||
fast_connect: on
|
|
||||||
|
|
||||||
captive_portal:
|
|
||||||
|
|
||||||
esp32_ble_tracker:
|
|
||||||
scan_parameters:
|
|
||||||
interval: 1100ms
|
|
||||||
window: 1100ms
|
|
||||||
active: true
|
|
||||||
|
|
||||||
ble_client:
|
|
||||||
- mac_address: "C8:47:80:29:0F:DB"
|
|
||||||
id: jk_ble0
|
|
||||||
|
|
||||||
jk_bms_ble:
|
|
||||||
- ble_client_id: jk_ble0
|
|
||||||
protocol_version: JK02_32S
|
|
||||||
throttle: 1s
|
|
||||||
id: jk_bms0
|
|
||||||
|
|
||||||
button:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
retrieve_settings:
|
|
||||||
name: "JK0 retrieve settings"
|
|
||||||
retrieve_device_info:
|
|
||||||
name: "JK0 retrieve device info"
|
|
||||||
|
|
||||||
sensor:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
jk_bms_ble_id: jk_bms0
|
|
||||||
total_voltage:
|
|
||||||
name: "JK0 Total Voltage"
|
|
||||||
state_of_charge:
|
|
||||||
name: "JK0 SoC"
|
|
||||||
charging_power:
|
|
||||||
name: "JK0 charging power"
|
|
||||||
discharging_power:
|
|
||||||
name: "JK0 discharging power"
|
|
||||||
temperature_sensor_1:
|
|
||||||
name: "JK0 Temp 1"
|
|
||||||
temperature_sensor_2:
|
|
||||||
name: "JK0 Temp 2"
|
|
||||||
balancing:
|
|
||||||
name: "JK0 balancing"
|
|
||||||
total_runtime:
|
|
||||||
name: "JK0 total runtime"
|
|
||||||
balancing_current:
|
|
||||||
name: "JK0 balancing current"
|
|
||||||
delta_cell_voltage:
|
|
||||||
name: "JK0 cell delta voltage"
|
|
||||||
average_cell_voltage:
|
|
||||||
name: "JK0 cell average voltage"
|
|
||||||
cell_voltage_1:
|
|
||||||
name: "JK0 cell voltage 1"
|
|
||||||
cell_voltage_2:
|
|
||||||
name: "JK0 cell voltage 2"
|
|
||||||
cell_voltage_3:
|
|
||||||
name: "JK0 cell voltage 3"
|
|
||||||
cell_voltage_4:
|
|
||||||
name: "JK0 cell voltage 4"
|
|
||||||
cell_voltage_5:
|
|
||||||
name: "JK0 cell voltage 5"
|
|
||||||
cell_voltage_6:
|
|
||||||
name: "JK0 cell voltage 6"
|
|
||||||
cell_voltage_7:
|
|
||||||
name: "JK0 cell voltage 7"
|
|
||||||
cell_voltage_8:
|
|
||||||
name: "JK0 cell voltage 8"
|
|
||||||
cell_resistance_1:
|
|
||||||
name: "JK0 cell resistance 1"
|
|
||||||
cell_resistance_2:
|
|
||||||
name: "JK0 cell resistance 2"
|
|
||||||
cell_resistance_3:
|
|
||||||
name: "JK0 cell resistance 3"
|
|
||||||
cell_resistance_4:
|
|
||||||
name: "JK0 cell resistance 4"
|
|
||||||
cell_resistance_5:
|
|
||||||
name: "JK0 cell resistance 5"
|
|
||||||
cell_resistance_6:
|
|
||||||
name: "JK0 cell resistance 6"
|
|
||||||
cell_resistance_7:
|
|
||||||
name: "JK0 cell resistance 7"
|
|
||||||
cell_resistance_8:
|
|
||||||
name: "JK0 cell resistance 8"
|
|
||||||
total_charging_cycle_capacity:
|
|
||||||
name: "JK0 total charging cycle capacity"
|
|
||||||
|
|
||||||
text_sensor:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
jk_bms_ble_id: jk_bms0
|
|
||||||
errors:
|
|
||||||
name: "JK0 Errors"
|
|
||||||
|
|
||||||
switch:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
jk_bms_ble_id: jk_bms0
|
|
||||||
charging:
|
|
||||||
name: "JK0 Charging"
|
|
||||||
discharging:
|
|
||||||
name: "JK0 Discharging"
|
|
||||||
balancer:
|
|
||||||
name: "JK0 Balancing"
|
|
||||||
|
|
||||||
- platform: ble_client
|
|
||||||
ble_client_id: jk_ble0
|
|
||||||
name: "JK0 enable bluetooth connection"
|
|
||||||
id: ble_client_switch0
|
|
||||||
@@ -1,132 +0,0 @@
|
|||||||
esphome:
|
|
||||||
name: battery1
|
|
||||||
friendly_name: battery1
|
|
||||||
|
|
||||||
esp32:
|
|
||||||
board: esp32dev
|
|
||||||
framework:
|
|
||||||
type: arduino
|
|
||||||
|
|
||||||
logger:
|
|
||||||
|
|
||||||
api:
|
|
||||||
encryption:
|
|
||||||
key: !secret api_key
|
|
||||||
|
|
||||||
external_components:
|
|
||||||
- source: github://syssi/esphome-jk-bms@main
|
|
||||||
|
|
||||||
ota:
|
|
||||||
- platform: esphome
|
|
||||||
password: !secret ota_password
|
|
||||||
|
|
||||||
wifi:
|
|
||||||
ssid: !secret wifi_ssid
|
|
||||||
password: !secret wifi_password
|
|
||||||
fast_connect: on
|
|
||||||
|
|
||||||
captive_portal:
|
|
||||||
|
|
||||||
esp32_ble_tracker:
|
|
||||||
scan_parameters:
|
|
||||||
interval: 1100ms
|
|
||||||
window: 1100ms
|
|
||||||
active: true
|
|
||||||
|
|
||||||
ble_client:
|
|
||||||
- mac_address: "C8:47:80:37:9D:DD"
|
|
||||||
id: jk_ble1
|
|
||||||
|
|
||||||
jk_bms_ble:
|
|
||||||
- ble_client_id: jk_ble1
|
|
||||||
protocol_version: JK02_32S
|
|
||||||
throttle: 1s
|
|
||||||
id: jk_bms1
|
|
||||||
|
|
||||||
button:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
retrieve_settings:
|
|
||||||
name: "JK1 retrieve settings"
|
|
||||||
retrieve_device_info:
|
|
||||||
name: "JK1 retrieve device info"
|
|
||||||
|
|
||||||
sensor:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
jk_bms_ble_id: jk_bms1
|
|
||||||
total_voltage:
|
|
||||||
name: "JK1 Total Voltage"
|
|
||||||
state_of_charge:
|
|
||||||
name: "JK1 SoC"
|
|
||||||
charging_power:
|
|
||||||
name: "JK1 charging power"
|
|
||||||
discharging_power:
|
|
||||||
name: "JK1 discharging power"
|
|
||||||
temperature_sensor_1:
|
|
||||||
name: "JK1 Temp 1"
|
|
||||||
temperature_sensor_2:
|
|
||||||
name: "JK1 Temp 2"
|
|
||||||
balancing:
|
|
||||||
name: "JK1 balancing"
|
|
||||||
total_runtime:
|
|
||||||
name: "JK1 total runtime"
|
|
||||||
balancing_current:
|
|
||||||
name: "JK1 balancing current"
|
|
||||||
delta_cell_voltage:
|
|
||||||
name: "JK1 cell delta voltage"
|
|
||||||
average_cell_voltage:
|
|
||||||
name: "JK1 cell average voltage"
|
|
||||||
cell_voltage_1:
|
|
||||||
name: "JK1 cell voltage 1"
|
|
||||||
cell_voltage_2:
|
|
||||||
name: "JK1 cell voltage 2"
|
|
||||||
cell_voltage_3:
|
|
||||||
name: "JK1 cell voltage 3"
|
|
||||||
cell_voltage_4:
|
|
||||||
name: "JK1 cell voltage 4"
|
|
||||||
cell_voltage_5:
|
|
||||||
name: "JK1 cell voltage 5"
|
|
||||||
cell_voltage_6:
|
|
||||||
name: "JK1 cell voltage 6"
|
|
||||||
cell_voltage_7:
|
|
||||||
name: "JK1 cell voltage 7"
|
|
||||||
cell_voltage_8:
|
|
||||||
name: "JK1 cell voltage 8"
|
|
||||||
cell_resistance_1:
|
|
||||||
name: "JK1 cell resistance 1"
|
|
||||||
cell_resistance_2:
|
|
||||||
name: "JK1 cell resistance 2"
|
|
||||||
cell_resistance_3:
|
|
||||||
name: "JK1 cell resistance 3"
|
|
||||||
cell_resistance_4:
|
|
||||||
name: "JK1 cell resistance 4"
|
|
||||||
cell_resistance_5:
|
|
||||||
name: "JK1 cell resistance 5"
|
|
||||||
cell_resistance_6:
|
|
||||||
name: "JK1 cell resistance 6"
|
|
||||||
cell_resistance_7:
|
|
||||||
name: "JK1 cell resistance 7"
|
|
||||||
cell_resistance_8:
|
|
||||||
name: "JK1 cell resistance 8"
|
|
||||||
total_charging_cycle_capacity:
|
|
||||||
name: "JK1 total charging cycle capacity"
|
|
||||||
|
|
||||||
text_sensor:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
jk_bms_ble_id: jk_bms1
|
|
||||||
errors:
|
|
||||||
name: "JK1 Errors"
|
|
||||||
|
|
||||||
switch:
|
|
||||||
- platform: jk_bms_ble
|
|
||||||
jk_bms_ble_id: jk_bms1
|
|
||||||
charging:
|
|
||||||
name: "JK1 Charging"
|
|
||||||
discharging:
|
|
||||||
name: "JK1 Discharging"
|
|
||||||
balancer:
|
|
||||||
name: "JK1 Balancing"
|
|
||||||
|
|
||||||
- platform: ble_client
|
|
||||||
ble_client_id: jk_ble1
|
|
||||||
name: "JK1 enable bluetooth connection"
|
|
||||||
id: ble_client_switch0
|
|
||||||
@@ -1,48 +0,0 @@
|
|||||||
esphome:
|
|
||||||
name: "environment"
|
|
||||||
friendly_name: "environment"
|
|
||||||
|
|
||||||
esp32:
|
|
||||||
board: esp32dev
|
|
||||||
framework:
|
|
||||||
type: arduino
|
|
||||||
|
|
||||||
i2c:
|
|
||||||
sda: GPIO21
|
|
||||||
scl: GPIO22
|
|
||||||
scan: True
|
|
||||||
id: bus_a
|
|
||||||
|
|
||||||
sensor:
|
|
||||||
- platform: aht10
|
|
||||||
i2c_id: bus_a
|
|
||||||
address: 0x38
|
|
||||||
variant: AHT20
|
|
||||||
temperature:
|
|
||||||
name: "environment Temperature"
|
|
||||||
id: aht10_temperature
|
|
||||||
humidity:
|
|
||||||
name: "environment Humidity"
|
|
||||||
id: aht10_humidity
|
|
||||||
update_interval: 5s
|
|
||||||
|
|
||||||
web_server:
|
|
||||||
port: 80
|
|
||||||
|
|
||||||
logger:
|
|
||||||
level: DEBUG
|
|
||||||
|
|
||||||
api:
|
|
||||||
encryption:
|
|
||||||
key: !secret api_key
|
|
||||||
|
|
||||||
ota:
|
|
||||||
- platform: esphome
|
|
||||||
password: !secret ota_password
|
|
||||||
|
|
||||||
wifi:
|
|
||||||
ssid: !secret wifi_ssid
|
|
||||||
password: !secret wifi_password
|
|
||||||
fast_connect: on
|
|
||||||
|
|
||||||
captive_portal:
|
|
||||||
36
flake.lock
generated
36
flake.lock
generated
@@ -8,11 +8,11 @@
|
|||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"dir": "pkgs/firefox-addons",
|
"dir": "pkgs/firefox-addons",
|
||||||
"lastModified": 1773979456,
|
"lastModified": 1763697825,
|
||||||
"narHash": "sha256-9kBMJ5IvxqNlkkj/swmE8uK1Sc7TL/LIRUI958m7uBM=",
|
"narHash": "sha256-AgCCcVPOi1tuzuW5/StlwqBjRWSX62oL97qWuxrq5UA=",
|
||||||
"owner": "rycee",
|
"owner": "rycee",
|
||||||
"repo": "nur-expressions",
|
"repo": "nur-expressions",
|
||||||
"rev": "81e28f47ac18d9e89513929c77e711e657b64851",
|
"rev": "cefce78793603231be226fa77e7ad58e0e4899b8",
|
||||||
"type": "gitlab"
|
"type": "gitlab"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -29,11 +29,11 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1774007980,
|
"lastModified": 1763748372,
|
||||||
"narHash": "sha256-FOnZjElEI8pqqCvB6K/1JRHTE8o4rer8driivTpq2uo=",
|
"narHash": "sha256-AUc78Qv3sWir0hvbmfXoZ7Jzq9VVL97l+sP9Jgms+JU=",
|
||||||
"owner": "nix-community",
|
"owner": "nix-community",
|
||||||
"repo": "home-manager",
|
"repo": "home-manager",
|
||||||
"rev": "9670de2921812bc4e0452f6e3efd8c859696c183",
|
"rev": "d10a9b16b2a3ee28433f3d1c603f4e9f1fecb8e1",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -44,11 +44,11 @@
|
|||||||
},
|
},
|
||||||
"nixos-hardware": {
|
"nixos-hardware": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1774018263,
|
"lastModified": 1762847253,
|
||||||
"narHash": "sha256-HHYEwK1A22aSaxv2ibhMMkKvrDGKGlA/qObG4smrSqc=",
|
"narHash": "sha256-BWWnUUT01lPwCWUvS0p6Px5UOBFeXJ8jR+ZdLX8IbrU=",
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixos-hardware",
|
"repo": "nixos-hardware",
|
||||||
"rev": "2d4b4717b2534fad5c715968c1cece04a172b365",
|
"rev": "899dc449bc6428b9ee6b3b8f771ca2b0ef945ab9",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -60,11 +60,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs": {
|
"nixpkgs": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1773821835,
|
"lastModified": 1763421233,
|
||||||
"narHash": "sha256-TJ3lSQtW0E2JrznGVm8hOQGVpXjJyXY2guAxku2O9A4=",
|
"narHash": "sha256-Stk9ZYRkGrnnpyJ4eqt9eQtdFWRRIvMxpNRf4sIegnw=",
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "b40629efe5d6ec48dd1efba650c797ddbd39ace0",
|
"rev": "89c2b2330e733d6cdb5eae7b899326930c2c0648",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -76,11 +76,11 @@
|
|||||||
},
|
},
|
||||||
"nixpkgs-master": {
|
"nixpkgs-master": {
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1774051532,
|
"lastModified": 1763774007,
|
||||||
"narHash": "sha256-d3CGMweyYIcPuTj5BKq+1Lx4zwlgL31nVtN647tOZKo=",
|
"narHash": "sha256-PPeHfKA11P09kBkBD5pS3tIAFjnG5muHQnODQGTY87g=",
|
||||||
"owner": "nixos",
|
"owner": "nixos",
|
||||||
"repo": "nixpkgs",
|
"repo": "nixpkgs",
|
||||||
"rev": "8620c0b5cc8fbe76502442181be1d0514bc3a1b7",
|
"rev": "8a7cf7e9e18384533d9ecd0bfbcf475ac1dc497e",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
@@ -125,11 +125,11 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"locked": {
|
"locked": {
|
||||||
"lastModified": 1773889674,
|
"lastModified": 1763607916,
|
||||||
"narHash": "sha256-+ycaiVAk3MEshJTg35cBTUa0MizGiS+bgpYw/f8ohkg=",
|
"narHash": "sha256-VefBA1JWRXM929mBAFohFUtQJLUnEwZ2vmYUNkFnSjE=",
|
||||||
"owner": "Mic92",
|
"owner": "Mic92",
|
||||||
"repo": "sops-nix",
|
"repo": "sops-nix",
|
||||||
"rev": "29b6519f3e0780452bca0ac0be4584f04ac16cc5",
|
"rev": "877bb495a6f8faf0d89fc10bd142c4b7ed2bcc0b",
|
||||||
"type": "github"
|
"type": "github"
|
||||||
},
|
},
|
||||||
"original": {
|
"original": {
|
||||||
|
|||||||
24
frontend/.gitignore
vendored
24
frontend/.gitignore
vendored
@@ -1,24 +0,0 @@
|
|||||||
# Logs
|
|
||||||
logs
|
|
||||||
*.log
|
|
||||||
npm-debug.log*
|
|
||||||
yarn-debug.log*
|
|
||||||
yarn-error.log*
|
|
||||||
pnpm-debug.log*
|
|
||||||
lerna-debug.log*
|
|
||||||
|
|
||||||
node_modules
|
|
||||||
dist
|
|
||||||
dist-ssr
|
|
||||||
*.local
|
|
||||||
|
|
||||||
# Editor directories and files
|
|
||||||
.vscode/*
|
|
||||||
!.vscode/extensions.json
|
|
||||||
.idea
|
|
||||||
.DS_Store
|
|
||||||
*.suo
|
|
||||||
*.ntvs*
|
|
||||||
*.njsproj
|
|
||||||
*.sln
|
|
||||||
*.sw?
|
|
||||||
@@ -3,46 +3,38 @@
|
|||||||
# When applied, the stable nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.stable'
|
# When applied, the stable nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.stable'
|
||||||
stable = final: _prev: {
|
stable = final: _prev: {
|
||||||
stable = import inputs.nixpkgs-stable {
|
stable = import inputs.nixpkgs-stable {
|
||||||
system = final.stdenv.hostPlatform.system;
|
system = final.system;
|
||||||
config.allowUnfree = true;
|
config.allowUnfree = true;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
# When applied, the master nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.master'
|
# When applied, the master nixpkgs set (declared in the flake inputs) will be accessible through 'pkgs.master'
|
||||||
master = final: _prev: {
|
master = final: _prev: {
|
||||||
master = import inputs.nixpkgs-master {
|
master = import inputs.nixpkgs-master {
|
||||||
system = final.stdenv.hostPlatform.system;
|
system = final.system;
|
||||||
config.allowUnfree = true;
|
config.allowUnfree = true;
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
||||||
python-env = final: _prev: {
|
python-env = final: _prev: {
|
||||||
my_python = final.python314.withPackages (
|
my_python = final.python313.withPackages (
|
||||||
ps: with ps; [
|
ps: with ps; [
|
||||||
alembic
|
|
||||||
apprise
|
apprise
|
||||||
apscheduler
|
apscheduler
|
||||||
fastapi
|
|
||||||
fastapi-cli
|
|
||||||
httpx
|
|
||||||
mypy
|
mypy
|
||||||
polars
|
polars
|
||||||
psycopg
|
psycopg
|
||||||
pydantic
|
|
||||||
pyfakefs
|
pyfakefs
|
||||||
pytest
|
pytest
|
||||||
pytest-cov
|
pytest-cov
|
||||||
pytest-mock
|
pytest-mock
|
||||||
pytest-xdist
|
pytest-xdist
|
||||||
python-multipart
|
requests
|
||||||
ruff
|
ruff
|
||||||
scalene
|
scalene
|
||||||
sqlalchemy
|
sqlalchemy
|
||||||
sqlalchemy
|
|
||||||
tenacity
|
|
||||||
textual
|
textual
|
||||||
tinytuya
|
|
||||||
typer
|
typer
|
||||||
websockets
|
types-requests
|
||||||
]
|
]
|
||||||
);
|
);
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -7,25 +7,7 @@ requires-python = "~=3.13.0"
|
|||||||
readme = "README.md"
|
readme = "README.md"
|
||||||
license = "MIT"
|
license = "MIT"
|
||||||
# these dependencies are a best effort and aren't guaranteed to work
|
# these dependencies are a best effort and aren't guaranteed to work
|
||||||
# for up-to-date dependencies, see overlays/default.nix
|
dependencies = ["apprise", "apscheduler", "polars", "requests", "typer"]
|
||||||
dependencies = [
|
|
||||||
"alembic",
|
|
||||||
"apprise",
|
|
||||||
"apscheduler",
|
|
||||||
"httpx",
|
|
||||||
"python-multipart",
|
|
||||||
"polars",
|
|
||||||
"psycopg[binary]",
|
|
||||||
"pydantic",
|
|
||||||
"pyyaml",
|
|
||||||
"sqlalchemy",
|
|
||||||
"typer",
|
|
||||||
"websockets",
|
|
||||||
]
|
|
||||||
|
|
||||||
[project.scripts]
|
|
||||||
database = "python.database_cli:app"
|
|
||||||
van-inventory = "python.van_inventory.main:serve"
|
|
||||||
|
|
||||||
[dependency-groups]
|
[dependency-groups]
|
||||||
dev = [
|
dev = [
|
||||||
@@ -36,6 +18,7 @@ dev = [
|
|||||||
"pytest-xdist",
|
"pytest-xdist",
|
||||||
"pytest",
|
"pytest",
|
||||||
"ruff",
|
"ruff",
|
||||||
|
"types-requests",
|
||||||
]
|
]
|
||||||
|
|
||||||
[tool.ruff]
|
[tool.ruff]
|
||||||
@@ -55,38 +38,21 @@ lint.ignore = [
|
|||||||
[tool.ruff.lint.per-file-ignores]
|
[tool.ruff.lint.per-file-ignores]
|
||||||
|
|
||||||
"tests/**" = [
|
"tests/**" = [
|
||||||
"ANN", # (perm) type annotations not needed in tests
|
"S101", # (perm) pytest needs asserts
|
||||||
"D", # (perm) docstrings not needed in tests
|
|
||||||
"PLR2004", # (perm) magic values are fine in test assertions
|
|
||||||
"S101", # (perm) pytest needs asserts
|
|
||||||
]
|
]
|
||||||
"python/stuff/**" = [
|
"python/random/**" = [
|
||||||
"T201", # (perm) I don't care about print statements dir
|
"T201", # (perm) I don't care about print statements dir
|
||||||
]
|
]
|
||||||
"python/testing/**" = [
|
"python/testing/**" = [
|
||||||
"T201", # (perm) I don't care about print statements dir
|
"T201", # (perm) I don't care about print statements dir
|
||||||
"ERA001", # (perm) I don't care about print statements dir
|
"ERA001", # (perm) I don't care about print statements dir
|
||||||
]
|
]
|
||||||
|
|
||||||
"python/splendor/**" = [
|
"python/splendor/**" = [
|
||||||
"S311", # (perm) there is no security issue here
|
"S311", # (perm) there is no security issue here
|
||||||
"T201", # (perm) I don't care about print statements dir
|
"T201", # (perm) I don't care about print statements dir
|
||||||
"PLR2004", # (temps) need to think about this
|
"PLR2004", # (temps) need to think about this
|
||||||
]
|
]
|
||||||
"python/orm/**" = [
|
|
||||||
"TC003", # (perm) this creates issues because sqlalchemy uses these at runtime
|
|
||||||
]
|
|
||||||
"python/congress_tracker/**" = [
|
|
||||||
"TC003", # (perm) this creates issues because sqlalchemy uses these at runtime
|
|
||||||
]
|
|
||||||
"python/eval_warnings/**" = [
|
|
||||||
"S607", # (perm) gh and git are expected on PATH in the runner environment
|
|
||||||
]
|
|
||||||
"python/alembic/**" = [
|
|
||||||
"INP001", # (perm) this creates LSP issues for alembic
|
|
||||||
]
|
|
||||||
"python/signal_bot/**" = [
|
|
||||||
"D107", # (perm) class docstrings cover __init__
|
|
||||||
]
|
|
||||||
|
|
||||||
[tool.ruff.lint.pydocstyle]
|
[tool.ruff.lint.pydocstyle]
|
||||||
convention = "google"
|
convention = "google"
|
||||||
@@ -110,5 +76,4 @@ exclude_lines = [
|
|||||||
|
|
||||||
[tool.pytest.ini_options]
|
[tool.pytest.ini_options]
|
||||||
addopts = "-n auto -ra"
|
addopts = "-n auto -ra"
|
||||||
testpaths = ["tests"]
|
|
||||||
# --cov=system_tools --cov-report=term-missing --cov-report=xml --cov-report=html --cov-branch
|
# --cov=system_tools --cov-report=term-missing --cov-report=xml --cov-report=html --cov-branch
|
||||||
|
|||||||
@@ -1,121 +0,0 @@
|
|||||||
"""Alembic."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import logging
|
|
||||||
import sys
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import TYPE_CHECKING, Any, Literal
|
|
||||||
|
|
||||||
from alembic import context
|
|
||||||
from alembic.script import write_hooks
|
|
||||||
from sqlalchemy.schema import CreateSchema
|
|
||||||
|
|
||||||
from python.common import bash_wrapper
|
|
||||||
from python.orm.common import get_postgres_engine
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import MutableMapping
|
|
||||||
|
|
||||||
from sqlalchemy.orm import DeclarativeBase
|
|
||||||
|
|
||||||
config = context.config
|
|
||||||
|
|
||||||
base_class: type[DeclarativeBase] = config.attributes.get("base")
|
|
||||||
if base_class is None:
|
|
||||||
error = "No base class provided. Use the database CLI to run alembic commands."
|
|
||||||
raise RuntimeError(error)
|
|
||||||
|
|
||||||
target_metadata = base_class.metadata
|
|
||||||
logging.basicConfig(
|
|
||||||
level="DEBUG",
|
|
||||||
datefmt="%Y-%m-%dT%H:%M:%S%z",
|
|
||||||
format="%(asctime)s %(levelname)s %(filename)s:%(lineno)d - %(message)s",
|
|
||||||
handlers=[logging.StreamHandler(sys.stdout)],
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@write_hooks.register("dynamic_schema")
|
|
||||||
def dynamic_schema(filename: str, _options: dict[Any, Any]) -> None:
|
|
||||||
"""Dynamic schema."""
|
|
||||||
original_file = Path(filename).read_text()
|
|
||||||
schema_name = base_class.schema_name
|
|
||||||
dynamic_schema_file_part1 = original_file.replace(f"schema='{schema_name}'", "schema=schema")
|
|
||||||
dynamic_schema_file = dynamic_schema_file_part1.replace(f"'{schema_name}.", "f'{schema}.")
|
|
||||||
Path(filename).write_text(dynamic_schema_file)
|
|
||||||
|
|
||||||
|
|
||||||
@write_hooks.register("import_postgresql")
|
|
||||||
def import_postgresql(filename: str, _options: dict[Any, Any]) -> None:
|
|
||||||
"""Add postgresql dialect import when postgresql types are used."""
|
|
||||||
content = Path(filename).read_text()
|
|
||||||
if "postgresql." in content and "from sqlalchemy.dialects import postgresql" not in content:
|
|
||||||
content = content.replace(
|
|
||||||
"import sqlalchemy as sa\n",
|
|
||||||
"import sqlalchemy as sa\nfrom sqlalchemy.dialects import postgresql\n",
|
|
||||||
)
|
|
||||||
Path(filename).write_text(content)
|
|
||||||
|
|
||||||
|
|
||||||
@write_hooks.register("ruff")
|
|
||||||
def ruff_check_and_format(filename: str, _options: dict[Any, Any]) -> None:
|
|
||||||
"""Docstring for ruff_check_and_format."""
|
|
||||||
bash_wrapper(f"ruff check --fix {filename}")
|
|
||||||
bash_wrapper(f"ruff format {filename}")
|
|
||||||
|
|
||||||
|
|
||||||
def include_name(
|
|
||||||
name: str | None,
|
|
||||||
type_: Literal["schema", "table", "column", "index", "unique_constraint", "foreign_key_constraint"],
|
|
||||||
_parent_names: MutableMapping[Literal["schema_name", "table_name", "schema_qualified_table_name"], str | None],
|
|
||||||
) -> bool:
|
|
||||||
"""Filter tables to be included in the migration.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
name (str): The name of the table.
|
|
||||||
type_ (str): The type of the table.
|
|
||||||
_parent_names (MutableMapping): The names of the parent tables.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
bool: True if the table should be included, False otherwise.
|
|
||||||
|
|
||||||
"""
|
|
||||||
if type_ == "schema":
|
|
||||||
return name == target_metadata.schema
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def run_migrations_online() -> None:
|
|
||||||
"""Run migrations in 'online' mode.
|
|
||||||
|
|
||||||
In this scenario we need to create an Engine
|
|
||||||
and associate a connection with the context.
|
|
||||||
|
|
||||||
"""
|
|
||||||
env_prefix = config.attributes.get("env_prefix", "POSTGRES")
|
|
||||||
connectable = get_postgres_engine(name=env_prefix)
|
|
||||||
|
|
||||||
with connectable.connect() as connection:
|
|
||||||
schema = base_class.schema_name
|
|
||||||
if not connectable.dialect.has_schema(connection, schema):
|
|
||||||
answer = input(f"Schema {schema!r} does not exist. Create it? [y/N] ")
|
|
||||||
if answer.lower() != "y":
|
|
||||||
error = f"Schema {schema!r} does not exist. Exiting."
|
|
||||||
raise SystemExit(error)
|
|
||||||
connection.execute(CreateSchema(schema))
|
|
||||||
connection.commit()
|
|
||||||
|
|
||||||
context.configure(
|
|
||||||
connection=connection,
|
|
||||||
target_metadata=target_metadata,
|
|
||||||
include_schemas=True,
|
|
||||||
version_table_schema=schema,
|
|
||||||
include_name=include_name,
|
|
||||||
)
|
|
||||||
|
|
||||||
with context.begin_transaction():
|
|
||||||
context.run_migrations()
|
|
||||||
connection.commit()
|
|
||||||
|
|
||||||
|
|
||||||
run_migrations_online()
|
|
||||||
@@ -1,113 +0,0 @@
|
|||||||
"""created contact api.
|
|
||||||
|
|
||||||
Revision ID: edd7dd61a3d2
|
|
||||||
Revises:
|
|
||||||
Create Date: 2026-01-11 15:45:59.909266
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "edd7dd61a3d2"
|
|
||||||
down_revision: str | None = None
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"contact",
|
|
||||||
sa.Column("name", sa.String(), nullable=False),
|
|
||||||
sa.Column("age", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("bio", sa.String(), nullable=True),
|
|
||||||
sa.Column("current_job", sa.String(), nullable=True),
|
|
||||||
sa.Column("gender", sa.String(), nullable=True),
|
|
||||||
sa.Column("goals", sa.String(), nullable=True),
|
|
||||||
sa.Column("legal_name", sa.String(), nullable=True),
|
|
||||||
sa.Column("profile_pic", sa.String(), nullable=True),
|
|
||||||
sa.Column("safe_conversation_starters", sa.String(), nullable=True),
|
|
||||||
sa.Column("self_sufficiency_score", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("social_structure_style", sa.String(), nullable=True),
|
|
||||||
sa.Column("ssn", sa.String(), nullable=True),
|
|
||||||
sa.Column("suffix", sa.String(), nullable=True),
|
|
||||||
sa.Column("timezone", sa.String(), nullable=True),
|
|
||||||
sa.Column("topics_to_avoid", sa.String(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_contact")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"need",
|
|
||||||
sa.Column("name", sa.String(), nullable=False),
|
|
||||||
sa.Column("description", sa.String(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_need")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"contact_need",
|
|
||||||
sa.Column("contact_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("need_id", sa.Integer(), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["contact_id"],
|
|
||||||
[f"{schema}.contact.id"],
|
|
||||||
name=op.f("fk_contact_need_contact_id_contact"),
|
|
||||||
ondelete="CASCADE",
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["need_id"], [f"{schema}.need.id"], name=op.f("fk_contact_need_need_id_need"), ondelete="CASCADE"
|
|
||||||
),
|
|
||||||
sa.PrimaryKeyConstraint("contact_id", "need_id", name=op.f("pk_contact_need")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"contact_relationship",
|
|
||||||
sa.Column("contact_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("related_contact_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("relationship_type", sa.String(length=100), nullable=False),
|
|
||||||
sa.Column("closeness_weight", sa.Integer(), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["contact_id"],
|
|
||||||
[f"{schema}.contact.id"],
|
|
||||||
name=op.f("fk_contact_relationship_contact_id_contact"),
|
|
||||||
ondelete="CASCADE",
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["related_contact_id"],
|
|
||||||
[f"{schema}.contact.id"],
|
|
||||||
name=op.f("fk_contact_relationship_related_contact_id_contact"),
|
|
||||||
ondelete="CASCADE",
|
|
||||||
),
|
|
||||||
sa.PrimaryKeyConstraint("contact_id", "related_contact_id", name=op.f("pk_contact_relationship")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("contact_relationship", schema=schema)
|
|
||||||
op.drop_table("contact_need", schema=schema)
|
|
||||||
op.drop_table("need", schema=schema)
|
|
||||||
op.drop_table("contact", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,135 +0,0 @@
|
|||||||
"""add congress tracker tables.
|
|
||||||
|
|
||||||
Revision ID: 3f71565e38de
|
|
||||||
Revises: edd7dd61a3d2
|
|
||||||
Create Date: 2026-02-12 16:36:09.457303
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "3f71565e38de"
|
|
||||||
down_revision: str | None = "edd7dd61a3d2"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"bill",
|
|
||||||
sa.Column("congress", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("bill_type", sa.String(), nullable=False),
|
|
||||||
sa.Column("number", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("title", sa.String(), nullable=True),
|
|
||||||
sa.Column("title_short", sa.String(), nullable=True),
|
|
||||||
sa.Column("official_title", sa.String(), nullable=True),
|
|
||||||
sa.Column("status", sa.String(), nullable=True),
|
|
||||||
sa.Column("status_at", sa.Date(), nullable=True),
|
|
||||||
sa.Column("sponsor_bioguide_id", sa.String(), nullable=True),
|
|
||||||
sa.Column("subjects_top_term", sa.String(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_bill")),
|
|
||||||
sa.UniqueConstraint("congress", "bill_type", "number", name="uq_bill_congress_type_number"),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_index("ix_bill_congress", "bill", ["congress"], unique=False, schema=schema)
|
|
||||||
op.create_table(
|
|
||||||
"legislator",
|
|
||||||
sa.Column("bioguide_id", sa.Text(), nullable=False),
|
|
||||||
sa.Column("thomas_id", sa.String(), nullable=True),
|
|
||||||
sa.Column("lis_id", sa.String(), nullable=True),
|
|
||||||
sa.Column("govtrack_id", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("opensecrets_id", sa.String(), nullable=True),
|
|
||||||
sa.Column("fec_ids", sa.String(), nullable=True),
|
|
||||||
sa.Column("first_name", sa.String(), nullable=False),
|
|
||||||
sa.Column("last_name", sa.String(), nullable=False),
|
|
||||||
sa.Column("official_full_name", sa.String(), nullable=True),
|
|
||||||
sa.Column("nickname", sa.String(), nullable=True),
|
|
||||||
sa.Column("birthday", sa.Date(), nullable=True),
|
|
||||||
sa.Column("gender", sa.String(), nullable=True),
|
|
||||||
sa.Column("current_party", sa.String(), nullable=True),
|
|
||||||
sa.Column("current_state", sa.String(), nullable=True),
|
|
||||||
sa.Column("current_district", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("current_chamber", sa.String(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_legislator")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_index(op.f("ix_legislator_bioguide_id"), "legislator", ["bioguide_id"], unique=True, schema=schema)
|
|
||||||
op.create_table(
|
|
||||||
"vote",
|
|
||||||
sa.Column("congress", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("chamber", sa.String(), nullable=False),
|
|
||||||
sa.Column("session", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("number", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("vote_type", sa.String(), nullable=True),
|
|
||||||
sa.Column("question", sa.String(), nullable=True),
|
|
||||||
sa.Column("result", sa.String(), nullable=True),
|
|
||||||
sa.Column("result_text", sa.String(), nullable=True),
|
|
||||||
sa.Column("vote_date", sa.Date(), nullable=False),
|
|
||||||
sa.Column("yea_count", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("nay_count", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("not_voting_count", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("present_count", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("bill_id", sa.Integer(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(["bill_id"], [f"{schema}.bill.id"], name=op.f("fk_vote_bill_id_bill")),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_vote")),
|
|
||||||
sa.UniqueConstraint("congress", "chamber", "session", "number", name="uq_vote_congress_chamber_session_number"),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_index("ix_vote_congress_chamber", "vote", ["congress", "chamber"], unique=False, schema=schema)
|
|
||||||
op.create_index("ix_vote_date", "vote", ["vote_date"], unique=False, schema=schema)
|
|
||||||
op.create_table(
|
|
||||||
"vote_record",
|
|
||||||
sa.Column("vote_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("legislator_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("position", sa.String(), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["legislator_id"],
|
|
||||||
[f"{schema}.legislator.id"],
|
|
||||||
name=op.f("fk_vote_record_legislator_id_legislator"),
|
|
||||||
ondelete="CASCADE",
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["vote_id"], [f"{schema}.vote.id"], name=op.f("fk_vote_record_vote_id_vote"), ondelete="CASCADE"
|
|
||||||
),
|
|
||||||
sa.PrimaryKeyConstraint("vote_id", "legislator_id", name=op.f("pk_vote_record")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("vote_record", schema=schema)
|
|
||||||
op.drop_index("ix_vote_date", table_name="vote", schema=schema)
|
|
||||||
op.drop_index("ix_vote_congress_chamber", table_name="vote", schema=schema)
|
|
||||||
op.drop_table("vote", schema=schema)
|
|
||||||
op.drop_index(op.f("ix_legislator_bioguide_id"), table_name="legislator", schema=schema)
|
|
||||||
op.drop_table("legislator", schema=schema)
|
|
||||||
op.drop_index("ix_bill_congress", table_name="bill", schema=schema)
|
|
||||||
op.drop_table("bill", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,58 +0,0 @@
|
|||||||
"""adding SignalDevice for DeviceRegistry for signal bot.
|
|
||||||
|
|
||||||
Revision ID: 4c410c16e39c
|
|
||||||
Revises: 3f71565e38de
|
|
||||||
Create Date: 2026-03-09 14:51:24.228976
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
from sqlalchemy.dialects import postgresql
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "4c410c16e39c"
|
|
||||||
down_revision: str | None = "3f71565e38de"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"signal_device",
|
|
||||||
sa.Column("phone_number", sa.String(length=50), nullable=False),
|
|
||||||
sa.Column("safety_number", sa.String(), nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"trust_level",
|
|
||||||
postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column("last_seen", sa.DateTime(timezone=True), nullable=False),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_signal_device")),
|
|
||||||
sa.UniqueConstraint("phone_number", name=op.f("uq_signal_device_phone_number")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("signal_device", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,41 +0,0 @@
|
|||||||
"""fixed safety number logic.
|
|
||||||
|
|
||||||
Revision ID: 99fec682516c
|
|
||||||
Revises: 4c410c16e39c
|
|
||||||
Create Date: 2026-03-09 16:25:25.085806
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "99fec682516c"
|
|
||||||
down_revision: str | None = "4c410c16e39c"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.alter_column("signal_device", "safety_number", existing_type=sa.VARCHAR(), nullable=True, schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.alter_column("signal_device", "safety_number", existing_type=sa.VARCHAR(), nullable=False, schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,54 +0,0 @@
|
|||||||
"""add dead_letter_message table.
|
|
||||||
|
|
||||||
Revision ID: a1b2c3d4e5f6
|
|
||||||
Revises: 99fec682516c
|
|
||||||
Create Date: 2026-03-10 12:00:00.000000
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
from sqlalchemy.dialects import postgresql
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "a1b2c3d4e5f6"
|
|
||||||
down_revision: str | None = "99fec682516c"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
op.create_table(
|
|
||||||
"dead_letter_message",
|
|
||||||
sa.Column("source", sa.String(), nullable=False),
|
|
||||||
sa.Column("message", sa.Text(), nullable=False),
|
|
||||||
sa.Column("received_at", sa.DateTime(timezone=True), nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"status",
|
|
||||||
postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_dead_letter_message")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
op.drop_table("dead_letter_message", schema=schema)
|
|
||||||
op.execute(sa.text(f"DROP TYPE IF EXISTS {schema}.message_status"))
|
|
||||||
@@ -1,66 +0,0 @@
|
|||||||
"""adding roles to signal devices.
|
|
||||||
|
|
||||||
Revision ID: 2ef7ba690159
|
|
||||||
Revises: a1b2c3d4e5f6
|
|
||||||
Create Date: 2026-03-16 19:22:38.020350
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "2ef7ba690159"
|
|
||||||
down_revision: str | None = "a1b2c3d4e5f6"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"role",
|
|
||||||
sa.Column("name", sa.String(length=50), nullable=False),
|
|
||||||
sa.Column("id", sa.SmallInteger(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_role")),
|
|
||||||
sa.UniqueConstraint("name", name=op.f("uq_role_name")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"device_role",
|
|
||||||
sa.Column("device_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("role_id", sa.SmallInteger(), nullable=False),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["device_id"], [f"{schema}.signal_device.id"], name=op.f("fk_device_role_device_id_signal_device")
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(["role_id"], [f"{schema}.role.id"], name=op.f("fk_device_role_role_id_role")),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_device_role")),
|
|
||||||
sa.UniqueConstraint("device_id", "role_id", name="uq_device_role_device_role"),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("device_role", schema=schema)
|
|
||||||
op.drop_table("role", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,171 +0,0 @@
|
|||||||
"""seprating signal_bot database.
|
|
||||||
|
|
||||||
Revision ID: 6b275323f435
|
|
||||||
Revises: 2ef7ba690159
|
|
||||||
Create Date: 2026-03-18 08:34:28.785885
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
from sqlalchemy.dialects import postgresql
|
|
||||||
|
|
||||||
from python.orm import RichieBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "6b275323f435"
|
|
||||||
down_revision: str | None = "2ef7ba690159"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = RichieBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("device_role", schema=schema)
|
|
||||||
op.drop_table("signal_device", schema=schema)
|
|
||||||
op.drop_table("role", schema=schema)
|
|
||||||
op.drop_table("dead_letter_message", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"dead_letter_message",
|
|
||||||
sa.Column("source", sa.VARCHAR(), autoincrement=False, nullable=False),
|
|
||||||
sa.Column("message", sa.TEXT(), autoincrement=False, nullable=False),
|
|
||||||
sa.Column("received_at", postgresql.TIMESTAMP(timezone=True), autoincrement=False, nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"status",
|
|
||||||
postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"created",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column(
|
|
||||||
"updated",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_dead_letter_message")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"role",
|
|
||||||
sa.Column("name", sa.VARCHAR(length=50), autoincrement=False, nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"id",
|
|
||||||
sa.SMALLINT(),
|
|
||||||
server_default=sa.text(f"nextval('{schema}.role_id_seq'::regclass)"),
|
|
||||||
autoincrement=True,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column(
|
|
||||||
"created",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column(
|
|
||||||
"updated",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_role")),
|
|
||||||
sa.UniqueConstraint(
|
|
||||||
"name", name=op.f("uq_role_name"), postgresql_include=[], postgresql_nulls_not_distinct=False
|
|
||||||
),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"signal_device",
|
|
||||||
sa.Column("phone_number", sa.VARCHAR(length=50), autoincrement=False, nullable=False),
|
|
||||||
sa.Column("safety_number", sa.VARCHAR(), autoincrement=False, nullable=True),
|
|
||||||
sa.Column(
|
|
||||||
"trust_level",
|
|
||||||
postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column("last_seen", postgresql.TIMESTAMP(timezone=True), autoincrement=False, nullable=False),
|
|
||||||
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"created",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column(
|
|
||||||
"updated",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_signal_device")),
|
|
||||||
sa.UniqueConstraint(
|
|
||||||
"phone_number",
|
|
||||||
name=op.f("uq_signal_device_phone_number"),
|
|
||||||
postgresql_include=[],
|
|
||||||
postgresql_nulls_not_distinct=False,
|
|
||||||
),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"device_role",
|
|
||||||
sa.Column("device_id", sa.INTEGER(), autoincrement=False, nullable=False),
|
|
||||||
sa.Column("role_id", sa.SMALLINT(), autoincrement=False, nullable=False),
|
|
||||||
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"created",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column(
|
|
||||||
"updated",
|
|
||||||
postgresql.TIMESTAMP(timezone=True),
|
|
||||||
server_default=sa.text("now()"),
|
|
||||||
autoincrement=False,
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["device_id"], [f"{schema}.signal_device.id"], name=op.f("fk_device_role_device_id_signal_device")
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(["role_id"], [f"{schema}.role.id"], name=op.f("fk_device_role_role_id_role")),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_device_role")),
|
|
||||||
sa.UniqueConstraint(
|
|
||||||
"device_id",
|
|
||||||
"role_id",
|
|
||||||
name=op.f("uq_device_role_device_role"),
|
|
||||||
postgresql_include=[],
|
|
||||||
postgresql_nulls_not_distinct=False,
|
|
||||||
),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,36 +0,0 @@
|
|||||||
"""${message}.
|
|
||||||
|
|
||||||
Revision ID: ${up_revision}
|
|
||||||
Revises: ${down_revision | comma,n}
|
|
||||||
Create Date: ${create_date}
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
|
|
||||||
from alembic import op
|
|
||||||
from python.orm import ${config.attributes["base"].__name__}
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = ${repr(up_revision)}
|
|
||||||
down_revision: str | None = ${repr(down_revision)}
|
|
||||||
branch_labels: str | Sequence[str] | None = ${repr(branch_labels)}
|
|
||||||
depends_on: str | Sequence[str] | None = ${repr(depends_on)}
|
|
||||||
|
|
||||||
schema=${config.attributes["base"].__name__}.schema_name
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
${upgrades if upgrades else "pass"}
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
${downgrades if downgrades else "pass"}
|
|
||||||
@@ -1,100 +0,0 @@
|
|||||||
"""seprating signal_bot database.
|
|
||||||
|
|
||||||
Revision ID: 6eaf696e07a5
|
|
||||||
Revises:
|
|
||||||
Create Date: 2026-03-17 21:35:37.612672
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
from sqlalchemy.dialects import postgresql
|
|
||||||
|
|
||||||
from python.orm import SignalBotBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "6eaf696e07a5"
|
|
||||||
down_revision: str | None = None
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = SignalBotBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"dead_letter_message",
|
|
||||||
sa.Column("source", sa.String(), nullable=False),
|
|
||||||
sa.Column("message", sa.Text(), nullable=False),
|
|
||||||
sa.Column("received_at", sa.DateTime(timezone=True), nullable=False),
|
|
||||||
sa.Column(
|
|
||||||
"status", postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema), nullable=False
|
|
||||||
),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_dead_letter_message")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"role",
|
|
||||||
sa.Column("name", sa.String(length=50), nullable=False),
|
|
||||||
sa.Column("id", sa.SmallInteger(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_role")),
|
|
||||||
sa.UniqueConstraint("name", name=op.f("uq_role_name")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"signal_device",
|
|
||||||
sa.Column("phone_number", sa.String(length=50), nullable=False),
|
|
||||||
sa.Column("safety_number", sa.String(), nullable=True),
|
|
||||||
sa.Column(
|
|
||||||
"trust_level",
|
|
||||||
postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
|
||||||
nullable=False,
|
|
||||||
),
|
|
||||||
sa.Column("last_seen", sa.DateTime(timezone=True), nullable=False),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_signal_device")),
|
|
||||||
sa.UniqueConstraint("phone_number", name=op.f("uq_signal_device_phone_number")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"device_role",
|
|
||||||
sa.Column("device_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("role_id", sa.SmallInteger(), nullable=False),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(
|
|
||||||
["device_id"], [f"{schema}.signal_device.id"], name=op.f("fk_device_role_device_id_signal_device")
|
|
||||||
),
|
|
||||||
sa.ForeignKeyConstraint(["role_id"], [f"{schema}.role.id"], name=op.f("fk_device_role_role_id_role")),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_device_role")),
|
|
||||||
sa.UniqueConstraint("device_id", "role_id", name="uq_device_role_device_role"),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("device_role", schema=schema)
|
|
||||||
op.drop_table("signal_device", schema=schema)
|
|
||||||
op.drop_table("role", schema=schema)
|
|
||||||
op.drop_table("dead_letter_message", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,72 +0,0 @@
|
|||||||
"""test.
|
|
||||||
|
|
||||||
Revision ID: 66bdd532bcab
|
|
||||||
Revises: 6eaf696e07a5
|
|
||||||
Create Date: 2026-03-18 19:21:14.561568
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
from sqlalchemy.dialects import postgresql
|
|
||||||
|
|
||||||
from python.orm import SignalBotBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "66bdd532bcab"
|
|
||||||
down_revision: str | None = "6eaf696e07a5"
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = SignalBotBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.alter_column(
|
|
||||||
"dead_letter_message",
|
|
||||||
"status",
|
|
||||||
existing_type=postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
|
||||||
type_=sa.Enum("UNPROCESSED", "PROCESSED", name="message_status", native_enum=False),
|
|
||||||
existing_nullable=False,
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.alter_column(
|
|
||||||
"signal_device",
|
|
||||||
"trust_level",
|
|
||||||
existing_type=postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
|
||||||
type_=sa.Enum("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", native_enum=False),
|
|
||||||
existing_nullable=False,
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.alter_column(
|
|
||||||
"signal_device",
|
|
||||||
"trust_level",
|
|
||||||
existing_type=sa.Enum("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", native_enum=False),
|
|
||||||
type_=postgresql.ENUM("VERIFIED", "UNVERIFIED", "BLOCKED", name="trust_level", schema=schema),
|
|
||||||
existing_nullable=False,
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.alter_column(
|
|
||||||
"dead_letter_message",
|
|
||||||
"status",
|
|
||||||
existing_type=sa.Enum("UNPROCESSED", "PROCESSED", name="message_status", native_enum=False),
|
|
||||||
type_=postgresql.ENUM("UNPROCESSED", "PROCESSED", name="message_status", schema=schema),
|
|
||||||
existing_nullable=False,
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1,80 +0,0 @@
|
|||||||
"""starting van invintory.
|
|
||||||
|
|
||||||
Revision ID: 15e733499804
|
|
||||||
Revises:
|
|
||||||
Create Date: 2026-03-08 00:18:20.759720
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
import sqlalchemy as sa
|
|
||||||
from alembic import op
|
|
||||||
|
|
||||||
from python.orm import VanInventoryBase
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Sequence
|
|
||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
|
||||||
revision: str = "15e733499804"
|
|
||||||
down_revision: str | None = None
|
|
||||||
branch_labels: str | Sequence[str] | None = None
|
|
||||||
depends_on: str | Sequence[str] | None = None
|
|
||||||
|
|
||||||
schema = VanInventoryBase.schema_name
|
|
||||||
|
|
||||||
|
|
||||||
def upgrade() -> None:
|
|
||||||
"""Upgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.create_table(
|
|
||||||
"items",
|
|
||||||
sa.Column("name", sa.String(), nullable=False),
|
|
||||||
sa.Column("quantity", sa.Float(), nullable=False),
|
|
||||||
sa.Column("unit", sa.String(), nullable=False),
|
|
||||||
sa.Column("category", sa.String(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_items")),
|
|
||||||
sa.UniqueConstraint("name", name=op.f("uq_items_name")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"meals",
|
|
||||||
sa.Column("name", sa.String(), nullable=False),
|
|
||||||
sa.Column("instructions", sa.String(), nullable=True),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_meals")),
|
|
||||||
sa.UniqueConstraint("name", name=op.f("uq_meals_name")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
op.create_table(
|
|
||||||
"meal_ingredients",
|
|
||||||
sa.Column("meal_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("item_id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("quantity_needed", sa.Float(), nullable=False),
|
|
||||||
sa.Column("id", sa.Integer(), nullable=False),
|
|
||||||
sa.Column("created", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.Column("updated", sa.DateTime(timezone=True), server_default=sa.text("now()"), nullable=False),
|
|
||||||
sa.ForeignKeyConstraint(["item_id"], [f"{schema}.items.id"], name=op.f("fk_meal_ingredients_item_id_items")),
|
|
||||||
sa.ForeignKeyConstraint(["meal_id"], [f"{schema}.meals.id"], name=op.f("fk_meal_ingredients_meal_id_meals")),
|
|
||||||
sa.PrimaryKeyConstraint("id", name=op.f("pk_meal_ingredients")),
|
|
||||||
sa.UniqueConstraint("meal_id", "item_id", name=op.f("uq_meal_ingredients_meal_id")),
|
|
||||||
schema=schema,
|
|
||||||
)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
|
|
||||||
|
|
||||||
def downgrade() -> None:
|
|
||||||
"""Downgrade."""
|
|
||||||
# ### commands auto generated by Alembic - please adjust! ###
|
|
||||||
op.drop_table("meal_ingredients", schema=schema)
|
|
||||||
op.drop_table("meals", schema=schema)
|
|
||||||
op.drop_table("items", schema=schema)
|
|
||||||
# ### end Alembic commands ###
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""FastAPI applications."""
|
|
||||||
@@ -1,16 +0,0 @@
|
|||||||
"""FastAPI dependencies."""
|
|
||||||
|
|
||||||
from collections.abc import Iterator
|
|
||||||
from typing import Annotated
|
|
||||||
|
|
||||||
from fastapi import Depends, Request
|
|
||||||
from sqlalchemy.orm import Session
|
|
||||||
|
|
||||||
|
|
||||||
def get_db(request: Request) -> Iterator[Session]:
|
|
||||||
"""Get database session from app state."""
|
|
||||||
with Session(request.app.state.engine) as session:
|
|
||||||
yield session
|
|
||||||
|
|
||||||
|
|
||||||
DbSession = Annotated[Session, Depends(get_db)]
|
|
||||||
@@ -1,52 +0,0 @@
|
|||||||
"""FastAPI interface for Contact database."""
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from collections.abc import AsyncIterator
|
|
||||||
from contextlib import asynccontextmanager
|
|
||||||
from typing import Annotated
|
|
||||||
|
|
||||||
import typer
|
|
||||||
import uvicorn
|
|
||||||
from fastapi import FastAPI
|
|
||||||
|
|
||||||
from python.api.middleware import ZstdMiddleware
|
|
||||||
from python.api.routers import contact_router, views_router
|
|
||||||
from python.common import configure_logger
|
|
||||||
from python.orm.common import get_postgres_engine
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def create_app() -> FastAPI:
|
|
||||||
"""Create and configure the FastAPI application."""
|
|
||||||
|
|
||||||
@asynccontextmanager
|
|
||||||
async def lifespan(app: FastAPI) -> AsyncIterator[None]:
|
|
||||||
"""Manage application lifespan."""
|
|
||||||
app.state.engine = get_postgres_engine()
|
|
||||||
yield
|
|
||||||
app.state.engine.dispose()
|
|
||||||
|
|
||||||
app = FastAPI(title="Contact Database API", lifespan=lifespan)
|
|
||||||
app.add_middleware(ZstdMiddleware)
|
|
||||||
|
|
||||||
app.include_router(contact_router)
|
|
||||||
app.include_router(views_router)
|
|
||||||
|
|
||||||
return app
|
|
||||||
|
|
||||||
|
|
||||||
def serve(
|
|
||||||
host: Annotated[str, typer.Option("--host", "-h", help="Host to bind to")],
|
|
||||||
port: Annotated[int, typer.Option("--port", "-p", help="Port to bind to")] = 8000,
|
|
||||||
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
|
||||||
) -> None:
|
|
||||||
"""Start the Contact API server."""
|
|
||||||
configure_logger(log_level)
|
|
||||||
|
|
||||||
app = create_app()
|
|
||||||
uvicorn.run(app, host=host, port=port)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
typer.run(serve)
|
|
||||||
@@ -1,49 +0,0 @@
|
|||||||
"""Middleware for the FastAPI application."""
|
|
||||||
|
|
||||||
from compression import zstd
|
|
||||||
from starlette.middleware.base import BaseHTTPMiddleware, RequestResponseEndpoint
|
|
||||||
from starlette.requests import Request
|
|
||||||
from starlette.responses import Response
|
|
||||||
|
|
||||||
MINIMUM_RESPONSE_SIZE = 500
|
|
||||||
|
|
||||||
|
|
||||||
class ZstdMiddleware(BaseHTTPMiddleware):
|
|
||||||
"""Middleware that compresses responses with zstd when the client supports it."""
|
|
||||||
|
|
||||||
async def dispatch(self, request: Request, call_next: RequestResponseEndpoint) -> Response:
|
|
||||||
"""Compress the response with zstd if the client accepts it."""
|
|
||||||
accepted_encodings = request.headers.get("accept-encoding", "")
|
|
||||||
if "zstd" not in accepted_encodings:
|
|
||||||
return await call_next(request)
|
|
||||||
|
|
||||||
response = await call_next(request)
|
|
||||||
|
|
||||||
if response.headers.get("content-encoding") or "text/event-stream" in response.headers.get("content-type", ""):
|
|
||||||
return response
|
|
||||||
|
|
||||||
body = b""
|
|
||||||
async for chunk in response.body_iterator:
|
|
||||||
body += chunk if isinstance(chunk, bytes) else chunk.encode()
|
|
||||||
|
|
||||||
if len(body) < MINIMUM_RESPONSE_SIZE:
|
|
||||||
return Response(
|
|
||||||
content=body,
|
|
||||||
status_code=response.status_code,
|
|
||||||
headers=dict(response.headers),
|
|
||||||
media_type=response.media_type,
|
|
||||||
)
|
|
||||||
|
|
||||||
compressed = zstd.compress(body)
|
|
||||||
|
|
||||||
headers = dict(response.headers)
|
|
||||||
headers["content-encoding"] = "zstd"
|
|
||||||
headers["content-length"] = str(len(compressed))
|
|
||||||
headers.pop("transfer-encoding", None)
|
|
||||||
|
|
||||||
return Response(
|
|
||||||
content=compressed,
|
|
||||||
status_code=response.status_code,
|
|
||||||
headers=headers,
|
|
||||||
media_type=response.media_type,
|
|
||||||
)
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
"""API routers."""
|
|
||||||
|
|
||||||
from python.api.routers.contact import router as contact_router
|
|
||||||
from python.api.routers.views import router as views_router
|
|
||||||
|
|
||||||
__all__ = ["contact_router", "views_router"]
|
|
||||||
@@ -1,481 +0,0 @@
|
|||||||
"""Contact API router."""
|
|
||||||
|
|
||||||
from pathlib import Path
|
|
||||||
|
|
||||||
from fastapi import APIRouter, HTTPException, Request
|
|
||||||
from fastapi.responses import HTMLResponse
|
|
||||||
from fastapi.templating import Jinja2Templates
|
|
||||||
from pydantic import BaseModel
|
|
||||||
from sqlalchemy import select
|
|
||||||
from sqlalchemy.orm import selectinload
|
|
||||||
|
|
||||||
from python.api.dependencies import DbSession
|
|
||||||
from python.orm.richie.contact import Contact, ContactRelationship, Need, RelationshipType
|
|
||||||
|
|
||||||
TEMPLATES_DIR = Path(__file__).parent.parent / "templates"
|
|
||||||
templates = Jinja2Templates(directory=TEMPLATES_DIR)
|
|
||||||
|
|
||||||
|
|
||||||
def _is_htmx(request: Request) -> bool:
|
|
||||||
"""Check if the request is from HTMX."""
|
|
||||||
return request.headers.get("HX-Request") == "true"
|
|
||||||
|
|
||||||
|
|
||||||
class NeedBase(BaseModel):
|
|
||||||
"""Base schema for Need."""
|
|
||||||
|
|
||||||
name: str
|
|
||||||
description: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class NeedCreate(NeedBase):
|
|
||||||
"""Schema for creating a Need."""
|
|
||||||
|
|
||||||
|
|
||||||
class NeedResponse(NeedBase):
|
|
||||||
"""Schema for Need response."""
|
|
||||||
|
|
||||||
id: int
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
|
|
||||||
class ContactRelationshipCreate(BaseModel):
|
|
||||||
"""Schema for creating a contact relationship."""
|
|
||||||
|
|
||||||
related_contact_id: int
|
|
||||||
relationship_type: RelationshipType
|
|
||||||
closeness_weight: int | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class ContactRelationshipUpdate(BaseModel):
|
|
||||||
"""Schema for updating a contact relationship."""
|
|
||||||
|
|
||||||
relationship_type: RelationshipType | None = None
|
|
||||||
closeness_weight: int | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class ContactRelationshipResponse(BaseModel):
|
|
||||||
"""Schema for contact relationship response."""
|
|
||||||
|
|
||||||
contact_id: int
|
|
||||||
related_contact_id: int
|
|
||||||
relationship_type: str
|
|
||||||
closeness_weight: int
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
|
|
||||||
class RelationshipTypeInfo(BaseModel):
|
|
||||||
"""Information about a relationship type."""
|
|
||||||
|
|
||||||
value: str
|
|
||||||
display_name: str
|
|
||||||
default_weight: int
|
|
||||||
|
|
||||||
|
|
||||||
class GraphNode(BaseModel):
|
|
||||||
"""Node in the relationship graph."""
|
|
||||||
|
|
||||||
id: int
|
|
||||||
name: str
|
|
||||||
current_job: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class GraphEdge(BaseModel):
|
|
||||||
"""Edge in the relationship graph."""
|
|
||||||
|
|
||||||
source: int
|
|
||||||
target: int
|
|
||||||
relationship_type: str
|
|
||||||
closeness_weight: int
|
|
||||||
|
|
||||||
|
|
||||||
class GraphData(BaseModel):
|
|
||||||
"""Complete graph data for visualization."""
|
|
||||||
|
|
||||||
nodes: list[GraphNode]
|
|
||||||
edges: list[GraphEdge]
|
|
||||||
|
|
||||||
|
|
||||||
class ContactBase(BaseModel):
|
|
||||||
"""Base schema for Contact."""
|
|
||||||
|
|
||||||
name: str
|
|
||||||
age: int | None = None
|
|
||||||
bio: str | None = None
|
|
||||||
current_job: str | None = None
|
|
||||||
gender: str | None = None
|
|
||||||
goals: str | None = None
|
|
||||||
legal_name: str | None = None
|
|
||||||
profile_pic: str | None = None
|
|
||||||
safe_conversation_starters: str | None = None
|
|
||||||
self_sufficiency_score: int | None = None
|
|
||||||
social_structure_style: str | None = None
|
|
||||||
ssn: str | None = None
|
|
||||||
suffix: str | None = None
|
|
||||||
timezone: str | None = None
|
|
||||||
topics_to_avoid: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class ContactCreate(ContactBase):
|
|
||||||
"""Schema for creating a Contact."""
|
|
||||||
|
|
||||||
need_ids: list[int] = []
|
|
||||||
|
|
||||||
|
|
||||||
class ContactUpdate(BaseModel):
|
|
||||||
"""Schema for updating a Contact."""
|
|
||||||
|
|
||||||
name: str | None = None
|
|
||||||
age: int | None = None
|
|
||||||
bio: str | None = None
|
|
||||||
current_job: str | None = None
|
|
||||||
gender: str | None = None
|
|
||||||
goals: str | None = None
|
|
||||||
legal_name: str | None = None
|
|
||||||
profile_pic: str | None = None
|
|
||||||
safe_conversation_starters: str | None = None
|
|
||||||
self_sufficiency_score: int | None = None
|
|
||||||
social_structure_style: str | None = None
|
|
||||||
ssn: str | None = None
|
|
||||||
suffix: str | None = None
|
|
||||||
timezone: str | None = None
|
|
||||||
topics_to_avoid: str | None = None
|
|
||||||
need_ids: list[int] | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class ContactResponse(ContactBase):
|
|
||||||
"""Schema for Contact response with relationships."""
|
|
||||||
|
|
||||||
id: int
|
|
||||||
needs: list[NeedResponse] = []
|
|
||||||
related_to: list[ContactRelationshipResponse] = []
|
|
||||||
related_from: list[ContactRelationshipResponse] = []
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
|
|
||||||
class ContactListResponse(ContactBase):
|
|
||||||
"""Schema for Contact list response."""
|
|
||||||
|
|
||||||
id: int
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
|
|
||||||
router = APIRouter(prefix="/api", tags=["contacts"])
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/needs", response_model=NeedResponse)
|
|
||||||
def create_need(need: NeedCreate, db: DbSession) -> Need:
|
|
||||||
"""Create a new need."""
|
|
||||||
db_need = Need(name=need.name, description=need.description)
|
|
||||||
db.add(db_need)
|
|
||||||
db.commit()
|
|
||||||
db.refresh(db_need)
|
|
||||||
return db_need
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/needs", response_model=list[NeedResponse])
|
|
||||||
def list_needs(db: DbSession) -> list[Need]:
|
|
||||||
"""List all needs."""
|
|
||||||
return list(db.scalars(select(Need)).all())
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/needs/{need_id}", response_model=NeedResponse)
|
|
||||||
def get_need(need_id: int, db: DbSession) -> Need:
|
|
||||||
"""Get a need by ID."""
|
|
||||||
need = db.get(Need, need_id)
|
|
||||||
if not need:
|
|
||||||
raise HTTPException(status_code=404, detail="Need not found")
|
|
||||||
return need
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/needs/{need_id}", response_model=None)
|
|
||||||
def delete_need(need_id: int, request: Request, db: DbSession) -> dict[str, bool] | HTMLResponse:
|
|
||||||
"""Delete a need by ID."""
|
|
||||||
need = db.get(Need, need_id)
|
|
||||||
if not need:
|
|
||||||
raise HTTPException(status_code=404, detail="Need not found")
|
|
||||||
db.delete(need)
|
|
||||||
db.commit()
|
|
||||||
if _is_htmx(request):
|
|
||||||
return HTMLResponse("")
|
|
||||||
return {"deleted": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/contacts", response_model=ContactResponse)
|
|
||||||
def create_contact(contact: ContactCreate, db: DbSession) -> Contact:
|
|
||||||
"""Create a new contact."""
|
|
||||||
need_ids = contact.need_ids
|
|
||||||
contact_data = contact.model_dump(exclude={"need_ids"})
|
|
||||||
db_contact = Contact(**contact_data)
|
|
||||||
|
|
||||||
if need_ids:
|
|
||||||
needs = list(db.scalars(select(Need).where(Need.id.in_(need_ids))).all())
|
|
||||||
db_contact.needs = needs
|
|
||||||
|
|
||||||
db.add(db_contact)
|
|
||||||
db.commit()
|
|
||||||
db.refresh(db_contact)
|
|
||||||
return db_contact
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/contacts", response_model=list[ContactListResponse])
|
|
||||||
def list_contacts(
|
|
||||||
db: DbSession,
|
|
||||||
skip: int = 0,
|
|
||||||
limit: int = 100,
|
|
||||||
) -> list[Contact]:
|
|
||||||
"""List all contacts with pagination."""
|
|
||||||
return list(db.scalars(select(Contact).offset(skip).limit(limit)).all())
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/contacts/{contact_id}", response_model=ContactResponse)
|
|
||||||
def get_contact(contact_id: int, db: DbSession) -> Contact:
|
|
||||||
"""Get a contact by ID with all relationships."""
|
|
||||||
contact = db.scalar(
|
|
||||||
select(Contact)
|
|
||||||
.where(Contact.id == contact_id)
|
|
||||||
.options(
|
|
||||||
selectinload(Contact.needs),
|
|
||||||
selectinload(Contact.related_to),
|
|
||||||
selectinload(Contact.related_from),
|
|
||||||
)
|
|
||||||
)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
return contact
|
|
||||||
|
|
||||||
|
|
||||||
@router.patch("/contacts/{contact_id}", response_model=ContactResponse)
|
|
||||||
def update_contact(
|
|
||||||
contact_id: int,
|
|
||||||
contact: ContactUpdate,
|
|
||||||
db: DbSession,
|
|
||||||
) -> Contact:
|
|
||||||
"""Update a contact by ID."""
|
|
||||||
db_contact = db.get(Contact, contact_id)
|
|
||||||
if not db_contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
update_data = contact.model_dump(exclude_unset=True)
|
|
||||||
need_ids = update_data.pop("need_ids", None)
|
|
||||||
|
|
||||||
for key, value in update_data.items():
|
|
||||||
setattr(db_contact, key, value)
|
|
||||||
|
|
||||||
if need_ids is not None:
|
|
||||||
needs = list(db.scalars(select(Need).where(Need.id.in_(need_ids))).all())
|
|
||||||
db_contact.needs = needs
|
|
||||||
|
|
||||||
db.commit()
|
|
||||||
db.refresh(db_contact)
|
|
||||||
return db_contact
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/contacts/{contact_id}", response_model=None)
|
|
||||||
def delete_contact(contact_id: int, request: Request, db: DbSession) -> dict[str, bool] | HTMLResponse:
|
|
||||||
"""Delete a contact by ID."""
|
|
||||||
contact = db.get(Contact, contact_id)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
db.delete(contact)
|
|
||||||
db.commit()
|
|
||||||
if _is_htmx(request):
|
|
||||||
return HTMLResponse("")
|
|
||||||
return {"deleted": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/contacts/{contact_id}/needs/{need_id}")
|
|
||||||
def add_need_to_contact(
|
|
||||||
contact_id: int,
|
|
||||||
need_id: int,
|
|
||||||
db: DbSession,
|
|
||||||
) -> dict[str, bool]:
|
|
||||||
"""Add a need to a contact."""
|
|
||||||
contact = db.get(Contact, contact_id)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
need = db.get(Need, need_id)
|
|
||||||
if not need:
|
|
||||||
raise HTTPException(status_code=404, detail="Need not found")
|
|
||||||
|
|
||||||
if need not in contact.needs:
|
|
||||||
contact.needs.append(need)
|
|
||||||
db.commit()
|
|
||||||
|
|
||||||
return {"added": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/contacts/{contact_id}/needs/{need_id}", response_model=None)
|
|
||||||
def remove_need_from_contact(
|
|
||||||
contact_id: int,
|
|
||||||
need_id: int,
|
|
||||||
request: Request,
|
|
||||||
db: DbSession,
|
|
||||||
) -> dict[str, bool] | HTMLResponse:
|
|
||||||
"""Remove a need from a contact."""
|
|
||||||
contact = db.get(Contact, contact_id)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
need = db.get(Need, need_id)
|
|
||||||
if not need:
|
|
||||||
raise HTTPException(status_code=404, detail="Need not found")
|
|
||||||
|
|
||||||
if need in contact.needs:
|
|
||||||
contact.needs.remove(need)
|
|
||||||
db.commit()
|
|
||||||
|
|
||||||
if _is_htmx(request):
|
|
||||||
return HTMLResponse("")
|
|
||||||
return {"removed": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.post(
|
|
||||||
"/contacts/{contact_id}/relationships",
|
|
||||||
response_model=ContactRelationshipResponse,
|
|
||||||
)
|
|
||||||
def add_contact_relationship(
|
|
||||||
contact_id: int,
|
|
||||||
relationship: ContactRelationshipCreate,
|
|
||||||
db: DbSession,
|
|
||||||
) -> ContactRelationship:
|
|
||||||
"""Add a relationship between two contacts."""
|
|
||||||
contact = db.get(Contact, contact_id)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
related_contact = db.get(Contact, relationship.related_contact_id)
|
|
||||||
if not related_contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Related contact not found")
|
|
||||||
|
|
||||||
if contact_id == relationship.related_contact_id:
|
|
||||||
raise HTTPException(status_code=400, detail="Cannot relate contact to itself")
|
|
||||||
|
|
||||||
# Use provided weight or default from relationship type
|
|
||||||
weight = relationship.closeness_weight
|
|
||||||
if weight is None:
|
|
||||||
weight = relationship.relationship_type.default_weight
|
|
||||||
|
|
||||||
db_relationship = ContactRelationship(
|
|
||||||
contact_id=contact_id,
|
|
||||||
related_contact_id=relationship.related_contact_id,
|
|
||||||
relationship_type=relationship.relationship_type.value,
|
|
||||||
closeness_weight=weight,
|
|
||||||
)
|
|
||||||
db.add(db_relationship)
|
|
||||||
db.commit()
|
|
||||||
db.refresh(db_relationship)
|
|
||||||
return db_relationship
|
|
||||||
|
|
||||||
|
|
||||||
@router.get(
|
|
||||||
"/contacts/{contact_id}/relationships",
|
|
||||||
response_model=list[ContactRelationshipResponse],
|
|
||||||
)
|
|
||||||
def get_contact_relationships(
|
|
||||||
contact_id: int,
|
|
||||||
db: DbSession,
|
|
||||||
) -> list[ContactRelationship]:
|
|
||||||
"""Get all relationships for a contact."""
|
|
||||||
contact = db.get(Contact, contact_id)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
outgoing = list(db.scalars(select(ContactRelationship).where(ContactRelationship.contact_id == contact_id)).all())
|
|
||||||
incoming = list(
|
|
||||||
db.scalars(select(ContactRelationship).where(ContactRelationship.related_contact_id == contact_id)).all()
|
|
||||||
)
|
|
||||||
return outgoing + incoming
|
|
||||||
|
|
||||||
|
|
||||||
@router.patch(
|
|
||||||
"/contacts/{contact_id}/relationships/{related_contact_id}",
|
|
||||||
response_model=ContactRelationshipResponse,
|
|
||||||
)
|
|
||||||
def update_contact_relationship(
|
|
||||||
contact_id: int,
|
|
||||||
related_contact_id: int,
|
|
||||||
update: ContactRelationshipUpdate,
|
|
||||||
db: DbSession,
|
|
||||||
) -> ContactRelationship:
|
|
||||||
"""Update a relationship between two contacts."""
|
|
||||||
relationship = db.scalar(
|
|
||||||
select(ContactRelationship).where(
|
|
||||||
ContactRelationship.contact_id == contact_id,
|
|
||||||
ContactRelationship.related_contact_id == related_contact_id,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
if not relationship:
|
|
||||||
raise HTTPException(status_code=404, detail="Relationship not found")
|
|
||||||
|
|
||||||
if update.relationship_type is not None:
|
|
||||||
relationship.relationship_type = update.relationship_type.value
|
|
||||||
if update.closeness_weight is not None:
|
|
||||||
relationship.closeness_weight = update.closeness_weight
|
|
||||||
|
|
||||||
db.commit()
|
|
||||||
db.refresh(relationship)
|
|
||||||
return relationship
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/contacts/{contact_id}/relationships/{related_contact_id}", response_model=None)
|
|
||||||
def remove_contact_relationship(
|
|
||||||
contact_id: int,
|
|
||||||
related_contact_id: int,
|
|
||||||
request: Request,
|
|
||||||
db: DbSession,
|
|
||||||
) -> dict[str, bool] | HTMLResponse:
|
|
||||||
"""Remove a relationship between two contacts."""
|
|
||||||
relationship = db.scalar(
|
|
||||||
select(ContactRelationship).where(
|
|
||||||
ContactRelationship.contact_id == contact_id,
|
|
||||||
ContactRelationship.related_contact_id == related_contact_id,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
if not relationship:
|
|
||||||
raise HTTPException(status_code=404, detail="Relationship not found")
|
|
||||||
|
|
||||||
db.delete(relationship)
|
|
||||||
db.commit()
|
|
||||||
if _is_htmx(request):
|
|
||||||
return HTMLResponse("")
|
|
||||||
return {"deleted": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/relationship-types")
|
|
||||||
def list_relationship_types() -> list[RelationshipTypeInfo]:
|
|
||||||
"""List all available relationship types with their default weights."""
|
|
||||||
return [
|
|
||||||
RelationshipTypeInfo(
|
|
||||||
value=rt.value,
|
|
||||||
display_name=rt.display_name,
|
|
||||||
default_weight=rt.default_weight,
|
|
||||||
)
|
|
||||||
for rt in RelationshipType
|
|
||||||
]
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/graph")
|
|
||||||
def get_relationship_graph(db: DbSession) -> GraphData:
|
|
||||||
"""Get all contacts and relationships as graph data for visualization."""
|
|
||||||
contacts = list(db.scalars(select(Contact)).all())
|
|
||||||
relationships = list(db.scalars(select(ContactRelationship)).all())
|
|
||||||
|
|
||||||
nodes = [GraphNode(id=c.id, name=c.name, current_job=c.current_job) for c in contacts]
|
|
||||||
|
|
||||||
edges = [
|
|
||||||
GraphEdge(
|
|
||||||
source=rel.contact_id,
|
|
||||||
target=rel.related_contact_id,
|
|
||||||
relationship_type=rel.relationship_type,
|
|
||||||
closeness_weight=rel.closeness_weight,
|
|
||||||
)
|
|
||||||
for rel in relationships
|
|
||||||
]
|
|
||||||
|
|
||||||
return GraphData(nodes=nodes, edges=edges)
|
|
||||||
@@ -1,345 +0,0 @@
|
|||||||
"""HTMX server-rendered view router."""
|
|
||||||
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import Annotated, Any
|
|
||||||
|
|
||||||
from fastapi import APIRouter, Form, HTTPException, Request
|
|
||||||
from fastapi.responses import HTMLResponse, RedirectResponse
|
|
||||||
from fastapi.templating import Jinja2Templates
|
|
||||||
from sqlalchemy import select
|
|
||||||
from sqlalchemy.orm import Session, selectinload
|
|
||||||
|
|
||||||
from python.api.dependencies import DbSession
|
|
||||||
from python.orm.richie.contact import Contact, ContactRelationship, Need, RelationshipType
|
|
||||||
|
|
||||||
TEMPLATES_DIR = Path(__file__).parent.parent / "templates"
|
|
||||||
templates = Jinja2Templates(directory=TEMPLATES_DIR)
|
|
||||||
|
|
||||||
router = APIRouter(tags=["views"])
|
|
||||||
|
|
||||||
FAMILIAL_TYPES = {
|
|
||||||
"parent",
|
|
||||||
"child",
|
|
||||||
"sibling",
|
|
||||||
"grandparent",
|
|
||||||
"grandchild",
|
|
||||||
"aunt_uncle",
|
|
||||||
"niece_nephew",
|
|
||||||
"cousin",
|
|
||||||
"in_law",
|
|
||||||
}
|
|
||||||
FRIEND_TYPES = {"best_friend", "close_friend", "friend", "acquaintance", "neighbor"}
|
|
||||||
PARTNER_TYPES = {"spouse", "partner"}
|
|
||||||
PROFESSIONAL_TYPES = {"mentor", "mentee", "business_partner", "colleague", "manager", "direct_report", "client"}
|
|
||||||
|
|
||||||
CONTACT_STRING_FIELDS = (
|
|
||||||
"name",
|
|
||||||
"legal_name",
|
|
||||||
"suffix",
|
|
||||||
"gender",
|
|
||||||
"current_job",
|
|
||||||
"timezone",
|
|
||||||
"profile_pic",
|
|
||||||
"bio",
|
|
||||||
"goals",
|
|
||||||
"social_structure_style",
|
|
||||||
"safe_conversation_starters",
|
|
||||||
"topics_to_avoid",
|
|
||||||
"ssn",
|
|
||||||
)
|
|
||||||
|
|
||||||
CONTACT_INT_FIELDS = ("age", "self_sufficiency_score")
|
|
||||||
|
|
||||||
|
|
||||||
def _group_relationships(relationships: list[ContactRelationship]) -> dict[str, list[ContactRelationship]]:
|
|
||||||
"""Group relationships by category."""
|
|
||||||
groups: dict[str, list[ContactRelationship]] = {
|
|
||||||
"familial": [],
|
|
||||||
"partners": [],
|
|
||||||
"friends": [],
|
|
||||||
"professional": [],
|
|
||||||
"other": [],
|
|
||||||
}
|
|
||||||
for rel in relationships:
|
|
||||||
if rel.relationship_type in FAMILIAL_TYPES:
|
|
||||||
groups["familial"].append(rel)
|
|
||||||
elif rel.relationship_type in PARTNER_TYPES:
|
|
||||||
groups["partners"].append(rel)
|
|
||||||
elif rel.relationship_type in FRIEND_TYPES:
|
|
||||||
groups["friends"].append(rel)
|
|
||||||
elif rel.relationship_type in PROFESSIONAL_TYPES:
|
|
||||||
groups["professional"].append(rel)
|
|
||||||
else:
|
|
||||||
groups["other"].append(rel)
|
|
||||||
return groups
|
|
||||||
|
|
||||||
|
|
||||||
def _build_contact_name_map(database: Session, contact: Contact) -> dict[int, str]:
|
|
||||||
"""Build a mapping of contact IDs to names for relationship display."""
|
|
||||||
related_ids = {rel.related_contact_id for rel in contact.related_to}
|
|
||||||
related_ids |= {rel.contact_id for rel in contact.related_from}
|
|
||||||
related_ids.discard(contact.id)
|
|
||||||
|
|
||||||
if not related_ids:
|
|
||||||
return {}
|
|
||||||
|
|
||||||
related_contacts = list(database.scalars(select(Contact).where(Contact.id.in_(related_ids))).all())
|
|
||||||
return {related.id: related.name for related in related_contacts}
|
|
||||||
|
|
||||||
|
|
||||||
def _get_relationship_type_display() -> dict[str, str]:
|
|
||||||
"""Build a mapping of relationship type values to display names."""
|
|
||||||
return {rel_type.value: rel_type.display_name for rel_type in RelationshipType}
|
|
||||||
|
|
||||||
|
|
||||||
async def _parse_contact_form(request: Request) -> dict[str, Any]:
|
|
||||||
"""Parse contact form data from a multipart/form request."""
|
|
||||||
form_data = await request.form()
|
|
||||||
result: dict[str, Any] = {}
|
|
||||||
|
|
||||||
for field in CONTACT_STRING_FIELDS:
|
|
||||||
value = form_data.get(field, "")
|
|
||||||
result[field] = str(value) if value else None
|
|
||||||
|
|
||||||
for field in CONTACT_INT_FIELDS:
|
|
||||||
value = form_data.get(field, "")
|
|
||||||
result[field] = int(value) if value else None
|
|
||||||
|
|
||||||
result["need_ids"] = [int(value) for value in form_data.getlist("need_ids")]
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def _save_contact_from_form(database: Session, contact: Contact, form_result: dict[str, Any]) -> None:
|
|
||||||
"""Apply parsed form data to a Contact and save associated needs."""
|
|
||||||
need_ids = form_result.pop("need_ids")
|
|
||||||
|
|
||||||
for key, value in form_result.items():
|
|
||||||
setattr(contact, key, value)
|
|
||||||
|
|
||||||
if need_ids:
|
|
||||||
contact.needs = list(database.scalars(select(Need).where(Need.id.in_(need_ids))).all())
|
|
||||||
else:
|
|
||||||
contact.needs = []
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/", response_class=HTMLResponse)
|
|
||||||
@router.get("/contacts", response_class=HTMLResponse)
|
|
||||||
def contact_list_page(request: Request, database: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the contacts list page."""
|
|
||||||
contacts = list(database.scalars(select(Contact)).all())
|
|
||||||
return templates.TemplateResponse(request, "contact_list.html", {"contacts": contacts})
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/contacts/new", response_class=HTMLResponse)
|
|
||||||
def new_contact_page(request: Request, database: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the new contact form page."""
|
|
||||||
all_needs = list(database.scalars(select(Need)).all())
|
|
||||||
return templates.TemplateResponse(request, "contact_form.html", {"contact": None, "all_needs": all_needs})
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/htmx/contacts/new")
|
|
||||||
async def create_contact_form(request: Request, database: DbSession) -> RedirectResponse:
|
|
||||||
"""Handle the create contact form submission."""
|
|
||||||
form_result = await _parse_contact_form(request)
|
|
||||||
contact = Contact()
|
|
||||||
_save_contact_from_form(database, contact, form_result)
|
|
||||||
|
|
||||||
database.add(contact)
|
|
||||||
database.commit()
|
|
||||||
database.refresh(contact)
|
|
||||||
return RedirectResponse(url=f"/contacts/{contact.id}", status_code=303)
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/contacts/{contact_id}", response_class=HTMLResponse)
|
|
||||||
def contact_detail_page(contact_id: int, request: Request, database: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the contact detail page."""
|
|
||||||
contact = database.scalar(
|
|
||||||
select(Contact)
|
|
||||||
.where(Contact.id == contact_id)
|
|
||||||
.options(
|
|
||||||
selectinload(Contact.needs),
|
|
||||||
selectinload(Contact.related_to),
|
|
||||||
selectinload(Contact.related_from),
|
|
||||||
)
|
|
||||||
)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
contact_names = _build_contact_name_map(database, contact)
|
|
||||||
grouped_relationships = _group_relationships(contact.related_to)
|
|
||||||
all_contacts = list(database.scalars(select(Contact)).all())
|
|
||||||
all_needs = list(database.scalars(select(Need)).all())
|
|
||||||
available_needs = [need for need in all_needs if need not in contact.needs]
|
|
||||||
|
|
||||||
return templates.TemplateResponse(
|
|
||||||
request,
|
|
||||||
"contact_detail.html",
|
|
||||||
{
|
|
||||||
"contact": contact,
|
|
||||||
"contact_names": contact_names,
|
|
||||||
"grouped_relationships": grouped_relationships,
|
|
||||||
"all_contacts": all_contacts,
|
|
||||||
"available_needs": available_needs,
|
|
||||||
"relationship_types": list(RelationshipType),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/contacts/{contact_id}/edit", response_class=HTMLResponse)
|
|
||||||
def edit_contact_page(contact_id: int, request: Request, database: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the edit contact form page."""
|
|
||||||
contact = database.scalar(select(Contact).where(Contact.id == contact_id).options(selectinload(Contact.needs)))
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
all_needs = list(database.scalars(select(Need)).all())
|
|
||||||
return templates.TemplateResponse(request, "contact_form.html", {"contact": contact, "all_needs": all_needs})
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/htmx/contacts/{contact_id}/edit")
|
|
||||||
async def update_contact_form(contact_id: int, request: Request, database: DbSession) -> RedirectResponse:
|
|
||||||
"""Handle the edit contact form submission."""
|
|
||||||
contact = database.get(Contact, contact_id)
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
form_result = await _parse_contact_form(request)
|
|
||||||
_save_contact_from_form(database, contact, form_result)
|
|
||||||
|
|
||||||
database.commit()
|
|
||||||
return RedirectResponse(url=f"/contacts/{contact_id}", status_code=303)
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/htmx/contacts/{contact_id}/add-need", response_class=HTMLResponse)
|
|
||||||
def add_need_to_contact_htmx(
|
|
||||||
contact_id: int,
|
|
||||||
request: Request,
|
|
||||||
database: DbSession,
|
|
||||||
need_id: Annotated[int, Form()],
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Add a need to a contact and return updated manage-needs partial."""
|
|
||||||
contact = database.scalar(select(Contact).where(Contact.id == contact_id).options(selectinload(Contact.needs)))
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
need = database.get(Need, need_id)
|
|
||||||
if not need:
|
|
||||||
raise HTTPException(status_code=404, detail="Need not found")
|
|
||||||
|
|
||||||
if need not in contact.needs:
|
|
||||||
contact.needs.append(need)
|
|
||||||
database.commit()
|
|
||||||
database.refresh(contact)
|
|
||||||
|
|
||||||
return templates.TemplateResponse(request, "partials/manage_needs.html", {"contact": contact})
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/htmx/contacts/{contact_id}/add-relationship", response_class=HTMLResponse)
|
|
||||||
def add_relationship_htmx(
|
|
||||||
contact_id: int,
|
|
||||||
request: Request,
|
|
||||||
database: DbSession,
|
|
||||||
related_contact_id: Annotated[int, Form()],
|
|
||||||
relationship_type: Annotated[str, Form()],
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Add a relationship and return updated manage-relationships partial."""
|
|
||||||
contact = database.scalar(select(Contact).where(Contact.id == contact_id).options(selectinload(Contact.related_to)))
|
|
||||||
if not contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Contact not found")
|
|
||||||
|
|
||||||
related_contact = database.get(Contact, related_contact_id)
|
|
||||||
if not related_contact:
|
|
||||||
raise HTTPException(status_code=404, detail="Related contact not found")
|
|
||||||
|
|
||||||
rel_type = RelationshipType(relationship_type)
|
|
||||||
weight = rel_type.default_weight
|
|
||||||
|
|
||||||
relationship = ContactRelationship(
|
|
||||||
contact_id=contact_id,
|
|
||||||
related_contact_id=related_contact_id,
|
|
||||||
relationship_type=relationship_type,
|
|
||||||
closeness_weight=weight,
|
|
||||||
)
|
|
||||||
database.add(relationship)
|
|
||||||
database.commit()
|
|
||||||
database.refresh(contact)
|
|
||||||
|
|
||||||
contact_names = _build_contact_name_map(database, contact)
|
|
||||||
return templates.TemplateResponse(
|
|
||||||
request,
|
|
||||||
"partials/manage_relationships.html",
|
|
||||||
{"contact": contact, "contact_names": contact_names},
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/htmx/contacts/{contact_id}/relationships/{related_contact_id}/weight")
|
|
||||||
def update_relationship_weight_htmx(
|
|
||||||
contact_id: int,
|
|
||||||
related_contact_id: int,
|
|
||||||
database: DbSession,
|
|
||||||
closeness_weight: Annotated[int, Form()],
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Update a relationship's closeness weight from HTMX range input."""
|
|
||||||
relationship = database.scalar(
|
|
||||||
select(ContactRelationship).where(
|
|
||||||
ContactRelationship.contact_id == contact_id,
|
|
||||||
ContactRelationship.related_contact_id == related_contact_id,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
if not relationship:
|
|
||||||
raise HTTPException(status_code=404, detail="Relationship not found")
|
|
||||||
|
|
||||||
relationship.closeness_weight = closeness_weight
|
|
||||||
database.commit()
|
|
||||||
return HTMLResponse("")
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/htmx/needs", response_class=HTMLResponse)
|
|
||||||
def create_need_htmx(
|
|
||||||
request: Request,
|
|
||||||
database: DbSession,
|
|
||||||
name: Annotated[str, Form()],
|
|
||||||
description: Annotated[str, Form()] = "",
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Create a need via form data and return updated needs list."""
|
|
||||||
need = Need(name=name, description=description or None)
|
|
||||||
database.add(need)
|
|
||||||
database.commit()
|
|
||||||
needs = list(database.scalars(select(Need)).all())
|
|
||||||
return templates.TemplateResponse(request, "partials/need_items.html", {"needs": needs})
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/needs", response_class=HTMLResponse)
|
|
||||||
def needs_page(request: Request, database: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the needs list page."""
|
|
||||||
needs = list(database.scalars(select(Need)).all())
|
|
||||||
return templates.TemplateResponse(request, "need_list.html", {"needs": needs})
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/graph", response_class=HTMLResponse)
|
|
||||||
def graph_page(request: Request, database: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the relationship graph page."""
|
|
||||||
contacts = list(database.scalars(select(Contact)).all())
|
|
||||||
relationships = list(database.scalars(select(ContactRelationship)).all())
|
|
||||||
|
|
||||||
graph_data = {
|
|
||||||
"nodes": [{"id": contact.id, "name": contact.name, "current_job": contact.current_job} for contact in contacts],
|
|
||||||
"edges": [
|
|
||||||
{
|
|
||||||
"source": rel.contact_id,
|
|
||||||
"target": rel.related_contact_id,
|
|
||||||
"relationship_type": rel.relationship_type,
|
|
||||||
"closeness_weight": rel.closeness_weight,
|
|
||||||
}
|
|
||||||
for rel in relationships
|
|
||||||
],
|
|
||||||
}
|
|
||||||
|
|
||||||
return templates.TemplateResponse(
|
|
||||||
request,
|
|
||||||
"graph.html",
|
|
||||||
{
|
|
||||||
"graph_data": graph_data,
|
|
||||||
"relationship_type_display": _get_relationship_type_display(),
|
|
||||||
},
|
|
||||||
)
|
|
||||||
@@ -1,198 +0,0 @@
|
|||||||
<!DOCTYPE html>
|
|
||||||
<html lang="en" data-theme="light">
|
|
||||||
<head>
|
|
||||||
<meta charset="UTF-8">
|
|
||||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
|
||||||
<title>{% block title %}Contact Database{% endblock %}</title>
|
|
||||||
<script src="https://unpkg.com/htmx.org@2.0.4"></script>
|
|
||||||
<style>
|
|
||||||
:root {
|
|
||||||
--color-bg: #f5f5f5;
|
|
||||||
--color-bg-card: #ffffff;
|
|
||||||
--color-bg-hover: #f0f0f0;
|
|
||||||
--color-bg-muted: #f9f9f9;
|
|
||||||
--color-bg-error: #ffe0e0;
|
|
||||||
--color-text: #333333;
|
|
||||||
--color-text-muted: #666666;
|
|
||||||
--color-text-error: #cc0000;
|
|
||||||
--color-border: #dddddd;
|
|
||||||
--color-border-light: #eeeeee;
|
|
||||||
--color-border-lighter: #f0f0f0;
|
|
||||||
--color-primary: #0066cc;
|
|
||||||
--color-primary-hover: #0055aa;
|
|
||||||
--color-danger: #cc3333;
|
|
||||||
--color-danger-hover: #aa2222;
|
|
||||||
--color-tag-bg: #e0e0e0;
|
|
||||||
--shadow: 0 1px 3px rgba(0, 0, 0, 0.1);
|
|
||||||
font-family: -apple-system, BlinkMacSystemFont, "Segoe UI", Roboto, sans-serif;
|
|
||||||
line-height: 1.5;
|
|
||||||
color: var(--color-text);
|
|
||||||
background-color: var(--color-bg);
|
|
||||||
}
|
|
||||||
[data-theme="dark"] {
|
|
||||||
--color-bg: #1a1a1a;
|
|
||||||
--color-bg-card: #2d2d2d;
|
|
||||||
--color-bg-hover: #3d3d3d;
|
|
||||||
--color-bg-muted: #252525;
|
|
||||||
--color-bg-error: #4a2020;
|
|
||||||
--color-text: #e0e0e0;
|
|
||||||
--color-text-muted: #a0a0a0;
|
|
||||||
--color-text-error: #ff6b6b;
|
|
||||||
--color-border: #404040;
|
|
||||||
--color-border-light: #353535;
|
|
||||||
--color-border-lighter: #303030;
|
|
||||||
--color-primary: #4da6ff;
|
|
||||||
--color-primary-hover: #7dbfff;
|
|
||||||
--color-danger: #ff6b6b;
|
|
||||||
--color-danger-hover: #ff8a8a;
|
|
||||||
--color-tag-bg: #404040;
|
|
||||||
--shadow: 0 1px 3px rgba(0, 0, 0, 0.3);
|
|
||||||
}
|
|
||||||
* { box-sizing: border-box; }
|
|
||||||
body { margin: 0; background: var(--color-bg); color: var(--color-text); }
|
|
||||||
.app { max-width: 1000px; margin: 0 auto; padding: 20px; }
|
|
||||||
nav { display: flex; align-items: center; gap: 20px; padding: 15px 0; border-bottom: 1px solid var(--color-border); margin-bottom: 20px; }
|
|
||||||
nav a { color: var(--color-primary); text-decoration: none; font-weight: 500; }
|
|
||||||
nav a:hover { text-decoration: underline; }
|
|
||||||
.theme-toggle { margin-left: auto; }
|
|
||||||
main { background: var(--color-bg-card); padding: 20px; border-radius: 8px; box-shadow: var(--shadow); }
|
|
||||||
.header { display: flex; justify-content: space-between; align-items: center; margin-bottom: 20px; }
|
|
||||||
.header h1 { margin: 0; }
|
|
||||||
a { color: var(--color-primary); }
|
|
||||||
a:hover { text-decoration: underline; }
|
|
||||||
|
|
||||||
.btn { display: inline-block; padding: 8px 16px; border: 1px solid var(--color-border); border-radius: 4px; background: var(--color-bg-card); color: var(--color-text); text-decoration: none; cursor: pointer; font-size: 14px; margin-left: 8px; }
|
|
||||||
.btn:hover { background: var(--color-bg-hover); }
|
|
||||||
.btn-primary { background: var(--color-primary); border-color: var(--color-primary); color: white; }
|
|
||||||
.btn-primary:hover { background: var(--color-primary-hover); }
|
|
||||||
.btn-danger { background: var(--color-danger); border-color: var(--color-danger); color: white; }
|
|
||||||
.btn-danger:hover { background: var(--color-danger-hover); }
|
|
||||||
.btn-small { padding: 4px 8px; font-size: 12px; }
|
|
||||||
.btn:disabled { opacity: 0.6; cursor: not-allowed; }
|
|
||||||
|
|
||||||
table { width: 100%; border-collapse: collapse; }
|
|
||||||
th, td { padding: 12px; text-align: left; border-bottom: 1px solid var(--color-border-light); }
|
|
||||||
th { font-weight: 600; background: var(--color-bg-muted); }
|
|
||||||
tr:hover { background: var(--color-bg-muted); }
|
|
||||||
|
|
||||||
.error { background: var(--color-bg-error); color: var(--color-text-error); padding: 10px; border-radius: 4px; margin-bottom: 20px; }
|
|
||||||
.tag { display: inline-block; background: var(--color-tag-bg); padding: 2px 8px; border-radius: 12px; font-size: 12px; color: var(--color-text-muted); }
|
|
||||||
|
|
||||||
.add-form { display: flex; gap: 10px; margin-top: 15px; flex-wrap: wrap; }
|
|
||||||
.add-form select, .add-form input { padding: 8px; border: 1px solid var(--color-border); border-radius: 4px; min-width: 200px; background: var(--color-bg-card); color: var(--color-text); }
|
|
||||||
|
|
||||||
.form-group { margin-bottom: 20px; }
|
|
||||||
.form-group label { display: block; font-weight: 500; margin-bottom: 5px; }
|
|
||||||
.form-group input, .form-group textarea, .form-group select { width: 100%; padding: 10px; border: 1px solid var(--color-border); border-radius: 4px; font-size: 14px; background: var(--color-bg-card); color: var(--color-text); }
|
|
||||||
.form-group textarea { resize: vertical; }
|
|
||||||
.form-row { display: grid; grid-template-columns: 1fr 1fr; gap: 20px; }
|
|
||||||
.checkbox-group { display: flex; flex-wrap: wrap; gap: 15px; }
|
|
||||||
.checkbox-label { display: flex; align-items: center; gap: 5px; cursor: pointer; }
|
|
||||||
.form-actions { display: flex; gap: 10px; margin-top: 30px; padding-top: 20px; border-top: 1px solid var(--color-border-light); }
|
|
||||||
|
|
||||||
.need-form { background: var(--color-bg-muted); padding: 20px; border-radius: 4px; margin-bottom: 20px; }
|
|
||||||
.need-items { list-style: none; padding: 0; }
|
|
||||||
.need-items li { display: flex; justify-content: space-between; align-items: flex-start; padding: 15px; border: 1px solid var(--color-border-light); border-radius: 4px; margin-bottom: 10px; }
|
|
||||||
.need-info p { margin: 5px 0 0; color: var(--color-text-muted); font-size: 14px; }
|
|
||||||
|
|
||||||
.graph-container { width: 100%; }
|
|
||||||
.graph-hint { color: var(--color-text-muted); font-size: 14px; margin-bottom: 15px; }
|
|
||||||
.selected-info { margin-top: 15px; padding: 15px; background: var(--color-bg-muted); border-radius: 8px; }
|
|
||||||
.selected-info h3 { margin: 0 0 10px; }
|
|
||||||
.selected-info p { margin: 5px 0; color: var(--color-text-muted); }
|
|
||||||
.legend { margin-top: 20px; padding: 15px; background: var(--color-bg-muted); border-radius: 8px; }
|
|
||||||
.legend h4 { margin: 0 0 10px; font-size: 14px; }
|
|
||||||
.legend-items { display: flex; flex-wrap: wrap; gap: 15px; }
|
|
||||||
.legend-item { display: flex; align-items: center; gap: 8px; font-size: 12px; color: var(--color-text-muted); }
|
|
||||||
.legend-line { width: 30px; border-radius: 2px; }
|
|
||||||
|
|
||||||
.id-card { width: 100%; }
|
|
||||||
.id-card-inner { background: linear-gradient(135deg, #0a0a0f 0%, #1a1a2e 50%, #0a0a0f 100%); background-image: radial-gradient(white 1px, transparent 1px), linear-gradient(135deg, #0a0a0f 0%, #1a1a2e 50%, #0a0a0f 100%); background-size: 50px 50px, 100% 100%; color: #fff; border-radius: 12px; padding: 25px; min-height: 500px; position: relative; overflow: hidden; }
|
|
||||||
.id-card-header { display: flex; justify-content: space-between; align-items: flex-start; margin-bottom: 15px; }
|
|
||||||
.id-card-header-left { flex: 1; }
|
|
||||||
.id-card-header-right { display: flex; flex-direction: column; align-items: flex-end; gap: 10px; }
|
|
||||||
.id-card-title { font-size: 2.5rem; font-weight: 700; margin: 0; color: #fff; text-shadow: 2px 2px 4px rgba(0,0,0,0.5); }
|
|
||||||
.id-profile-pic { width: 80px; height: 80px; border-radius: 8px; object-fit: cover; border: 2px solid rgba(255,255,255,0.3); }
|
|
||||||
.id-profile-placeholder { width: 80px; height: 80px; border-radius: 8px; background: linear-gradient(135deg, #4ecdc4 0%, #44a8a0 100%); display: flex; align-items: center; justify-content: center; border: 2px solid rgba(255,255,255,0.3); }
|
|
||||||
.id-profile-placeholder span { font-size: 2rem; font-weight: 700; color: #fff; text-shadow: 1px 1px 2px rgba(0,0,0,0.3); }
|
|
||||||
.id-card-actions { display: flex; gap: 8px; }
|
|
||||||
.id-card-actions .btn { background: rgba(255,255,255,0.1); border-color: rgba(255,255,255,0.3); color: #fff; }
|
|
||||||
.id-card-actions .btn:hover { background: rgba(255,255,255,0.2); }
|
|
||||||
.id-card-body { display: grid; grid-template-columns: 1fr 1.5fr; gap: 30px; }
|
|
||||||
.id-card-left { display: flex; flex-direction: column; gap: 8px; }
|
|
||||||
.id-field { font-size: 1rem; line-height: 1.4; }
|
|
||||||
.id-field-block { margin-top: 15px; font-size: 0.95rem; line-height: 1.5; }
|
|
||||||
.id-label { color: #4ecdc4; font-weight: 500; }
|
|
||||||
.id-card-right { display: flex; flex-direction: column; gap: 20px; }
|
|
||||||
.id-bio { font-size: 0.9rem; line-height: 1.6; color: #e0e0e0; }
|
|
||||||
.id-relationships { margin-top: 10px; }
|
|
||||||
.id-section-title { font-size: 1.5rem; margin: 0 0 15px; color: #fff; border-bottom: 1px solid rgba(255,255,255,0.2); padding-bottom: 8px; }
|
|
||||||
.id-rel-group { margin-bottom: 12px; font-size: 0.9rem; line-height: 1.6; }
|
|
||||||
.id-rel-label { color: #a0a0a0; }
|
|
||||||
.id-rel-group a { color: #4ecdc4; text-decoration: none; }
|
|
||||||
.id-rel-group a:hover { text-decoration: underline; }
|
|
||||||
.id-rel-type { color: #888; font-size: 0.85em; }
|
|
||||||
.id-card-warnings { margin-top: 30px; padding-top: 20px; border-top: 1px solid rgba(255,255,255,0.2); display: flex; flex-wrap: wrap; gap: 20px; }
|
|
||||||
.id-warning { display: flex; align-items: center; gap: 8px; font-size: 0.9rem; color: #ff6b6b; }
|
|
||||||
.warning-dot { width: 8px; height: 8px; background: #ff6b6b; border-radius: 50%; flex-shrink: 0; }
|
|
||||||
.warning-desc { color: #ccc; }
|
|
||||||
|
|
||||||
.id-card-manage { margin-top: 20px; background: var(--color-bg-muted); border-radius: 8px; padding: 15px; }
|
|
||||||
.id-card-manage summary { cursor: pointer; font-weight: 600; font-size: 1.1rem; padding: 5px 0; }
|
|
||||||
.id-card-manage[open] summary { margin-bottom: 15px; border-bottom: 1px solid var(--color-border-light); padding-bottom: 10px; }
|
|
||||||
.manage-section { margin-bottom: 25px; }
|
|
||||||
.manage-section h3 { margin: 0 0 15px; font-size: 1rem; }
|
|
||||||
.manage-relationships { display: flex; flex-direction: column; gap: 10px; margin-bottom: 15px; }
|
|
||||||
.manage-rel-item { display: flex; align-items: center; gap: 12px; padding: 10px; background: var(--color-bg-card); border-radius: 6px; flex-wrap: wrap; }
|
|
||||||
.manage-rel-item a { font-weight: 500; min-width: 120px; }
|
|
||||||
.weight-control { display: flex; align-items: center; gap: 8px; font-size: 12px; color: var(--color-text-muted); }
|
|
||||||
.weight-control input[type="range"] { width: 80px; cursor: pointer; }
|
|
||||||
.weight-value { min-width: 20px; text-align: center; font-weight: 600; }
|
|
||||||
.manage-needs-list { list-style: none; padding: 0; margin: 0 0 15px; }
|
|
||||||
.manage-needs-list li { display: flex; align-items: center; gap: 12px; padding: 10px; background: var(--color-bg-card); border-radius: 6px; margin-bottom: 8px; }
|
|
||||||
.manage-needs-list li .btn { margin-left: auto; }
|
|
||||||
|
|
||||||
.htmx-indicator { display: none; }
|
|
||||||
.htmx-request .htmx-indicator { display: inline; }
|
|
||||||
.htmx-request.htmx-indicator { display: inline; }
|
|
||||||
|
|
||||||
@media (max-width: 768px) {
|
|
||||||
.id-card-body { grid-template-columns: 1fr; }
|
|
||||||
.id-card-title { font-size: 1.8rem; }
|
|
||||||
.id-card-header { flex-direction: column; gap: 15px; }
|
|
||||||
}
|
|
||||||
</style>
|
|
||||||
</head>
|
|
||||||
<body>
|
|
||||||
<div class="app">
|
|
||||||
<nav>
|
|
||||||
<a href="/contacts">Contacts</a>
|
|
||||||
<a href="/graph">Graph</a>
|
|
||||||
<a href="/needs">Needs</a>
|
|
||||||
<button class="btn btn-small theme-toggle" onclick="toggleTheme()">
|
|
||||||
<span id="theme-label">Dark</span>
|
|
||||||
</button>
|
|
||||||
</nav>
|
|
||||||
|
|
||||||
<main id="main-content">
|
|
||||||
{% block content %}{% endblock %}
|
|
||||||
</main>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<script>
|
|
||||||
function toggleTheme() {
|
|
||||||
const html = document.documentElement;
|
|
||||||
const current = html.getAttribute('data-theme');
|
|
||||||
const next = current === 'light' ? 'dark' : 'light';
|
|
||||||
html.setAttribute('data-theme', next);
|
|
||||||
localStorage.setItem('theme', next);
|
|
||||||
document.getElementById('theme-label').textContent = next === 'light' ? 'Dark' : 'Light';
|
|
||||||
}
|
|
||||||
(function() {
|
|
||||||
const saved = localStorage.getItem('theme') || 'light';
|
|
||||||
document.documentElement.setAttribute('data-theme', saved);
|
|
||||||
document.getElementById('theme-label').textContent = saved === 'light' ? 'Dark' : 'Light';
|
|
||||||
})();
|
|
||||||
</script>
|
|
||||||
</body>
|
|
||||||
</html>
|
|
||||||
@@ -1,204 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}{{ contact.name }}{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<div class="id-card">
|
|
||||||
<div class="id-card-inner">
|
|
||||||
<div class="id-card-header">
|
|
||||||
<div class="id-card-header-left">
|
|
||||||
<h1 class="id-card-title">I.D.: {{ contact.name }}</h1>
|
|
||||||
</div>
|
|
||||||
<div class="id-card-header-right">
|
|
||||||
{% if contact.profile_pic %}
|
|
||||||
<img src="{{ contact.profile_pic }}" alt="{{ contact.name }}'s profile" class="id-profile-pic">
|
|
||||||
{% else %}
|
|
||||||
<div class="id-profile-placeholder">
|
|
||||||
<span>{{ contact.name[0]|upper }}</span>
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
<div class="id-card-actions">
|
|
||||||
<a href="/contacts/{{ contact.id }}/edit" class="btn btn-small">Edit</a>
|
|
||||||
<a href="/contacts" class="btn btn-small">Back</a>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="id-card-body">
|
|
||||||
<div class="id-card-left">
|
|
||||||
{% if contact.legal_name %}
|
|
||||||
<div class="id-field">Legal name: {{ contact.legal_name }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.suffix %}
|
|
||||||
<div class="id-field">Suffix: {{ contact.suffix }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.gender %}
|
|
||||||
<div class="id-field">Gender: {{ contact.gender }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.age %}
|
|
||||||
<div class="id-field">Age: {{ contact.age }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.current_job %}
|
|
||||||
<div class="id-field">Job: {{ contact.current_job }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.social_structure_style %}
|
|
||||||
<div class="id-field">Social style: {{ contact.social_structure_style }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.self_sufficiency_score is not none %}
|
|
||||||
<div class="id-field">Self-Sufficiency: {{ contact.self_sufficiency_score }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.timezone %}
|
|
||||||
<div class="id-field">Timezone: {{ contact.timezone }}</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.safe_conversation_starters %}
|
|
||||||
<div class="id-field-block">
|
|
||||||
<span class="id-label">Safe con starters:</span> {{ contact.safe_conversation_starters }}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.topics_to_avoid %}
|
|
||||||
<div class="id-field-block">
|
|
||||||
<span class="id-label">Topics to avoid:</span> {{ contact.topics_to_avoid }}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
{% if contact.goals %}
|
|
||||||
<div class="id-field-block">
|
|
||||||
<span class="id-label">Goals:</span> {{ contact.goals }}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="id-card-right">
|
|
||||||
{% if contact.bio %}
|
|
||||||
<div class="id-bio">
|
|
||||||
<span class="id-label">Bio:</span> {{ contact.bio }}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
<div class="id-relationships">
|
|
||||||
<h2 class="id-section-title">Relationships</h2>
|
|
||||||
|
|
||||||
{% if grouped_relationships.familial %}
|
|
||||||
<div class="id-rel-group">
|
|
||||||
<span class="id-rel-label">Familial:</span>
|
|
||||||
{% for rel in grouped_relationships.familial %}
|
|
||||||
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a><span class="id-rel-type">({{ rel.relationship_type|replace("_", " ")|title }})</span>{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if grouped_relationships.partners %}
|
|
||||||
<div class="id-rel-group">
|
|
||||||
<span class="id-rel-label">Partners:</span>
|
|
||||||
{% for rel in grouped_relationships.partners %}
|
|
||||||
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a>{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if grouped_relationships.friends %}
|
|
||||||
<div class="id-rel-group">
|
|
||||||
<span class="id-rel-label">Friends:</span>
|
|
||||||
{% for rel in grouped_relationships.friends %}
|
|
||||||
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a>{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if grouped_relationships.professional %}
|
|
||||||
<div class="id-rel-group">
|
|
||||||
<span class="id-rel-label">Professional:</span>
|
|
||||||
{% for rel in grouped_relationships.professional %}
|
|
||||||
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a><span class="id-rel-type">({{ rel.relationship_type|replace("_", " ")|title }})</span>{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if grouped_relationships.other %}
|
|
||||||
<div class="id-rel-group">
|
|
||||||
<span class="id-rel-label">Other:</span>
|
|
||||||
{% for rel in grouped_relationships.other %}
|
|
||||||
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a><span class="id-rel-type">({{ rel.relationship_type|replace("_", " ")|title }})</span>{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
{% if contact.related_from %}
|
|
||||||
<div class="id-rel-group">
|
|
||||||
<span class="id-rel-label">Known by:</span>
|
|
||||||
{% for rel in contact.related_from %}
|
|
||||||
<a href="/contacts/{{ rel.contact_id }}">{{ contact_names[rel.contact_id] }}</a>{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{% if contact.needs %}
|
|
||||||
<div class="id-card-warnings">
|
|
||||||
{% for need in contact.needs %}
|
|
||||||
<div class="id-warning">
|
|
||||||
<span class="warning-dot"></span>
|
|
||||||
Warning: {{ need.name }}
|
|
||||||
{% if need.description %}<span class="warning-desc"> - {{ need.description }}</span>{% endif %}
|
|
||||||
</div>
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<details class="id-card-manage">
|
|
||||||
<summary>Manage Contact</summary>
|
|
||||||
|
|
||||||
<div class="manage-section">
|
|
||||||
<h3>Manage Relationships</h3>
|
|
||||||
<div id="manage-relationships" class="manage-relationships">
|
|
||||||
{% include "partials/manage_relationships.html" %}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{% if all_contacts %}
|
|
||||||
<form hx-post="/htmx/contacts/{{ contact.id }}/add-relationship"
|
|
||||||
hx-target="#manage-relationships"
|
|
||||||
hx-swap="innerHTML"
|
|
||||||
class="add-form">
|
|
||||||
<select name="related_contact_id" required>
|
|
||||||
<option value="">Select contact...</option>
|
|
||||||
{% for other in all_contacts %}
|
|
||||||
{% if other.id != contact.id %}
|
|
||||||
<option value="{{ other.id }}">{{ other.name }}</option>
|
|
||||||
{% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</select>
|
|
||||||
<select name="relationship_type" required>
|
|
||||||
<option value="">Select relationship type...</option>
|
|
||||||
{% for rel_type in relationship_types %}
|
|
||||||
<option value="{{ rel_type.value }}">{{ rel_type.display_name }}</option>
|
|
||||||
{% endfor %}
|
|
||||||
</select>
|
|
||||||
<button type="submit" class="btn btn-primary">Add Relationship</button>
|
|
||||||
</form>
|
|
||||||
{% endif %}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="manage-section">
|
|
||||||
<h3>Manage Needs/Warnings</h3>
|
|
||||||
<div id="manage-needs">
|
|
||||||
{% include "partials/manage_needs.html" %}
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{% if available_needs %}
|
|
||||||
<form hx-post="/htmx/contacts/{{ contact.id }}/add-need"
|
|
||||||
hx-target="#manage-needs"
|
|
||||||
hx-swap="innerHTML"
|
|
||||||
class="add-form">
|
|
||||||
<select name="need_id" required>
|
|
||||||
<option value="">Select a need...</option>
|
|
||||||
{% for need in available_needs %}
|
|
||||||
<option value="{{ need.id }}">{{ need.name }}</option>
|
|
||||||
{% endfor %}
|
|
||||||
</select>
|
|
||||||
<button type="submit" class="btn btn-primary">Add Need</button>
|
|
||||||
</form>
|
|
||||||
{% endif %}
|
|
||||||
</div>
|
|
||||||
</details>
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,115 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}{{ "Edit " + contact.name if contact else "New Contact" }}{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<div class="contact-form">
|
|
||||||
<h1>{{ "Edit Contact" if contact else "New Contact" }}</h1>
|
|
||||||
|
|
||||||
{% if contact %}
|
|
||||||
<form method="post" action="/htmx/contacts/{{ contact.id }}/edit">
|
|
||||||
{% else %}
|
|
||||||
<form method="post" action="/htmx/contacts/new">
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="name">Name *</label>
|
|
||||||
<input id="name" name="name" type="text" value="{{ contact.name if contact else '' }}" required>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-row">
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="legal_name">Legal Name</label>
|
|
||||||
<input id="legal_name" name="legal_name" type="text" value="{{ contact.legal_name or '' }}">
|
|
||||||
</div>
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="suffix">Suffix</label>
|
|
||||||
<input id="suffix" name="suffix" type="text" value="{{ contact.suffix or '' }}">
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-row">
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="age">Age</label>
|
|
||||||
<input id="age" name="age" type="number" value="{{ contact.age if contact and contact.age is not none else '' }}">
|
|
||||||
</div>
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="gender">Gender</label>
|
|
||||||
<input id="gender" name="gender" type="text" value="{{ contact.gender or '' }}">
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="current_job">Current Job</label>
|
|
||||||
<input id="current_job" name="current_job" type="text" value="{{ contact.current_job or '' }}">
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="timezone">Timezone</label>
|
|
||||||
<input id="timezone" name="timezone" type="text" value="{{ contact.timezone or '' }}">
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="profile_pic">Profile Picture URL</label>
|
|
||||||
<input id="profile_pic" name="profile_pic" type="url" placeholder="https://example.com/photo.jpg" value="{{ contact.profile_pic or '' }}">
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="bio">Bio</label>
|
|
||||||
<textarea id="bio" name="bio" rows="3">{{ contact.bio or '' }}</textarea>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="goals">Goals</label>
|
|
||||||
<textarea id="goals" name="goals" rows="3">{{ contact.goals or '' }}</textarea>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="social_structure_style">Social Structure Style</label>
|
|
||||||
<input id="social_structure_style" name="social_structure_style" type="text" value="{{ contact.social_structure_style or '' }}">
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="self_sufficiency_score">Self-Sufficiency Score (1-10)</label>
|
|
||||||
<input id="self_sufficiency_score" name="self_sufficiency_score" type="number" min="1" max="10" value="{{ contact.self_sufficiency_score if contact and contact.self_sufficiency_score is not none else '' }}">
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="safe_conversation_starters">Safe Conversation Starters</label>
|
|
||||||
<textarea id="safe_conversation_starters" name="safe_conversation_starters" rows="2">{{ contact.safe_conversation_starters or '' }}</textarea>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="topics_to_avoid">Topics to Avoid</label>
|
|
||||||
<textarea id="topics_to_avoid" name="topics_to_avoid" rows="2">{{ contact.topics_to_avoid or '' }}</textarea>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="ssn">SSN</label>
|
|
||||||
<input id="ssn" name="ssn" type="text" value="{{ contact.ssn or '' }}">
|
|
||||||
</div>
|
|
||||||
|
|
||||||
{% if all_needs %}
|
|
||||||
<div class="form-group">
|
|
||||||
<label>Needs/Accommodations</label>
|
|
||||||
<div class="checkbox-group">
|
|
||||||
{% for need in all_needs %}
|
|
||||||
<label class="checkbox-label">
|
|
||||||
<input type="checkbox" name="need_ids" value="{{ need.id }}"
|
|
||||||
{% if contact and need in contact.needs %}checked{% endif %}>
|
|
||||||
{{ need.name }}
|
|
||||||
</label>
|
|
||||||
{% endfor %}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
{% endif %}
|
|
||||||
|
|
||||||
<div class="form-actions">
|
|
||||||
<button type="submit" class="btn btn-primary">Save</button>
|
|
||||||
{% if contact %}
|
|
||||||
<a href="/contacts/{{ contact.id }}" class="btn">Cancel</a>
|
|
||||||
{% else %}
|
|
||||||
<a href="/contacts" class="btn">Cancel</a>
|
|
||||||
{% endif %}
|
|
||||||
</div>
|
|
||||||
</form>
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}Contacts{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<div class="contact-list">
|
|
||||||
<div class="header">
|
|
||||||
<h1>Contacts</h1>
|
|
||||||
<a href="/contacts/new" class="btn btn-primary">Add Contact</a>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<div id="contact-table">
|
|
||||||
{% include "partials/contact_table.html" %}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,198 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}Relationship Graph{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<div class="graph-container">
|
|
||||||
<div class="header">
|
|
||||||
<h1>Relationship Graph</h1>
|
|
||||||
</div>
|
|
||||||
<p class="graph-hint">Drag nodes to reposition. Closer relationships have shorter, darker edges.</p>
|
|
||||||
<canvas id="graph-canvas" width="900" height="600"
|
|
||||||
style="border: 1px solid var(--color-border); border-radius: 8px; background: var(--color-bg); cursor: grab;">
|
|
||||||
</canvas>
|
|
||||||
<div id="selected-info"></div>
|
|
||||||
<div class="legend">
|
|
||||||
<h4>Relationship Closeness (1-10)</h4>
|
|
||||||
<div class="legend-items">
|
|
||||||
<div class="legend-item">
|
|
||||||
<span class="legend-line" style="background: hsl(220, 70%, 40%); height: 4px; display: inline-block;"></span>
|
|
||||||
<span>10 - Very Close (Spouse, Partner)</span>
|
|
||||||
</div>
|
|
||||||
<div class="legend-item">
|
|
||||||
<span class="legend-line" style="background: hsl(220, 70%, 52%); height: 3px; display: inline-block;"></span>
|
|
||||||
<span>7 - Close (Family, Best Friend)</span>
|
|
||||||
</div>
|
|
||||||
<div class="legend-item">
|
|
||||||
<span class="legend-line" style="background: hsl(220, 70%, 64%); height: 2px; display: inline-block;"></span>
|
|
||||||
<span>4 - Moderate (Friend, Colleague)</span>
|
|
||||||
</div>
|
|
||||||
<div class="legend-item">
|
|
||||||
<span class="legend-line" style="background: hsl(220, 70%, 72%); height: 1px; display: inline-block;"></span>
|
|
||||||
<span>2 - Distant (Acquaintance)</span>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<script>
|
|
||||||
(function() {
|
|
||||||
const RELATIONSHIP_DISPLAY = {{ relationship_type_display|tojson }};
|
|
||||||
const graphData = {{ graph_data|tojson }};
|
|
||||||
|
|
||||||
const canvas = document.getElementById('graph-canvas');
|
|
||||||
const ctx = canvas.getContext('2d');
|
|
||||||
const width = canvas.width;
|
|
||||||
const height = canvas.height;
|
|
||||||
const centerX = width / 2;
|
|
||||||
const centerY = height / 2;
|
|
||||||
|
|
||||||
const nodes = graphData.nodes.map(function(node) {
|
|
||||||
return Object.assign({}, node, {
|
|
||||||
x: centerX + (Math.random() - 0.5) * 300,
|
|
||||||
y: centerY + (Math.random() - 0.5) * 300,
|
|
||||||
vx: 0,
|
|
||||||
vy: 0
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
const nodeMap = new Map(nodes.map(function(node) { return [node.id, node]; }));
|
|
||||||
|
|
||||||
const edges = graphData.edges.map(function(edge) {
|
|
||||||
const sourceNode = nodeMap.get(edge.source);
|
|
||||||
const targetNode = nodeMap.get(edge.target);
|
|
||||||
if (!sourceNode || !targetNode) return null;
|
|
||||||
return Object.assign({}, edge, { sourceNode: sourceNode, targetNode: targetNode });
|
|
||||||
}).filter(function(edge) { return edge !== null; });
|
|
||||||
|
|
||||||
let dragNode = null;
|
|
||||||
let selectedNode = null;
|
|
||||||
|
|
||||||
const repulsion = 5000;
|
|
||||||
const springStrength = 0.05;
|
|
||||||
const baseSpringLength = 150;
|
|
||||||
const damping = 0.9;
|
|
||||||
const centerPull = 0.01;
|
|
||||||
|
|
||||||
function simulate() {
|
|
||||||
for (const node of nodes) { node.vx = 0; node.vy = 0; }
|
|
||||||
for (let i = 0; i < nodes.length; i++) {
|
|
||||||
for (let j = i + 1; j < nodes.length; j++) {
|
|
||||||
const dx = nodes[j].x - nodes[i].x;
|
|
||||||
const dy = nodes[j].y - nodes[i].y;
|
|
||||||
const dist = Math.sqrt(dx * dx + dy * dy) || 1;
|
|
||||||
const force = repulsion / (dist * dist);
|
|
||||||
const fx = (dx / dist) * force;
|
|
||||||
const fy = (dy / dist) * force;
|
|
||||||
nodes[i].vx -= fx; nodes[i].vy -= fy;
|
|
||||||
nodes[j].vx += fx; nodes[j].vy += fy;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for (const edge of edges) {
|
|
||||||
const dx = edge.targetNode.x - edge.sourceNode.x;
|
|
||||||
const dy = edge.targetNode.y - edge.sourceNode.y;
|
|
||||||
const dist = Math.sqrt(dx * dx + dy * dy) || 1;
|
|
||||||
const normalizedWeight = edge.closeness_weight / 10;
|
|
||||||
const idealLength = baseSpringLength * (1.5 - normalizedWeight);
|
|
||||||
const displacement = dist - idealLength;
|
|
||||||
const force = springStrength * displacement;
|
|
||||||
const fx = (dx / dist) * force;
|
|
||||||
const fy = (dy / dist) * force;
|
|
||||||
edge.sourceNode.vx += fx; edge.sourceNode.vy += fy;
|
|
||||||
edge.targetNode.vx -= fx; edge.targetNode.vy -= fy;
|
|
||||||
}
|
|
||||||
for (const node of nodes) {
|
|
||||||
node.vx += (centerX - node.x) * centerPull;
|
|
||||||
node.vy += (centerY - node.y) * centerPull;
|
|
||||||
}
|
|
||||||
for (const node of nodes) {
|
|
||||||
if (node === dragNode) continue;
|
|
||||||
node.x += node.vx * damping;
|
|
||||||
node.y += node.vy * damping;
|
|
||||||
node.x = Math.max(30, Math.min(width - 30, node.x));
|
|
||||||
node.y = Math.max(30, Math.min(height - 30, node.y));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
function getEdgeColor(weight) {
|
|
||||||
const normalized = weight / 10;
|
|
||||||
return 'hsl(220, 70%, ' + (80 - normalized * 40) + '%)';
|
|
||||||
}
|
|
||||||
|
|
||||||
function draw() {
|
|
||||||
ctx.clearRect(0, 0, width, height);
|
|
||||||
for (const edge of edges) {
|
|
||||||
const lineWidth = 1 + (edge.closeness_weight / 10) * 3;
|
|
||||||
ctx.strokeStyle = getEdgeColor(edge.closeness_weight);
|
|
||||||
ctx.lineWidth = lineWidth;
|
|
||||||
ctx.beginPath();
|
|
||||||
ctx.moveTo(edge.sourceNode.x, edge.sourceNode.y);
|
|
||||||
ctx.lineTo(edge.targetNode.x, edge.targetNode.y);
|
|
||||||
ctx.stroke();
|
|
||||||
const midX = (edge.sourceNode.x + edge.targetNode.x) / 2;
|
|
||||||
const midY = (edge.sourceNode.y + edge.targetNode.y) / 2;
|
|
||||||
ctx.fillStyle = '#666';
|
|
||||||
ctx.font = '10px sans-serif';
|
|
||||||
ctx.textAlign = 'center';
|
|
||||||
const label = RELATIONSHIP_DISPLAY[edge.relationship_type] || edge.relationship_type;
|
|
||||||
ctx.fillText(label, midX, midY - 5);
|
|
||||||
}
|
|
||||||
for (const node of nodes) {
|
|
||||||
const isSelected = node === selectedNode;
|
|
||||||
const radius = isSelected ? 25 : 20;
|
|
||||||
ctx.beginPath();
|
|
||||||
ctx.arc(node.x, node.y, radius, 0, Math.PI * 2);
|
|
||||||
ctx.fillStyle = isSelected ? '#0066cc' : '#fff';
|
|
||||||
ctx.fill();
|
|
||||||
ctx.strokeStyle = '#0066cc';
|
|
||||||
ctx.lineWidth = 2;
|
|
||||||
ctx.stroke();
|
|
||||||
ctx.fillStyle = isSelected ? '#fff' : '#333';
|
|
||||||
ctx.font = '12px sans-serif';
|
|
||||||
ctx.textAlign = 'center';
|
|
||||||
ctx.textBaseline = 'middle';
|
|
||||||
const name = node.name.length > 10 ? node.name.slice(0, 9) + '\u2026' : node.name;
|
|
||||||
ctx.fillText(name, node.x, node.y);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
function animate() {
|
|
||||||
simulate();
|
|
||||||
draw();
|
|
||||||
requestAnimationFrame(animate);
|
|
||||||
}
|
|
||||||
animate();
|
|
||||||
|
|
||||||
function getNodeAt(x, y) {
|
|
||||||
for (const node of nodes) {
|
|
||||||
const dx = x - node.x;
|
|
||||||
const dy = y - node.y;
|
|
||||||
if (dx * dx + dy * dy < 400) return node;
|
|
||||||
}
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|
||||||
canvas.addEventListener('mousedown', function(event) {
|
|
||||||
const rect = canvas.getBoundingClientRect();
|
|
||||||
const node = getNodeAt(event.clientX - rect.left, event.clientY - rect.top);
|
|
||||||
if (node) {
|
|
||||||
dragNode = node;
|
|
||||||
selectedNode = node;
|
|
||||||
const infoDiv = document.getElementById('selected-info');
|
|
||||||
let html = '<div class="selected-info"><h3>' + node.name + '</h3>';
|
|
||||||
if (node.current_job) html += '<p>Job: ' + node.current_job + '</p>';
|
|
||||||
html += '<a href="/contacts/' + node.id + '">View details</a></div>';
|
|
||||||
infoDiv.innerHTML = html;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
canvas.addEventListener('mousemove', function(event) {
|
|
||||||
if (!dragNode) return;
|
|
||||||
const rect = canvas.getBoundingClientRect();
|
|
||||||
dragNode.x = event.clientX - rect.left;
|
|
||||||
dragNode.y = event.clientY - rect.top;
|
|
||||||
});
|
|
||||||
|
|
||||||
canvas.addEventListener('mouseup', function() { dragNode = null; });
|
|
||||||
canvas.addEventListener('mouseleave', function() { dragNode = null; });
|
|
||||||
})();
|
|
||||||
</script>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,31 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}Needs{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<div class="need-list">
|
|
||||||
<div class="header">
|
|
||||||
<h1>Needs / Accommodations</h1>
|
|
||||||
<button class="btn btn-primary" onclick="document.getElementById('need-form').toggleAttribute('hidden')">Add Need</button>
|
|
||||||
</div>
|
|
||||||
|
|
||||||
<form id="need-form" hidden
|
|
||||||
hx-post="/htmx/needs"
|
|
||||||
hx-target="#need-items"
|
|
||||||
hx-swap="innerHTML"
|
|
||||||
hx-on::after-request="if(event.detail.successful) this.reset()"
|
|
||||||
class="need-form">
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="name">Name *</label>
|
|
||||||
<input id="name" name="name" type="text" placeholder="e.g., Light Sensitive, ADHD" required>
|
|
||||||
</div>
|
|
||||||
<div class="form-group">
|
|
||||||
<label for="description">Description</label>
|
|
||||||
<textarea id="description" name="description" placeholder="Optional description..." rows="2"></textarea>
|
|
||||||
</div>
|
|
||||||
<button type="submit" class="btn btn-primary">Create</button>
|
|
||||||
</form>
|
|
||||||
|
|
||||||
<div id="need-items">
|
|
||||||
{% include "partials/need_items.html" %}
|
|
||||||
</div>
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,33 +0,0 @@
|
|||||||
{% if contacts %}
|
|
||||||
<table>
|
|
||||||
<thead>
|
|
||||||
<tr>
|
|
||||||
<th>Name</th>
|
|
||||||
<th>Job</th>
|
|
||||||
<th>Timezone</th>
|
|
||||||
<th>Actions</th>
|
|
||||||
</tr>
|
|
||||||
</thead>
|
|
||||||
<tbody>
|
|
||||||
{% for contact in contacts %}
|
|
||||||
<tr id="contact-row-{{ contact.id }}">
|
|
||||||
<td><a href="/contacts/{{ contact.id }}">{{ contact.name }}</a></td>
|
|
||||||
<td>{{ contact.current_job or "-" }}</td>
|
|
||||||
<td>{{ contact.timezone or "-" }}</td>
|
|
||||||
<td>
|
|
||||||
<a href="/contacts/{{ contact.id }}/edit" class="btn">Edit</a>
|
|
||||||
<button class="btn btn-danger"
|
|
||||||
hx-delete="/api/contacts/{{ contact.id }}"
|
|
||||||
hx-target="#contact-row-{{ contact.id }}"
|
|
||||||
hx-swap="outerHTML"
|
|
||||||
hx-confirm="Delete this contact?">
|
|
||||||
Delete
|
|
||||||
</button>
|
|
||||||
</td>
|
|
||||||
</tr>
|
|
||||||
{% endfor %}
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
{% else %}
|
|
||||||
<p>No contacts yet.</p>
|
|
||||||
{% endif %}
|
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
<ul class="manage-needs-list">
|
|
||||||
{% for need in contact.needs %}
|
|
||||||
<li id="contact-need-{{ need.id }}">
|
|
||||||
<strong>{{ need.name }}</strong>
|
|
||||||
{% if need.description %}<span> - {{ need.description }}</span>{% endif %}
|
|
||||||
<button class="btn btn-small btn-danger"
|
|
||||||
hx-delete="/api/contacts/{{ contact.id }}/needs/{{ need.id }}"
|
|
||||||
hx-target="#contact-need-{{ need.id }}"
|
|
||||||
hx-swap="outerHTML">
|
|
||||||
Remove
|
|
||||||
</button>
|
|
||||||
</li>
|
|
||||||
{% endfor %}
|
|
||||||
</ul>
|
|
||||||
@@ -1,23 +0,0 @@
|
|||||||
{% for rel in contact.related_to %}
|
|
||||||
<div class="manage-rel-item" id="rel-{{ contact.id }}-{{ rel.related_contact_id }}">
|
|
||||||
<a href="/contacts/{{ rel.related_contact_id }}">{{ contact_names[rel.related_contact_id] }}</a>
|
|
||||||
<span class="tag">{{ rel.relationship_type|replace("_", " ")|title }}</span>
|
|
||||||
<label class="weight-control">
|
|
||||||
<span>Closeness:</span>
|
|
||||||
<input type="range" min="1" max="10" value="{{ rel.closeness_weight }}"
|
|
||||||
hx-post="/htmx/contacts/{{ contact.id }}/relationships/{{ rel.related_contact_id }}/weight"
|
|
||||||
hx-trigger="change"
|
|
||||||
hx-include="this"
|
|
||||||
name="closeness_weight"
|
|
||||||
hx-swap="none"
|
|
||||||
oninput="this.nextElementSibling.textContent = this.value">
|
|
||||||
<span class="weight-value">{{ rel.closeness_weight }}</span>
|
|
||||||
</label>
|
|
||||||
<button class="btn btn-small btn-danger"
|
|
||||||
hx-delete="/api/contacts/{{ contact.id }}/relationships/{{ rel.related_contact_id }}"
|
|
||||||
hx-target="#rel-{{ contact.id }}-{{ rel.related_contact_id }}"
|
|
||||||
hx-swap="outerHTML">
|
|
||||||
Remove
|
|
||||||
</button>
|
|
||||||
</div>
|
|
||||||
{% endfor %}
|
|
||||||
@@ -1,21 +0,0 @@
|
|||||||
{% if needs %}
|
|
||||||
<ul class="need-items">
|
|
||||||
{% for need in needs %}
|
|
||||||
<li id="need-item-{{ need.id }}">
|
|
||||||
<div class="need-info">
|
|
||||||
<strong>{{ need.name }}</strong>
|
|
||||||
{% if need.description %}<p>{{ need.description }}</p>{% endif %}
|
|
||||||
</div>
|
|
||||||
<button class="btn btn-danger"
|
|
||||||
hx-delete="/api/needs/{{ need.id }}"
|
|
||||||
hx-target="#need-item-{{ need.id }}"
|
|
||||||
hx-swap="outerHTML"
|
|
||||||
hx-confirm="Delete this need?">
|
|
||||||
Delete
|
|
||||||
</button>
|
|
||||||
</li>
|
|
||||||
{% endfor %}
|
|
||||||
</ul>
|
|
||||||
{% else %}
|
|
||||||
<p>No needs defined yet.</p>
|
|
||||||
{% endif %}
|
|
||||||
@@ -1,122 +0,0 @@
|
|||||||
"""CLI wrapper around alembic for multi-database support.
|
|
||||||
|
|
||||||
Usage:
|
|
||||||
database <db_name> <command> [args...]
|
|
||||||
|
|
||||||
Examples:
|
|
||||||
database van_inventory upgrade head
|
|
||||||
database van_inventory downgrade head-1
|
|
||||||
database van_inventory revision --autogenerate -m "add meals table"
|
|
||||||
database van_inventory check
|
|
||||||
database richie check
|
|
||||||
database richie upgrade head
|
|
||||||
"""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from dataclasses import dataclass
|
|
||||||
from importlib import import_module
|
|
||||||
from typing import TYPE_CHECKING, Annotated
|
|
||||||
|
|
||||||
import typer
|
|
||||||
from alembic.config import CommandLine, Config
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from sqlalchemy.orm import DeclarativeBase
|
|
||||||
|
|
||||||
|
|
||||||
@dataclass(frozen=True)
|
|
||||||
class DatabaseConfig:
|
|
||||||
"""Configuration for a database."""
|
|
||||||
|
|
||||||
env_prefix: str
|
|
||||||
version_location: str
|
|
||||||
base_module: str
|
|
||||||
base_class_name: str
|
|
||||||
models_module: str
|
|
||||||
script_location: str = "python/alembic"
|
|
||||||
file_template: str = "%%(year)d_%%(month).2d_%%(day).2d-%%(slug)s_%%(rev)s"
|
|
||||||
|
|
||||||
def get_base(self) -> type[DeclarativeBase]:
|
|
||||||
"""Import and return the Base class."""
|
|
||||||
module = import_module(self.base_module)
|
|
||||||
return getattr(module, self.base_class_name)
|
|
||||||
|
|
||||||
def import_models(self) -> None:
|
|
||||||
"""Import ORM models so alembic autogenerate can detect them."""
|
|
||||||
import_module(self.models_module)
|
|
||||||
|
|
||||||
def alembic_config(self) -> Config:
|
|
||||||
"""Build an alembic Config for this database."""
|
|
||||||
# Runtime import needed — Config is in TYPE_CHECKING for the return type annotation
|
|
||||||
from alembic.config import Config as AlembicConfig # noqa: PLC0415
|
|
||||||
|
|
||||||
cfg = AlembicConfig()
|
|
||||||
cfg.set_main_option("script_location", self.script_location)
|
|
||||||
cfg.set_main_option("file_template", self.file_template)
|
|
||||||
cfg.set_main_option("prepend_sys_path", ".")
|
|
||||||
cfg.set_main_option("version_path_separator", "os")
|
|
||||||
cfg.set_main_option("version_locations", self.version_location)
|
|
||||||
cfg.set_main_option("revision_environment", "true")
|
|
||||||
cfg.set_section_option("post_write_hooks", "hooks", "dynamic_schema,import_postgresql,ruff")
|
|
||||||
cfg.set_section_option("post_write_hooks", "dynamic_schema.type", "dynamic_schema")
|
|
||||||
cfg.set_section_option("post_write_hooks", "import_postgresql.type", "import_postgresql")
|
|
||||||
cfg.set_section_option("post_write_hooks", "ruff.type", "ruff")
|
|
||||||
cfg.attributes["base"] = self.get_base()
|
|
||||||
cfg.attributes["env_prefix"] = self.env_prefix
|
|
||||||
self.import_models()
|
|
||||||
return cfg
|
|
||||||
|
|
||||||
|
|
||||||
DATABASES: dict[str, DatabaseConfig] = {
|
|
||||||
"richie": DatabaseConfig(
|
|
||||||
env_prefix="RICHIE",
|
|
||||||
version_location="python/alembic/richie/versions",
|
|
||||||
base_module="python.orm.richie.base",
|
|
||||||
base_class_name="RichieBase",
|
|
||||||
models_module="python.orm.richie",
|
|
||||||
),
|
|
||||||
"van_inventory": DatabaseConfig(
|
|
||||||
env_prefix="VAN_INVENTORY",
|
|
||||||
version_location="python/alembic/van_inventory/versions",
|
|
||||||
base_module="python.orm.van_inventory.base",
|
|
||||||
base_class_name="VanInventoryBase",
|
|
||||||
models_module="python.orm.van_inventory.models",
|
|
||||||
),
|
|
||||||
"signal_bot": DatabaseConfig(
|
|
||||||
env_prefix="SIGNALBOT",
|
|
||||||
version_location="python/alembic/signal_bot/versions",
|
|
||||||
base_module="python.orm.signal_bot.base",
|
|
||||||
base_class_name="SignalBotBase",
|
|
||||||
models_module="python.orm.signal_bot.models",
|
|
||||||
),
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
app = typer.Typer(help="Multi-database alembic wrapper.")
|
|
||||||
|
|
||||||
|
|
||||||
@app.command(
|
|
||||||
context_settings={"allow_extra_args": True, "ignore_unknown_options": True},
|
|
||||||
)
|
|
||||||
def main(
|
|
||||||
ctx: typer.Context,
|
|
||||||
db_name: Annotated[str, typer.Argument(help=f"Database name. Options: {', '.join(DATABASES)}")],
|
|
||||||
command: Annotated[str, typer.Argument(help="Alembic command (upgrade, downgrade, revision, check, etc.)")],
|
|
||||||
) -> None:
|
|
||||||
"""Run an alembic command against the specified database."""
|
|
||||||
db_config = DATABASES.get(db_name)
|
|
||||||
if not db_config:
|
|
||||||
typer.echo(f"Unknown database: {db_name!r}. Available: {', '.join(DATABASES)}", err=True)
|
|
||||||
raise typer.Exit(code=1)
|
|
||||||
|
|
||||||
alembic_cfg = db_config.alembic_config()
|
|
||||||
|
|
||||||
cmd_line = CommandLine()
|
|
||||||
options = cmd_line.parser.parse_args([command, *ctx.args])
|
|
||||||
cmd_line.run_cmd(alembic_cfg, options)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
app()
|
|
||||||
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Detect Nix evaluation warnings from build logs and create PRs with LLM-suggested fixes."""
|
|
||||||
@@ -1,449 +0,0 @@
|
|||||||
"""Detect Nix evaluation warnings and create PRs with LLM-suggested fixes."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import hashlib
|
|
||||||
import logging
|
|
||||||
import re
|
|
||||||
import subprocess
|
|
||||||
from dataclasses import dataclass
|
|
||||||
from io import BytesIO
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import Annotated
|
|
||||||
from zipfile import ZipFile
|
|
||||||
|
|
||||||
import typer
|
|
||||||
from httpx import HTTPError, post
|
|
||||||
|
|
||||||
from python.common import configure_logger
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
@dataclass(frozen=True)
|
|
||||||
class EvalWarning:
|
|
||||||
"""A single Nix evaluation warning."""
|
|
||||||
|
|
||||||
system: str
|
|
||||||
message: str
|
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
|
||||||
class FileChange:
|
|
||||||
"""A file change suggested by the LLM."""
|
|
||||||
|
|
||||||
file_path: str
|
|
||||||
original: str
|
|
||||||
fixed: str
|
|
||||||
|
|
||||||
|
|
||||||
def run_cmd(cmd: list[str], *, check: bool = True) -> subprocess.CompletedProcess[str]:
|
|
||||||
"""Run a subprocess command and return the result.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
cmd: Command and arguments.
|
|
||||||
check: Whether to raise on non-zero exit.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
CompletedProcess with captured stdout/stderr.
|
|
||||||
"""
|
|
||||||
logger.debug("Running: %s", " ".join(cmd))
|
|
||||||
return subprocess.run(cmd, capture_output=True, text=True, check=check)
|
|
||||||
|
|
||||||
|
|
||||||
def download_logs(run_id: str, repo: str) -> dict[str, str]:
|
|
||||||
"""Download build logs for a GitHub Actions run.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
run_id: The workflow run ID.
|
|
||||||
repo: The GitHub repository (owner/repo).
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Dict mapping zip entry names to their text content, filtered to build log files.
|
|
||||||
|
|
||||||
Raises:
|
|
||||||
RuntimeError: If log download fails.
|
|
||||||
"""
|
|
||||||
result = subprocess.run(
|
|
||||||
["gh", "api", f"repos/{repo}/actions/runs/{run_id}/logs"],
|
|
||||||
capture_output=True,
|
|
||||||
check=False,
|
|
||||||
)
|
|
||||||
if result.returncode != 0:
|
|
||||||
msg = f"Failed to download logs: {result.stderr.decode(errors='replace')}"
|
|
||||||
raise RuntimeError(msg)
|
|
||||||
|
|
||||||
logs: dict[str, str] = {}
|
|
||||||
with ZipFile(BytesIO(result.stdout)) as zip_file:
|
|
||||||
for name in zip_file.namelist():
|
|
||||||
if name.startswith("build-") and name.endswith(".txt"):
|
|
||||||
logs[name] = zip_file.read(name).decode(errors="replace")
|
|
||||||
|
|
||||||
return logs
|
|
||||||
|
|
||||||
|
|
||||||
def parse_warnings(logs: dict[str, str]) -> set[EvalWarning]:
|
|
||||||
"""Parse Nix evaluation warnings from build log contents.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
logs: Dict mapping zip entry names (e.g. "build-bob/2_Build.txt") to their text.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Deduplicated set of warnings.
|
|
||||||
"""
|
|
||||||
warnings: set[EvalWarning] = set()
|
|
||||||
warning_pattern = re.compile(r"(?:^[\d\-T:.Z]+ )?(warning:|trace: warning:)")
|
|
||||||
timestamp_prefix = re.compile(r"^[\d\-T:.Z]+ ")
|
|
||||||
|
|
||||||
for name, content in sorted(logs.items()):
|
|
||||||
system = name.split("/")[0].removeprefix("build-")
|
|
||||||
for line in content.splitlines():
|
|
||||||
if warning_pattern.search(line):
|
|
||||||
message = timestamp_prefix.sub("", line).strip()
|
|
||||||
if message.startswith("warning: ignoring untrusted flake configuration setting"):
|
|
||||||
continue
|
|
||||||
logger.debug(f"Found warning: {line}")
|
|
||||||
warnings.add(EvalWarning(system=system, message=message))
|
|
||||||
|
|
||||||
logger.info("Found %d unique warnings", len(warnings))
|
|
||||||
return warnings
|
|
||||||
|
|
||||||
|
|
||||||
def extract_referenced_files(warnings: set[EvalWarning]) -> dict[str, str]:
|
|
||||||
"""Extract file paths referenced in warnings and read their contents.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
warnings: List of parsed warnings.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Dict mapping repo-relative file paths to their contents.
|
|
||||||
"""
|
|
||||||
paths: set[str] = set()
|
|
||||||
warning_text = "\n".join(w.message for w in warnings)
|
|
||||||
|
|
||||||
nix_store_path = re.compile(r"/nix/store/[^/]+-source/([^:]+\.nix)")
|
|
||||||
for match in nix_store_path.finditer(warning_text):
|
|
||||||
paths.add(match.group(1))
|
|
||||||
|
|
||||||
repo_relative_path = re.compile(r"(?<![/\w])(systems|common|users|overlays)/[^:\s]+\.nix")
|
|
||||||
for match in repo_relative_path.finditer(warning_text):
|
|
||||||
paths.add(match.group(0))
|
|
||||||
|
|
||||||
files: dict[str, str] = {}
|
|
||||||
for path_str in sorted(paths):
|
|
||||||
path = Path(path_str)
|
|
||||||
if path.is_file():
|
|
||||||
files[path_str] = path.read_text()
|
|
||||||
|
|
||||||
if not files and Path("flake.nix").is_file():
|
|
||||||
files["flake.nix"] = Path("flake.nix").read_text()
|
|
||||||
|
|
||||||
logger.info("Extracted %d referenced files", len(files))
|
|
||||||
return files
|
|
||||||
|
|
||||||
|
|
||||||
def compute_warning_hash(warnings: set[EvalWarning]) -> str:
|
|
||||||
"""Compute a short hash of the warning set for deduplication.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
warnings: List of warnings.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
8-character hex hash.
|
|
||||||
"""
|
|
||||||
text = "\n".join(sorted(f"[{w.system}] {w.message}" for w in warnings))
|
|
||||||
return hashlib.sha256(text.encode()).hexdigest()[:8]
|
|
||||||
|
|
||||||
|
|
||||||
def check_duplicate_pr(warning_hash: str) -> bool:
|
|
||||||
"""Check if an open PR already exists for this warning hash.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
warning_hash: The hash to check.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
True if a duplicate PR exists.
|
|
||||||
|
|
||||||
Raises:
|
|
||||||
RuntimeError: If the gh CLI call fails.
|
|
||||||
"""
|
|
||||||
result = run_cmd(
|
|
||||||
[
|
|
||||||
"gh",
|
|
||||||
"pr",
|
|
||||||
"list",
|
|
||||||
"--state",
|
|
||||||
"open",
|
|
||||||
"--label",
|
|
||||||
"eval-warning-fix",
|
|
||||||
"--json",
|
|
||||||
"title",
|
|
||||||
"--jq",
|
|
||||||
".[].title",
|
|
||||||
],
|
|
||||||
check=False,
|
|
||||||
)
|
|
||||||
if result.returncode != 0:
|
|
||||||
msg = f"Failed to check for duplicate PRs: {result.stderr}"
|
|
||||||
raise RuntimeError(msg)
|
|
||||||
|
|
||||||
for title in result.stdout.splitlines():
|
|
||||||
if warning_hash in title:
|
|
||||||
logger.info("Duplicate PR found for hash %s", warning_hash)
|
|
||||||
return True
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def query_ollama(
|
|
||||||
warnings: set[EvalWarning],
|
|
||||||
files: dict[str, str],
|
|
||||||
ollama_url: str,
|
|
||||||
) -> str | None:
|
|
||||||
"""Query Ollama for a fix suggestion.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
warnings: List of warnings.
|
|
||||||
files: Referenced file contents.
|
|
||||||
ollama_url: Ollama API base URL.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
LLM response text, or None on failure.
|
|
||||||
"""
|
|
||||||
warning_text = "\n".join(f"[{w.system}] {w.message}" for w in warnings)
|
|
||||||
file_context = "\n".join(f"--- FILE: {path} ---\n{content}\n--- END FILE ---" for path, content in files.items())
|
|
||||||
|
|
||||||
prompt = f"""You are a NixOS configuration expert. \
|
|
||||||
Analyze the following Nix evaluation warnings and suggest fixes.
|
|
||||||
|
|
||||||
## Warnings
|
|
||||||
{warning_text}
|
|
||||||
|
|
||||||
## Referenced Files
|
|
||||||
{file_context}
|
|
||||||
|
|
||||||
## Instructions
|
|
||||||
- Identify the root cause of each warning
|
|
||||||
- Provide the exact file changes needed to fix the warnings
|
|
||||||
- Output your response in two clearly separated sections:
|
|
||||||
1. **REASONING**: Brief explanation of what causes each warning and how to fix it
|
|
||||||
2. **CHANGES**: For each file that needs changes, output a block like:
|
|
||||||
FILE: path/to/file.nix
|
|
||||||
<<<<<<< ORIGINAL
|
|
||||||
the original lines to replace
|
|
||||||
=======
|
|
||||||
the replacement lines
|
|
||||||
>>>>>>> FIXED
|
|
||||||
- Only suggest changes for files that exist in the repository
|
|
||||||
- Do not add unnecessary complexity
|
|
||||||
- Preserve the existing code style
|
|
||||||
- If a warning comes from upstream nixpkgs and cannot be fixed in this repo, \
|
|
||||||
say so in REASONING and do not suggest changes"""
|
|
||||||
|
|
||||||
try:
|
|
||||||
response = post(
|
|
||||||
f"{ollama_url}/api/generate",
|
|
||||||
json={
|
|
||||||
"model": "qwen3-coder:30b",
|
|
||||||
"prompt": prompt,
|
|
||||||
"stream": False,
|
|
||||||
"options": {"num_predict": 4096},
|
|
||||||
},
|
|
||||||
timeout=300,
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
|
||||||
except HTTPError:
|
|
||||||
logger.exception("Ollama request failed")
|
|
||||||
return None
|
|
||||||
|
|
||||||
return response.json().get("response")
|
|
||||||
|
|
||||||
|
|
||||||
def parse_changes(response: str) -> list[FileChange]:
|
|
||||||
"""Parse file changes from the **CHANGES** section of the LLM response.
|
|
||||||
|
|
||||||
Expects blocks in the format:
|
|
||||||
FILE: path/to/file.nix
|
|
||||||
<<<<<<< ORIGINAL
|
|
||||||
...
|
|
||||||
=======
|
|
||||||
...
|
|
||||||
>>>>>>> FIXED
|
|
||||||
|
|
||||||
Args:
|
|
||||||
response: Raw LLM response text.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
List of parsed file changes.
|
|
||||||
"""
|
|
||||||
if "**CHANGES**" not in response:
|
|
||||||
logger.warning("LLM response missing **CHANGES** section")
|
|
||||||
return []
|
|
||||||
|
|
||||||
changes_section = response.split("**CHANGES**", 1)[1]
|
|
||||||
|
|
||||||
changes: list[FileChange] = []
|
|
||||||
current_file = ""
|
|
||||||
section: str | None = None
|
|
||||||
original_lines: list[str] = []
|
|
||||||
fixed_lines: list[str] = []
|
|
||||||
|
|
||||||
for line in changes_section.splitlines():
|
|
||||||
stripped = line.strip()
|
|
||||||
if stripped.startswith("FILE:"):
|
|
||||||
current_file = stripped.removeprefix("FILE:").strip()
|
|
||||||
elif stripped == "<<<<<<< ORIGINAL":
|
|
||||||
section = "original"
|
|
||||||
original_lines = []
|
|
||||||
elif stripped == "=======" and section == "original":
|
|
||||||
section = "fixed"
|
|
||||||
fixed_lines = []
|
|
||||||
elif stripped == ">>>>>>> FIXED" and section == "fixed":
|
|
||||||
section = None
|
|
||||||
if current_file:
|
|
||||||
changes.append(FileChange(current_file, "\n".join(original_lines), "\n".join(fixed_lines)))
|
|
||||||
elif section == "original":
|
|
||||||
original_lines.append(line)
|
|
||||||
elif section == "fixed":
|
|
||||||
fixed_lines.append(line)
|
|
||||||
|
|
||||||
logger.info("Parsed %d file changes", len(changes))
|
|
||||||
return changes
|
|
||||||
|
|
||||||
|
|
||||||
def apply_changes(changes: list[FileChange]) -> int:
|
|
||||||
"""Apply file changes to the working directory.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
changes: List of changes to apply.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Number of changes successfully applied.
|
|
||||||
"""
|
|
||||||
applied = 0
|
|
||||||
cwd = Path.cwd().resolve()
|
|
||||||
for change in changes:
|
|
||||||
path = Path(change.file_path).resolve()
|
|
||||||
if not path.is_relative_to(cwd):
|
|
||||||
logger.warning("Path traversal blocked: %s", change.file_path)
|
|
||||||
continue
|
|
||||||
if not path.is_file():
|
|
||||||
logger.warning("File not found: %s", change.file_path)
|
|
||||||
continue
|
|
||||||
|
|
||||||
content = path.read_text()
|
|
||||||
if change.original not in content:
|
|
||||||
logger.warning("Original text not found in %s", change.file_path)
|
|
||||||
continue
|
|
||||||
|
|
||||||
path.write_text(content.replace(change.original, change.fixed, 1))
|
|
||||||
logger.info("Applied fix to %s", change.file_path)
|
|
||||||
applied += 1
|
|
||||||
|
|
||||||
return applied
|
|
||||||
|
|
||||||
|
|
||||||
def create_pr(
|
|
||||||
warning_hash: str,
|
|
||||||
warnings: set[EvalWarning],
|
|
||||||
llm_response: str,
|
|
||||||
run_url: str,
|
|
||||||
) -> None:
|
|
||||||
"""Create a git branch and PR with the applied fixes.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
warning_hash: Short hash for branch naming and deduplication.
|
|
||||||
warnings: Original warnings for the PR body.
|
|
||||||
llm_response: Full LLM response for extracting reasoning.
|
|
||||||
run_url: URL to the triggering build run.
|
|
||||||
"""
|
|
||||||
branch = f"fix/eval-warning-{warning_hash}"
|
|
||||||
warning_text = "\n".join(f"[{w.system}] {w.message}" for w in warnings)
|
|
||||||
|
|
||||||
if "**REASONING**" not in llm_response:
|
|
||||||
logger.warning("LLM response missing **REASONING** section")
|
|
||||||
reasoning = ""
|
|
||||||
else:
|
|
||||||
_, after = llm_response.split("**REASONING**", 1)
|
|
||||||
reasoning = "\n".join(after.split("**CHANGES**", 1)[0].strip().splitlines()[:50])
|
|
||||||
|
|
||||||
run_cmd(["git", "config", "user.name", "github-actions[bot]"])
|
|
||||||
run_cmd(["git", "config", "user.email", "github-actions[bot]@users.noreply.github.com"])
|
|
||||||
run_cmd(["git", "checkout", "-b", branch])
|
|
||||||
run_cmd(["git", "add", "-A"])
|
|
||||||
|
|
||||||
diff_result = run_cmd(["git", "diff", "--cached", "--quiet"], check=False)
|
|
||||||
if diff_result.returncode == 0:
|
|
||||||
logger.info("No file changes to commit")
|
|
||||||
return
|
|
||||||
|
|
||||||
run_cmd(["git", "commit", "-m", f"fix: resolve nix evaluation warnings ({warning_hash})"])
|
|
||||||
run_cmd(["git", "push", "origin", branch, "--force"])
|
|
||||||
|
|
||||||
body = f"""## Nix Evaluation Warnings
|
|
||||||
|
|
||||||
Detected in [build_systems run]({run_url}):
|
|
||||||
|
|
||||||
```
|
|
||||||
{warning_text}
|
|
||||||
```
|
|
||||||
|
|
||||||
## LLM Analysis (qwen3-coder:30b)
|
|
||||||
|
|
||||||
{reasoning}
|
|
||||||
|
|
||||||
---
|
|
||||||
*Auto-generated by fix_eval_warnings. Review carefully before merging.*"""
|
|
||||||
|
|
||||||
run_cmd(
|
|
||||||
[
|
|
||||||
"gh",
|
|
||||||
"pr",
|
|
||||||
"create",
|
|
||||||
"--title",
|
|
||||||
f"fix: resolve nix eval warnings ({warning_hash})",
|
|
||||||
"--label",
|
|
||||||
"automated",
|
|
||||||
"--label",
|
|
||||||
"eval-warning-fix",
|
|
||||||
"--body",
|
|
||||||
body,
|
|
||||||
]
|
|
||||||
)
|
|
||||||
logger.info("PR created on branch %s", branch)
|
|
||||||
|
|
||||||
|
|
||||||
def main(
|
|
||||||
run_id: Annotated[str, typer.Option("--run-id", help="GitHub Actions run ID")],
|
|
||||||
repo: Annotated[str, typer.Option("--repo", help="GitHub repository (owner/repo)")],
|
|
||||||
ollama_url: Annotated[str, typer.Option("--ollama-url", help="Ollama API base URL")],
|
|
||||||
run_url: Annotated[str, typer.Option("--run-url", help="URL to the triggering build run")],
|
|
||||||
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
|
||||||
) -> None:
|
|
||||||
"""Detect Nix evaluation warnings and create PRs with LLM-suggested fixes."""
|
|
||||||
configure_logger(log_level)
|
|
||||||
|
|
||||||
logs = download_logs(run_id, repo)
|
|
||||||
warnings = parse_warnings(logs)
|
|
||||||
if not warnings:
|
|
||||||
return
|
|
||||||
|
|
||||||
warning_hash = compute_warning_hash(warnings)
|
|
||||||
if check_duplicate_pr(warning_hash):
|
|
||||||
return
|
|
||||||
|
|
||||||
files = extract_referenced_files(warnings)
|
|
||||||
llm_response = query_ollama(warnings, files, ollama_url)
|
|
||||||
if not llm_response:
|
|
||||||
return
|
|
||||||
|
|
||||||
changes = parse_changes(llm_response)
|
|
||||||
applied = apply_changes(changes)
|
|
||||||
if applied == 0:
|
|
||||||
logger.info("No changes could be applied")
|
|
||||||
return
|
|
||||||
|
|
||||||
create_pr(warning_hash, warnings, llm_response, run_url)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
typer.run(main)
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Tuya heater control service."""
|
|
||||||
@@ -1,69 +0,0 @@
|
|||||||
"""TinyTuya device controller for heater."""
|
|
||||||
|
|
||||||
import logging
|
|
||||||
|
|
||||||
import tinytuya
|
|
||||||
|
|
||||||
from python.heater.models import ActionResult, DeviceConfig, HeaterStatus
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
# DPS mapping for heater
|
|
||||||
DPS_POWER = "1" # bool: on/off
|
|
||||||
DPS_SETPOINT = "101" # int: target temp (read-only)
|
|
||||||
DPS_STATE = "102" # str: "Stop", "Heat", etc.
|
|
||||||
DPS_UNKNOWN = "104" # int: unknown
|
|
||||||
DPS_ERROR = "108" # int: last error code
|
|
||||||
|
|
||||||
|
|
||||||
class HeaterController:
|
|
||||||
"""Controls a Tuya heater device via local network."""
|
|
||||||
|
|
||||||
def __init__(self, config: DeviceConfig) -> None:
|
|
||||||
"""Initialize the controller."""
|
|
||||||
self.device = tinytuya.Device(config.device_id, config.ip, config.local_key)
|
|
||||||
self.device.set_version(config.version)
|
|
||||||
self.device.set_socketTimeout(0.5)
|
|
||||||
self.device.set_socketRetryLimit(1)
|
|
||||||
|
|
||||||
def status(self) -> HeaterStatus:
|
|
||||||
"""Get current heater status."""
|
|
||||||
data = self.device.status()
|
|
||||||
|
|
||||||
if "Error" in data:
|
|
||||||
logger.error("Device error: %s", data)
|
|
||||||
return HeaterStatus(power=False, raw_dps={"error": data["Error"]})
|
|
||||||
|
|
||||||
dps = data.get("dps", {})
|
|
||||||
return HeaterStatus(
|
|
||||||
power=bool(dps.get(DPS_POWER, False)),
|
|
||||||
setpoint=dps.get(DPS_SETPOINT),
|
|
||||||
state=dps.get(DPS_STATE),
|
|
||||||
error_code=dps.get(DPS_ERROR),
|
|
||||||
raw_dps=dps,
|
|
||||||
)
|
|
||||||
|
|
||||||
def turn_on(self) -> ActionResult:
|
|
||||||
"""Turn heater on."""
|
|
||||||
try:
|
|
||||||
self.device.set_value(index=DPS_POWER, value=True)
|
|
||||||
return ActionResult(success=True, action="on", power=True)
|
|
||||||
except Exception as error:
|
|
||||||
logger.exception("Failed to turn on")
|
|
||||||
return ActionResult(success=False, action="on", error=str(error))
|
|
||||||
|
|
||||||
def turn_off(self) -> ActionResult:
|
|
||||||
"""Turn heater off."""
|
|
||||||
try:
|
|
||||||
self.device.set_value(index=DPS_POWER, value=False)
|
|
||||||
return ActionResult(success=True, action="off", power=False)
|
|
||||||
except Exception as error:
|
|
||||||
logger.exception("Failed to turn off")
|
|
||||||
return ActionResult(success=False, action="off", error=str(error))
|
|
||||||
|
|
||||||
def toggle(self) -> ActionResult:
|
|
||||||
"""Toggle heater power state."""
|
|
||||||
status = self.status()
|
|
||||||
if status.power:
|
|
||||||
return self.turn_off()
|
|
||||||
return self.turn_on()
|
|
||||||
@@ -1,85 +0,0 @@
|
|||||||
"""FastAPI heater control service."""
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from collections.abc import AsyncIterator
|
|
||||||
from contextlib import asynccontextmanager
|
|
||||||
from typing import Annotated
|
|
||||||
|
|
||||||
import typer
|
|
||||||
import uvicorn
|
|
||||||
from fastapi import FastAPI, HTTPException
|
|
||||||
|
|
||||||
from python.common import configure_logger
|
|
||||||
from python.heater.controller import HeaterController
|
|
||||||
from python.heater.models import ActionResult, DeviceConfig, HeaterStatus
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def create_app(config: DeviceConfig) -> FastAPI:
|
|
||||||
"""Create FastAPI application."""
|
|
||||||
|
|
||||||
@asynccontextmanager
|
|
||||||
async def lifespan(app: FastAPI) -> AsyncIterator[None]:
|
|
||||||
app.state.controller = HeaterController(config)
|
|
||||||
yield
|
|
||||||
|
|
||||||
app = FastAPI(
|
|
||||||
title="Heater Control API",
|
|
||||||
description="Fast local control for Tuya heater",
|
|
||||||
lifespan=lifespan,
|
|
||||||
)
|
|
||||||
|
|
||||||
@app.get("/status")
|
|
||||||
def get_status() -> HeaterStatus:
|
|
||||||
return app.state.controller.status()
|
|
||||||
|
|
||||||
@app.post("/on")
|
|
||||||
def heater_on() -> ActionResult:
|
|
||||||
result = app.state.controller.turn_on()
|
|
||||||
if not result.success:
|
|
||||||
raise HTTPException(status_code=500, detail=result.error)
|
|
||||||
return result
|
|
||||||
|
|
||||||
@app.post("/off")
|
|
||||||
def heater_off() -> ActionResult:
|
|
||||||
result = app.state.controller.turn_off()
|
|
||||||
if not result.success:
|
|
||||||
raise HTTPException(status_code=500, detail=result.error)
|
|
||||||
return result
|
|
||||||
|
|
||||||
@app.post("/toggle")
|
|
||||||
def heater_toggle() -> ActionResult:
|
|
||||||
result = app.state.controller.toggle()
|
|
||||||
if not result.success:
|
|
||||||
raise HTTPException(status_code=500, detail=result.error)
|
|
||||||
return result
|
|
||||||
|
|
||||||
return app
|
|
||||||
|
|
||||||
|
|
||||||
def serve(
|
|
||||||
host: Annotated[str, typer.Option("--host", "-h", help="Host to bind to")],
|
|
||||||
port: Annotated[int, typer.Option("--port", "-p", help="Port to bind to")] = 8124,
|
|
||||||
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
|
||||||
device_id: Annotated[str | None, typer.Option("--device-id", envvar="TUYA_DEVICE_ID")] = None,
|
|
||||||
device_ip: Annotated[str | None, typer.Option("--device-ip", envvar="TUYA_DEVICE_IP")] = None,
|
|
||||||
local_key: Annotated[str | None, typer.Option("--local-key", envvar="TUYA_LOCAL_KEY")] = None,
|
|
||||||
) -> None:
|
|
||||||
"""Start the heater control API server."""
|
|
||||||
configure_logger(log_level)
|
|
||||||
|
|
||||||
logger.info("Starting heater control API server")
|
|
||||||
|
|
||||||
if not device_id or not device_ip or not local_key:
|
|
||||||
error = "Must provide device ID, IP, and local key"
|
|
||||||
raise typer.Exit(error)
|
|
||||||
|
|
||||||
config = DeviceConfig(device_id=device_id, ip=device_ip, local_key=local_key)
|
|
||||||
|
|
||||||
app = create_app(config)
|
|
||||||
uvicorn.run(app, host=host, port=port)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
typer.run(serve)
|
|
||||||
@@ -1,31 +0,0 @@
|
|||||||
"""Pydantic models for heater API."""
|
|
||||||
|
|
||||||
from pydantic import BaseModel, Field
|
|
||||||
|
|
||||||
|
|
||||||
class DeviceConfig(BaseModel):
|
|
||||||
"""Tuya device configuration."""
|
|
||||||
|
|
||||||
device_id: str
|
|
||||||
ip: str
|
|
||||||
local_key: str
|
|
||||||
version: float = 3.5
|
|
||||||
|
|
||||||
|
|
||||||
class HeaterStatus(BaseModel):
|
|
||||||
"""Current heater status."""
|
|
||||||
|
|
||||||
power: bool
|
|
||||||
setpoint: int | None = None
|
|
||||||
state: str | None = None # "Stop", "Heat", etc.
|
|
||||||
error_code: int | None = None
|
|
||||||
raw_dps: dict[str, object] = Field(default_factory=dict)
|
|
||||||
|
|
||||||
|
|
||||||
class ActionResult(BaseModel):
|
|
||||||
"""Result of a heater action."""
|
|
||||||
|
|
||||||
success: bool
|
|
||||||
action: str
|
|
||||||
power: bool | None = None
|
|
||||||
error: str | None = None
|
|
||||||
@@ -1,11 +0,0 @@
|
|||||||
"""ORM package exports."""
|
|
||||||
|
|
||||||
from python.orm.richie.base import RichieBase
|
|
||||||
from python.orm.signal_bot.base import SignalBotBase
|
|
||||||
from python.orm.van_inventory.base import VanInventoryBase
|
|
||||||
|
|
||||||
__all__ = [
|
|
||||||
"RichieBase",
|
|
||||||
"SignalBotBase",
|
|
||||||
"VanInventoryBase",
|
|
||||||
]
|
|
||||||
@@ -1,51 +0,0 @@
|
|||||||
"""Shared ORM definitions."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from os import getenv
|
|
||||||
from typing import cast
|
|
||||||
|
|
||||||
from sqlalchemy import create_engine
|
|
||||||
from sqlalchemy.engine import URL, Engine
|
|
||||||
|
|
||||||
NAMING_CONVENTION = {
|
|
||||||
"ix": "ix_%(table_name)s_%(column_0_name)s",
|
|
||||||
"uq": "uq_%(table_name)s_%(column_0_name)s",
|
|
||||||
"ck": "ck_%(table_name)s_%(constraint_name)s",
|
|
||||||
"fk": "fk_%(table_name)s_%(column_0_name)s_%(referred_table_name)s",
|
|
||||||
"pk": "pk_%(table_name)s",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def get_connection_info(name: str) -> tuple[str, str, str, str, str | None]:
|
|
||||||
"""Get connection info from environment variables."""
|
|
||||||
database = getenv(f"{name}_DB")
|
|
||||||
host = getenv(f"{name}_HOST")
|
|
||||||
port = getenv(f"{name}_PORT")
|
|
||||||
username = getenv(f"{name}_USER")
|
|
||||||
password = getenv(f"{name}_PASSWORD")
|
|
||||||
|
|
||||||
if None in (database, host, port, username):
|
|
||||||
error = f"Missing environment variables for Postgres connection.\n{database=}\n{host=}\n{port=}\n{username=}\n"
|
|
||||||
raise ValueError(error)
|
|
||||||
return cast("tuple[str, str, str, str, str | None]", (database, host, port, username, password))
|
|
||||||
|
|
||||||
|
|
||||||
def get_postgres_engine(*, name: str = "POSTGRES", pool_pre_ping: bool = True) -> Engine:
|
|
||||||
"""Create a SQLAlchemy engine from environment variables."""
|
|
||||||
database, host, port, username, password = get_connection_info(name)
|
|
||||||
|
|
||||||
url = URL.create(
|
|
||||||
drivername="postgresql+psycopg",
|
|
||||||
username=username,
|
|
||||||
password=password,
|
|
||||||
host=host,
|
|
||||||
port=int(port),
|
|
||||||
database=database,
|
|
||||||
)
|
|
||||||
|
|
||||||
return create_engine(
|
|
||||||
url=url,
|
|
||||||
pool_pre_ping=pool_pre_ping,
|
|
||||||
pool_recycle=1800,
|
|
||||||
)
|
|
||||||
@@ -1,29 +0,0 @@
|
|||||||
"""Richie database ORM exports."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from python.orm.richie.base import RichieBase, TableBase, TableBaseBig, TableBaseSmall
|
|
||||||
from python.orm.richie.congress import Bill, Legislator, Vote, VoteRecord
|
|
||||||
from python.orm.richie.contact import (
|
|
||||||
Contact,
|
|
||||||
ContactNeed,
|
|
||||||
ContactRelationship,
|
|
||||||
Need,
|
|
||||||
RelationshipType,
|
|
||||||
)
|
|
||||||
|
|
||||||
__all__ = [
|
|
||||||
"Bill",
|
|
||||||
"Contact",
|
|
||||||
"ContactNeed",
|
|
||||||
"ContactRelationship",
|
|
||||||
"Legislator",
|
|
||||||
"Need",
|
|
||||||
"RelationshipType",
|
|
||||||
"RichieBase",
|
|
||||||
"TableBase",
|
|
||||||
"TableBaseBig",
|
|
||||||
"TableBaseSmall",
|
|
||||||
"Vote",
|
|
||||||
"VoteRecord",
|
|
||||||
]
|
|
||||||
@@ -1,60 +0,0 @@
|
|||||||
"""Richie database ORM base."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
from sqlalchemy import BigInteger, DateTime, MetaData, SmallInteger, func
|
|
||||||
from sqlalchemy.ext.declarative import AbstractConcreteBase
|
|
||||||
from sqlalchemy.orm import DeclarativeBase, Mapped, mapped_column
|
|
||||||
|
|
||||||
from python.orm.common import NAMING_CONVENTION
|
|
||||||
|
|
||||||
|
|
||||||
class RichieBase(DeclarativeBase):
|
|
||||||
"""Base class for richie database ORM models."""
|
|
||||||
|
|
||||||
schema_name = "main"
|
|
||||||
|
|
||||||
metadata = MetaData(
|
|
||||||
schema=schema_name,
|
|
||||||
naming_convention=NAMING_CONVENTION,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class _TableMixin:
|
|
||||||
"""Shared timestamp columns for all table bases."""
|
|
||||||
|
|
||||||
created: Mapped[datetime] = mapped_column(
|
|
||||||
DateTime(timezone=True),
|
|
||||||
server_default=func.now(),
|
|
||||||
)
|
|
||||||
updated: Mapped[datetime] = mapped_column(
|
|
||||||
DateTime(timezone=True),
|
|
||||||
server_default=func.now(),
|
|
||||||
onupdate=func.now(),
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class TableBaseSmall(_TableMixin, AbstractConcreteBase, RichieBase):
|
|
||||||
"""Table with SmallInteger primary key."""
|
|
||||||
|
|
||||||
__abstract__ = True
|
|
||||||
|
|
||||||
id: Mapped[int] = mapped_column(SmallInteger, primary_key=True)
|
|
||||||
|
|
||||||
|
|
||||||
class TableBase(_TableMixin, AbstractConcreteBase, RichieBase):
|
|
||||||
"""Table with Integer primary key."""
|
|
||||||
|
|
||||||
__abstract__ = True
|
|
||||||
|
|
||||||
id: Mapped[int] = mapped_column(primary_key=True)
|
|
||||||
|
|
||||||
|
|
||||||
class TableBaseBig(_TableMixin, AbstractConcreteBase, RichieBase):
|
|
||||||
"""Table with BigInteger primary key."""
|
|
||||||
|
|
||||||
__abstract__ = True
|
|
||||||
|
|
||||||
id: Mapped[int] = mapped_column(BigInteger, primary_key=True)
|
|
||||||
@@ -1,150 +0,0 @@
|
|||||||
"""Congress Tracker database models."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from datetime import date
|
|
||||||
|
|
||||||
from sqlalchemy import ForeignKey, Index, Text, UniqueConstraint
|
|
||||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
|
||||||
|
|
||||||
from python.orm.richie.base import RichieBase, TableBase
|
|
||||||
|
|
||||||
|
|
||||||
class Legislator(TableBase):
|
|
||||||
"""Legislator model - members of Congress."""
|
|
||||||
|
|
||||||
__tablename__ = "legislator"
|
|
||||||
|
|
||||||
# Natural key - bioguide ID is the authoritative identifier
|
|
||||||
bioguide_id: Mapped[str] = mapped_column(Text, unique=True, index=True)
|
|
||||||
|
|
||||||
# Other IDs for cross-referencing
|
|
||||||
thomas_id: Mapped[str | None]
|
|
||||||
lis_id: Mapped[str | None]
|
|
||||||
govtrack_id: Mapped[int | None]
|
|
||||||
opensecrets_id: Mapped[str | None]
|
|
||||||
fec_ids: Mapped[str | None] # JSON array stored as string
|
|
||||||
|
|
||||||
# Name info
|
|
||||||
first_name: Mapped[str]
|
|
||||||
last_name: Mapped[str]
|
|
||||||
official_full_name: Mapped[str | None]
|
|
||||||
nickname: Mapped[str | None]
|
|
||||||
|
|
||||||
# Bio
|
|
||||||
birthday: Mapped[date | None]
|
|
||||||
gender: Mapped[str | None] # M/F
|
|
||||||
|
|
||||||
# Current term info (denormalized for query efficiency)
|
|
||||||
current_party: Mapped[str | None]
|
|
||||||
current_state: Mapped[str | None]
|
|
||||||
current_district: Mapped[int | None] # House only
|
|
||||||
current_chamber: Mapped[str | None] # rep/sen
|
|
||||||
|
|
||||||
# Relationships
|
|
||||||
vote_records: Mapped[list[VoteRecord]] = relationship(
|
|
||||||
"VoteRecord",
|
|
||||||
back_populates="legislator",
|
|
||||||
cascade="all, delete-orphan",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class Bill(TableBase):
|
|
||||||
"""Bill model - legislation introduced in Congress."""
|
|
||||||
|
|
||||||
__tablename__ = "bill"
|
|
||||||
|
|
||||||
# Composite natural key: congress + bill_type + number
|
|
||||||
congress: Mapped[int]
|
|
||||||
bill_type: Mapped[str] # hr, s, hres, sres, hjres, sjres
|
|
||||||
number: Mapped[int]
|
|
||||||
|
|
||||||
# Bill info
|
|
||||||
title: Mapped[str | None]
|
|
||||||
title_short: Mapped[str | None]
|
|
||||||
official_title: Mapped[str | None]
|
|
||||||
|
|
||||||
# Status
|
|
||||||
status: Mapped[str | None]
|
|
||||||
status_at: Mapped[date | None]
|
|
||||||
|
|
||||||
# Sponsor
|
|
||||||
sponsor_bioguide_id: Mapped[str | None]
|
|
||||||
|
|
||||||
# Subjects
|
|
||||||
subjects_top_term: Mapped[str | None]
|
|
||||||
|
|
||||||
# Relationships
|
|
||||||
votes: Mapped[list[Vote]] = relationship(
|
|
||||||
"Vote",
|
|
||||||
back_populates="bill",
|
|
||||||
)
|
|
||||||
|
|
||||||
__table_args__ = (
|
|
||||||
UniqueConstraint("congress", "bill_type", "number", name="uq_bill_congress_type_number"),
|
|
||||||
Index("ix_bill_congress", "congress"),
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class Vote(TableBase):
|
|
||||||
"""Vote model - roll call votes in Congress."""
|
|
||||||
|
|
||||||
__tablename__ = "vote"
|
|
||||||
|
|
||||||
# Composite natural key: congress + chamber + session + number
|
|
||||||
congress: Mapped[int]
|
|
||||||
chamber: Mapped[str] # house/senate
|
|
||||||
session: Mapped[int]
|
|
||||||
number: Mapped[int]
|
|
||||||
|
|
||||||
# Vote details
|
|
||||||
vote_type: Mapped[str | None]
|
|
||||||
question: Mapped[str | None]
|
|
||||||
result: Mapped[str | None]
|
|
||||||
result_text: Mapped[str | None]
|
|
||||||
|
|
||||||
# Timing
|
|
||||||
vote_date: Mapped[date]
|
|
||||||
|
|
||||||
# Vote counts (denormalized for efficiency)
|
|
||||||
yea_count: Mapped[int | None]
|
|
||||||
nay_count: Mapped[int | None]
|
|
||||||
not_voting_count: Mapped[int | None]
|
|
||||||
present_count: Mapped[int | None]
|
|
||||||
|
|
||||||
# Related bill (optional - not all votes are on bills)
|
|
||||||
bill_id: Mapped[int | None] = mapped_column(ForeignKey("main.bill.id"))
|
|
||||||
|
|
||||||
# Relationships
|
|
||||||
bill: Mapped[Bill | None] = relationship("Bill", back_populates="votes")
|
|
||||||
vote_records: Mapped[list[VoteRecord]] = relationship(
|
|
||||||
"VoteRecord",
|
|
||||||
back_populates="vote",
|
|
||||||
cascade="all, delete-orphan",
|
|
||||||
)
|
|
||||||
|
|
||||||
__table_args__ = (
|
|
||||||
UniqueConstraint("congress", "chamber", "session", "number", name="uq_vote_congress_chamber_session_number"),
|
|
||||||
Index("ix_vote_date", "vote_date"),
|
|
||||||
Index("ix_vote_congress_chamber", "congress", "chamber"),
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class VoteRecord(RichieBase):
|
|
||||||
"""Association table: Vote <-> Legislator with position."""
|
|
||||||
|
|
||||||
__tablename__ = "vote_record"
|
|
||||||
|
|
||||||
vote_id: Mapped[int] = mapped_column(
|
|
||||||
ForeignKey("main.vote.id", ondelete="CASCADE"),
|
|
||||||
primary_key=True,
|
|
||||||
)
|
|
||||||
legislator_id: Mapped[int] = mapped_column(
|
|
||||||
ForeignKey("main.legislator.id", ondelete="CASCADE"),
|
|
||||||
primary_key=True,
|
|
||||||
)
|
|
||||||
position: Mapped[str] # Yea, Nay, Not Voting, Present
|
|
||||||
|
|
||||||
# Relationships
|
|
||||||
vote: Mapped[Vote] = relationship("Vote", back_populates="vote_records")
|
|
||||||
legislator: Mapped[Legislator] = relationship("Legislator", back_populates="vote_records")
|
|
||||||
@@ -1,168 +0,0 @@
|
|||||||
"""Contact database models."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from enum import StrEnum
|
|
||||||
|
|
||||||
from sqlalchemy import ForeignKey, String
|
|
||||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
|
||||||
|
|
||||||
from python.orm.richie.base import RichieBase, TableBase
|
|
||||||
|
|
||||||
|
|
||||||
class RelationshipType(StrEnum):
|
|
||||||
"""Relationship types with default closeness weights.
|
|
||||||
|
|
||||||
Default weight is an integer 1-10 where 10 = closest relationship.
|
|
||||||
Users can override this per-relationship in the UI.
|
|
||||||
"""
|
|
||||||
|
|
||||||
SPOUSE = "spouse"
|
|
||||||
PARTNER = "partner"
|
|
||||||
PARENT = "parent"
|
|
||||||
CHILD = "child"
|
|
||||||
SIBLING = "sibling"
|
|
||||||
BEST_FRIEND = "best_friend"
|
|
||||||
GRANDPARENT = "grandparent"
|
|
||||||
GRANDCHILD = "grandchild"
|
|
||||||
AUNT_UNCLE = "aunt_uncle"
|
|
||||||
NIECE_NEPHEW = "niece_nephew"
|
|
||||||
COUSIN = "cousin"
|
|
||||||
IN_LAW = "in_law"
|
|
||||||
CLOSE_FRIEND = "close_friend"
|
|
||||||
FRIEND = "friend"
|
|
||||||
MENTOR = "mentor"
|
|
||||||
MENTEE = "mentee"
|
|
||||||
BUSINESS_PARTNER = "business_partner"
|
|
||||||
COLLEAGUE = "colleague"
|
|
||||||
MANAGER = "manager"
|
|
||||||
DIRECT_REPORT = "direct_report"
|
|
||||||
CLIENT = "client"
|
|
||||||
ACQUAINTANCE = "acquaintance"
|
|
||||||
NEIGHBOR = "neighbor"
|
|
||||||
EX = "ex"
|
|
||||||
OTHER = "other"
|
|
||||||
|
|
||||||
@property
|
|
||||||
def default_weight(self) -> int:
|
|
||||||
"""Return the default closeness weight (1-10) for this relationship type."""
|
|
||||||
weights = {
|
|
||||||
RelationshipType.SPOUSE: 10,
|
|
||||||
RelationshipType.PARTNER: 10,
|
|
||||||
RelationshipType.PARENT: 9,
|
|
||||||
RelationshipType.CHILD: 9,
|
|
||||||
RelationshipType.SIBLING: 9,
|
|
||||||
RelationshipType.BEST_FRIEND: 8,
|
|
||||||
RelationshipType.GRANDPARENT: 7,
|
|
||||||
RelationshipType.GRANDCHILD: 7,
|
|
||||||
RelationshipType.AUNT_UNCLE: 7,
|
|
||||||
RelationshipType.NIECE_NEPHEW: 7,
|
|
||||||
RelationshipType.COUSIN: 7,
|
|
||||||
RelationshipType.IN_LAW: 7,
|
|
||||||
RelationshipType.CLOSE_FRIEND: 6,
|
|
||||||
RelationshipType.FRIEND: 6,
|
|
||||||
RelationshipType.MENTOR: 5,
|
|
||||||
RelationshipType.MENTEE: 5,
|
|
||||||
RelationshipType.BUSINESS_PARTNER: 5,
|
|
||||||
RelationshipType.COLLEAGUE: 4,
|
|
||||||
RelationshipType.MANAGER: 4,
|
|
||||||
RelationshipType.DIRECT_REPORT: 4,
|
|
||||||
RelationshipType.CLIENT: 4,
|
|
||||||
RelationshipType.ACQUAINTANCE: 3,
|
|
||||||
RelationshipType.NEIGHBOR: 3,
|
|
||||||
RelationshipType.EX: 2,
|
|
||||||
RelationshipType.OTHER: 2,
|
|
||||||
}
|
|
||||||
return weights.get(self, 5)
|
|
||||||
|
|
||||||
@property
|
|
||||||
def display_name(self) -> str:
|
|
||||||
"""Return a human-readable display name."""
|
|
||||||
return self.value.replace("_", " ").title()
|
|
||||||
|
|
||||||
|
|
||||||
class ContactNeed(RichieBase):
|
|
||||||
"""Association table: Contact <-> Need."""
|
|
||||||
|
|
||||||
__tablename__ = "contact_need"
|
|
||||||
|
|
||||||
contact_id: Mapped[int] = mapped_column(
|
|
||||||
ForeignKey("main.contact.id", ondelete="CASCADE"),
|
|
||||||
primary_key=True,
|
|
||||||
)
|
|
||||||
need_id: Mapped[int] = mapped_column(
|
|
||||||
ForeignKey("main.need.id", ondelete="CASCADE"),
|
|
||||||
primary_key=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class ContactRelationship(RichieBase):
|
|
||||||
"""Association table: Contact <-> Contact with relationship type and weight."""
|
|
||||||
|
|
||||||
__tablename__ = "contact_relationship"
|
|
||||||
|
|
||||||
contact_id: Mapped[int] = mapped_column(
|
|
||||||
ForeignKey("main.contact.id", ondelete="CASCADE"),
|
|
||||||
primary_key=True,
|
|
||||||
)
|
|
||||||
related_contact_id: Mapped[int] = mapped_column(
|
|
||||||
ForeignKey("main.contact.id", ondelete="CASCADE"),
|
|
||||||
primary_key=True,
|
|
||||||
)
|
|
||||||
relationship_type: Mapped[str] = mapped_column(String(100))
|
|
||||||
closeness_weight: Mapped[int] = mapped_column(default=5)
|
|
||||||
|
|
||||||
|
|
||||||
class Contact(TableBase):
|
|
||||||
"""Contact model."""
|
|
||||||
|
|
||||||
__tablename__ = "contact"
|
|
||||||
|
|
||||||
name: Mapped[str]
|
|
||||||
|
|
||||||
age: Mapped[int | None]
|
|
||||||
bio: Mapped[str | None]
|
|
||||||
current_job: Mapped[str | None]
|
|
||||||
gender: Mapped[str | None]
|
|
||||||
goals: Mapped[str | None]
|
|
||||||
legal_name: Mapped[str | None]
|
|
||||||
profile_pic: Mapped[str | None]
|
|
||||||
safe_conversation_starters: Mapped[str | None]
|
|
||||||
self_sufficiency_score: Mapped[int | None]
|
|
||||||
social_structure_style: Mapped[str | None]
|
|
||||||
ssn: Mapped[str | None]
|
|
||||||
suffix: Mapped[str | None]
|
|
||||||
timezone: Mapped[str | None]
|
|
||||||
topics_to_avoid: Mapped[str | None]
|
|
||||||
|
|
||||||
needs: Mapped[list[Need]] = relationship(
|
|
||||||
"Need",
|
|
||||||
secondary=ContactNeed.__table__,
|
|
||||||
back_populates="contacts",
|
|
||||||
)
|
|
||||||
|
|
||||||
related_to: Mapped[list[ContactRelationship]] = relationship(
|
|
||||||
"ContactRelationship",
|
|
||||||
foreign_keys=[ContactRelationship.contact_id],
|
|
||||||
cascade="all, delete-orphan",
|
|
||||||
)
|
|
||||||
related_from: Mapped[list[ContactRelationship]] = relationship(
|
|
||||||
"ContactRelationship",
|
|
||||||
foreign_keys=[ContactRelationship.related_contact_id],
|
|
||||||
cascade="all, delete-orphan",
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class Need(TableBase):
|
|
||||||
"""Need/accommodation model (e.g., light sensitive, ADHD)."""
|
|
||||||
|
|
||||||
__tablename__ = "need"
|
|
||||||
|
|
||||||
name: Mapped[str]
|
|
||||||
description: Mapped[str | None]
|
|
||||||
|
|
||||||
contacts: Mapped[list[Contact]] = relationship(
|
|
||||||
"Contact",
|
|
||||||
secondary=ContactNeed.__table__,
|
|
||||||
back_populates="needs",
|
|
||||||
)
|
|
||||||
@@ -1,16 +0,0 @@
|
|||||||
"""Signal bot database ORM exports."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from python.orm.signal_bot.base import SignalBotBase, SignalBotTableBase, SignalBotTableBaseSmall
|
|
||||||
from python.orm.signal_bot.models import DeadLetterMessage, DeviceRole, RoleRecord, SignalDevice
|
|
||||||
|
|
||||||
__all__ = [
|
|
||||||
"DeadLetterMessage",
|
|
||||||
"DeviceRole",
|
|
||||||
"RoleRecord",
|
|
||||||
"SignalBotBase",
|
|
||||||
"SignalBotTableBase",
|
|
||||||
"SignalBotTableBaseSmall",
|
|
||||||
"SignalDevice",
|
|
||||||
]
|
|
||||||
@@ -1,52 +0,0 @@
|
|||||||
"""Signal bot database ORM base."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
from sqlalchemy import DateTime, MetaData, SmallInteger, func
|
|
||||||
from sqlalchemy.ext.declarative import AbstractConcreteBase
|
|
||||||
from sqlalchemy.orm import DeclarativeBase, Mapped, mapped_column
|
|
||||||
|
|
||||||
from python.orm.common import NAMING_CONVENTION
|
|
||||||
|
|
||||||
|
|
||||||
class SignalBotBase(DeclarativeBase):
|
|
||||||
"""Base class for signal_bot database ORM models."""
|
|
||||||
|
|
||||||
schema_name = "main"
|
|
||||||
|
|
||||||
metadata = MetaData(
|
|
||||||
schema=schema_name,
|
|
||||||
naming_convention=NAMING_CONVENTION,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class _TableMixin:
|
|
||||||
"""Shared timestamp columns for all table bases."""
|
|
||||||
|
|
||||||
created: Mapped[datetime] = mapped_column(
|
|
||||||
DateTime(timezone=True),
|
|
||||||
server_default=func.now(),
|
|
||||||
)
|
|
||||||
updated: Mapped[datetime] = mapped_column(
|
|
||||||
DateTime(timezone=True),
|
|
||||||
server_default=func.now(),
|
|
||||||
onupdate=func.now(),
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class SignalBotTableBaseSmall(_TableMixin, AbstractConcreteBase, SignalBotBase):
|
|
||||||
"""Table with SmallInteger primary key."""
|
|
||||||
|
|
||||||
__abstract__ = True
|
|
||||||
|
|
||||||
id: Mapped[int] = mapped_column(SmallInteger, primary_key=True)
|
|
||||||
|
|
||||||
|
|
||||||
class SignalBotTableBase(_TableMixin, AbstractConcreteBase, SignalBotBase):
|
|
||||||
"""Table with Integer primary key."""
|
|
||||||
|
|
||||||
__abstract__ = True
|
|
||||||
|
|
||||||
id: Mapped[int] = mapped_column(primary_key=True)
|
|
||||||
@@ -1,62 +0,0 @@
|
|||||||
"""Signal bot device, role, and dead letter ORM models."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
from sqlalchemy import DateTime, Enum, ForeignKey, SmallInteger, String, Text, UniqueConstraint
|
|
||||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
|
||||||
|
|
||||||
from python.orm.signal_bot.base import SignalBotTableBase, SignalBotTableBaseSmall
|
|
||||||
from python.signal_bot.models import MessageStatus, TrustLevel
|
|
||||||
|
|
||||||
|
|
||||||
class RoleRecord(SignalBotTableBaseSmall):
|
|
||||||
"""Lookup table for RBAC roles, keyed by smallint."""
|
|
||||||
|
|
||||||
__tablename__ = "role"
|
|
||||||
|
|
||||||
name: Mapped[str] = mapped_column(String(50), unique=True)
|
|
||||||
|
|
||||||
|
|
||||||
class DeviceRole(SignalBotTableBase):
|
|
||||||
"""Association between a device and a role."""
|
|
||||||
|
|
||||||
__tablename__ = "device_role"
|
|
||||||
__table_args__ = (
|
|
||||||
UniqueConstraint("device_id", "role_id", name="uq_device_role_device_role"),
|
|
||||||
{"schema": "main"},
|
|
||||||
)
|
|
||||||
|
|
||||||
device_id: Mapped[int] = mapped_column(ForeignKey("main.signal_device.id"))
|
|
||||||
role_id: Mapped[int] = mapped_column(SmallInteger, ForeignKey("main.role.id"))
|
|
||||||
|
|
||||||
|
|
||||||
class SignalDevice(SignalBotTableBase):
|
|
||||||
"""A Signal device tracked by phone number and safety number."""
|
|
||||||
|
|
||||||
__tablename__ = "signal_device"
|
|
||||||
|
|
||||||
phone_number: Mapped[str] = mapped_column(String(50), unique=True)
|
|
||||||
safety_number: Mapped[str | None]
|
|
||||||
trust_level: Mapped[TrustLevel] = mapped_column(
|
|
||||||
Enum(TrustLevel, name="trust_level", create_constraint=False, native_enum=False),
|
|
||||||
default=TrustLevel.UNVERIFIED,
|
|
||||||
)
|
|
||||||
last_seen: Mapped[datetime] = mapped_column(DateTime(timezone=True))
|
|
||||||
|
|
||||||
roles: Mapped[list[RoleRecord]] = relationship(secondary=DeviceRole.__table__)
|
|
||||||
|
|
||||||
|
|
||||||
class DeadLetterMessage(SignalBotTableBase):
|
|
||||||
"""A Signal message that failed processing and was sent to the dead letter queue."""
|
|
||||||
|
|
||||||
__tablename__ = "dead_letter_message"
|
|
||||||
|
|
||||||
source: Mapped[str]
|
|
||||||
message: Mapped[str] = mapped_column(Text)
|
|
||||||
received_at: Mapped[datetime] = mapped_column(DateTime(timezone=True))
|
|
||||||
status: Mapped[MessageStatus] = mapped_column(
|
|
||||||
Enum(MessageStatus, name="message_status", create_constraint=False, native_enum=False),
|
|
||||||
default=MessageStatus.UNPROCESSED,
|
|
||||||
)
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Van inventory database ORM exports."""
|
|
||||||
@@ -1,39 +0,0 @@
|
|||||||
"""Van inventory database ORM base."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
from sqlalchemy import DateTime, MetaData, func
|
|
||||||
from sqlalchemy.ext.declarative import AbstractConcreteBase
|
|
||||||
from sqlalchemy.orm import DeclarativeBase, Mapped, mapped_column
|
|
||||||
|
|
||||||
from python.orm.common import NAMING_CONVENTION
|
|
||||||
|
|
||||||
|
|
||||||
class VanInventoryBase(DeclarativeBase):
|
|
||||||
"""Base class for van_inventory database ORM models."""
|
|
||||||
|
|
||||||
schema_name = "main"
|
|
||||||
|
|
||||||
metadata = MetaData(
|
|
||||||
schema=schema_name,
|
|
||||||
naming_convention=NAMING_CONVENTION,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class VanTableBase(AbstractConcreteBase, VanInventoryBase):
|
|
||||||
"""Abstract concrete base for van_inventory tables with IDs and timestamps."""
|
|
||||||
|
|
||||||
__abstract__ = True
|
|
||||||
|
|
||||||
id: Mapped[int] = mapped_column(primary_key=True)
|
|
||||||
created: Mapped[datetime] = mapped_column(
|
|
||||||
DateTime(timezone=True),
|
|
||||||
server_default=func.now(),
|
|
||||||
)
|
|
||||||
updated: Mapped[datetime] = mapped_column(
|
|
||||||
DateTime(timezone=True),
|
|
||||||
server_default=func.now(),
|
|
||||||
onupdate=func.now(),
|
|
||||||
)
|
|
||||||
@@ -1,46 +0,0 @@
|
|||||||
"""Van inventory ORM models."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from sqlalchemy import ForeignKey, UniqueConstraint
|
|
||||||
from sqlalchemy.orm import Mapped, mapped_column, relationship
|
|
||||||
|
|
||||||
from python.orm.van_inventory.base import VanTableBase
|
|
||||||
|
|
||||||
|
|
||||||
class Item(VanTableBase):
|
|
||||||
"""A food item in the van."""
|
|
||||||
|
|
||||||
__tablename__ = "items"
|
|
||||||
|
|
||||||
name: Mapped[str] = mapped_column(unique=True)
|
|
||||||
quantity: Mapped[float] = mapped_column(default=0)
|
|
||||||
unit: Mapped[str]
|
|
||||||
category: Mapped[str | None]
|
|
||||||
|
|
||||||
meal_ingredients: Mapped[list[MealIngredient]] = relationship(back_populates="item")
|
|
||||||
|
|
||||||
|
|
||||||
class Meal(VanTableBase):
|
|
||||||
"""A meal that can be made from items in the van."""
|
|
||||||
|
|
||||||
__tablename__ = "meals"
|
|
||||||
|
|
||||||
name: Mapped[str] = mapped_column(unique=True)
|
|
||||||
instructions: Mapped[str | None]
|
|
||||||
|
|
||||||
ingredients: Mapped[list[MealIngredient]] = relationship(back_populates="meal")
|
|
||||||
|
|
||||||
|
|
||||||
class MealIngredient(VanTableBase):
|
|
||||||
"""Links a meal to the items it requires, with quantities."""
|
|
||||||
|
|
||||||
__tablename__ = "meal_ingredients"
|
|
||||||
__table_args__ = (UniqueConstraint("meal_id", "item_id"),)
|
|
||||||
|
|
||||||
meal_id: Mapped[int] = mapped_column(ForeignKey("meals.id"))
|
|
||||||
item_id: Mapped[int] = mapped_column(ForeignKey("items.id"))
|
|
||||||
quantity_needed: Mapped[float]
|
|
||||||
|
|
||||||
meal: Mapped[Meal] = relationship(back_populates="ingredients")
|
|
||||||
item: Mapped[Item] = relationship(back_populates="meal_ingredients")
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Signal command and control bot."""
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Signal bot commands."""
|
|
||||||
@@ -1,137 +0,0 @@
|
|||||||
"""Van inventory command — parse receipts and item lists via LLM, push to API."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import json
|
|
||||||
import logging
|
|
||||||
from typing import TYPE_CHECKING, Any
|
|
||||||
|
|
||||||
import httpx
|
|
||||||
|
|
||||||
from python.signal_bot.models import InventoryItem, InventoryUpdate
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from python.signal_bot.llm_client import LLMClient
|
|
||||||
from python.signal_bot.models import SignalMessage
|
|
||||||
from python.signal_bot.signal_client import SignalClient
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
SYSTEM_PROMPT = """\
|
|
||||||
You are an inventory assistant. Extract items from the input and return ONLY
|
|
||||||
a JSON array. Each element must have these fields:
|
|
||||||
- "name": item name (string)
|
|
||||||
- "quantity": numeric count or amount (default 1)
|
|
||||||
- "unit": unit of measure (e.g. "each", "lb", "oz", "gallon", "bag", "box")
|
|
||||||
- "category": category like "food", "tools", "supplies", etc.
|
|
||||||
- "notes": any extra detail (empty string if none)
|
|
||||||
|
|
||||||
Example output:
|
|
||||||
[{"name": "water bottles", "quantity": 6, "unit": "gallon", "category": "supplies", "notes": "1 gallon each"}]
|
|
||||||
|
|
||||||
Return ONLY the JSON array, no other text.\
|
|
||||||
"""
|
|
||||||
|
|
||||||
IMAGE_PROMPT = "Extract all items from this receipt or inventory photo."
|
|
||||||
TEXT_PROMPT = "Extract all items from this inventory list."
|
|
||||||
|
|
||||||
|
|
||||||
def parse_llm_response(raw: str) -> list[InventoryItem]:
|
|
||||||
"""Parse the LLM JSON response into InventoryItem list."""
|
|
||||||
text = raw.strip()
|
|
||||||
# Strip markdown code fences if present
|
|
||||||
if text.startswith("```"):
|
|
||||||
lines = text.split("\n")
|
|
||||||
lines = [line for line in lines if not line.startswith("```")]
|
|
||||||
text = "\n".join(lines)
|
|
||||||
|
|
||||||
items_data: list[dict[str, Any]] = json.loads(text)
|
|
||||||
return [InventoryItem.model_validate(item) for item in items_data]
|
|
||||||
|
|
||||||
|
|
||||||
def _upsert_item(api_url: str, item: InventoryItem) -> None:
|
|
||||||
"""Create or update an item via the van_inventory API.
|
|
||||||
|
|
||||||
Fetches existing items, and if one with the same name exists,
|
|
||||||
patches its quantity (summing). Otherwise creates a new item.
|
|
||||||
"""
|
|
||||||
base = api_url.rstrip("/")
|
|
||||||
response = httpx.get(f"{base}/api/items", timeout=10)
|
|
||||||
response.raise_for_status()
|
|
||||||
existing: list[dict[str, Any]] = response.json()
|
|
||||||
|
|
||||||
match = next((e for e in existing if e["name"].lower() == item.name.lower()), None)
|
|
||||||
|
|
||||||
if match:
|
|
||||||
new_qty = match["quantity"] + item.quantity
|
|
||||||
patch = {"quantity": new_qty}
|
|
||||||
if item.category:
|
|
||||||
patch["category"] = item.category
|
|
||||||
response = httpx.patch(f"{base}/api/items/{match['id']}", json=patch, timeout=10)
|
|
||||||
response.raise_for_status()
|
|
||||||
return
|
|
||||||
payload = {
|
|
||||||
"name": item.name,
|
|
||||||
"quantity": item.quantity,
|
|
||||||
"unit": item.unit,
|
|
||||||
"category": item.category or None,
|
|
||||||
}
|
|
||||||
response = httpx.post(f"{base}/api/items", json=payload, timeout=10)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
|
|
||||||
def handle_inventory_update(
|
|
||||||
message: SignalMessage,
|
|
||||||
signal: SignalClient,
|
|
||||||
llm: LLMClient,
|
|
||||||
api_url: str,
|
|
||||||
) -> InventoryUpdate:
|
|
||||||
"""Process an inventory update from a Signal message.
|
|
||||||
|
|
||||||
Accepts either an image (receipt photo) or text list.
|
|
||||||
Uses the LLM to extract structured items, then pushes to the van_inventory API.
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
logger.info(f"Processing inventory update from {message.source}")
|
|
||||||
if message.attachments:
|
|
||||||
image_data = signal.get_attachment(message.attachments[0])
|
|
||||||
raw_response = llm.chat(
|
|
||||||
IMAGE_PROMPT,
|
|
||||||
image_data=image_data,
|
|
||||||
system=SYSTEM_PROMPT,
|
|
||||||
)
|
|
||||||
source_type = "receipt_photo"
|
|
||||||
elif message.message.strip():
|
|
||||||
raw_response = llm.chat(
|
|
||||||
f"{TEXT_PROMPT}\n\n{message.message}",
|
|
||||||
system=SYSTEM_PROMPT,
|
|
||||||
)
|
|
||||||
source_type = "text_list"
|
|
||||||
else:
|
|
||||||
signal.reply(message, "Send a photo of a receipt or a text list of items to update inventory.")
|
|
||||||
return InventoryUpdate()
|
|
||||||
|
|
||||||
logger.info(f"{raw_response=}")
|
|
||||||
|
|
||||||
new_items = parse_llm_response(raw_response)
|
|
||||||
|
|
||||||
logger.info(f"{new_items=}")
|
|
||||||
|
|
||||||
for item in new_items:
|
|
||||||
_upsert_item(api_url, item)
|
|
||||||
|
|
||||||
summary = _format_summary(new_items)
|
|
||||||
signal.reply(message, f"Inventory updated with {len(new_items)} item(s):\n{summary}")
|
|
||||||
|
|
||||||
return InventoryUpdate(items=new_items, raw_response=raw_response, source_type=source_type)
|
|
||||||
|
|
||||||
except Exception:
|
|
||||||
logger.exception("Failed to process inventory update")
|
|
||||||
signal.reply(message, "Failed to process inventory update. Check logs for details.")
|
|
||||||
return InventoryUpdate()
|
|
||||||
|
|
||||||
|
|
||||||
def _format_summary(items: list[InventoryItem]) -> str:
|
|
||||||
"""Format items into a readable summary."""
|
|
||||||
lines = [f" - {item.name} x{item.quantity} {item.unit} [{item.category}]" for item in items]
|
|
||||||
return "\n".join(lines)
|
|
||||||
@@ -1,64 +0,0 @@
|
|||||||
"""Location command for the Signal bot."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from typing import TYPE_CHECKING, Any
|
|
||||||
|
|
||||||
import httpx
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from python.signal_bot.models import SignalMessage
|
|
||||||
from python.signal_bot.signal_client import SignalClient
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def _get_entity_state(ha_url: str, ha_token: str, entity_id: str) -> dict[str, Any]:
|
|
||||||
"""Fetch an entity's state from Home Assistant."""
|
|
||||||
entity_url = f"{ha_url}/api/states/{entity_id}"
|
|
||||||
logger.debug(f"Fetching {entity_url=}")
|
|
||||||
response = httpx.get(
|
|
||||||
entity_url,
|
|
||||||
headers={"Authorization": f"Bearer {ha_token}"},
|
|
||||||
timeout=30,
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
|
||||||
return response.json()
|
|
||||||
|
|
||||||
|
|
||||||
def _format_location(latitude: str, longitude: str) -> str:
|
|
||||||
"""Render a friendly location response."""
|
|
||||||
return f"Van location: {latitude}, {longitude}\nhttps://maps.google.com/?q={latitude},{longitude}"
|
|
||||||
|
|
||||||
|
|
||||||
def handle_location_request(
|
|
||||||
message: SignalMessage,
|
|
||||||
signal: SignalClient,
|
|
||||||
ha_url: str | None,
|
|
||||||
ha_token: str | None,
|
|
||||||
) -> None:
|
|
||||||
"""Reply with van location from Home Assistant."""
|
|
||||||
if ha_url is None or ha_token is None:
|
|
||||||
signal.reply(message, "Location command is not configured (missing HA_URL or HA_TOKEN).")
|
|
||||||
return
|
|
||||||
|
|
||||||
lat_payload = None
|
|
||||||
lon_payload = None
|
|
||||||
try:
|
|
||||||
lat_payload = _get_entity_state(ha_url, ha_token, "sensor.van_last_known_latitude")
|
|
||||||
lon_payload = _get_entity_state(ha_url, ha_token, "sensor.van_last_known_longitude")
|
|
||||||
except httpx.HTTPError:
|
|
||||||
logger.exception("Couldn't fetch van location from Home Assistant right now.")
|
|
||||||
logger.debug(f"{ha_url=} {lat_payload=} {lon_payload=}")
|
|
||||||
signal.reply(message, "Couldn't fetch van location from Home Assistant right now.")
|
|
||||||
return
|
|
||||||
|
|
||||||
latitude = lat_payload.get("state", "")
|
|
||||||
longitude = lon_payload.get("state", "")
|
|
||||||
|
|
||||||
if not latitude or not longitude or latitude == "unavailable" or longitude == "unavailable":
|
|
||||||
signal.reply(message, "Van location is unavailable in Home Assistant right now.")
|
|
||||||
return
|
|
||||||
|
|
||||||
signal.reply(message, _format_location(latitude, longitude))
|
|
||||||
@@ -1,286 +0,0 @@
|
|||||||
"""Device registry — tracks verified/unverified devices by safety number."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from datetime import datetime, timedelta
|
|
||||||
from typing import TYPE_CHECKING, NamedTuple
|
|
||||||
|
|
||||||
from sqlalchemy import delete, select
|
|
||||||
from sqlalchemy.orm import Session
|
|
||||||
|
|
||||||
from python.common import utcnow
|
|
||||||
from python.orm.signal_bot.models import RoleRecord, SignalDevice
|
|
||||||
from python.signal_bot.models import Role, TrustLevel
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from sqlalchemy.engine import Engine
|
|
||||||
|
|
||||||
from python.signal_bot.signal_client import SignalClient
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
_BLOCKED_TTL = timedelta(minutes=60)
|
|
||||||
_DEFAULT_TTL = timedelta(minutes=5)
|
|
||||||
|
|
||||||
|
|
||||||
class _CacheEntry(NamedTuple):
|
|
||||||
expires: datetime
|
|
||||||
trust_level: TrustLevel
|
|
||||||
has_safety_number: bool
|
|
||||||
safety_number: str | None
|
|
||||||
roles: list[Role]
|
|
||||||
|
|
||||||
|
|
||||||
class DeviceRegistry:
|
|
||||||
"""Manage device trust based on Signal safety numbers.
|
|
||||||
|
|
||||||
Devices start as UNVERIFIED. An admin verifies them over SSH by calling
|
|
||||||
``verify(phone_number)`` which marks the device VERIFIED and also tells
|
|
||||||
signal-cli to trust the identity.
|
|
||||||
|
|
||||||
Only VERIFIED devices may execute commands.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, signal_client: SignalClient, engine: Engine) -> None:
|
|
||||||
self.signal_client = signal_client
|
|
||||||
self.engine = engine
|
|
||||||
self._contact_cache: dict[str, _CacheEntry] = {}
|
|
||||||
|
|
||||||
def is_verified(self, phone_number: str) -> bool:
|
|
||||||
"""Check if a phone number is verified."""
|
|
||||||
if entry := self._cached(phone_number):
|
|
||||||
return entry.trust_level == TrustLevel.VERIFIED
|
|
||||||
device = self._load_device(phone_number)
|
|
||||||
return device is not None and device.trust_level == TrustLevel.VERIFIED
|
|
||||||
|
|
||||||
def record_contact(self, phone_number: str, safety_number: str | None = None) -> None:
|
|
||||||
"""Record seeing a device. Creates entry if new, updates last_seen."""
|
|
||||||
now = utcnow()
|
|
||||||
|
|
||||||
entry = self._cached(phone_number)
|
|
||||||
if entry and entry.safety_number == safety_number:
|
|
||||||
return
|
|
||||||
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
device = session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
if device:
|
|
||||||
if device.safety_number != safety_number and device.trust_level != TrustLevel.BLOCKED:
|
|
||||||
logger.warning(f"Safety number changed for {phone_number}, resetting to UNVERIFIED")
|
|
||||||
device.safety_number = safety_number
|
|
||||||
device.trust_level = TrustLevel.UNVERIFIED
|
|
||||||
device.last_seen = now
|
|
||||||
else:
|
|
||||||
device = SignalDevice(
|
|
||||||
phone_number=phone_number,
|
|
||||||
safety_number=safety_number,
|
|
||||||
trust_level=TrustLevel.UNVERIFIED,
|
|
||||||
last_seen=now,
|
|
||||||
)
|
|
||||||
session.add(device)
|
|
||||||
logger.info(f"New device registered: {phone_number}")
|
|
||||||
|
|
||||||
session.commit()
|
|
||||||
self._update_cache(phone_number, device)
|
|
||||||
|
|
||||||
def has_safety_number(self, phone_number: str) -> bool:
|
|
||||||
"""Check if a device has a safety number on file."""
|
|
||||||
if entry := self._cached(phone_number):
|
|
||||||
return entry.has_safety_number
|
|
||||||
device = self._load_device(phone_number)
|
|
||||||
return device is not None and device.safety_number is not None
|
|
||||||
|
|
||||||
def verify(self, phone_number: str) -> bool:
|
|
||||||
"""Mark a device as verified. Called by admin over SSH.
|
|
||||||
|
|
||||||
Returns True if the device was found and verified.
|
|
||||||
"""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
device = session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
if not device:
|
|
||||||
logger.warning(f"Cannot verify unknown device: {phone_number}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
device.trust_level = TrustLevel.VERIFIED
|
|
||||||
self.signal_client.trust_identity(phone_number, trust_all_known_keys=True)
|
|
||||||
session.commit()
|
|
||||||
self._update_cache(phone_number, device)
|
|
||||||
logger.info(f"Device verified: {phone_number}")
|
|
||||||
return True
|
|
||||||
|
|
||||||
def block(self, phone_number: str) -> bool:
|
|
||||||
"""Block a device."""
|
|
||||||
return self._set_trust(phone_number, TrustLevel.BLOCKED, "Device blocked")
|
|
||||||
|
|
||||||
def unverify(self, phone_number: str) -> bool:
|
|
||||||
"""Reset a device to unverified."""
|
|
||||||
return self._set_trust(phone_number, TrustLevel.UNVERIFIED)
|
|
||||||
|
|
||||||
# -- role management ------------------------------------------------------
|
|
||||||
|
|
||||||
def get_roles(self, phone_number: str) -> list[Role]:
|
|
||||||
"""Return the roles for a device, defaulting to empty."""
|
|
||||||
if entry := self._cached(phone_number):
|
|
||||||
return entry.roles
|
|
||||||
device = self._load_device(phone_number)
|
|
||||||
return _extract_roles(device) if device else []
|
|
||||||
|
|
||||||
def has_role(self, phone_number: str, role: Role) -> bool:
|
|
||||||
"""Check if a device has a specific role or is admin."""
|
|
||||||
roles = self.get_roles(phone_number)
|
|
||||||
return Role.ADMIN in roles or role in roles
|
|
||||||
|
|
||||||
def grant_role(self, phone_number: str, role: Role) -> bool:
|
|
||||||
"""Add a role to a device. Called by admin over SSH."""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
device = session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
if not device:
|
|
||||||
logger.warning(f"Cannot grant role for unknown device: {phone_number}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
if any(record.name == role for record in device.roles):
|
|
||||||
return True
|
|
||||||
|
|
||||||
role_record = session.execute(select(RoleRecord).where(RoleRecord.name == role)).scalar_one_or_none()
|
|
||||||
|
|
||||||
if not role_record:
|
|
||||||
logger.warning(f"Unknown role: {role}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
device.roles.append(role_record)
|
|
||||||
session.commit()
|
|
||||||
self._update_cache(phone_number, device)
|
|
||||||
logger.info(f"Device {phone_number} granted role {role}")
|
|
||||||
return True
|
|
||||||
|
|
||||||
def revoke_role(self, phone_number: str, role: Role) -> bool:
|
|
||||||
"""Remove a role from a device. Called by admin over SSH."""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
device = session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
if not device:
|
|
||||||
logger.warning(f"Cannot revoke role for unknown device: {phone_number}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
device.roles = [record for record in device.roles if record.name != role]
|
|
||||||
session.commit()
|
|
||||||
self._update_cache(phone_number, device)
|
|
||||||
logger.info(f"Device {phone_number} revoked role {role}")
|
|
||||||
return True
|
|
||||||
|
|
||||||
def set_roles(self, phone_number: str, roles: list[Role]) -> bool:
|
|
||||||
"""Replace all roles for a device. Called by admin over SSH."""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
device = session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
if not device:
|
|
||||||
logger.warning(f"Cannot set roles for unknown device: {phone_number}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
role_names = [str(role) for role in roles]
|
|
||||||
records = list(session.execute(select(RoleRecord).where(RoleRecord.name.in_(role_names))).scalars().all())
|
|
||||||
device.roles = records
|
|
||||||
session.commit()
|
|
||||||
self._update_cache(phone_number, device)
|
|
||||||
logger.info(f"Device {phone_number} roles set to {role_names}")
|
|
||||||
return True
|
|
||||||
|
|
||||||
# -- queries --------------------------------------------------------------
|
|
||||||
|
|
||||||
def list_devices(self) -> list[SignalDevice]:
|
|
||||||
"""Return all known devices."""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
return list(session.execute(select(SignalDevice)).scalars().all())
|
|
||||||
|
|
||||||
def sync_identities(self) -> None:
|
|
||||||
"""Pull identity list from signal-cli and record any new ones."""
|
|
||||||
identities = self.signal_client.get_identities()
|
|
||||||
for identity in identities:
|
|
||||||
number = identity.get("number", "")
|
|
||||||
safety = identity.get("safety_number", identity.get("fingerprint", ""))
|
|
||||||
if number:
|
|
||||||
self.record_contact(number, safety)
|
|
||||||
|
|
||||||
# -- internals ------------------------------------------------------------
|
|
||||||
|
|
||||||
def _cached(self, phone_number: str) -> _CacheEntry | None:
|
|
||||||
"""Return the cache entry if it exists and hasn't expired."""
|
|
||||||
entry = self._contact_cache.get(phone_number)
|
|
||||||
if entry and utcnow() < entry.expires:
|
|
||||||
return entry
|
|
||||||
return None
|
|
||||||
|
|
||||||
def _load_device(self, phone_number: str) -> SignalDevice | None:
|
|
||||||
"""Fetch a device by phone number (with joined roles)."""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
return session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
def _update_cache(self, phone_number: str, device: SignalDevice) -> None:
|
|
||||||
"""Refresh the cache entry for a device."""
|
|
||||||
ttl = _BLOCKED_TTL if device.trust_level == TrustLevel.BLOCKED else _DEFAULT_TTL
|
|
||||||
self._contact_cache[phone_number] = _CacheEntry(
|
|
||||||
expires=utcnow() + ttl,
|
|
||||||
trust_level=device.trust_level,
|
|
||||||
has_safety_number=device.safety_number is not None,
|
|
||||||
safety_number=device.safety_number,
|
|
||||||
roles=_extract_roles(device),
|
|
||||||
)
|
|
||||||
|
|
||||||
def _set_trust(self, phone_number: str, level: str, log_msg: str | None = None) -> bool:
|
|
||||||
"""Update the trust level for a device."""
|
|
||||||
with Session(self.engine) as session:
|
|
||||||
device = session.execute(
|
|
||||||
select(SignalDevice).where(SignalDevice.phone_number == phone_number)
|
|
||||||
).scalar_one_or_none()
|
|
||||||
|
|
||||||
if not device:
|
|
||||||
return False
|
|
||||||
|
|
||||||
device.trust_level = level
|
|
||||||
session.commit()
|
|
||||||
self._update_cache(phone_number, device)
|
|
||||||
if log_msg:
|
|
||||||
logger.info(f"{log_msg}: {phone_number}")
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
def _extract_roles(device: SignalDevice) -> list[Role]:
|
|
||||||
"""Convert a device's RoleRecord objects to a list of Role enums."""
|
|
||||||
return [Role(record.name) for record in device.roles]
|
|
||||||
|
|
||||||
|
|
||||||
def sync_roles(engine: Engine) -> None:
|
|
||||||
"""Sync the Role enum to the role table, adding new and removing stale entries."""
|
|
||||||
expected = {role.value for role in Role}
|
|
||||||
|
|
||||||
with Session(engine) as session:
|
|
||||||
existing = {record.name for record in session.execute(select(RoleRecord)).scalars().all()}
|
|
||||||
|
|
||||||
to_add = expected - existing
|
|
||||||
to_remove = existing - expected
|
|
||||||
|
|
||||||
for name in to_add:
|
|
||||||
session.add(RoleRecord(name=name))
|
|
||||||
logger.info(f"Role added: {name}")
|
|
||||||
|
|
||||||
if to_remove:
|
|
||||||
session.execute(delete(RoleRecord).where(RoleRecord.name.in_(to_remove)))
|
|
||||||
for name in to_remove:
|
|
||||||
logger.info(f"Role removed: {name}")
|
|
||||||
|
|
||||||
session.commit()
|
|
||||||
@@ -1,80 +0,0 @@
|
|||||||
"""Flexible LLM client for ollama backends."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import base64
|
|
||||||
import logging
|
|
||||||
from typing import Any, Self
|
|
||||||
|
|
||||||
import httpx
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
class LLMClient:
|
|
||||||
"""Talk to an ollama instance.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
model: Ollama model name.
|
|
||||||
host: Ollama host.
|
|
||||||
port: Ollama port.
|
|
||||||
temperature: Sampling temperature.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
*,
|
|
||||||
model: str,
|
|
||||||
host: str,
|
|
||||||
port: int = 11434,
|
|
||||||
temperature: float = 0.1,
|
|
||||||
timeout: int = 300,
|
|
||||||
) -> None:
|
|
||||||
self.model = model
|
|
||||||
self.temperature = temperature
|
|
||||||
self._client = httpx.Client(base_url=f"http://{host}:{port}", timeout=timeout)
|
|
||||||
|
|
||||||
def chat(self, prompt: str, image_data: bytes | None = None, system: str | None = None) -> str:
|
|
||||||
"""Send a text prompt and return the response."""
|
|
||||||
messages: list[dict[str, Any]] = []
|
|
||||||
if system:
|
|
||||||
messages.append({"role": "system", "content": system})
|
|
||||||
|
|
||||||
user_msg = {"role": "user", "content": prompt}
|
|
||||||
if image_data:
|
|
||||||
user_msg["images"] = [base64.b64encode(image_data).decode()]
|
|
||||||
|
|
||||||
messages.append(user_msg)
|
|
||||||
return self._generate(messages)
|
|
||||||
|
|
||||||
def _generate(self, messages: list[dict[str, Any]]) -> str:
|
|
||||||
"""Call the ollama chat API."""
|
|
||||||
payload = {
|
|
||||||
"model": self.model,
|
|
||||||
"messages": messages,
|
|
||||||
"stream": False,
|
|
||||||
"options": {"temperature": self.temperature},
|
|
||||||
}
|
|
||||||
logger.info(f"LLM request to {self.model}")
|
|
||||||
response = self._client.post("/api/chat", json=payload)
|
|
||||||
response.raise_for_status()
|
|
||||||
data = response.json()
|
|
||||||
return data["message"]["content"]
|
|
||||||
|
|
||||||
def list_models(self) -> list[str]:
|
|
||||||
"""List available models on the ollama instance."""
|
|
||||||
response = self._client.get("/api/tags")
|
|
||||||
response.raise_for_status()
|
|
||||||
return [m["name"] for m in response.json().get("models", [])]
|
|
||||||
|
|
||||||
def __enter__(self) -> Self:
|
|
||||||
"""Enter the context manager."""
|
|
||||||
return self
|
|
||||||
|
|
||||||
def __exit__(self, *args: object) -> None:
|
|
||||||
"""Close the HTTP client on exit."""
|
|
||||||
self.close()
|
|
||||||
|
|
||||||
def close(self) -> None:
|
|
||||||
"""Close the HTTP client."""
|
|
||||||
self._client.close()
|
|
||||||
@@ -1,239 +0,0 @@
|
|||||||
"""Signal command and control bot — main entry point."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from dataclasses import dataclass
|
|
||||||
from os import getenv
|
|
||||||
from typing import TYPE_CHECKING, Annotated
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Callable
|
|
||||||
|
|
||||||
import typer
|
|
||||||
from alembic.command import upgrade
|
|
||||||
from sqlalchemy.orm import Session
|
|
||||||
from tenacity import before_sleep_log, retry, stop_after_attempt, wait_exponential
|
|
||||||
|
|
||||||
from python.common import configure_logger, utcnow
|
|
||||||
from python.database_cli import DATABASES
|
|
||||||
from python.orm.common import get_postgres_engine
|
|
||||||
from python.orm.signal_bot.models import DeadLetterMessage
|
|
||||||
from python.signal_bot.commands.inventory import handle_inventory_update
|
|
||||||
from python.signal_bot.commands.location import handle_location_request
|
|
||||||
from python.signal_bot.device_registry import DeviceRegistry, sync_roles
|
|
||||||
from python.signal_bot.llm_client import LLMClient
|
|
||||||
from python.signal_bot.models import BotConfig, MessageStatus, Role, SignalMessage
|
|
||||||
from python.signal_bot.signal_client import SignalClient
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
@dataclass(frozen=True, slots=True)
|
|
||||||
class Command:
|
|
||||||
"""A registered bot command."""
|
|
||||||
|
|
||||||
action: Callable[[SignalMessage, str], None]
|
|
||||||
help_text: str
|
|
||||||
role: Role | None # None = no role required (always allowed)
|
|
||||||
|
|
||||||
|
|
||||||
class Bot:
|
|
||||||
"""Holds shared resources and dispatches incoming messages to command handlers."""
|
|
||||||
|
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
signal: SignalClient,
|
|
||||||
llm: LLMClient,
|
|
||||||
registry: DeviceRegistry,
|
|
||||||
config: BotConfig,
|
|
||||||
) -> None:
|
|
||||||
self.signal = signal
|
|
||||||
self.llm = llm
|
|
||||||
self.registry = registry
|
|
||||||
self.config = config
|
|
||||||
self.commands: dict[str, Command] = {
|
|
||||||
"help": Command(action=self._help, help_text="show this help message", role=None),
|
|
||||||
"status": Command(action=self._status, help_text="show bot status", role=Role.STATUS),
|
|
||||||
"inventory": Command(
|
|
||||||
action=self._inventory,
|
|
||||||
help_text="update van inventory from a text list or receipt photo",
|
|
||||||
role=Role.INVENTORY,
|
|
||||||
),
|
|
||||||
"location": Command(
|
|
||||||
action=self._location,
|
|
||||||
help_text="get current van location",
|
|
||||||
role=Role.LOCATION,
|
|
||||||
),
|
|
||||||
}
|
|
||||||
|
|
||||||
# -- actions --------------------------------------------------------------
|
|
||||||
|
|
||||||
def _help(self, message: SignalMessage, _cmd: str) -> None:
|
|
||||||
"""Return help text filtered to the sender's roles."""
|
|
||||||
self.signal.reply(message, self._build_help(self.registry.get_roles(message.source)))
|
|
||||||
|
|
||||||
def _status(self, message: SignalMessage, _cmd: str) -> None:
|
|
||||||
"""Return the status of the bot."""
|
|
||||||
models = self.llm.list_models()
|
|
||||||
model_list = ", ".join(models[:10])
|
|
||||||
device_count = len(self.registry.list_devices())
|
|
||||||
self.signal.reply(
|
|
||||||
message,
|
|
||||||
f"Bot online.\nLLM: {self.llm.model}\nAvailable models: {model_list}\nKnown devices: {device_count}",
|
|
||||||
)
|
|
||||||
|
|
||||||
def _inventory(self, message: SignalMessage, _cmd: str) -> None:
|
|
||||||
"""Process an inventory update."""
|
|
||||||
handle_inventory_update(message, self.signal, self.llm, self.config.inventory_api_url)
|
|
||||||
|
|
||||||
def _location(self, message: SignalMessage, _cmd: str) -> None:
|
|
||||||
"""Reply with current van location."""
|
|
||||||
handle_location_request(message, self.signal, self.config.ha_url, self.config.ha_token)
|
|
||||||
|
|
||||||
# -- dispatch -------------------------------------------------------------
|
|
||||||
|
|
||||||
def _build_help(self, roles: list[Role]) -> str:
|
|
||||||
"""Build help text showing only the commands the user can access."""
|
|
||||||
is_admin = Role.ADMIN in roles
|
|
||||||
lines = ["Available commands:"]
|
|
||||||
for name, cmd in self.commands.items():
|
|
||||||
if cmd.role is None or is_admin or cmd.role in roles:
|
|
||||||
lines.append(f" {name:20s} — {cmd.help_text}")
|
|
||||||
return "\n".join(lines)
|
|
||||||
|
|
||||||
def dispatch(self, message: SignalMessage) -> None:
|
|
||||||
"""Route an incoming message to the right command handler."""
|
|
||||||
source = message.source
|
|
||||||
|
|
||||||
if not self.registry.is_verified(source):
|
|
||||||
logger.info(f"Device {source} not verified, ignoring message")
|
|
||||||
return
|
|
||||||
|
|
||||||
if not self.registry.has_safety_number(source) and self.registry.has_role(source, Role.ADMIN):
|
|
||||||
logger.warning(f"Admin device {source} missing safety number, ignoring message")
|
|
||||||
return
|
|
||||||
|
|
||||||
text = message.message.strip()
|
|
||||||
parts = text.split()
|
|
||||||
|
|
||||||
if not parts and not message.attachments:
|
|
||||||
return
|
|
||||||
|
|
||||||
cmd = parts[0].lower() if parts else ""
|
|
||||||
|
|
||||||
logger.info(f"f{source=} running {cmd=} with {message=}")
|
|
||||||
|
|
||||||
command = self.commands.get(cmd)
|
|
||||||
if command is None:
|
|
||||||
if message.attachments:
|
|
||||||
command = self.commands["inventory"]
|
|
||||||
cmd = "inventory"
|
|
||||||
else:
|
|
||||||
return
|
|
||||||
|
|
||||||
if command.role is not None and not self.registry.has_role(source, command.role):
|
|
||||||
logger.warning(f"Device {source} denied access to {cmd!r}")
|
|
||||||
self.signal.reply(message, f"Permission denied: you do not have the '{command.role}' role.")
|
|
||||||
return
|
|
||||||
|
|
||||||
command.action(message, cmd)
|
|
||||||
|
|
||||||
def process_message(self, message: SignalMessage) -> None:
|
|
||||||
"""Process a single message, sending it to the dead letter queue after repeated failures."""
|
|
||||||
max_attempts = self.config.max_message_attempts
|
|
||||||
for attempt in range(1, max_attempts + 1):
|
|
||||||
try:
|
|
||||||
safety_number = self.signal.get_safety_number(message.source)
|
|
||||||
self.registry.record_contact(message.source, safety_number)
|
|
||||||
self.dispatch(message)
|
|
||||||
except Exception:
|
|
||||||
logger.exception(f"Failed to process message (attempt {attempt}/{max_attempts})")
|
|
||||||
else:
|
|
||||||
return
|
|
||||||
|
|
||||||
logger.error(f"Message from {message.source} failed {max_attempts} times, sending to dead letter queue")
|
|
||||||
with Session(self.config.engine) as session:
|
|
||||||
session.add(
|
|
||||||
DeadLetterMessage(
|
|
||||||
source=message.source,
|
|
||||||
message=message.message,
|
|
||||||
received_at=utcnow(),
|
|
||||||
status=MessageStatus.UNPROCESSED,
|
|
||||||
)
|
|
||||||
)
|
|
||||||
session.commit()
|
|
||||||
|
|
||||||
def run(self) -> None:
|
|
||||||
"""Listen for messages via WebSocket, reconnecting on failure."""
|
|
||||||
logger.info("Bot started — listening via WebSocket")
|
|
||||||
|
|
||||||
@retry(
|
|
||||||
stop=stop_after_attempt(self.config.max_retries),
|
|
||||||
wait=wait_exponential(multiplier=self.config.reconnect_delay, max=self.config.max_reconnect_delay),
|
|
||||||
before_sleep=before_sleep_log(logger, logging.WARNING),
|
|
||||||
reraise=True,
|
|
||||||
)
|
|
||||||
def _listen() -> None:
|
|
||||||
for message in self.signal.listen():
|
|
||||||
logger.info(f"Message from {message.source}: {message.message[:80]}")
|
|
||||||
self.process_message(message)
|
|
||||||
|
|
||||||
try:
|
|
||||||
_listen()
|
|
||||||
except Exception:
|
|
||||||
logger.critical("Max retries exceeded, shutting down")
|
|
||||||
raise
|
|
||||||
|
|
||||||
|
|
||||||
def main(
|
|
||||||
log_level: Annotated[str, typer.Option()] = "DEBUG",
|
|
||||||
llm_timeout: Annotated[int, typer.Option()] = 600,
|
|
||||||
) -> None:
|
|
||||||
"""Run the Signal command and control bot."""
|
|
||||||
configure_logger(log_level)
|
|
||||||
signal_api_url = getenv("SIGNAL_API_URL")
|
|
||||||
phone_number = getenv("SIGNAL_PHONE_NUMBER")
|
|
||||||
inventory_api_url = getenv("INVENTORY_API_URL")
|
|
||||||
|
|
||||||
if signal_api_url is None:
|
|
||||||
error = "SIGNAL_API_URL environment variable not set"
|
|
||||||
raise ValueError(error)
|
|
||||||
if phone_number is None:
|
|
||||||
error = "SIGNAL_PHONE_NUMBER environment variable not set"
|
|
||||||
raise ValueError(error)
|
|
||||||
if inventory_api_url is None:
|
|
||||||
error = "INVENTORY_API_URL environment variable not set"
|
|
||||||
raise ValueError(error)
|
|
||||||
|
|
||||||
signal_bot_config = DATABASES["signal_bot"].alembic_config()
|
|
||||||
upgrade(signal_bot_config, "head")
|
|
||||||
engine = get_postgres_engine(name="SIGNALBOT")
|
|
||||||
sync_roles(engine)
|
|
||||||
config = BotConfig(
|
|
||||||
signal_api_url=signal_api_url,
|
|
||||||
phone_number=phone_number,
|
|
||||||
inventory_api_url=inventory_api_url,
|
|
||||||
ha_url=getenv("HA_URL"),
|
|
||||||
ha_token=getenv("HA_TOKEN"),
|
|
||||||
engine=engine,
|
|
||||||
)
|
|
||||||
|
|
||||||
llm_host = getenv("LLM_HOST")
|
|
||||||
llm_model = getenv("LLM_MODEL", "qwen3-vl:32b")
|
|
||||||
llm_port = int(getenv("LLM_PORT", "11434"))
|
|
||||||
if llm_host is None:
|
|
||||||
error = "LLM_HOST environment variable not set"
|
|
||||||
raise ValueError(error)
|
|
||||||
|
|
||||||
with (
|
|
||||||
SignalClient(config.signal_api_url, config.phone_number) as signal,
|
|
||||||
LLMClient(model=llm_model, host=llm_host, port=llm_port, timeout=llm_timeout) as llm,
|
|
||||||
):
|
|
||||||
registry = DeviceRegistry(signal, engine)
|
|
||||||
bot = Bot(signal, llm, registry, config)
|
|
||||||
bot.run()
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
typer.run(main)
|
|
||||||
@@ -1,97 +0,0 @@
|
|||||||
"""Models for the Signal command and control bot."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from datetime import datetime # noqa: TC003 - pydantic needs this at runtime
|
|
||||||
from enum import StrEnum
|
|
||||||
from typing import Any
|
|
||||||
|
|
||||||
from pydantic import BaseModel, ConfigDict
|
|
||||||
from sqlalchemy.engine import Engine # noqa: TC002 - pydantic needs this at runtime
|
|
||||||
|
|
||||||
|
|
||||||
class TrustLevel(StrEnum):
|
|
||||||
"""Device trust level."""
|
|
||||||
|
|
||||||
VERIFIED = "verified"
|
|
||||||
UNVERIFIED = "unverified"
|
|
||||||
BLOCKED = "blocked"
|
|
||||||
|
|
||||||
|
|
||||||
class Role(StrEnum):
|
|
||||||
"""RBAC roles — one per command, plus admin which grants all."""
|
|
||||||
|
|
||||||
ADMIN = "admin"
|
|
||||||
STATUS = "status"
|
|
||||||
INVENTORY = "inventory"
|
|
||||||
LOCATION = "location"
|
|
||||||
|
|
||||||
|
|
||||||
class MessageStatus(StrEnum):
|
|
||||||
"""Dead letter queue message status."""
|
|
||||||
|
|
||||||
UNPROCESSED = "unprocessed"
|
|
||||||
PROCESSED = "processed"
|
|
||||||
|
|
||||||
|
|
||||||
class Device(BaseModel):
|
|
||||||
"""A registered device tracked by safety number."""
|
|
||||||
|
|
||||||
phone_number: str
|
|
||||||
safety_number: str
|
|
||||||
trust_level: TrustLevel = TrustLevel.UNVERIFIED
|
|
||||||
first_seen: datetime
|
|
||||||
last_seen: datetime
|
|
||||||
|
|
||||||
|
|
||||||
class SignalMessage(BaseModel):
|
|
||||||
"""An incoming Signal message."""
|
|
||||||
|
|
||||||
source: str
|
|
||||||
timestamp: int
|
|
||||||
message: str = ""
|
|
||||||
attachments: list[str] = []
|
|
||||||
group_id: str | None = None
|
|
||||||
is_receipt: bool = False
|
|
||||||
|
|
||||||
|
|
||||||
class SignalEnvelope(BaseModel):
|
|
||||||
"""Raw envelope from signal-cli-rest-api."""
|
|
||||||
|
|
||||||
envelope: dict[str, Any]
|
|
||||||
account: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class InventoryItem(BaseModel):
|
|
||||||
"""An item in the van inventory."""
|
|
||||||
|
|
||||||
name: str
|
|
||||||
quantity: float = 1
|
|
||||||
unit: str = "each"
|
|
||||||
category: str = ""
|
|
||||||
notes: str = ""
|
|
||||||
|
|
||||||
|
|
||||||
class InventoryUpdate(BaseModel):
|
|
||||||
"""Result of processing an inventory update."""
|
|
||||||
|
|
||||||
items: list[InventoryItem] = []
|
|
||||||
raw_response: str = ""
|
|
||||||
source_type: str = "" # "receipt_photo" or "text_list"
|
|
||||||
|
|
||||||
|
|
||||||
class BotConfig(BaseModel):
|
|
||||||
"""Top-level bot configuration."""
|
|
||||||
|
|
||||||
model_config = ConfigDict(arbitrary_types_allowed=True)
|
|
||||||
|
|
||||||
signal_api_url: str
|
|
||||||
phone_number: str
|
|
||||||
inventory_api_url: str
|
|
||||||
ha_url: str | None = None
|
|
||||||
ha_token: str | None = None
|
|
||||||
engine: Engine
|
|
||||||
reconnect_delay: int = 5
|
|
||||||
max_reconnect_delay: int = 300
|
|
||||||
max_retries: int = 10
|
|
||||||
max_message_attempts: int = 3
|
|
||||||
@@ -1,141 +0,0 @@
|
|||||||
"""Client for the signal-cli-rest-api."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import json
|
|
||||||
import logging
|
|
||||||
from typing import TYPE_CHECKING, Any, Self
|
|
||||||
|
|
||||||
import httpx
|
|
||||||
import websockets.sync.client
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import Generator
|
|
||||||
|
|
||||||
from python.signal_bot.models import SignalMessage
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def _parse_envelope(envelope: dict[str, Any]) -> SignalMessage | None:
|
|
||||||
"""Parse a signal-cli envelope into a SignalMessage, or None if not a data message."""
|
|
||||||
data_message = envelope.get("dataMessage")
|
|
||||||
if not data_message:
|
|
||||||
return None
|
|
||||||
|
|
||||||
attachment_ids = [att["id"] for att in data_message.get("attachments", []) if "id" in att]
|
|
||||||
|
|
||||||
group_info = data_message.get("groupInfo")
|
|
||||||
group_id = group_info.get("groupId") if group_info else None
|
|
||||||
|
|
||||||
return SignalMessage(
|
|
||||||
source=envelope.get("source", ""),
|
|
||||||
timestamp=envelope.get("timestamp", 0),
|
|
||||||
message=data_message.get("message", "") or "",
|
|
||||||
attachments=attachment_ids,
|
|
||||||
group_id=group_id,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class SignalClient:
|
|
||||||
"""Communicate with signal-cli-rest-api.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
base_url: URL of the signal-cli-rest-api (e.g. http://localhost:8989).
|
|
||||||
phone_number: The registered phone number to send/receive as.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, base_url: str, phone_number: str) -> None:
|
|
||||||
self.base_url = base_url.rstrip("/")
|
|
||||||
self.phone_number = phone_number
|
|
||||||
self._client = httpx.Client(base_url=self.base_url, timeout=30)
|
|
||||||
|
|
||||||
def _ws_url(self) -> str:
|
|
||||||
"""Build the WebSocket URL from the base HTTP URL."""
|
|
||||||
url = self.base_url.replace("http://", "ws://").replace("https://", "wss://")
|
|
||||||
return f"{url}/v1/receive/{self.phone_number}"
|
|
||||||
|
|
||||||
def listen(self) -> Generator[SignalMessage]:
|
|
||||||
"""Connect via WebSocket and yield messages as they arrive."""
|
|
||||||
ws_url = self._ws_url()
|
|
||||||
logger.info(f"Connecting to WebSocket: {ws_url}")
|
|
||||||
|
|
||||||
with websockets.sync.client.connect(ws_url) as ws:
|
|
||||||
for raw in ws:
|
|
||||||
try:
|
|
||||||
data = json.loads(raw)
|
|
||||||
envelope = data.get("envelope", {})
|
|
||||||
message = _parse_envelope(envelope)
|
|
||||||
if message:
|
|
||||||
yield message
|
|
||||||
except json.JSONDecodeError:
|
|
||||||
logger.warning(f"Non-JSON WebSocket frame: {raw[:200]}")
|
|
||||||
|
|
||||||
def send(self, recipient: str, message: str) -> None:
|
|
||||||
"""Send a text message."""
|
|
||||||
payload = {
|
|
||||||
"message": message,
|
|
||||||
"number": self.phone_number,
|
|
||||||
"recipients": [recipient],
|
|
||||||
}
|
|
||||||
response = self._client.post("/v2/send", json=payload)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
def send_to_group(self, group_id: str, message: str) -> None:
|
|
||||||
"""Send a message to a group."""
|
|
||||||
payload = {
|
|
||||||
"message": message,
|
|
||||||
"number": self.phone_number,
|
|
||||||
"recipients": [group_id],
|
|
||||||
}
|
|
||||||
response = self._client.post("/v2/send", json=payload)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
def get_attachment(self, attachment_id: str) -> bytes:
|
|
||||||
"""Download an attachment by ID."""
|
|
||||||
response = self._client.get(f"/v1/attachments/{attachment_id}")
|
|
||||||
response.raise_for_status()
|
|
||||||
return response.content
|
|
||||||
|
|
||||||
def get_identities(self) -> list[dict[str, Any]]:
|
|
||||||
"""List known identities and their trust levels."""
|
|
||||||
response = self._client.get(f"/v1/identities/{self.phone_number}")
|
|
||||||
response.raise_for_status()
|
|
||||||
return response.json()
|
|
||||||
|
|
||||||
def get_safety_number(self, phone_number: str) -> str | None:
|
|
||||||
"""Look up the safety number for a contact from signal-cli's local store."""
|
|
||||||
for identity in self.get_identities():
|
|
||||||
if identity.get("number") == phone_number:
|
|
||||||
return identity.get("safety_number", identity.get("fingerprint", ""))
|
|
||||||
return None
|
|
||||||
|
|
||||||
def trust_identity(self, number_to_trust: str, *, trust_all_known_keys: bool = False) -> None:
|
|
||||||
"""Trust an identity (verify safety number)."""
|
|
||||||
payload: dict[str, Any] = {}
|
|
||||||
if trust_all_known_keys:
|
|
||||||
payload["trust_all_known_keys"] = True
|
|
||||||
response = self._client.put(
|
|
||||||
f"/v1/identities/{self.phone_number}/trust/{number_to_trust}",
|
|
||||||
json=payload,
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
def reply(self, message: SignalMessage, text: str) -> None:
|
|
||||||
"""Reply to a message, routing to group or individual."""
|
|
||||||
if message.group_id:
|
|
||||||
self.send_to_group(message.group_id, text)
|
|
||||||
else:
|
|
||||||
self.send(message.source, text)
|
|
||||||
|
|
||||||
def __enter__(self) -> Self:
|
|
||||||
"""Enter the context manager."""
|
|
||||||
return self
|
|
||||||
|
|
||||||
def __exit__(self, *args: object) -> None:
|
|
||||||
"""Close the HTTP client on exit."""
|
|
||||||
self.close()
|
|
||||||
|
|
||||||
def close(self) -> None:
|
|
||||||
"""Close the HTTP client."""
|
|
||||||
self._client.close()
|
|
||||||
111
python/tools/safe_reboot.py
Normal file
111
python/tools/safe_reboot.py
Normal file
@@ -0,0 +1,111 @@
|
|||||||
|
"""Safe reboot helper."""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
|
||||||
|
import logging
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
from typing import TYPE_CHECKING, Annotated
|
||||||
|
|
||||||
|
import typer
|
||||||
|
|
||||||
|
from python.common import bash_wrapper, configure_logger
|
||||||
|
from python.zfs import Dataset, get_datasets
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from collections.abc import Sequence
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def get_root_pool_datasets(dataset_prefix: str) -> list[Dataset]:
|
||||||
|
"""Return datasets that start with the provided prefix."""
|
||||||
|
return [dataset for dataset in get_datasets() if dataset.name.startswith(dataset_prefix)]
|
||||||
|
|
||||||
|
|
||||||
|
def get_non_executable_datasets(datasets: Sequence[Dataset]) -> list[str]:
|
||||||
|
"""Return dataset names that have exec disabled."""
|
||||||
|
return [dataset.name for dataset in datasets if dataset.exec.lower() != "on"]
|
||||||
|
|
||||||
|
|
||||||
|
def drive_present(drive: str) -> bool:
|
||||||
|
"""Check whether the provided drive exists."""
|
||||||
|
drive_path = drive.strip()
|
||||||
|
if not drive_path:
|
||||||
|
error = "Drive path cannot be empty"
|
||||||
|
raise ValueError(error)
|
||||||
|
|
||||||
|
return Path(drive_path).exists()
|
||||||
|
|
||||||
|
|
||||||
|
def reboot_system() -> None:
|
||||||
|
"""Call systemctl reboot."""
|
||||||
|
output, return_code = bash_wrapper("systemctl reboot")
|
||||||
|
if return_code != 0:
|
||||||
|
raise RuntimeError(output.strip() or "Failed to issue reboot command")
|
||||||
|
|
||||||
|
|
||||||
|
def validate_state(drive: str | None, dataset_prefix: str) -> list[str]:
|
||||||
|
"""Validate dataset and drive state."""
|
||||||
|
datasets = get_root_pool_datasets(dataset_prefix)
|
||||||
|
|
||||||
|
errors: list[str] = []
|
||||||
|
if not datasets:
|
||||||
|
errors.append(f"No datasets found with prefix {dataset_prefix}")
|
||||||
|
else:
|
||||||
|
non_exec_datasets = get_non_executable_datasets(datasets)
|
||||||
|
if non_exec_datasets:
|
||||||
|
errors.append(f"Datasets missing exec=on: {', '.join(non_exec_datasets)}")
|
||||||
|
|
||||||
|
if drive:
|
||||||
|
try:
|
||||||
|
if not drive_present(drive):
|
||||||
|
errors.append(f"Drive {drive} is not present")
|
||||||
|
except ValueError as err:
|
||||||
|
errors.append(str(err))
|
||||||
|
|
||||||
|
return errors
|
||||||
|
|
||||||
|
|
||||||
|
def reboot(
|
||||||
|
drive: Annotated[str | None, typer.Argument(help="Drive that must exist before rebooting.")] = None,
|
||||||
|
dataset_prefix: Annotated[
|
||||||
|
str,
|
||||||
|
typer.Option(
|
||||||
|
"--dataset-prefix",
|
||||||
|
"-p",
|
||||||
|
help="Datasets with this prefix are validated.",
|
||||||
|
),
|
||||||
|
] = "root_pool/",
|
||||||
|
dry_run: Annotated[
|
||||||
|
bool,
|
||||||
|
typer.Option(
|
||||||
|
"--check-only",
|
||||||
|
help="Only validate state without issuing the reboot command.",
|
||||||
|
),
|
||||||
|
] = False,
|
||||||
|
) -> None:
|
||||||
|
"""Validate datasets and drive before rebooting."""
|
||||||
|
configure_logger()
|
||||||
|
logger.info("Starting safe reboot checks")
|
||||||
|
|
||||||
|
if errors := validate_state(drive, dataset_prefix):
|
||||||
|
for error in errors:
|
||||||
|
logger.error(error)
|
||||||
|
sys.exit(1)
|
||||||
|
|
||||||
|
if dry_run:
|
||||||
|
logger.info("All checks passed")
|
||||||
|
return
|
||||||
|
|
||||||
|
logger.info("All checks passed, issuing reboot")
|
||||||
|
reboot_system()
|
||||||
|
|
||||||
|
|
||||||
|
def cli() -> None:
|
||||||
|
"""CLI entry point."""
|
||||||
|
typer.run(reboot)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
cli()
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Van inventory FastAPI application."""
|
|
||||||
@@ -1,16 +0,0 @@
|
|||||||
"""FastAPI dependencies for van inventory."""
|
|
||||||
|
|
||||||
from collections.abc import Iterator
|
|
||||||
from typing import Annotated
|
|
||||||
|
|
||||||
from fastapi import Depends, Request
|
|
||||||
from sqlalchemy.orm import Session
|
|
||||||
|
|
||||||
|
|
||||||
def get_db(request: Request) -> Iterator[Session]:
|
|
||||||
"""Get database session from app state."""
|
|
||||||
with Session(request.app.state.engine) as session:
|
|
||||||
yield session
|
|
||||||
|
|
||||||
|
|
||||||
DbSession = Annotated[Session, Depends(get_db)]
|
|
||||||
@@ -1,56 +0,0 @@
|
|||||||
"""FastAPI app for van inventory."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from contextlib import asynccontextmanager
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import TYPE_CHECKING, Annotated
|
|
||||||
|
|
||||||
import typer
|
|
||||||
import uvicorn
|
|
||||||
from fastapi import FastAPI
|
|
||||||
from fastapi.staticfiles import StaticFiles
|
|
||||||
|
|
||||||
from python.common import configure_logger
|
|
||||||
from python.orm.common import get_postgres_engine
|
|
||||||
from python.van_inventory.routers import api_router, frontend_router
|
|
||||||
|
|
||||||
STATIC_DIR = Path(__file__).resolve().parent / "static"
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from collections.abc import AsyncIterator
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
def create_app() -> FastAPI:
|
|
||||||
"""Create and configure the FastAPI application."""
|
|
||||||
|
|
||||||
@asynccontextmanager
|
|
||||||
async def lifespan(app: FastAPI) -> AsyncIterator[None]:
|
|
||||||
app.state.engine = get_postgres_engine(name="VAN_INVENTORY")
|
|
||||||
yield
|
|
||||||
app.state.engine.dispose()
|
|
||||||
|
|
||||||
app = FastAPI(title="Van Inventory", lifespan=lifespan)
|
|
||||||
app.mount("/static", StaticFiles(directory=STATIC_DIR), name="static")
|
|
||||||
app.include_router(api_router)
|
|
||||||
app.include_router(frontend_router)
|
|
||||||
return app
|
|
||||||
|
|
||||||
|
|
||||||
def serve(
|
|
||||||
# Intentionally binds all interfaces — this is a LAN-only van server
|
|
||||||
host: Annotated[str, typer.Option("--host", "-h", help="Host to bind to")] = "0.0.0.0", # noqa: S104
|
|
||||||
port: Annotated[int, typer.Option("--port", "-p", help="Port to bind to")] = 8001,
|
|
||||||
log_level: Annotated[str, typer.Option("--log-level", "-l", help="Log level")] = "INFO",
|
|
||||||
) -> None:
|
|
||||||
"""Start the Van Inventory server."""
|
|
||||||
configure_logger(log_level)
|
|
||||||
app = create_app()
|
|
||||||
uvicorn.run(app, host=host, port=port)
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
typer.run(serve)
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
"""Van inventory API routers."""
|
|
||||||
|
|
||||||
from python.van_inventory.routers.api import router as api_router
|
|
||||||
from python.van_inventory.routers.frontend import router as frontend_router
|
|
||||||
|
|
||||||
__all__ = ["api_router", "frontend_router"]
|
|
||||||
@@ -1,314 +0,0 @@
|
|||||||
"""Van inventory API router."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from typing import TYPE_CHECKING
|
|
||||||
|
|
||||||
from fastapi import APIRouter, HTTPException
|
|
||||||
from pydantic import BaseModel, Field
|
|
||||||
from sqlalchemy import select
|
|
||||||
from sqlalchemy.orm import selectinload
|
|
||||||
|
|
||||||
from python.orm.van_inventory.models import Item, Meal, MealIngredient
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
|
||||||
from python.van_inventory.dependencies import DbSession
|
|
||||||
|
|
||||||
|
|
||||||
# --- Schemas ---
|
|
||||||
|
|
||||||
|
|
||||||
class ItemCreate(BaseModel):
|
|
||||||
"""Schema for creating an item."""
|
|
||||||
|
|
||||||
name: str
|
|
||||||
quantity: float = Field(default=0, ge=0)
|
|
||||||
unit: str
|
|
||||||
category: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class ItemUpdate(BaseModel):
|
|
||||||
"""Schema for updating an item."""
|
|
||||||
|
|
||||||
name: str | None = None
|
|
||||||
quantity: float | None = Field(default=None, ge=0)
|
|
||||||
unit: str | None = None
|
|
||||||
category: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class ItemResponse(BaseModel):
|
|
||||||
"""Schema for item response."""
|
|
||||||
|
|
||||||
id: int
|
|
||||||
name: str
|
|
||||||
quantity: float
|
|
||||||
unit: str
|
|
||||||
category: str | None
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
|
|
||||||
class IngredientCreate(BaseModel):
|
|
||||||
"""Schema for adding an ingredient to a meal."""
|
|
||||||
|
|
||||||
item_id: int
|
|
||||||
quantity_needed: float = Field(gt=0)
|
|
||||||
|
|
||||||
|
|
||||||
class MealCreate(BaseModel):
|
|
||||||
"""Schema for creating a meal."""
|
|
||||||
|
|
||||||
name: str
|
|
||||||
instructions: str | None = None
|
|
||||||
ingredients: list[IngredientCreate] = []
|
|
||||||
|
|
||||||
|
|
||||||
class MealUpdate(BaseModel):
|
|
||||||
"""Schema for updating a meal."""
|
|
||||||
|
|
||||||
name: str | None = None
|
|
||||||
instructions: str | None = None
|
|
||||||
|
|
||||||
|
|
||||||
class IngredientResponse(BaseModel):
|
|
||||||
"""Schema for ingredient response."""
|
|
||||||
|
|
||||||
item_id: int
|
|
||||||
item_name: str
|
|
||||||
quantity_needed: float
|
|
||||||
unit: str
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
|
|
||||||
class MealResponse(BaseModel):
|
|
||||||
"""Schema for meal response."""
|
|
||||||
|
|
||||||
id: int
|
|
||||||
name: str
|
|
||||||
instructions: str | None
|
|
||||||
ingredients: list[IngredientResponse] = []
|
|
||||||
|
|
||||||
model_config = {"from_attributes": True}
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def from_meal(cls, meal: Meal) -> MealResponse:
|
|
||||||
"""Build a MealResponse from an ORM Meal with loaded ingredients."""
|
|
||||||
return cls(
|
|
||||||
id=meal.id,
|
|
||||||
name=meal.name,
|
|
||||||
instructions=meal.instructions,
|
|
||||||
ingredients=[
|
|
||||||
IngredientResponse(
|
|
||||||
item_id=mi.item_id,
|
|
||||||
item_name=mi.item.name,
|
|
||||||
quantity_needed=mi.quantity_needed,
|
|
||||||
unit=mi.item.unit,
|
|
||||||
)
|
|
||||||
for mi in meal.ingredients
|
|
||||||
],
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
class ShoppingItem(BaseModel):
|
|
||||||
"""An item needed for a meal that is short on stock."""
|
|
||||||
|
|
||||||
item_name: str
|
|
||||||
unit: str
|
|
||||||
needed: float
|
|
||||||
have: float
|
|
||||||
short: float
|
|
||||||
|
|
||||||
|
|
||||||
class MealAvailability(BaseModel):
|
|
||||||
"""Availability status for a meal."""
|
|
||||||
|
|
||||||
meal_id: int
|
|
||||||
meal_name: str
|
|
||||||
can_make: bool
|
|
||||||
missing: list[ShoppingItem] = []
|
|
||||||
|
|
||||||
|
|
||||||
# --- Routes ---
|
|
||||||
|
|
||||||
router = APIRouter(prefix="/api", tags=["van_inventory"])
|
|
||||||
|
|
||||||
|
|
||||||
# Items
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/items", response_model=ItemResponse)
|
|
||||||
def create_item(item: ItemCreate, db: DbSession) -> Item:
|
|
||||||
"""Create a new inventory item."""
|
|
||||||
db_item = Item(**item.model_dump())
|
|
||||||
db.add(db_item)
|
|
||||||
db.commit()
|
|
||||||
db.refresh(db_item)
|
|
||||||
return db_item
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/items", response_model=list[ItemResponse])
|
|
||||||
def list_items(db: DbSession) -> list[Item]:
|
|
||||||
"""List all inventory items."""
|
|
||||||
return list(db.scalars(select(Item).order_by(Item.name)).all())
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/items/{item_id}", response_model=ItemResponse)
|
|
||||||
def get_item(item_id: int, db: DbSession) -> Item:
|
|
||||||
"""Get an item by ID."""
|
|
||||||
item = db.get(Item, item_id)
|
|
||||||
if not item:
|
|
||||||
raise HTTPException(status_code=404, detail="Item not found")
|
|
||||||
return item
|
|
||||||
|
|
||||||
|
|
||||||
@router.patch("/items/{item_id}", response_model=ItemResponse)
|
|
||||||
def update_item(item_id: int, item: ItemUpdate, db: DbSession) -> Item:
|
|
||||||
"""Update an item by ID."""
|
|
||||||
db_item = db.get(Item, item_id)
|
|
||||||
if not db_item:
|
|
||||||
raise HTTPException(status_code=404, detail="Item not found")
|
|
||||||
for key, value in item.model_dump(exclude_unset=True).items():
|
|
||||||
setattr(db_item, key, value)
|
|
||||||
db.commit()
|
|
||||||
db.refresh(db_item)
|
|
||||||
return db_item
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/items/{item_id}")
|
|
||||||
def delete_item(item_id: int, db: DbSession) -> dict[str, bool]:
|
|
||||||
"""Delete an item by ID."""
|
|
||||||
item = db.get(Item, item_id)
|
|
||||||
if not item:
|
|
||||||
raise HTTPException(status_code=404, detail="Item not found")
|
|
||||||
db.delete(item)
|
|
||||||
db.commit()
|
|
||||||
return {"deleted": True}
|
|
||||||
|
|
||||||
|
|
||||||
# Meals
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/meals", response_model=MealResponse)
|
|
||||||
def create_meal(meal: MealCreate, db: DbSession) -> MealResponse:
|
|
||||||
"""Create a new meal with optional ingredients."""
|
|
||||||
for ing in meal.ingredients:
|
|
||||||
if not db.get(Item, ing.item_id):
|
|
||||||
raise HTTPException(status_code=422, detail=f"Item {ing.item_id} not found")
|
|
||||||
db_meal = Meal(name=meal.name, instructions=meal.instructions)
|
|
||||||
db.add(db_meal)
|
|
||||||
db.flush()
|
|
||||||
for ing in meal.ingredients:
|
|
||||||
db.add(MealIngredient(meal_id=db_meal.id, item_id=ing.item_id, quantity_needed=ing.quantity_needed))
|
|
||||||
db.commit()
|
|
||||||
db_meal = db.scalar(
|
|
||||||
select(Meal)
|
|
||||||
.where(Meal.id == db_meal.id)
|
|
||||||
.options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))
|
|
||||||
)
|
|
||||||
return MealResponse.from_meal(db_meal)
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/meals", response_model=list[MealResponse])
|
|
||||||
def list_meals(db: DbSession) -> list[MealResponse]:
|
|
||||||
"""List all meals with ingredients."""
|
|
||||||
meals = list(
|
|
||||||
db.scalars(
|
|
||||||
select(Meal).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item)).order_by(Meal.name)
|
|
||||||
).all()
|
|
||||||
)
|
|
||||||
return [MealResponse.from_meal(m) for m in meals]
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/meals/availability", response_model=list[MealAvailability])
|
|
||||||
def check_all_meals(db: DbSession) -> list[MealAvailability]:
|
|
||||||
"""Check which meals can be made with current inventory."""
|
|
||||||
meals = list(
|
|
||||||
db.scalars(select(Meal).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))).all()
|
|
||||||
)
|
|
||||||
return [_check_meal(m) for m in meals]
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/meals/{meal_id}", response_model=MealResponse)
|
|
||||||
def get_meal(meal_id: int, db: DbSession) -> MealResponse:
|
|
||||||
"""Get a meal by ID with ingredients."""
|
|
||||||
meal = db.scalar(
|
|
||||||
select(Meal).where(Meal.id == meal_id).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))
|
|
||||||
)
|
|
||||||
if not meal:
|
|
||||||
raise HTTPException(status_code=404, detail="Meal not found")
|
|
||||||
return MealResponse.from_meal(meal)
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/meals/{meal_id}")
|
|
||||||
def delete_meal(meal_id: int, db: DbSession) -> dict[str, bool]:
|
|
||||||
"""Delete a meal by ID."""
|
|
||||||
meal = db.get(Meal, meal_id)
|
|
||||||
if not meal:
|
|
||||||
raise HTTPException(status_code=404, detail="Meal not found")
|
|
||||||
db.delete(meal)
|
|
||||||
db.commit()
|
|
||||||
return {"deleted": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/meals/{meal_id}/ingredients", response_model=MealResponse)
|
|
||||||
def add_ingredient(meal_id: int, ingredient: IngredientCreate, db: DbSession) -> MealResponse:
|
|
||||||
"""Add an ingredient to a meal."""
|
|
||||||
meal = db.get(Meal, meal_id)
|
|
||||||
if not meal:
|
|
||||||
raise HTTPException(status_code=404, detail="Meal not found")
|
|
||||||
if not db.get(Item, ingredient.item_id):
|
|
||||||
raise HTTPException(status_code=422, detail="Item not found")
|
|
||||||
existing = db.scalar(
|
|
||||||
select(MealIngredient).where(MealIngredient.meal_id == meal_id, MealIngredient.item_id == ingredient.item_id)
|
|
||||||
)
|
|
||||||
if existing:
|
|
||||||
raise HTTPException(status_code=409, detail="Ingredient already exists for this meal")
|
|
||||||
db.add(MealIngredient(meal_id=meal_id, item_id=ingredient.item_id, quantity_needed=ingredient.quantity_needed))
|
|
||||||
db.commit()
|
|
||||||
meal = db.scalar(
|
|
||||||
select(Meal).where(Meal.id == meal_id).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))
|
|
||||||
)
|
|
||||||
return MealResponse.from_meal(meal)
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/meals/{meal_id}/ingredients/{item_id}")
|
|
||||||
def remove_ingredient(meal_id: int, item_id: int, db: DbSession) -> dict[str, bool]:
|
|
||||||
"""Remove an ingredient from a meal."""
|
|
||||||
mi = db.scalar(select(MealIngredient).where(MealIngredient.meal_id == meal_id, MealIngredient.item_id == item_id))
|
|
||||||
if not mi:
|
|
||||||
raise HTTPException(status_code=404, detail="Ingredient not found")
|
|
||||||
db.delete(mi)
|
|
||||||
db.commit()
|
|
||||||
return {"deleted": True}
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/meals/{meal_id}/availability", response_model=MealAvailability)
|
|
||||||
def check_meal(meal_id: int, db: DbSession) -> MealAvailability:
|
|
||||||
"""Check if a specific meal can be made and what's missing."""
|
|
||||||
meal = db.scalar(
|
|
||||||
select(Meal).where(Meal.id == meal_id).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))
|
|
||||||
)
|
|
||||||
if not meal:
|
|
||||||
raise HTTPException(status_code=404, detail="Meal not found")
|
|
||||||
return _check_meal(meal)
|
|
||||||
|
|
||||||
|
|
||||||
def _check_meal(meal: Meal) -> MealAvailability:
|
|
||||||
missing = [
|
|
||||||
ShoppingItem(
|
|
||||||
item_name=mi.item.name,
|
|
||||||
unit=mi.item.unit,
|
|
||||||
needed=mi.quantity_needed,
|
|
||||||
have=mi.item.quantity,
|
|
||||||
short=mi.quantity_needed - mi.item.quantity,
|
|
||||||
)
|
|
||||||
for mi in meal.ingredients
|
|
||||||
if mi.item.quantity < mi.quantity_needed
|
|
||||||
]
|
|
||||||
return MealAvailability(
|
|
||||||
meal_id=meal.id,
|
|
||||||
meal_name=meal.name,
|
|
||||||
can_make=len(missing) == 0,
|
|
||||||
missing=missing,
|
|
||||||
)
|
|
||||||
@@ -1,198 +0,0 @@
|
|||||||
"""HTMX frontend routes for van inventory."""
|
|
||||||
|
|
||||||
from __future__ import annotations
|
|
||||||
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import Annotated
|
|
||||||
|
|
||||||
from fastapi import APIRouter, Form, HTTPException, Request
|
|
||||||
from fastapi.responses import HTMLResponse
|
|
||||||
from fastapi.templating import Jinja2Templates
|
|
||||||
from sqlalchemy import select
|
|
||||||
from sqlalchemy.orm import selectinload
|
|
||||||
|
|
||||||
from python.orm.van_inventory.models import Item, Meal, MealIngredient
|
|
||||||
|
|
||||||
# FastAPI needs DbSession at runtime to resolve the Depends() annotation
|
|
||||||
from python.van_inventory.dependencies import DbSession # noqa: TC001
|
|
||||||
from python.van_inventory.routers.api import _check_meal
|
|
||||||
|
|
||||||
TEMPLATE_DIR = Path(__file__).resolve().parent.parent / "templates"
|
|
||||||
templates = Jinja2Templates(directory=TEMPLATE_DIR)
|
|
||||||
|
|
||||||
router = APIRouter(tags=["frontend"])
|
|
||||||
|
|
||||||
|
|
||||||
# --- Items ---
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/", response_class=HTMLResponse)
|
|
||||||
def items_page(request: Request, db: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the inventory page."""
|
|
||||||
items = list(db.scalars(select(Item).order_by(Item.name)).all())
|
|
||||||
return templates.TemplateResponse(request, "items.html", {"items": items})
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/items", response_class=HTMLResponse)
|
|
||||||
def htmx_create_item(
|
|
||||||
request: Request,
|
|
||||||
db: DbSession,
|
|
||||||
name: Annotated[str, Form()],
|
|
||||||
quantity: Annotated[float, Form()] = 0,
|
|
||||||
unit: Annotated[str, Form()] = "",
|
|
||||||
category: Annotated[str | None, Form()] = None,
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Create an item and return updated item rows."""
|
|
||||||
if quantity < 0:
|
|
||||||
raise HTTPException(status_code=422, detail="Quantity must not be negative")
|
|
||||||
db.add(Item(name=name, quantity=quantity, unit=unit, category=category or None))
|
|
||||||
db.commit()
|
|
||||||
items = list(db.scalars(select(Item).order_by(Item.name)).all())
|
|
||||||
return templates.TemplateResponse(request, "partials/item_rows.html", {"items": items})
|
|
||||||
|
|
||||||
|
|
||||||
@router.patch("/items/{item_id}", response_class=HTMLResponse)
|
|
||||||
def htmx_update_item(
|
|
||||||
request: Request,
|
|
||||||
item_id: int,
|
|
||||||
db: DbSession,
|
|
||||||
quantity: Annotated[float, Form()],
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Update an item's quantity and return updated item rows."""
|
|
||||||
if quantity < 0:
|
|
||||||
raise HTTPException(status_code=422, detail="Quantity must not be negative")
|
|
||||||
item = db.get(Item, item_id)
|
|
||||||
if item:
|
|
||||||
item.quantity = quantity
|
|
||||||
db.commit()
|
|
||||||
items = list(db.scalars(select(Item).order_by(Item.name)).all())
|
|
||||||
return templates.TemplateResponse(request, "partials/item_rows.html", {"items": items})
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/items/{item_id}", response_class=HTMLResponse)
|
|
||||||
def htmx_delete_item(request: Request, item_id: int, db: DbSession) -> HTMLResponse:
|
|
||||||
"""Delete an item and return updated item rows."""
|
|
||||||
item = db.get(Item, item_id)
|
|
||||||
if item:
|
|
||||||
db.delete(item)
|
|
||||||
db.commit()
|
|
||||||
items = list(db.scalars(select(Item).order_by(Item.name)).all())
|
|
||||||
return templates.TemplateResponse(request, "partials/item_rows.html", {"items": items})
|
|
||||||
|
|
||||||
|
|
||||||
# --- Meals ---
|
|
||||||
|
|
||||||
|
|
||||||
def _load_meals(db: DbSession) -> list[Meal]:
|
|
||||||
return list(
|
|
||||||
db.scalars(
|
|
||||||
select(Meal).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item)).order_by(Meal.name)
|
|
||||||
).all()
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/meals", response_class=HTMLResponse)
|
|
||||||
def meals_page(request: Request, db: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the meals page."""
|
|
||||||
meals = _load_meals(db)
|
|
||||||
return templates.TemplateResponse(request, "meals.html", {"meals": meals})
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/meals", response_class=HTMLResponse)
|
|
||||||
def htmx_create_meal(
|
|
||||||
request: Request,
|
|
||||||
db: DbSession,
|
|
||||||
name: Annotated[str, Form()],
|
|
||||||
instructions: Annotated[str | None, Form()] = None,
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Create a meal and return updated meal rows."""
|
|
||||||
db.add(Meal(name=name, instructions=instructions or None))
|
|
||||||
db.commit()
|
|
||||||
meals = _load_meals(db)
|
|
||||||
return templates.TemplateResponse(request, "partials/meal_rows.html", {"meals": meals})
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/meals/{meal_id}", response_class=HTMLResponse)
|
|
||||||
def htmx_delete_meal(request: Request, meal_id: int, db: DbSession) -> HTMLResponse:
|
|
||||||
"""Delete a meal and return updated meal rows."""
|
|
||||||
meal = db.get(Meal, meal_id)
|
|
||||||
if meal:
|
|
||||||
db.delete(meal)
|
|
||||||
db.commit()
|
|
||||||
meals = _load_meals(db)
|
|
||||||
return templates.TemplateResponse(request, "partials/meal_rows.html", {"meals": meals})
|
|
||||||
|
|
||||||
|
|
||||||
# --- Meal detail ---
|
|
||||||
|
|
||||||
|
|
||||||
def _load_meal(db: DbSession, meal_id: int) -> Meal | None:
|
|
||||||
return db.scalar(
|
|
||||||
select(Meal).where(Meal.id == meal_id).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/meals/{meal_id}", response_class=HTMLResponse)
|
|
||||||
def meal_detail_page(request: Request, meal_id: int, db: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the meal detail page."""
|
|
||||||
meal = _load_meal(db, meal_id)
|
|
||||||
if not meal:
|
|
||||||
raise HTTPException(status_code=404, detail="Meal not found")
|
|
||||||
items = list(db.scalars(select(Item).order_by(Item.name)).all())
|
|
||||||
return templates.TemplateResponse(request, "meal_detail.html", {"meal": meal, "items": items})
|
|
||||||
|
|
||||||
|
|
||||||
@router.post("/meals/{meal_id}/ingredients", response_class=HTMLResponse)
|
|
||||||
def htmx_add_ingredient(
|
|
||||||
request: Request,
|
|
||||||
meal_id: int,
|
|
||||||
db: DbSession,
|
|
||||||
item_id: Annotated[int, Form()],
|
|
||||||
quantity_needed: Annotated[float, Form()],
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Add an ingredient to a meal and return updated ingredient rows."""
|
|
||||||
if quantity_needed <= 0:
|
|
||||||
raise HTTPException(status_code=422, detail="Quantity must be positive")
|
|
||||||
meal = db.get(Meal, meal_id)
|
|
||||||
if not meal:
|
|
||||||
raise HTTPException(status_code=404, detail="Meal not found")
|
|
||||||
if not db.get(Item, item_id):
|
|
||||||
raise HTTPException(status_code=422, detail="Item not found")
|
|
||||||
existing = db.scalar(
|
|
||||||
select(MealIngredient).where(MealIngredient.meal_id == meal_id, MealIngredient.item_id == item_id)
|
|
||||||
)
|
|
||||||
if existing:
|
|
||||||
raise HTTPException(status_code=409, detail="Ingredient already exists for this meal")
|
|
||||||
db.add(MealIngredient(meal_id=meal_id, item_id=item_id, quantity_needed=quantity_needed))
|
|
||||||
db.commit()
|
|
||||||
meal = _load_meal(db, meal_id)
|
|
||||||
return templates.TemplateResponse(request, "partials/ingredient_rows.html", {"meal": meal})
|
|
||||||
|
|
||||||
|
|
||||||
@router.delete("/meals/{meal_id}/ingredients/{item_id}", response_class=HTMLResponse)
|
|
||||||
def htmx_remove_ingredient(
|
|
||||||
request: Request,
|
|
||||||
meal_id: int,
|
|
||||||
item_id: int,
|
|
||||||
db: DbSession,
|
|
||||||
) -> HTMLResponse:
|
|
||||||
"""Remove an ingredient from a meal and return updated ingredient rows."""
|
|
||||||
mi = db.scalar(select(MealIngredient).where(MealIngredient.meal_id == meal_id, MealIngredient.item_id == item_id))
|
|
||||||
if mi:
|
|
||||||
db.delete(mi)
|
|
||||||
db.commit()
|
|
||||||
meal = _load_meal(db, meal_id)
|
|
||||||
return templates.TemplateResponse(request, "partials/ingredient_rows.html", {"meal": meal})
|
|
||||||
|
|
||||||
|
|
||||||
# --- Availability ---
|
|
||||||
|
|
||||||
|
|
||||||
@router.get("/availability", response_class=HTMLResponse)
|
|
||||||
def availability_page(request: Request, db: DbSession) -> HTMLResponse:
|
|
||||||
"""Render the meal availability page."""
|
|
||||||
meals = list(
|
|
||||||
db.scalars(select(Meal).options(selectinload(Meal.ingredients).selectinload(MealIngredient.item))).all()
|
|
||||||
)
|
|
||||||
availability = [_check_meal(m) for m in meals]
|
|
||||||
return templates.TemplateResponse(request, "availability.html", {"availability": availability})
|
|
||||||
@@ -1,212 +0,0 @@
|
|||||||
:root {
|
|
||||||
--neon-pink: #ff2a6d;
|
|
||||||
--neon-cyan: #05d9e8;
|
|
||||||
--neon-yellow: #f9f002;
|
|
||||||
--neon-purple: #d300c5;
|
|
||||||
--bg-dark: #0a0a0f;
|
|
||||||
--bg-panel: #0d0d1a;
|
|
||||||
--bg-input: #111128;
|
|
||||||
--border: #1a1a3e;
|
|
||||||
--text: #c0c0d0;
|
|
||||||
--text-dim: #8e8ea0;
|
|
||||||
}
|
|
||||||
|
|
||||||
* { box-sizing: border-box; margin: 0; padding: 0; }
|
|
||||||
|
|
||||||
body {
|
|
||||||
font-family: 'Share Tech Mono', monospace;
|
|
||||||
max-width: 900px;
|
|
||||||
margin: 0 auto;
|
|
||||||
padding: 1rem;
|
|
||||||
background: var(--bg-dark);
|
|
||||||
color: var(--text);
|
|
||||||
position: relative;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* Scanline overlay */
|
|
||||||
body::before {
|
|
||||||
content: '';
|
|
||||||
position: fixed;
|
|
||||||
top: 0; left: 0; right: 0; bottom: 0;
|
|
||||||
background: repeating-linear-gradient(
|
|
||||||
0deg,
|
|
||||||
transparent,
|
|
||||||
transparent 2px,
|
|
||||||
rgba(0, 0, 0, 0.08) 2px,
|
|
||||||
rgba(0, 0, 0, 0.08) 4px
|
|
||||||
);
|
|
||||||
pointer-events: none;
|
|
||||||
z-index: 9999;
|
|
||||||
}
|
|
||||||
|
|
||||||
h1, h2, h3 {
|
|
||||||
font-family: 'Orbitron', sans-serif;
|
|
||||||
margin-bottom: 0.5rem;
|
|
||||||
color: var(--neon-cyan);
|
|
||||||
text-shadow: 0 0 10px rgba(5, 217, 232, 0.5), 0 0 40px rgba(5, 217, 232, 0.2);
|
|
||||||
text-transform: uppercase;
|
|
||||||
letter-spacing: 2px;
|
|
||||||
}
|
|
||||||
|
|
||||||
a { color: var(--neon-pink); text-decoration: none; transition: all 0.2s; }
|
|
||||||
a:hover {
|
|
||||||
text-shadow: 0 0 8px rgba(255, 42, 109, 0.8), 0 0 20px rgba(255, 42, 109, 0.4);
|
|
||||||
}
|
|
||||||
|
|
||||||
nav {
|
|
||||||
display: flex;
|
|
||||||
gap: 1.5rem;
|
|
||||||
padding: 1rem 0;
|
|
||||||
border-bottom: 1px solid var(--border);
|
|
||||||
margin-bottom: 1.5rem;
|
|
||||||
position: relative;
|
|
||||||
}
|
|
||||||
nav::after {
|
|
||||||
content: '';
|
|
||||||
position: absolute;
|
|
||||||
bottom: -1px;
|
|
||||||
left: 0;
|
|
||||||
right: 0;
|
|
||||||
height: 1px;
|
|
||||||
background: linear-gradient(90deg, var(--neon-pink), var(--neon-cyan), var(--neon-purple));
|
|
||||||
opacity: 0.6;
|
|
||||||
}
|
|
||||||
nav a {
|
|
||||||
font-family: 'Orbitron', sans-serif;
|
|
||||||
font-weight: 700;
|
|
||||||
font-size: 0.85rem;
|
|
||||||
letter-spacing: 1px;
|
|
||||||
text-transform: uppercase;
|
|
||||||
padding: 0.3rem 0;
|
|
||||||
border-bottom: 2px solid transparent;
|
|
||||||
transition: all 0.2s;
|
|
||||||
}
|
|
||||||
nav a:hover {
|
|
||||||
border-bottom-color: var(--neon-pink);
|
|
||||||
text-shadow: 0 0 8px rgba(255, 42, 109, 0.8);
|
|
||||||
}
|
|
||||||
|
|
||||||
table {
|
|
||||||
width: 100%;
|
|
||||||
border-collapse: collapse;
|
|
||||||
margin: 1rem 0;
|
|
||||||
border: 1px solid var(--border);
|
|
||||||
}
|
|
||||||
th, td {
|
|
||||||
text-align: left;
|
|
||||||
padding: 0.6rem 0.75rem;
|
|
||||||
border-bottom: 1px solid var(--border);
|
|
||||||
}
|
|
||||||
th {
|
|
||||||
font-family: 'Orbitron', sans-serif;
|
|
||||||
color: var(--neon-cyan);
|
|
||||||
font-size: 0.7rem;
|
|
||||||
text-transform: uppercase;
|
|
||||||
letter-spacing: 2px;
|
|
||||||
background: var(--bg-panel);
|
|
||||||
border-bottom: 1px solid var(--neon-cyan);
|
|
||||||
text-shadow: 0 0 6px rgba(5, 217, 232, 0.3);
|
|
||||||
}
|
|
||||||
tr:hover td {
|
|
||||||
background: rgba(5, 217, 232, 0.03);
|
|
||||||
}
|
|
||||||
|
|
||||||
form {
|
|
||||||
display: flex;
|
|
||||||
flex-wrap: wrap;
|
|
||||||
gap: 0.5rem;
|
|
||||||
align-items: end;
|
|
||||||
margin: 1rem 0;
|
|
||||||
padding: 1rem;
|
|
||||||
border: 1px solid var(--border);
|
|
||||||
background: var(--bg-panel);
|
|
||||||
}
|
|
||||||
|
|
||||||
input, select {
|
|
||||||
padding: 0.5rem 0.6rem;
|
|
||||||
border: 1px solid var(--border);
|
|
||||||
border-radius: 2px;
|
|
||||||
background: var(--bg-input);
|
|
||||||
color: var(--neon-cyan);
|
|
||||||
font-family: 'Share Tech Mono', monospace;
|
|
||||||
transition: all 0.2s;
|
|
||||||
}
|
|
||||||
input:focus, select:focus {
|
|
||||||
outline: none;
|
|
||||||
border-color: var(--neon-cyan);
|
|
||||||
box-shadow: 0 0 8px rgba(5, 217, 232, 0.3), inset 0 0 8px rgba(5, 217, 232, 0.05);
|
|
||||||
}
|
|
||||||
|
|
||||||
button {
|
|
||||||
padding: 0.5rem 1.2rem;
|
|
||||||
border: 1px solid var(--neon-pink);
|
|
||||||
border-radius: 2px;
|
|
||||||
background: transparent;
|
|
||||||
color: var(--neon-pink);
|
|
||||||
cursor: pointer;
|
|
||||||
font-family: 'Orbitron', sans-serif;
|
|
||||||
font-weight: 700;
|
|
||||||
font-size: 0.7rem;
|
|
||||||
letter-spacing: 1px;
|
|
||||||
text-transform: uppercase;
|
|
||||||
transition: all 0.2s;
|
|
||||||
}
|
|
||||||
button:hover {
|
|
||||||
background: var(--neon-pink);
|
|
||||||
color: var(--bg-dark);
|
|
||||||
box-shadow: 0 0 15px rgba(255, 42, 109, 0.5), 0 0 30px rgba(255, 42, 109, 0.2);
|
|
||||||
}
|
|
||||||
button.danger {
|
|
||||||
border-color: var(--text-dim);
|
|
||||||
color: var(--text-dim);
|
|
||||||
}
|
|
||||||
button.danger:hover {
|
|
||||||
border-color: var(--neon-pink);
|
|
||||||
background: var(--neon-pink);
|
|
||||||
color: var(--bg-dark);
|
|
||||||
box-shadow: 0 0 15px rgba(255, 42, 109, 0.5);
|
|
||||||
}
|
|
||||||
|
|
||||||
.badge {
|
|
||||||
display: inline-block;
|
|
||||||
padding: 0.2rem 0.6rem;
|
|
||||||
border-radius: 2px;
|
|
||||||
font-family: 'Orbitron', sans-serif;
|
|
||||||
font-size: 0.65rem;
|
|
||||||
font-weight: 700;
|
|
||||||
letter-spacing: 1px;
|
|
||||||
text-transform: uppercase;
|
|
||||||
}
|
|
||||||
.badge.yes {
|
|
||||||
background: rgba(5, 217, 232, 0.1);
|
|
||||||
color: var(--neon-cyan);
|
|
||||||
border: 1px solid var(--neon-cyan);
|
|
||||||
text-shadow: 0 0 6px rgba(5, 217, 232, 0.5);
|
|
||||||
}
|
|
||||||
.badge.no {
|
|
||||||
background: rgba(255, 42, 109, 0.1);
|
|
||||||
color: var(--neon-pink);
|
|
||||||
border: 1px solid var(--neon-pink);
|
|
||||||
text-shadow: 0 0 6px rgba(255, 42, 109, 0.5);
|
|
||||||
}
|
|
||||||
|
|
||||||
.missing-list { font-size: 0.85rem; color: var(--text-dim); }
|
|
||||||
|
|
||||||
label {
|
|
||||||
font-size: 0.75rem;
|
|
||||||
color: var(--text-dim);
|
|
||||||
display: flex;
|
|
||||||
flex-direction: column;
|
|
||||||
gap: 0.2rem;
|
|
||||||
text-transform: uppercase;
|
|
||||||
letter-spacing: 1px;
|
|
||||||
}
|
|
||||||
|
|
||||||
.flash {
|
|
||||||
padding: 0.5rem 1rem;
|
|
||||||
margin: 0.5rem 0;
|
|
||||||
border-radius: 2px;
|
|
||||||
background: rgba(5, 217, 232, 0.1);
|
|
||||||
color: var(--neon-cyan);
|
|
||||||
border: 1px solid var(--neon-cyan);
|
|
||||||
}
|
|
||||||
@@ -1,30 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}What Can I Make? - Van{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<h1>What Can I Make?</h1>
|
|
||||||
|
|
||||||
<table>
|
|
||||||
<thead>
|
|
||||||
<tr><th>Meal</th><th>Status</th><th>Missing</th></tr>
|
|
||||||
</thead>
|
|
||||||
<tbody>
|
|
||||||
{% for meal in availability %}
|
|
||||||
<tr>
|
|
||||||
<td><a href="/meals/{{ meal.meal_id }}">{{ meal.meal_name }}</a></td>
|
|
||||||
<td>
|
|
||||||
{% if meal.can_make %}
|
|
||||||
<span class="badge yes">Ready</span>
|
|
||||||
{% else %}
|
|
||||||
<span class="badge no">Missing items</span>
|
|
||||||
{% endif %}
|
|
||||||
</td>
|
|
||||||
<td class="missing-list">
|
|
||||||
{% for m in meal.missing %}
|
|
||||||
{{ m.item_name }}: need {{ m.short }} more {{ m.unit }}{% if not loop.last %}, {% endif %}
|
|
||||||
{% endfor %}
|
|
||||||
</td>
|
|
||||||
</tr>
|
|
||||||
{% endfor %}
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,20 +0,0 @@
|
|||||||
<!DOCTYPE html>
|
|
||||||
<html lang="en">
|
|
||||||
<head>
|
|
||||||
<meta charset="UTF-8">
|
|
||||||
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
|
||||||
<title>{% block title %}Van Inventory{% endblock %}</title>
|
|
||||||
<script src="https://unpkg.com/htmx.org@2.0.4"></script>
|
|
||||||
<link rel="preconnect" href="https://fonts.googleapis.com">
|
|
||||||
<link href="https://fonts.googleapis.com/css2?family=Orbitron:wght@400;700;900&family=Share+Tech+Mono&display=swap" rel="stylesheet">
|
|
||||||
<link rel="stylesheet" href="/static/style.css">
|
|
||||||
</head>
|
|
||||||
<body>
|
|
||||||
<nav>
|
|
||||||
<a href="/">Inventory</a>
|
|
||||||
<a href="/meals">Meals</a>
|
|
||||||
<a href="/availability">What Can I Make?</a>
|
|
||||||
</nav>
|
|
||||||
{% block content %}{% endblock %}
|
|
||||||
</body>
|
|
||||||
</html>
|
|
||||||
@@ -1,17 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}Inventory - Van{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<h1>Van Inventory</h1>
|
|
||||||
|
|
||||||
<form hx-post="/items" hx-target="#item-list" hx-swap="innerHTML" hx-on::after-request="if(event.detail.successful) this.reset()">
|
|
||||||
<label>Name <input type="text" name="name" required></label>
|
|
||||||
<label>Qty <input type="number" name="quantity" step="any" value="0" min="0" required></label>
|
|
||||||
<label>Unit <input type="text" name="unit" required placeholder="lbs, cans, etc"></label>
|
|
||||||
<label>Category <input type="text" name="category" placeholder="optional"></label>
|
|
||||||
<button type="submit">Add Item</button>
|
|
||||||
</form>
|
|
||||||
|
|
||||||
<div id="item-list">
|
|
||||||
{% include "partials/item_rows.html" %}
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,24 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}{{ meal.name }} - Van{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<h1>{{ meal.name }}</h1>
|
|
||||||
{% if meal.instructions %}<p>{{ meal.instructions }}</p>{% endif %}
|
|
||||||
|
|
||||||
<h2>Ingredients</h2>
|
|
||||||
<form hx-post="/meals/{{ meal.id }}/ingredients" hx-target="#ingredient-list" hx-swap="innerHTML" hx-on::after-request="if(event.detail.successful) this.reset()">
|
|
||||||
<label>Item
|
|
||||||
<select name="item_id" required>
|
|
||||||
<option value="">--</option>
|
|
||||||
{% for item in items %}
|
|
||||||
<option value="{{ item.id }}">{{ item.name }} ({{ item.unit }})</option>
|
|
||||||
{% endfor %}
|
|
||||||
</select>
|
|
||||||
</label>
|
|
||||||
<label>Qty needed <input type="number" name="quantity_needed" step="any" min="0.01" required></label>
|
|
||||||
<button type="submit">Add</button>
|
|
||||||
</form>
|
|
||||||
|
|
||||||
<div id="ingredient-list">
|
|
||||||
{% include "partials/ingredient_rows.html" %}
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,15 +0,0 @@
|
|||||||
{% extends "base.html" %}
|
|
||||||
{% block title %}Meals - Van{% endblock %}
|
|
||||||
{% block content %}
|
|
||||||
<h1>Meals</h1>
|
|
||||||
|
|
||||||
<form hx-post="/meals" hx-target="#meal-list" hx-swap="innerHTML" hx-on::after-request="if(event.detail.successful) this.reset()">
|
|
||||||
<label>Name <input type="text" name="name" required></label>
|
|
||||||
<label>Instructions <input type="text" name="instructions" placeholder="optional"></label>
|
|
||||||
<button type="submit">Add Meal</button>
|
|
||||||
</form>
|
|
||||||
|
|
||||||
<div id="meal-list">
|
|
||||||
{% include "partials/meal_rows.html" %}
|
|
||||||
</div>
|
|
||||||
{% endblock %}
|
|
||||||
@@ -1,16 +0,0 @@
|
|||||||
<table>
|
|
||||||
<thead>
|
|
||||||
<tr><th>Item</th><th>Needed</th><th>Have</th><th>Unit</th><th></th></tr>
|
|
||||||
</thead>
|
|
||||||
<tbody>
|
|
||||||
{% for mi in meal.ingredients %}
|
|
||||||
<tr>
|
|
||||||
<td>{{ mi.item.name }}</td>
|
|
||||||
<td>{{ mi.quantity_needed }}</td>
|
|
||||||
<td>{{ mi.item.quantity }}</td>
|
|
||||||
<td>{{ mi.item.unit }}</td>
|
|
||||||
<td><button class="danger" hx-delete="/meals/{{ meal.id }}/ingredients/{{ mi.item_id }}" hx-target="#ingredient-list" hx-swap="innerHTML" hx-confirm="Remove {{ mi.item.name }}?">X</button></td>
|
|
||||||
</tr>
|
|
||||||
{% endfor %}
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
@@ -1,21 +0,0 @@
|
|||||||
<table>
|
|
||||||
<thead>
|
|
||||||
<tr><th>Name</th><th>Qty</th><th>Unit</th><th>Category</th><th></th></tr>
|
|
||||||
</thead>
|
|
||||||
<tbody>
|
|
||||||
{% for item in items %}
|
|
||||||
<tr>
|
|
||||||
<td>{{ item.name }}</td>
|
|
||||||
<td>
|
|
||||||
<form hx-patch="/items/{{ item.id }}" hx-target="#item-list" hx-swap="innerHTML" style="display:inline; margin:0;">
|
|
||||||
<input type="number" name="quantity" value="{{ item.quantity }}" step="any" min="0" style="width:5rem">
|
|
||||||
<button type="submit" style="padding:0.2rem 0.5rem; font-size:0.8rem;">Update</button>
|
|
||||||
</form>
|
|
||||||
</td>
|
|
||||||
<td>{{ item.unit }}</td>
|
|
||||||
<td>{{ item.category or "" }}</td>
|
|
||||||
<td><button class="danger" hx-delete="/items/{{ item.id }}" hx-target="#item-list" hx-swap="innerHTML" hx-confirm="Delete {{ item.name }}?">X</button></td>
|
|
||||||
</tr>
|
|
||||||
{% endfor %}
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
@@ -1,15 +0,0 @@
|
|||||||
<table>
|
|
||||||
<thead>
|
|
||||||
<tr><th>Name</th><th>Ingredients</th><th>Instructions</th><th></th></tr>
|
|
||||||
</thead>
|
|
||||||
<tbody>
|
|
||||||
{% for meal in meals %}
|
|
||||||
<tr>
|
|
||||||
<td><a href="/meals/{{ meal.id }}">{{ meal.name }}</a></td>
|
|
||||||
<td>{{ meal.ingredients | length }}</td>
|
|
||||||
<td>{{ (meal.instructions or "")[:50] }}</td>
|
|
||||||
<td><button class="danger" hx-delete="/meals/{{ meal.id }}" hx-target="#meal-list" hx-swap="innerHTML" hx-confirm="Delete {{ meal.name }}?">X</button></td>
|
|
||||||
</tr>
|
|
||||||
{% endfor %}
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
@@ -1 +0,0 @@
|
|||||||
"""Van weather service - fetches weather with masked GPS location."""
|
|
||||||
@@ -1,293 +0,0 @@
|
|||||||
"""Van weather service - fetches weather with masked GPS for privacy."""
|
|
||||||
|
|
||||||
import logging
|
|
||||||
from datetime import UTC, datetime
|
|
||||||
from typing import Annotated, Any
|
|
||||||
|
|
||||||
import httpx
|
|
||||||
import typer
|
|
||||||
from apscheduler.schedulers.blocking import BlockingScheduler
|
|
||||||
from tenacity import before_sleep_log, retry, stop_after_attempt, wait_fixed
|
|
||||||
|
|
||||||
from python.common import configure_logger
|
|
||||||
from python.van_weather.models import Config, DailyForecast, HourlyForecast, Weather
|
|
||||||
|
|
||||||
# Map Pirate Weather icons to Home Assistant conditions
|
|
||||||
CONDITION_MAP = {
|
|
||||||
"clear-day": "sunny",
|
|
||||||
"clear-night": "clear-night",
|
|
||||||
"rain": "rainy",
|
|
||||||
"snow": "snowy",
|
|
||||||
"sleet": "snowy-rainy",
|
|
||||||
"wind": "windy",
|
|
||||||
"fog": "fog",
|
|
||||||
"cloudy": "cloudy",
|
|
||||||
"partly-cloudy-day": "partlycloudy",
|
|
||||||
"partly-cloudy-night": "partlycloudy",
|
|
||||||
}
|
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
|
|
||||||
@retry(
|
|
||||||
stop=stop_after_attempt(3),
|
|
||||||
wait=wait_fixed(5),
|
|
||||||
before_sleep=before_sleep_log(logger, logging.WARNING),
|
|
||||||
reraise=True,
|
|
||||||
)
|
|
||||||
def get_ha_state(url: str, token: str, entity_id: str) -> float:
|
|
||||||
"""Get numeric state from Home Asasistant entity."""
|
|
||||||
response = httpx.get(
|
|
||||||
f"{url}/api/states/{entity_id}",
|
|
||||||
headers={"Authorization": f"Bearer {token}"},
|
|
||||||
timeout=30,
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
|
||||||
state = response.json()["state"]
|
|
||||||
if state in ("unavailable", "unknown"):
|
|
||||||
error = f"{entity_id} is {state}"
|
|
||||||
raise ValueError(error)
|
|
||||||
return float(state)
|
|
||||||
|
|
||||||
|
|
||||||
def parse_daily_forecast(data: dict[str, dict[str, Any]]) -> list[DailyForecast]:
|
|
||||||
"""Parse daily forecast from Pirate Weather API."""
|
|
||||||
daily = data.get("daily", {}).get("data", [])
|
|
||||||
daily_forecasts = []
|
|
||||||
for day in daily[:8]: # Up to 8 days
|
|
||||||
time_stamp = day.get("time")
|
|
||||||
if time_stamp:
|
|
||||||
date_time = datetime.fromtimestamp(time_stamp, tz=UTC).isoformat()
|
|
||||||
daily_forecasts.append(
|
|
||||||
DailyForecast(
|
|
||||||
date_time=date_time,
|
|
||||||
condition=CONDITION_MAP.get(day.get("icon", ""), "cloudy"),
|
|
||||||
temperature=day.get("temperatureHigh"),
|
|
||||||
templow=day.get("temperatureLow"),
|
|
||||||
precipitation_probability=day.get("precipProbability"),
|
|
||||||
moon_phase=day.get("moonPhase"),
|
|
||||||
wind_gust=day.get("windGust"),
|
|
||||||
cloud_cover=day.get("cloudCover"),
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
return daily_forecasts
|
|
||||||
|
|
||||||
|
|
||||||
def parse_hourly_forecast(data: dict[str, dict[str, Any]]) -> list[HourlyForecast]:
|
|
||||||
"""Parse hourly forecast from Pirate Weather API."""
|
|
||||||
hourly = data.get("hourly", {}).get("data", [])
|
|
||||||
hourly_forecasts = []
|
|
||||||
for hour in hourly[:48]: # Up to 48 hours
|
|
||||||
time_stamp = hour.get("time")
|
|
||||||
if time_stamp:
|
|
||||||
date_time = datetime.fromtimestamp(time_stamp, tz=UTC).isoformat()
|
|
||||||
hourly_forecasts.append(
|
|
||||||
HourlyForecast(
|
|
||||||
date_time=date_time,
|
|
||||||
condition=CONDITION_MAP.get(hour.get("icon", ""), "cloudy"),
|
|
||||||
temperature=hour.get("temperature"),
|
|
||||||
precipitation_probability=hour.get("precipProbability"),
|
|
||||||
)
|
|
||||||
)
|
|
||||||
return hourly_forecasts
|
|
||||||
|
|
||||||
|
|
||||||
@retry(
|
|
||||||
stop=stop_after_attempt(3),
|
|
||||||
wait=wait_fixed(5),
|
|
||||||
before_sleep=before_sleep_log(logger, logging.WARNING),
|
|
||||||
reraise=True,
|
|
||||||
)
|
|
||||||
def fetch_weather(api_key: str, lat: float, lon: float) -> Weather:
|
|
||||||
"""Fetch weather from Pirate Weather API."""
|
|
||||||
url = f"https://api.pirateweather.net/forecast/{api_key}/{lat},{lon}"
|
|
||||||
response = httpx.get(url, params={"units": "us"}, timeout=30)
|
|
||||||
response.raise_for_status()
|
|
||||||
data = response.json()
|
|
||||||
|
|
||||||
daily_forecasts = parse_daily_forecast(data)
|
|
||||||
hourly_forecasts = parse_hourly_forecast(data)
|
|
||||||
|
|
||||||
current = data.get("currently", {})
|
|
||||||
icon = current.get("icon", "")
|
|
||||||
return Weather(
|
|
||||||
temperature=current.get("temperature"),
|
|
||||||
feels_like=current.get("apparentTemperature"),
|
|
||||||
humidity=current.get("humidity"),
|
|
||||||
wind_speed=current.get("windSpeed"),
|
|
||||||
wind_bearing=current.get("windBearing"),
|
|
||||||
condition=CONDITION_MAP.get(icon, "cloudy"),
|
|
||||||
summary=current.get("summary"),
|
|
||||||
pressure=current.get("pressure"),
|
|
||||||
visibility=current.get("visibility"),
|
|
||||||
uv_index=current.get("uvIndex"),
|
|
||||||
ozone=current.get("ozone"),
|
|
||||||
nearest_storm_distance=current.get("nearestStormDistance"),
|
|
||||||
nearest_storm_bearing=current.get("nearestStormBearing"),
|
|
||||||
precip_probability=current.get("precipProbability"),
|
|
||||||
cloud_cover=current.get("cloudCover"),
|
|
||||||
daily_forecasts=daily_forecasts,
|
|
||||||
hourly_forecasts=hourly_forecasts,
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@retry(
|
|
||||||
stop=stop_after_attempt(3),
|
|
||||||
wait=wait_fixed(5),
|
|
||||||
before_sleep=before_sleep_log(logger, logging.WARNING),
|
|
||||||
reraise=True,
|
|
||||||
)
|
|
||||||
def post_to_ha(url: str, token: str, weather: Weather) -> None:
|
|
||||||
"""Post weather data to Home Assistant as sensor entities."""
|
|
||||||
headers = {"Authorization": f"Bearer {token}"}
|
|
||||||
|
|
||||||
# Post current weather as individual sensors
|
|
||||||
sensors = {
|
|
||||||
"sensor.van_weather_condition": {
|
|
||||||
"state": weather.condition or "unknown",
|
|
||||||
"attributes": {"friendly_name": "Van Weather Condition"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_temperature": {
|
|
||||||
"state": weather.temperature,
|
|
||||||
"attributes": {"unit_of_measurement": "°F", "device_class": "temperature"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_apparent_temperature": {
|
|
||||||
"state": weather.feels_like,
|
|
||||||
"attributes": {"unit_of_measurement": "°F", "device_class": "temperature"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_humidity": {
|
|
||||||
"state": int((weather.humidity or 0) * 100),
|
|
||||||
"attributes": {"unit_of_measurement": "%", "device_class": "humidity"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_pressure": {
|
|
||||||
"state": weather.pressure,
|
|
||||||
"attributes": {"unit_of_measurement": "mbar", "device_class": "pressure"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_wind_speed": {
|
|
||||||
"state": weather.wind_speed,
|
|
||||||
"attributes": {"unit_of_measurement": "mph", "device_class": "wind_speed"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_wind_bearing": {
|
|
||||||
"state": weather.wind_bearing,
|
|
||||||
"attributes": {"unit_of_measurement": "°"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_visibility": {
|
|
||||||
"state": weather.visibility,
|
|
||||||
"attributes": {"unit_of_measurement": "mi"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_uv_index": {
|
|
||||||
"state": weather.uv_index,
|
|
||||||
"attributes": {"friendly_name": "Van Weather UV Index", "icon": "mdi:sun-wireless"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_ozone": {
|
|
||||||
"state": weather.ozone,
|
|
||||||
"attributes": {"unit_of_measurement": "DU", "icon": "mdi:earth"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_nearest_storm_distance": {
|
|
||||||
"state": weather.nearest_storm_distance,
|
|
||||||
"attributes": {"unit_of_measurement": "mi", "icon": "mdi:weather-lightning"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_nearest_storm_bearing": {
|
|
||||||
"state": weather.nearest_storm_bearing,
|
|
||||||
"attributes": {"unit_of_measurement": "°", "icon": "mdi:weather-lightning"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_precip_probability": {
|
|
||||||
"state": int((weather.precip_probability or 0) * 100),
|
|
||||||
"attributes": {"unit_of_measurement": "%", "icon": "mdi:weather-rainy"},
|
|
||||||
},
|
|
||||||
"sensor.van_weather_cloud_cover": {
|
|
||||||
"state": int((weather.cloud_cover or 0) * 100),
|
|
||||||
"attributes": {"unit_of_measurement": "%", "icon": "mdi:weather-cloudy"},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
|
|
||||||
for entity_id, data in sensors.items():
|
|
||||||
if data["state"] is not None:
|
|
||||||
response = httpx.post(f"{url}/api/states/{entity_id}", headers=headers, json=data, timeout=30)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
# Post daily forecast as JSON attribute sensor
|
|
||||||
daily_forecast = [
|
|
||||||
{
|
|
||||||
"datetime": daily_forecast.date_time.isoformat(),
|
|
||||||
"condition": daily_forecast.condition,
|
|
||||||
"temperature": daily_forecast.temperature,
|
|
||||||
"templow": daily_forecast.templow,
|
|
||||||
"precipitation_probability": int((daily_forecast.precipitation_probability or 0) * 100),
|
|
||||||
}
|
|
||||||
for daily_forecast in weather.daily_forecasts
|
|
||||||
]
|
|
||||||
|
|
||||||
response = httpx.post(
|
|
||||||
f"{url}/api/states/sensor.van_weather_forecast_daily",
|
|
||||||
headers=headers,
|
|
||||||
json={"state": len(daily_forecast), "attributes": {"forecast": daily_forecast}},
|
|
||||||
timeout=30,
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
# Post hourly forecast as JSON attribute sensor
|
|
||||||
hourly_forecast = [
|
|
||||||
{
|
|
||||||
"datetime": hourly_forecast.date_time.isoformat(),
|
|
||||||
"condition": hourly_forecast.condition,
|
|
||||||
"temperature": hourly_forecast.temperature,
|
|
||||||
"precipitation_probability": int((hourly_forecast.precipitation_probability or 0) * 100),
|
|
||||||
}
|
|
||||||
for hourly_forecast in weather.hourly_forecasts
|
|
||||||
]
|
|
||||||
|
|
||||||
response = httpx.post(
|
|
||||||
f"{url}/api/states/sensor.van_weather_forecast_hourly",
|
|
||||||
headers=headers,
|
|
||||||
json={"state": len(hourly_forecast), "attributes": {"forecast": hourly_forecast}},
|
|
||||||
timeout=30,
|
|
||||||
)
|
|
||||||
response.raise_for_status()
|
|
||||||
|
|
||||||
|
|
||||||
def update_weather(config: Config) -> None:
|
|
||||||
"""Fetch weather using last-known location, post to HA."""
|
|
||||||
lat = get_ha_state(config.ha_url, config.ha_token, config.lat_entity)
|
|
||||||
lon = get_ha_state(config.ha_url, config.ha_token, config.lon_entity)
|
|
||||||
|
|
||||||
masked_lat = round(lat, config.mask_decimals)
|
|
||||||
masked_lon = round(lon, config.mask_decimals)
|
|
||||||
|
|
||||||
logger.info(f"Masked location: {masked_lat}, {masked_lon}")
|
|
||||||
|
|
||||||
weather = fetch_weather(config.pirate_weather_api_key, lat, lon)
|
|
||||||
logger.info(f"Weather: {weather.temperature}°F, {weather.condition}")
|
|
||||||
|
|
||||||
post_to_ha(config.ha_url, config.ha_token, weather)
|
|
||||||
logger.info("Posted weather to HA")
|
|
||||||
|
|
||||||
|
|
||||||
def main(
|
|
||||||
ha_url: Annotated[str, typer.Option(envvar="HA_URL")],
|
|
||||||
ha_token: Annotated[str, typer.Option(envvar="HA_TOKEN")],
|
|
||||||
api_key: Annotated[str, typer.Option(envvar="PIRATE_WEATHER_API_KEY")],
|
|
||||||
interval: Annotated[int, typer.Option(help="Poll interval in seconds")] = 900,
|
|
||||||
log_level: Annotated[str, typer.Option()] = "INFO",
|
|
||||||
) -> None:
|
|
||||||
"""Fetch weather for van using masked GPS location."""
|
|
||||||
configure_logger(log_level)
|
|
||||||
|
|
||||||
config = Config(ha_url=ha_url, ha_token=ha_token, pirate_weather_api_key=api_key)
|
|
||||||
|
|
||||||
logger.info(f"Starting van weather service, polling every {interval}s")
|
|
||||||
|
|
||||||
scheduler = BlockingScheduler()
|
|
||||||
scheduler.add_job(
|
|
||||||
update_weather,
|
|
||||||
"interval",
|
|
||||||
seconds=interval,
|
|
||||||
args=[config],
|
|
||||||
next_run_time=datetime.now(UTC),
|
|
||||||
)
|
|
||||||
scheduler.start()
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
typer.run(main)
|
|
||||||
@@ -1,70 +0,0 @@
|
|||||||
"""Models for van weather service."""
|
|
||||||
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
from pydantic import BaseModel, field_serializer
|
|
||||||
|
|
||||||
|
|
||||||
class Config(BaseModel):
|
|
||||||
"""Service configuration."""
|
|
||||||
|
|
||||||
ha_url: str
|
|
||||||
ha_token: str
|
|
||||||
pirate_weather_api_key: str
|
|
||||||
lat_entity: str = "sensor.van_last_known_latitude"
|
|
||||||
lon_entity: str = "sensor.van_last_known_longitude"
|
|
||||||
mask_decimals: int = 1 # ~11km accuracy
|
|
||||||
|
|
||||||
|
|
||||||
class DailyForecast(BaseModel):
|
|
||||||
"""Daily forecast entry."""
|
|
||||||
|
|
||||||
date_time: datetime
|
|
||||||
condition: str | None = None
|
|
||||||
temperature: float | None = None # High
|
|
||||||
templow: float | None = None # Low
|
|
||||||
precipitation_probability: float | None = None
|
|
||||||
moon_phase: float | None = None
|
|
||||||
wind_gust: float | None = None
|
|
||||||
cloud_cover: float | None = None
|
|
||||||
|
|
||||||
@field_serializer("date_time")
|
|
||||||
def serialize_date_time(self, date_time: datetime) -> str:
|
|
||||||
"""Serialize datetime to ISO format."""
|
|
||||||
return date_time.isoformat()
|
|
||||||
|
|
||||||
|
|
||||||
class HourlyForecast(BaseModel):
|
|
||||||
"""Hourly forecast entry."""
|
|
||||||
|
|
||||||
date_time: datetime
|
|
||||||
condition: str | None = None
|
|
||||||
temperature: float | None = None
|
|
||||||
precipitation_probability: float | None = None
|
|
||||||
|
|
||||||
@field_serializer("date_time")
|
|
||||||
def serialize_date_time(self, date_time: datetime) -> str:
|
|
||||||
"""Serialize datetime to ISO format."""
|
|
||||||
return date_time.isoformat()
|
|
||||||
|
|
||||||
|
|
||||||
class Weather(BaseModel):
|
|
||||||
"""Weather data from Pirate Weather."""
|
|
||||||
|
|
||||||
temperature: float | None = None
|
|
||||||
feels_like: float | None = None
|
|
||||||
humidity: float | None = None
|
|
||||||
wind_speed: float | None = None
|
|
||||||
wind_bearing: float | None = None
|
|
||||||
condition: str | None = None
|
|
||||||
summary: str | None = None
|
|
||||||
pressure: float | None = None
|
|
||||||
visibility: float | None = None
|
|
||||||
uv_index: float | None = None
|
|
||||||
ozone: float | None = None
|
|
||||||
nearest_storm_distance: float | None = None
|
|
||||||
nearest_storm_bearing: float | None = None
|
|
||||||
precip_probability: float | None = None
|
|
||||||
cloud_cover: float | None = None
|
|
||||||
daily_forecasts: list[DailyForecast] = []
|
|
||||||
hourly_forecasts: list[HourlyForecast] = []
|
|
||||||
@@ -2,6 +2,7 @@
|
|||||||
{
|
{
|
||||||
imports = [
|
imports = [
|
||||||
"${inputs.self}/users/richie"
|
"${inputs.self}/users/richie"
|
||||||
|
"${inputs.self}/users/gaming"
|
||||||
"${inputs.self}/common/global"
|
"${inputs.self}/common/global"
|
||||||
"${inputs.self}/common/optional/desktop.nix"
|
"${inputs.self}/common/optional/desktop.nix"
|
||||||
"${inputs.self}/common/optional/docker.nix"
|
"${inputs.self}/common/optional/docker.nix"
|
||||||
@@ -12,7 +13,6 @@
|
|||||||
"${inputs.self}/common/optional/update.nix"
|
"${inputs.self}/common/optional/update.nix"
|
||||||
"${inputs.self}/common/optional/yubikey.nix"
|
"${inputs.self}/common/optional/yubikey.nix"
|
||||||
"${inputs.self}/common/optional/zerotier.nix"
|
"${inputs.self}/common/optional/zerotier.nix"
|
||||||
"${inputs.self}/common/optional/brain_substituter.nix"
|
|
||||||
"${inputs.self}/common/optional/nvidia.nix"
|
"${inputs.self}/common/optional/nvidia.nix"
|
||||||
./hardware.nix
|
./hardware.nix
|
||||||
./syncthing.nix
|
./syncthing.nix
|
||||||
@@ -27,6 +27,15 @@
|
|||||||
};
|
};
|
||||||
|
|
||||||
services = {
|
services = {
|
||||||
|
displayManager = {
|
||||||
|
enable = true;
|
||||||
|
autoLogin = {
|
||||||
|
user = "gaming";
|
||||||
|
enable = true;
|
||||||
|
};
|
||||||
|
defaultSession = "plasma";
|
||||||
|
};
|
||||||
|
|
||||||
openssh.ports = [ 262 ];
|
openssh.ports = [ 262 ];
|
||||||
|
|
||||||
snapshot_manager.path = ./snapshot_config.toml;
|
snapshot_manager.path = ./snapshot_config.toml;
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user