10 Commits

Author SHA1 Message Date
Natan Keddem 36ee1f94cd added arbitrary property entry 2023-11-20 21:57:10 -05:00
Natan Keddem 07ce7e0bae fixed edge case for ErrorAggregator 2023-11-20 21:56:20 -05:00
Natan Keddem ffbc9b71c0 optimize date and time displays 2023-11-19 17:05:58 -05:00
Natan Keddem f3ef97a342 improved result timestamp 2023-11-18 21:46:56 -05:00
Natan Keddem 94fba0b925 added test to http pipe 2023-11-18 20:46:06 -05:00
Natan Keddem 8572ad766b improved browse and find 2023-11-18 20:45:53 -05:00
Natan Keddem 8a2922262e added output truncate to cli 2023-11-18 16:06:01 -05:00
Natan Keddem d322612fc8 refactor and fix table display and sorting 2023-11-17 23:18:03 -05:00
Natan Keddem 3d13876804 optimize startup 2023-11-17 23:16:15 -05:00
Natan Keddem 4685939cae fixed arbitrary target path recall 2023-11-16 21:37:21 -05:00
12 changed files with 197 additions and 110 deletions
+5 -2
View File
@@ -71,8 +71,11 @@ class ErrorAggregator:
@property
def no_errors(self) -> bool:
validators = all(validation(element.value) for element in self.elements for validation in element.validation.values())
return self.enable and validators
if len(self.elements) > 0:
validators = all(validation(element.value) for element in self.elements for validation in element.validation.values())
return self.enable and validators
else:
return True
class WColumn(ui.column):
+11 -4
View File
@@ -34,6 +34,7 @@ class Cli:
self.stderr: List[str] = []
self._terminate: asyncio.Event = asyncio.Event()
self._busy: bool = False
self._truncated: bool = False
self.prefix_line: str = ""
self._stdout_terminals: List[Terminal] = []
self._stderr_terminals: List[Terminal] = []
@@ -70,8 +71,11 @@ class Cli:
else:
break
async def _controller(self, process: Process) -> None:
async def _controller(self, process: Process, max_output_lines) -> None:
while process.returncode is None:
if max_output_lines > 0 and len(self.stderr) + len(self.stdout) > max_output_lines:
self._truncated = True
process.terminate()
if self._terminate.is_set():
process.terminate()
try:
@@ -83,7 +87,7 @@ class Cli:
def terminate(self) -> None:
self._terminate.set()
async def execute(self, command: str) -> Result:
async def execute(self, command: str, max_output_lines: int = 0) -> Result:
self._busy = True
c = shlex.split(command, posix=False)
try:
@@ -92,13 +96,14 @@ class Cli:
self.stdout.clear()
self.stderr.clear()
self._terminate.clear()
self._truncated = False
terminated = False
now = datetime.now().strftime("%Y/%m/%d %H:%M:%S")
self.prefix_line = f"<{now}> {command}\n"
for terminal in self._stdout_terminals:
terminal.call_terminal_method("write", "\n" + self.prefix_line)
await asyncio.gather(
self._controller(process=process),
self._controller(process=process, max_output_lines=max_output_lines),
self._read_stdout(stream=process.stdout),
self._read_stderr(stream=process.stderr),
)
@@ -110,7 +115,9 @@ class Cli:
finally:
self._terminate.clear()
self._busy = False
return Result(command=command, return_code=process.returncode, stdout_lines=self.stdout.copy(), stderr_lines=self.stderr.copy(), terminated=terminated)
return Result(
command=command, return_code=process.returncode, stdout_lines=self.stdout.copy(), stderr_lines=self.stderr.copy(), terminated=terminated, truncated=self._truncated
)
async def shell(self, command: str) -> Result:
self._busy = True
+2 -2
View File
@@ -90,10 +90,10 @@ class Ssh(Cli):
del self._config[self.host]
self.write_config()
async def execute(self, command: str) -> Result:
async def execute(self, command: str, max_output_lines: int = 0) -> Result:
self._base_cmd = f"{'' if self.use_key else f'sshpass -p {self.password} '} ssh -F {self._config_path} {self.host}"
self._full_cmd = f"{self._base_cmd} {command}"
return await super().execute(self._full_cmd)
return await super().execute(self._full_cmd, max_output_lines)
async def send_key(self) -> Result:
await get_public_key(self._raw_path)
+27 -19
View File
@@ -3,7 +3,7 @@ from pathlib import Path
import stat
from datetime import datetime
import uuid
from nicegui import app, events, ui
from nicegui import app, background_tasks, events, ui
from fastapi.responses import StreamingResponse
import asyncssh
from bale import elements as el
@@ -87,7 +87,7 @@ class SshFileBrowse(ui.dialog):
row.tailwind.height("[40px]")
el.DButton("Download", on_click=self._start_download)
ui.button("Exit", on_click=lambda: self.submit("exit"))
await self._update_grid()
await self._update_handler()
async def _connect(self) -> Tuple[asyncssh.SSHClientConnection, asyncssh.SFTPClient]:
ssh = await asyncssh.connect(self._zfs.hostname, username=self._zfs.username, client_keys=[self._zfs.key_path])
@@ -137,7 +137,7 @@ class SshFileBrowse(ui.dialog):
"permissions": attributes.permissions,
}
async def _update_grid(self) -> None:
async def _update_handler(self) -> None:
self._grid.call_api_method("showLoadingOverlay")
if self._ssh is None or self._sftp is None:
self._ssh, self._sftp = await self._connect()
@@ -165,7 +165,7 @@ class SshFileBrowse(ui.dialog):
async def _handle_double_click(self, e: events.GenericEventArguments) -> None:
self.path = e.args["data"]["path"]
if e.args["data"]["type"] == "directory":
await self._update_grid()
await self._update_handler()
else:
await self._start_download(e)
@@ -226,10 +226,10 @@ class SshFileFind(SshFileBrowse):
with el.DBody(height="fit", width="[90vw]"):
with el.WColumn().classes("col"):
filesystems = await self._zfs.filesystems
self._filesystem = el.DSelect(
list(filesystems.data.keys()), label="filesystem", with_input=True, on_change=self._update_grid
)
self._pattern = el.DInput("Pattern", on_change=self._update_grid)
self._filesystem = el.DSelect(list(filesystems.data.keys()), label="filesystem", with_input=True, on_change=self._update_handler)
with el.WRow():
self._pattern = ui.input("Pattern").classes("col").on("keydown.enter", handler=self._update_handler)
el.LgButton(icon="search", on_click=self._update_handler)
self._grid = ui.aggrid(
{
"defaultColDef": {"flex": 1, "sortable": True, "suppressMovable": True, "sortingOrder": ["asc", "desc"]},
@@ -237,12 +237,14 @@ class SshFileFind(SshFileBrowse):
{"field": "name", "headerName": "Name", "flex": 1, "sort": "desc", "resizable": True},
{"field": "location", "headerName": "Location", "flex": 1, "resizable": True},
{
"field": "modified_datetime",
"headerName": "Modified",
"maxWidth": 200,
":comparator": """(valueA, valueB, nodeA, nodeB, isInverted) => {
return (nodeA.data.modified_timestamp > nodeB.data.modified_timestamp) ? -1 : 1;
}""",
"field": "modified_timestamp",
"filter": "agTextColumnFilter",
"maxWidth": 125,
":cellRenderer": """(data) => {
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
return date;
}""",
},
{
"field": "size",
@@ -264,15 +266,21 @@ class SshFileFind(SshFileBrowse):
row.tailwind.height("[40px]")
el.DButton("Download", on_click=self._start_download)
ui.button("Exit", on_click=lambda: self.submit("exit"))
await self._update_grid()
self._grid.call_api_method("hideOverlay")
async def _update_grid(self) -> None:
self._grid.call_api_method("showLoadingOverlay")
if self._filesystem is not None:
async def _update_handler(self) -> None:
if len(self._pattern.value) > 0 and self._filesystem is not None:
self._grid.call_api_method("showLoadingOverlay")
self._filesystem.props("readonly")
self._pattern.props("readonly")
files = await self._zfs.find_files_in_snapshots(filesystem=self._filesystem.value, pattern=self._pattern.value)
self._grid.options["rowData"] = files.data
self._grid.update()
self._grid.call_api_method("hideOverlay")
if files.truncated is True:
el.notify("Too many files found, truncating list.", type="warning")
self._grid.update()
self._filesystem.props(remove="readonly")
self._pattern.props(remove="readonly")
self._grid.call_api_method("hideOverlay")
async def _handle_double_click(self, e: events.GenericEventArguments) -> None:
await self._start_download(e)
+30 -25
View File
@@ -81,7 +81,7 @@ class Zfs:
command = command if len(command) < 160 else command[:160] + "..."
el.notify(command)
async def execute(self, command: str, notify: bool = True) -> Result:
async def execute(self, command: str, max_output_lines: int = 0, notify: bool = True) -> Result:
if notify:
self.notify(command)
return Result(command=command)
@@ -166,25 +166,28 @@ class Zfs:
return result
async def find_files_in_snapshots(self, filesystem: str, pattern: str) -> Result:
filesystems = await self.filesystems
if filesystem in filesystems.data.keys():
if "mountpoint" in filesystems.data[filesystem]:
command = f"find {filesystems.data[filesystem]['mountpoint']}/.zfs/snapshot -type f -name '{pattern}' -printf '%h\t%f\t%s\t%T@\n'"
result = await self.execute(command=command, notify=False)
files = []
for line in result.stdout_lines:
matches = re.match(
"^(?P<location>[^\t]+)\t(?P<name>[^\t]+)\t(?P<bytes>[^\t]+)\t(?P<modified_timestamp>[^\n]+)",
line,
)
if matches is not None:
md = matches.groupdict()
md["path"] = f"{md['location']}/{md['name']}"
md["size"] = format_bytes(int(md["bytes"]))
md["modified_datetime"] = datetime.fromtimestamp(float(md["modified_timestamp"])).strftime("%Y/%m/%d %H:%M:%S")
files.append(md)
result.data = files
return result
try:
filesystems = await self.filesystems
command = f"find {filesystems.data[filesystem]['mountpoint']}/.zfs/snapshot -type f -name '{pattern}' -printf '%h\t%f\t%s\t%T@\n'"
result = await self.execute(command=command, notify=False, max_output_lines=1000)
files = []
for line in result.stdout_lines:
matches = re.match(
"^(?P<location>[^\t]+)\t(?P<name>[^\t]+)\t(?P<bytes>[^\t]+)\t(?P<modified_timestamp>[^\n]+)",
line,
)
if matches is not None:
md = matches.groupdict()
md["path"] = f"{md['location']}/{md['name']}"
md["bytes"] = int(md["bytes"])
md["size"] = format_bytes(md["bytes"])
md["modified_datetime"] = datetime.fromtimestamp(float(md["modified_timestamp"])).strftime("%Y/%m/%d %H:%M:%S")
md["modified_timestamp"] = float(md["modified_timestamp"])
files.append(md)
result.data = files
return result
except KeyError:
pass
return Result()
@property
@@ -219,9 +222,11 @@ class Zfs:
matches = re.match("^(?P<filesystem>[^@]+)@(?P<name>[^\t]+)\t(?P<used_bytes>[^\t]+)\t(?P<creation>[^\t]+)\t(?P<userrefs>[^\n]+)", line)
if matches is not None:
md = matches.groupdict()
md["creation_date"] = datetime.fromtimestamp(int(md["creation"])).strftime("%Y/%m/%d")
md["creation_time"] = datetime.fromtimestamp(int(md["creation"])).strftime("%H:%M")
md["used"] = format_bytes(int(md["used_bytes"]))
md["used_bytes"] = int(md["used_bytes"])
md["creation"] = int(md["creation"])
md["creation_date"] = datetime.fromtimestamp(md["creation"]).strftime("%Y/%m/%d")
md["creation_time"] = datetime.fromtimestamp(md["creation"]).strftime("%H:%M")
md["used"] = format_bytes(md["used_bytes"])
snapshot = f"{md['filesystem']}@{md['name']}"
snapshots[snapshot] = md
self._last_data[query] = snapshots
@@ -240,10 +245,10 @@ class Ssh(ssh.Ssh, Zfs):
def notify(self, command: str):
super().notify(f"<{self.host}> {command}")
async def execute(self, command: str, notify: bool = True) -> Result:
async def execute(self, command: str, max_output_lines: int = 0, notify: bool = True) -> Result:
if notify:
self.notify(command)
result = await super().execute(command)
result = await super().execute(command, max_output_lines)
if result.stderr != "":
el.notify(result.stderr, type="negative")
result.name = self.host
+1 -1
View File
@@ -10,7 +10,7 @@ logger = logging.getLogger(__name__)
def build():
@ui.page("/")
@ui.page("/", response_timeout=30)
async def index(client: Client) -> None:
app.add_static_files("/static", "static")
el.load_element_css()
+1
View File
@@ -13,6 +13,7 @@ class Result:
stdout_lines: List[str] = field(default_factory=list)
stderr_lines: List[str] = field(default_factory=list)
terminated: bool = False
truncated: bool = False
data: Any = None
trace: str = ""
cached: bool = False
+2
View File
@@ -11,6 +11,7 @@ from apscheduler.schedulers.asyncio import AsyncIOScheduler
@dataclass(kw_only=True)
class Automation:
id: str
name: str
app: str
hosts: List[str]
host: str
@@ -30,6 +31,7 @@ class Automation:
class Zfs_Autobackup(Automation):
app: str = "zfs_autobackup"
execute_mode: str = "local"
prop: str
target_host: str
target_path: str
target_paths: List[str]
+4 -2
View File
@@ -97,12 +97,14 @@ class Tab:
col.tailwind.max_width("lg")
ui.label(f"Host Name: {result.name}").classes("text-secondary")
ui.label(f"Command: {result.command}").classes("text-secondary")
ui.label(f"Date: {result.date}").classes("text-secondary")
timestamp = await ui.run_javascript(
f"new Date({result.timestamp} * 1000).toLocaleString(undefined, {{dateStyle: 'short', timeStyle: 'short', hour12: 'false'}});"
)
ui.label(f"Timestamp: {timestamp}").classes("text-secondary")
with ui.column() as col:
col.tailwind.max_width("lg")
ui.label(f"Task has failed: {result.failed}").classes("text-secondary")
ui.label(f"Data is cached: {result.cached}").classes("text-secondary")
ui.label(f"Time: {result.time}").classes("text-secondary")
with el.Card() as card:
with el.WColumn():
terminal = cli.Terminal(options={"rows": 18, "cols": 120, "convertEol": True})
+60 -39
View File
@@ -1,4 +1,4 @@
from typing import Any, Dict, List, Union
from typing import Any, Callable, Dict, List, Union
import asyncio
from datetime import datetime
import json
@@ -36,20 +36,20 @@ def populate_job_handler(app: str, job_id: str, host: str):
return job_handlers[job_id]
class CommandTemplate(string.Template):
class AutomationTemplate(string.Template):
delimiter = ""
async def automation_job(**kwargs) -> None:
if "data" in kwargs:
jd = json.loads(kwargs["data"])
command = CommandTemplate(jd["command"])
command = AutomationTemplate(jd["command"])
tab = Tab(host=None, spinner=None)
if jd["app"] == "zfs_autobackup":
d = scheduler.Zfs_Autobackup(**jd)
populate_job_handler(app=d.app, job_id=d.id, host=d.host)
if job_handlers[d.id].is_busy is False:
result = await job_handlers[d.id].execute(command.safe_substitute(host=d.host))
result = await job_handlers[d.id].execute(command.safe_substitute(name=d.name, host=d.host))
result.name = d.host
result.status = "success" if result.return_code == 0 else "error"
if d.pipe_success is True and result.status == "success":
@@ -63,7 +63,7 @@ async def automation_job(**kwargs) -> None:
d = scheduler.Automation(**jd)
populate_job_handler(app=d.app, job_id=d.id, host=d.host)
if job_handlers[d.id].is_busy is False:
result = await job_handlers[d.id].execute(command.safe_substitute(host=d.host))
result = await job_handlers[d.id].execute(command.safe_substitute(name=d.name, host=d.host))
result.name = d.host
if d.pipe_success is True and result.status == "success":
tab.pipe_result(result=result)
@@ -76,7 +76,7 @@ async def automation_job(**kwargs) -> None:
d = scheduler.Automation(**jd)
populate_job_handler(app=d.app, job_id=d.id, host=d.host)
if job_handlers[d.id].is_busy is False:
result = await job_handlers[d.id].execute(command.safe_substitute(host=d.host))
result = await job_handlers[d.id].execute(command.safe_substitute(name=d.name, host=d.host))
result.name = d.host
if d.pipe_success is True and result.status == "success":
tab.pipe_result(result=result)
@@ -99,7 +99,7 @@ class Automation(Tab):
self.job_data: Dict[str, str] = {}
self.job_names: List[str] = []
self.default_options: Dict[str, str] = {}
self.build_command: str = ""
self.build_command: Callable
self.target_host: el.DSelect
self.target_paths: List[str] = [""]
self.target_path: el.DSelect
@@ -161,8 +161,16 @@ class Automation(Tab):
"maxWidth": 150,
},
{"headerName": "Command", "field": "command", "filter": "agTextColumnFilter"},
{"headerName": "Next Date", "field": "next_run_date", "filter": "agDateColumnFilter", "maxWidth": 100},
{"headerName": "Next Time", "field": "next_run_time", "maxWidth": 100},
{
"headerName": "Next Run",
"field": "next_run",
"filter": "agTextColumnFilter",
"maxWidth": 125,
":cellRenderer": """(data) => {
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
return date;
}""",
},
{
"headerName": "Status",
"field": "status",
@@ -227,11 +235,9 @@ class Automation(Tab):
self._automations.clear()
for job in self.scheduler.scheduler.get_jobs():
if job.next_run_time is not None:
next_run_date = job.next_run_time.strftime("%Y/%m/%d")
next_run_time = job.next_run_time.strftime("%H:%M")
next_run = job.next_run_time.timestamp()
else:
next_run_date = "NA"
next_run_time = "NA"
next_run = "NA"
if "data" in job.kwargs:
jd = json.loads(job.kwargs["data"])
if self.host == jd["host"]:
@@ -239,8 +245,7 @@ class Automation(Tab):
{
"name": job.id.split("@")[0],
"command": jd["command"],
"next_run_date": next_run_date,
"next_run_time": next_run_time,
"next_run": next_run,
"status": "",
}
)
@@ -303,26 +308,17 @@ class Automation(Tab):
await self._create_automation(rows[0]["name"])
self._set_selection()
async def _add_prop_to_fs(
self,
host: str,
prop: str,
value: str,
module: str = "autobackup",
filesystems: Union[List[str], None] = None,
) -> None:
async def _add_prop_to_fs(self, host: str, prop: str, value: str, filesystems: Union[List[str], None] = None) -> None:
if filesystems is not None:
full_prop = f"{module}:{prop}"
for fs in filesystems:
result = await self._zfs[host].add_filesystem_prop(filesystem=fs, prop=full_prop, value=value)
result = await self._zfs[host].add_filesystem_prop(filesystem=fs, prop=prop, value=value)
self.add_history(result=result)
async def _remove_prop_from_all_fs(self, host: str, prop: str, module: str = "autobackup") -> None:
full_prop = f"{module}:{prop}"
filesystems_with_prop_result = await self._zfs[host].filesystems_with_prop(full_prop)
async def _remove_prop_from_all_fs(self, host: str, prop: str) -> None:
filesystems_with_prop_result = await self._zfs[host].filesystems_with_prop(prop)
filesystems_with_prop = list(filesystems_with_prop_result.data)
for fs in filesystems_with_prop:
result = await self._zfs[host].remove_filesystem_prop(filesystem=fs, prop=full_prop)
result = await self._zfs[host].remove_filesystem_prop(filesystem=fs, prop=prop)
self.add_history(result=result)
async def _create_automation(self, name: str = "") -> None:
@@ -416,15 +412,16 @@ class Automation(Tab):
self.target_path.update()
self.target_path.value = ""
async def target_path_selected() -> None:
self.build_command()
def build_command() -> None:
try:
prop_suffix = self.prop.value.split(":")[1]
except IndexError:
prop_suffix = ""
base = ""
for key, value in self.picked_options.items():
base = base + f" --{key}{f' {value}' if value != '' else ''}"
target_path = f"{f' {self.target_path.value}' if self.target_path.value != '' else ''}"
base = base + f" {self.auto_name.value.lower()}" + target_path
base = base + f" {prop_suffix}" + target_path
self.command.value = base
def all_fs_to_lists():
@@ -471,6 +468,17 @@ class Automation(Tab):
self.children.update()
self.exclude.update()
def validate_prop(value):
parts = value.split(":")
for part in parts:
if part.find(" ") != -1:
return False
if len(part) < 1:
return False
if len(parts) != 2:
return False
return True
if name == "":
self.default_options = {
"verbose": "",
@@ -491,9 +499,11 @@ class Automation(Tab):
row.tailwind.width("[860px]").justify_content("center")
with ui.column() as col:
col.tailwind.height("full").width("[420px]")
self.prop = el.DInput(label="Property", value="autobackup:{name}", on_change=build_command, validation=validate_prop)
self.app_em.append(self.prop)
self.target_host = el.DSelect(target_host, label="Target Host", on_change=target_host_selected)
self.target_paths = [""]
self.target_path = el.DSelect(self.target_paths, value="", label="Target Path", new_value_mode="add-unique", on_change=target_path_selected)
self.target_path = el.DSelect(self.target_paths, value="", label="Target Path", new_value_mode="add-unique", on_change=build_command)
self.hosts = el.DSelect(source_hosts, label="Source Host(s)", multiple=True, with_input=True)
all_fs_to_lists()
with ui.scroll_area().classes("col"):
@@ -529,12 +539,15 @@ class Automation(Tab):
col.tailwind.height("full").width("[420px]")
options_controls()
if name != "":
self.prop.value = self.job_data.get("prop", "autobackup:{name}")
self.target_host.value = self.job_data.get("target_host", "")
target_path = self.job_data.get("target_path", "")
tries = 0
while target_path not in self.target_path.options and tries < 20:
await asyncio.sleep(0.1)
tries = tries + 1
if target_path not in self.target_paths:
self.target_paths.append(target_path)
self.target_path.value = target_path
self.parentchildren.value = self.fs["values"].get("parentchildren", None)
self.parent.value = self.fs["values"].get("parent", None)
@@ -751,7 +764,9 @@ class Automation(Tab):
with el.WRow() as row:
row.tailwind.height("[40px]")
self.as_spinner = el.Spinner()
self.app_em = el.ErrorAggregator()
self.save = el.DButton("SAVE", on_click=lambda: automation_dialog.submit("save"))
self.save.bind_enabled_from(self.app_em, "no_errors")
el.Spinner(master=self.as_spinner)
self.auto_name.value = name
if name != "":
@@ -772,11 +787,13 @@ class Automation(Tab):
self.scheduler.scheduler.remove_job(job.id)
for host in hosts:
auto_id = f"{auto_name}@{host}"
await self._remove_prop_from_all_fs(host=host, prop=auto_name)
await self._add_prop_to_fs(host=host, prop=auto_name, value="true", filesystems=self.parentchildren.value)
await self._add_prop_to_fs(host=host, prop=auto_name, value="parent", filesystems=self.parent.value)
await self._add_prop_to_fs(host=host, prop=auto_name, value="child", filesystems=self.children.value)
await self._add_prop_to_fs(host=host, prop=auto_name, value="false", filesystems=self.exclude.value)
command = AutomationTemplate(self.prop.value)
prop = command.safe_substitute(name=auto_name, host=host)
await self._remove_prop_from_all_fs(host=host, prop=prop)
await self._add_prop_to_fs(host=host, prop=prop, value="true", filesystems=self.parentchildren.value)
await self._add_prop_to_fs(host=host, prop=prop, value="parent", filesystems=self.parent.value)
await self._add_prop_to_fs(host=host, prop=prop, value="child", filesystems=self.children.value)
await self._add_prop_to_fs(host=host, prop=prop, value="false", filesystems=self.exclude.value)
self.fs["values"] = {}
self.fs["values"]["parentchildren"] = self.parentchildren.value
self.fs["values"]["parent"] = self.parent.value
@@ -784,6 +801,7 @@ class Automation(Tab):
self.fs["values"]["exclude"] = self.exclude.value
auto = scheduler.Zfs_Autobackup(
id=auto_id,
name=auto_name,
hosts=hosts,
host=host,
command="python -m zfs_autobackup.ZfsAutobackup" + self.command.value,
@@ -796,6 +814,7 @@ class Automation(Tab):
filesystems=self.fs,
pipe_success=self.pipe_success.value,
pipe_error=self.pipe_error.value,
prop=self.prop.value,
)
self.scheduler.scheduler.add_job(
automation_job,
@@ -815,6 +834,7 @@ class Automation(Tab):
auto_id = f"{auto_name}@{host}"
auto = scheduler.Automation(
id=auto_id,
name=auto_name,
app=self.app.value,
hosts=hosts,
host=host,
@@ -837,6 +857,7 @@ class Automation(Tab):
auto_id = f"{auto_name}@{self.host}"
auto = scheduler.Automation(
id=auto_id,
name=auto_name,
app=self.app.value,
hosts=hosts,
host=self.host,
+33 -6
View File
@@ -1,7 +1,8 @@
from datetime import datetime
import json
from . import SelectionConfirm, Tab
from nicegui import ui, events
import httpx
from . import SelectionConfirm, Tab
from bale import elements as el
from bale.result import Result
from bale.interfaces import zfs
@@ -47,8 +48,16 @@ class History(Tab):
"filter": "agTextColumnFilter",
"flex": 1,
},
{"headerName": "Date", "field": "date", "filter": "agDateColumnFilter", "maxWidth": 100},
{"headerName": "Time", "field": "time", "maxWidth": 100},
{
"headerName": "Timestamp",
"field": "timestamp",
"filter": "agTextColumnFilter",
"maxWidth": 125,
":cellRenderer": """(data) => {
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
return date;
}""",
},
{
"headerName": "Status",
"field": "status",
@@ -90,6 +99,20 @@ class History(Tab):
http[status]["data"] = e.content["json"]["data"]
http[status]["headers"] = e.content["json"]["headers"]
def test(status):
try:
url = http[status]["url"]
data = self.process_pipe_data(result=Result(name=self.host, command="TEST COMMAND", status=status), data=http[status]["data"])
headers = http[status]["headers"]
post = httpx.post(url=url, json=data, headers=headers)
print(post.status_code)
if post.status_code == 200:
el.notify("Test successful!", type="positive")
else:
el.notify(f"Test failed with status code {post.status_code}!", type="negative")
except:
el.notify("Test failed!", type="negative")
def show_controls(status):
if status not in http:
http[status] = {}
@@ -107,7 +130,7 @@ class History(Tab):
"topic": "mytopic",
"tags": ["turtle"],
"title": "Successful Automation Run for {name}",
"message": "{stdout}",
"message": "{command}",
},
)
editor.properties["content"]["json"]["headers"] = self.get_pipe_status("http", status).get("headers", {"Authorization": "Bearer tk_..."})
@@ -126,11 +149,15 @@ class History(Tab):
with ui.step("On Success"):
with el.WColumn().classes("col justify-start"):
show_controls(status="success")
el.LgButton("NEXT", on_click=lambda _: stepper.next())
with el.WRow():
el.LgButton("TEST", on_click=lambda _: test(status="success"))
el.LgButton("NEXT", on_click=lambda _: stepper.next())
with ui.step("On Error"):
with el.WColumn().classes("col justify-start"):
show_controls(status="error")
el.DButton("SAVE", on_click=lambda: host_dialog.submit("save"))
with el.WRow():
el.LgButton("TEST", on_click=lambda _: test(status="error"))
el.DButton("SAVE", on_click=lambda: host_dialog.submit("save"))
result = await host_dialog
if result == "save":
+21 -10
View File
@@ -1,6 +1,6 @@
import asyncio
from copy import deepcopy
from nicegui import ui
from nicegui import background_tasks, ui
from . import SelectionConfirm, Tab, Task
from bale.result import Result
from bale import elements as el
@@ -75,11 +75,19 @@ class Manage(Tab):
"field": "used",
"maxWidth": 100,
":comparator": """(valueA, valueB, nodeA, nodeB, isInverted) => {
return (nodeA.data.used_bytes > nodeB.data.used_bytes) ? -1 : 1;
}""",
return (nodeA.data.used_bytes > nodeB.data.used_bytes) ? -1 : 1;
}""",
},
{
"headerName": "Created",
"field": "creation",
"filter": "agTextColumnFilter",
"maxWidth": 125,
":cellRenderer": """(data) => {
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
return date;
}""",
},
{"headerName": "Creation Date", "field": "creation_date", "filter": "agDateColumnFilter", "maxWidth": 150},
{"headerName": "Creation Time", "field": "creation_time", "maxWidth": 150},
{"headerName": "Holds", "field": "userrefs", "filter": "agNumberColumnFilter", "maxWidth": 100},
],
"rowData": [],
@@ -92,8 +100,8 @@ class Manage(Tab):
self._spinner.visible = True
self.zfs.invalidate_query()
snapshots = await self.zfs.snapshots
await self.zfs.filesystems
await self.zfs.holds_for_snapshot()
background_tasks.create(self.zfs.filesystems, name="zfs_filesystems")
background_tasks.create(self.zfs.holds_for_snapshot(), name="zfs_holds")
self._grid.options["rowData"] = list(snapshots.data.values())
self._grid.update()
self._spinner.visible = False
@@ -103,9 +111,12 @@ class Manage(Tab):
result = await SelectionConfirm(container=self._confirm, label=">BROWSE<")
if result == "confirm":
rows = await self._grid.get_selected_rows()
filesystems = await self.zfs.filesystems
mount_path = filesystems.data[rows[0]["filesystem"]]["mountpoint"]
await sshdl.SshFileBrowse(zfs=self.zfs, path=f"{mount_path}/.zfs/snapshot/{rows[0]['name']}")
try:
filesystems = await self.zfs.filesystems
mount_path = filesystems.data[rows[0]["filesystem"]]["mountpoint"]
await sshdl.SshFileBrowse(zfs=self.zfs, path=f"{mount_path}/.zfs/snapshot/{rows[0]['name']}")
except KeyError:
el.notify(f"Unable to browse {rows[0]['filesystem']}", type="warning")
self._set_selection()
async def _find(self) -> None: