mirror of
https://github.com/natankeddem/bale.git
synced 2026-05-03 06:02:54 +00:00
Compare commits
10 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 36ee1f94cd | |||
| 07ce7e0bae | |||
| ffbc9b71c0 | |||
| f3ef97a342 | |||
| 94fba0b925 | |||
| 8572ad766b | |||
| 8a2922262e | |||
| d322612fc8 | |||
| 3d13876804 | |||
| 4685939cae |
+5
-2
@@ -71,8 +71,11 @@ class ErrorAggregator:
|
||||
|
||||
@property
|
||||
def no_errors(self) -> bool:
|
||||
validators = all(validation(element.value) for element in self.elements for validation in element.validation.values())
|
||||
return self.enable and validators
|
||||
if len(self.elements) > 0:
|
||||
validators = all(validation(element.value) for element in self.elements for validation in element.validation.values())
|
||||
return self.enable and validators
|
||||
else:
|
||||
return True
|
||||
|
||||
|
||||
class WColumn(ui.column):
|
||||
|
||||
+11
-4
@@ -34,6 +34,7 @@ class Cli:
|
||||
self.stderr: List[str] = []
|
||||
self._terminate: asyncio.Event = asyncio.Event()
|
||||
self._busy: bool = False
|
||||
self._truncated: bool = False
|
||||
self.prefix_line: str = ""
|
||||
self._stdout_terminals: List[Terminal] = []
|
||||
self._stderr_terminals: List[Terminal] = []
|
||||
@@ -70,8 +71,11 @@ class Cli:
|
||||
else:
|
||||
break
|
||||
|
||||
async def _controller(self, process: Process) -> None:
|
||||
async def _controller(self, process: Process, max_output_lines) -> None:
|
||||
while process.returncode is None:
|
||||
if max_output_lines > 0 and len(self.stderr) + len(self.stdout) > max_output_lines:
|
||||
self._truncated = True
|
||||
process.terminate()
|
||||
if self._terminate.is_set():
|
||||
process.terminate()
|
||||
try:
|
||||
@@ -83,7 +87,7 @@ class Cli:
|
||||
def terminate(self) -> None:
|
||||
self._terminate.set()
|
||||
|
||||
async def execute(self, command: str) -> Result:
|
||||
async def execute(self, command: str, max_output_lines: int = 0) -> Result:
|
||||
self._busy = True
|
||||
c = shlex.split(command, posix=False)
|
||||
try:
|
||||
@@ -92,13 +96,14 @@ class Cli:
|
||||
self.stdout.clear()
|
||||
self.stderr.clear()
|
||||
self._terminate.clear()
|
||||
self._truncated = False
|
||||
terminated = False
|
||||
now = datetime.now().strftime("%Y/%m/%d %H:%M:%S")
|
||||
self.prefix_line = f"<{now}> {command}\n"
|
||||
for terminal in self._stdout_terminals:
|
||||
terminal.call_terminal_method("write", "\n" + self.prefix_line)
|
||||
await asyncio.gather(
|
||||
self._controller(process=process),
|
||||
self._controller(process=process, max_output_lines=max_output_lines),
|
||||
self._read_stdout(stream=process.stdout),
|
||||
self._read_stderr(stream=process.stderr),
|
||||
)
|
||||
@@ -110,7 +115,9 @@ class Cli:
|
||||
finally:
|
||||
self._terminate.clear()
|
||||
self._busy = False
|
||||
return Result(command=command, return_code=process.returncode, stdout_lines=self.stdout.copy(), stderr_lines=self.stderr.copy(), terminated=terminated)
|
||||
return Result(
|
||||
command=command, return_code=process.returncode, stdout_lines=self.stdout.copy(), stderr_lines=self.stderr.copy(), terminated=terminated, truncated=self._truncated
|
||||
)
|
||||
|
||||
async def shell(self, command: str) -> Result:
|
||||
self._busy = True
|
||||
|
||||
@@ -90,10 +90,10 @@ class Ssh(Cli):
|
||||
del self._config[self.host]
|
||||
self.write_config()
|
||||
|
||||
async def execute(self, command: str) -> Result:
|
||||
async def execute(self, command: str, max_output_lines: int = 0) -> Result:
|
||||
self._base_cmd = f"{'' if self.use_key else f'sshpass -p {self.password} '} ssh -F {self._config_path} {self.host}"
|
||||
self._full_cmd = f"{self._base_cmd} {command}"
|
||||
return await super().execute(self._full_cmd)
|
||||
return await super().execute(self._full_cmd, max_output_lines)
|
||||
|
||||
async def send_key(self) -> Result:
|
||||
await get_public_key(self._raw_path)
|
||||
|
||||
+27
-19
@@ -3,7 +3,7 @@ from pathlib import Path
|
||||
import stat
|
||||
from datetime import datetime
|
||||
import uuid
|
||||
from nicegui import app, events, ui
|
||||
from nicegui import app, background_tasks, events, ui
|
||||
from fastapi.responses import StreamingResponse
|
||||
import asyncssh
|
||||
from bale import elements as el
|
||||
@@ -87,7 +87,7 @@ class SshFileBrowse(ui.dialog):
|
||||
row.tailwind.height("[40px]")
|
||||
el.DButton("Download", on_click=self._start_download)
|
||||
ui.button("Exit", on_click=lambda: self.submit("exit"))
|
||||
await self._update_grid()
|
||||
await self._update_handler()
|
||||
|
||||
async def _connect(self) -> Tuple[asyncssh.SSHClientConnection, asyncssh.SFTPClient]:
|
||||
ssh = await asyncssh.connect(self._zfs.hostname, username=self._zfs.username, client_keys=[self._zfs.key_path])
|
||||
@@ -137,7 +137,7 @@ class SshFileBrowse(ui.dialog):
|
||||
"permissions": attributes.permissions,
|
||||
}
|
||||
|
||||
async def _update_grid(self) -> None:
|
||||
async def _update_handler(self) -> None:
|
||||
self._grid.call_api_method("showLoadingOverlay")
|
||||
if self._ssh is None or self._sftp is None:
|
||||
self._ssh, self._sftp = await self._connect()
|
||||
@@ -165,7 +165,7 @@ class SshFileBrowse(ui.dialog):
|
||||
async def _handle_double_click(self, e: events.GenericEventArguments) -> None:
|
||||
self.path = e.args["data"]["path"]
|
||||
if e.args["data"]["type"] == "directory":
|
||||
await self._update_grid()
|
||||
await self._update_handler()
|
||||
else:
|
||||
await self._start_download(e)
|
||||
|
||||
@@ -226,10 +226,10 @@ class SshFileFind(SshFileBrowse):
|
||||
with el.DBody(height="fit", width="[90vw]"):
|
||||
with el.WColumn().classes("col"):
|
||||
filesystems = await self._zfs.filesystems
|
||||
self._filesystem = el.DSelect(
|
||||
list(filesystems.data.keys()), label="filesystem", with_input=True, on_change=self._update_grid
|
||||
)
|
||||
self._pattern = el.DInput("Pattern", on_change=self._update_grid)
|
||||
self._filesystem = el.DSelect(list(filesystems.data.keys()), label="filesystem", with_input=True, on_change=self._update_handler)
|
||||
with el.WRow():
|
||||
self._pattern = ui.input("Pattern").classes("col").on("keydown.enter", handler=self._update_handler)
|
||||
el.LgButton(icon="search", on_click=self._update_handler)
|
||||
self._grid = ui.aggrid(
|
||||
{
|
||||
"defaultColDef": {"flex": 1, "sortable": True, "suppressMovable": True, "sortingOrder": ["asc", "desc"]},
|
||||
@@ -237,12 +237,14 @@ class SshFileFind(SshFileBrowse):
|
||||
{"field": "name", "headerName": "Name", "flex": 1, "sort": "desc", "resizable": True},
|
||||
{"field": "location", "headerName": "Location", "flex": 1, "resizable": True},
|
||||
{
|
||||
"field": "modified_datetime",
|
||||
"headerName": "Modified",
|
||||
"maxWidth": 200,
|
||||
":comparator": """(valueA, valueB, nodeA, nodeB, isInverted) => {
|
||||
return (nodeA.data.modified_timestamp > nodeB.data.modified_timestamp) ? -1 : 1;
|
||||
}""",
|
||||
"field": "modified_timestamp",
|
||||
"filter": "agTextColumnFilter",
|
||||
"maxWidth": 125,
|
||||
":cellRenderer": """(data) => {
|
||||
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
|
||||
return date;
|
||||
}""",
|
||||
},
|
||||
{
|
||||
"field": "size",
|
||||
@@ -264,15 +266,21 @@ class SshFileFind(SshFileBrowse):
|
||||
row.tailwind.height("[40px]")
|
||||
el.DButton("Download", on_click=self._start_download)
|
||||
ui.button("Exit", on_click=lambda: self.submit("exit"))
|
||||
await self._update_grid()
|
||||
self._grid.call_api_method("hideOverlay")
|
||||
|
||||
async def _update_grid(self) -> None:
|
||||
self._grid.call_api_method("showLoadingOverlay")
|
||||
if self._filesystem is not None:
|
||||
async def _update_handler(self) -> None:
|
||||
if len(self._pattern.value) > 0 and self._filesystem is not None:
|
||||
self._grid.call_api_method("showLoadingOverlay")
|
||||
self._filesystem.props("readonly")
|
||||
self._pattern.props("readonly")
|
||||
files = await self._zfs.find_files_in_snapshots(filesystem=self._filesystem.value, pattern=self._pattern.value)
|
||||
self._grid.options["rowData"] = files.data
|
||||
self._grid.update()
|
||||
self._grid.call_api_method("hideOverlay")
|
||||
if files.truncated is True:
|
||||
el.notify("Too many files found, truncating list.", type="warning")
|
||||
self._grid.update()
|
||||
self._filesystem.props(remove="readonly")
|
||||
self._pattern.props(remove="readonly")
|
||||
self._grid.call_api_method("hideOverlay")
|
||||
|
||||
async def _handle_double_click(self, e: events.GenericEventArguments) -> None:
|
||||
await self._start_download(e)
|
||||
|
||||
+30
-25
@@ -81,7 +81,7 @@ class Zfs:
|
||||
command = command if len(command) < 160 else command[:160] + "..."
|
||||
el.notify(command)
|
||||
|
||||
async def execute(self, command: str, notify: bool = True) -> Result:
|
||||
async def execute(self, command: str, max_output_lines: int = 0, notify: bool = True) -> Result:
|
||||
if notify:
|
||||
self.notify(command)
|
||||
return Result(command=command)
|
||||
@@ -166,25 +166,28 @@ class Zfs:
|
||||
return result
|
||||
|
||||
async def find_files_in_snapshots(self, filesystem: str, pattern: str) -> Result:
|
||||
filesystems = await self.filesystems
|
||||
if filesystem in filesystems.data.keys():
|
||||
if "mountpoint" in filesystems.data[filesystem]:
|
||||
command = f"find {filesystems.data[filesystem]['mountpoint']}/.zfs/snapshot -type f -name '{pattern}' -printf '%h\t%f\t%s\t%T@\n'"
|
||||
result = await self.execute(command=command, notify=False)
|
||||
files = []
|
||||
for line in result.stdout_lines:
|
||||
matches = re.match(
|
||||
"^(?P<location>[^\t]+)\t(?P<name>[^\t]+)\t(?P<bytes>[^\t]+)\t(?P<modified_timestamp>[^\n]+)",
|
||||
line,
|
||||
)
|
||||
if matches is not None:
|
||||
md = matches.groupdict()
|
||||
md["path"] = f"{md['location']}/{md['name']}"
|
||||
md["size"] = format_bytes(int(md["bytes"]))
|
||||
md["modified_datetime"] = datetime.fromtimestamp(float(md["modified_timestamp"])).strftime("%Y/%m/%d %H:%M:%S")
|
||||
files.append(md)
|
||||
result.data = files
|
||||
return result
|
||||
try:
|
||||
filesystems = await self.filesystems
|
||||
command = f"find {filesystems.data[filesystem]['mountpoint']}/.zfs/snapshot -type f -name '{pattern}' -printf '%h\t%f\t%s\t%T@\n'"
|
||||
result = await self.execute(command=command, notify=False, max_output_lines=1000)
|
||||
files = []
|
||||
for line in result.stdout_lines:
|
||||
matches = re.match(
|
||||
"^(?P<location>[^\t]+)\t(?P<name>[^\t]+)\t(?P<bytes>[^\t]+)\t(?P<modified_timestamp>[^\n]+)",
|
||||
line,
|
||||
)
|
||||
if matches is not None:
|
||||
md = matches.groupdict()
|
||||
md["path"] = f"{md['location']}/{md['name']}"
|
||||
md["bytes"] = int(md["bytes"])
|
||||
md["size"] = format_bytes(md["bytes"])
|
||||
md["modified_datetime"] = datetime.fromtimestamp(float(md["modified_timestamp"])).strftime("%Y/%m/%d %H:%M:%S")
|
||||
md["modified_timestamp"] = float(md["modified_timestamp"])
|
||||
files.append(md)
|
||||
result.data = files
|
||||
return result
|
||||
except KeyError:
|
||||
pass
|
||||
return Result()
|
||||
|
||||
@property
|
||||
@@ -219,9 +222,11 @@ class Zfs:
|
||||
matches = re.match("^(?P<filesystem>[^@]+)@(?P<name>[^\t]+)\t(?P<used_bytes>[^\t]+)\t(?P<creation>[^\t]+)\t(?P<userrefs>[^\n]+)", line)
|
||||
if matches is not None:
|
||||
md = matches.groupdict()
|
||||
md["creation_date"] = datetime.fromtimestamp(int(md["creation"])).strftime("%Y/%m/%d")
|
||||
md["creation_time"] = datetime.fromtimestamp(int(md["creation"])).strftime("%H:%M")
|
||||
md["used"] = format_bytes(int(md["used_bytes"]))
|
||||
md["used_bytes"] = int(md["used_bytes"])
|
||||
md["creation"] = int(md["creation"])
|
||||
md["creation_date"] = datetime.fromtimestamp(md["creation"]).strftime("%Y/%m/%d")
|
||||
md["creation_time"] = datetime.fromtimestamp(md["creation"]).strftime("%H:%M")
|
||||
md["used"] = format_bytes(md["used_bytes"])
|
||||
snapshot = f"{md['filesystem']}@{md['name']}"
|
||||
snapshots[snapshot] = md
|
||||
self._last_data[query] = snapshots
|
||||
@@ -240,10 +245,10 @@ class Ssh(ssh.Ssh, Zfs):
|
||||
def notify(self, command: str):
|
||||
super().notify(f"<{self.host}> {command}")
|
||||
|
||||
async def execute(self, command: str, notify: bool = True) -> Result:
|
||||
async def execute(self, command: str, max_output_lines: int = 0, notify: bool = True) -> Result:
|
||||
if notify:
|
||||
self.notify(command)
|
||||
result = await super().execute(command)
|
||||
result = await super().execute(command, max_output_lines)
|
||||
if result.stderr != "":
|
||||
el.notify(result.stderr, type="negative")
|
||||
result.name = self.host
|
||||
|
||||
+1
-1
@@ -10,7 +10,7 @@ logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def build():
|
||||
@ui.page("/")
|
||||
@ui.page("/", response_timeout=30)
|
||||
async def index(client: Client) -> None:
|
||||
app.add_static_files("/static", "static")
|
||||
el.load_element_css()
|
||||
|
||||
@@ -13,6 +13,7 @@ class Result:
|
||||
stdout_lines: List[str] = field(default_factory=list)
|
||||
stderr_lines: List[str] = field(default_factory=list)
|
||||
terminated: bool = False
|
||||
truncated: bool = False
|
||||
data: Any = None
|
||||
trace: str = ""
|
||||
cached: bool = False
|
||||
|
||||
@@ -11,6 +11,7 @@ from apscheduler.schedulers.asyncio import AsyncIOScheduler
|
||||
@dataclass(kw_only=True)
|
||||
class Automation:
|
||||
id: str
|
||||
name: str
|
||||
app: str
|
||||
hosts: List[str]
|
||||
host: str
|
||||
@@ -30,6 +31,7 @@ class Automation:
|
||||
class Zfs_Autobackup(Automation):
|
||||
app: str = "zfs_autobackup"
|
||||
execute_mode: str = "local"
|
||||
prop: str
|
||||
target_host: str
|
||||
target_path: str
|
||||
target_paths: List[str]
|
||||
|
||||
@@ -97,12 +97,14 @@ class Tab:
|
||||
col.tailwind.max_width("lg")
|
||||
ui.label(f"Host Name: {result.name}").classes("text-secondary")
|
||||
ui.label(f"Command: {result.command}").classes("text-secondary")
|
||||
ui.label(f"Date: {result.date}").classes("text-secondary")
|
||||
timestamp = await ui.run_javascript(
|
||||
f"new Date({result.timestamp} * 1000).toLocaleString(undefined, {{dateStyle: 'short', timeStyle: 'short', hour12: 'false'}});"
|
||||
)
|
||||
ui.label(f"Timestamp: {timestamp}").classes("text-secondary")
|
||||
with ui.column() as col:
|
||||
col.tailwind.max_width("lg")
|
||||
ui.label(f"Task has failed: {result.failed}").classes("text-secondary")
|
||||
ui.label(f"Data is cached: {result.cached}").classes("text-secondary")
|
||||
ui.label(f"Time: {result.time}").classes("text-secondary")
|
||||
with el.Card() as card:
|
||||
with el.WColumn():
|
||||
terminal = cli.Terminal(options={"rows": 18, "cols": 120, "convertEol": True})
|
||||
|
||||
+60
-39
@@ -1,4 +1,4 @@
|
||||
from typing import Any, Dict, List, Union
|
||||
from typing import Any, Callable, Dict, List, Union
|
||||
import asyncio
|
||||
from datetime import datetime
|
||||
import json
|
||||
@@ -36,20 +36,20 @@ def populate_job_handler(app: str, job_id: str, host: str):
|
||||
return job_handlers[job_id]
|
||||
|
||||
|
||||
class CommandTemplate(string.Template):
|
||||
class AutomationTemplate(string.Template):
|
||||
delimiter = ""
|
||||
|
||||
|
||||
async def automation_job(**kwargs) -> None:
|
||||
if "data" in kwargs:
|
||||
jd = json.loads(kwargs["data"])
|
||||
command = CommandTemplate(jd["command"])
|
||||
command = AutomationTemplate(jd["command"])
|
||||
tab = Tab(host=None, spinner=None)
|
||||
if jd["app"] == "zfs_autobackup":
|
||||
d = scheduler.Zfs_Autobackup(**jd)
|
||||
populate_job_handler(app=d.app, job_id=d.id, host=d.host)
|
||||
if job_handlers[d.id].is_busy is False:
|
||||
result = await job_handlers[d.id].execute(command.safe_substitute(host=d.host))
|
||||
result = await job_handlers[d.id].execute(command.safe_substitute(name=d.name, host=d.host))
|
||||
result.name = d.host
|
||||
result.status = "success" if result.return_code == 0 else "error"
|
||||
if d.pipe_success is True and result.status == "success":
|
||||
@@ -63,7 +63,7 @@ async def automation_job(**kwargs) -> None:
|
||||
d = scheduler.Automation(**jd)
|
||||
populate_job_handler(app=d.app, job_id=d.id, host=d.host)
|
||||
if job_handlers[d.id].is_busy is False:
|
||||
result = await job_handlers[d.id].execute(command.safe_substitute(host=d.host))
|
||||
result = await job_handlers[d.id].execute(command.safe_substitute(name=d.name, host=d.host))
|
||||
result.name = d.host
|
||||
if d.pipe_success is True and result.status == "success":
|
||||
tab.pipe_result(result=result)
|
||||
@@ -76,7 +76,7 @@ async def automation_job(**kwargs) -> None:
|
||||
d = scheduler.Automation(**jd)
|
||||
populate_job_handler(app=d.app, job_id=d.id, host=d.host)
|
||||
if job_handlers[d.id].is_busy is False:
|
||||
result = await job_handlers[d.id].execute(command.safe_substitute(host=d.host))
|
||||
result = await job_handlers[d.id].execute(command.safe_substitute(name=d.name, host=d.host))
|
||||
result.name = d.host
|
||||
if d.pipe_success is True and result.status == "success":
|
||||
tab.pipe_result(result=result)
|
||||
@@ -99,7 +99,7 @@ class Automation(Tab):
|
||||
self.job_data: Dict[str, str] = {}
|
||||
self.job_names: List[str] = []
|
||||
self.default_options: Dict[str, str] = {}
|
||||
self.build_command: str = ""
|
||||
self.build_command: Callable
|
||||
self.target_host: el.DSelect
|
||||
self.target_paths: List[str] = [""]
|
||||
self.target_path: el.DSelect
|
||||
@@ -161,8 +161,16 @@ class Automation(Tab):
|
||||
"maxWidth": 150,
|
||||
},
|
||||
{"headerName": "Command", "field": "command", "filter": "agTextColumnFilter"},
|
||||
{"headerName": "Next Date", "field": "next_run_date", "filter": "agDateColumnFilter", "maxWidth": 100},
|
||||
{"headerName": "Next Time", "field": "next_run_time", "maxWidth": 100},
|
||||
{
|
||||
"headerName": "Next Run",
|
||||
"field": "next_run",
|
||||
"filter": "agTextColumnFilter",
|
||||
"maxWidth": 125,
|
||||
":cellRenderer": """(data) => {
|
||||
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
|
||||
return date;
|
||||
}""",
|
||||
},
|
||||
{
|
||||
"headerName": "Status",
|
||||
"field": "status",
|
||||
@@ -227,11 +235,9 @@ class Automation(Tab):
|
||||
self._automations.clear()
|
||||
for job in self.scheduler.scheduler.get_jobs():
|
||||
if job.next_run_time is not None:
|
||||
next_run_date = job.next_run_time.strftime("%Y/%m/%d")
|
||||
next_run_time = job.next_run_time.strftime("%H:%M")
|
||||
next_run = job.next_run_time.timestamp()
|
||||
else:
|
||||
next_run_date = "NA"
|
||||
next_run_time = "NA"
|
||||
next_run = "NA"
|
||||
if "data" in job.kwargs:
|
||||
jd = json.loads(job.kwargs["data"])
|
||||
if self.host == jd["host"]:
|
||||
@@ -239,8 +245,7 @@ class Automation(Tab):
|
||||
{
|
||||
"name": job.id.split("@")[0],
|
||||
"command": jd["command"],
|
||||
"next_run_date": next_run_date,
|
||||
"next_run_time": next_run_time,
|
||||
"next_run": next_run,
|
||||
"status": "",
|
||||
}
|
||||
)
|
||||
@@ -303,26 +308,17 @@ class Automation(Tab):
|
||||
await self._create_automation(rows[0]["name"])
|
||||
self._set_selection()
|
||||
|
||||
async def _add_prop_to_fs(
|
||||
self,
|
||||
host: str,
|
||||
prop: str,
|
||||
value: str,
|
||||
module: str = "autobackup",
|
||||
filesystems: Union[List[str], None] = None,
|
||||
) -> None:
|
||||
async def _add_prop_to_fs(self, host: str, prop: str, value: str, filesystems: Union[List[str], None] = None) -> None:
|
||||
if filesystems is not None:
|
||||
full_prop = f"{module}:{prop}"
|
||||
for fs in filesystems:
|
||||
result = await self._zfs[host].add_filesystem_prop(filesystem=fs, prop=full_prop, value=value)
|
||||
result = await self._zfs[host].add_filesystem_prop(filesystem=fs, prop=prop, value=value)
|
||||
self.add_history(result=result)
|
||||
|
||||
async def _remove_prop_from_all_fs(self, host: str, prop: str, module: str = "autobackup") -> None:
|
||||
full_prop = f"{module}:{prop}"
|
||||
filesystems_with_prop_result = await self._zfs[host].filesystems_with_prop(full_prop)
|
||||
async def _remove_prop_from_all_fs(self, host: str, prop: str) -> None:
|
||||
filesystems_with_prop_result = await self._zfs[host].filesystems_with_prop(prop)
|
||||
filesystems_with_prop = list(filesystems_with_prop_result.data)
|
||||
for fs in filesystems_with_prop:
|
||||
result = await self._zfs[host].remove_filesystem_prop(filesystem=fs, prop=full_prop)
|
||||
result = await self._zfs[host].remove_filesystem_prop(filesystem=fs, prop=prop)
|
||||
self.add_history(result=result)
|
||||
|
||||
async def _create_automation(self, name: str = "") -> None:
|
||||
@@ -416,15 +412,16 @@ class Automation(Tab):
|
||||
self.target_path.update()
|
||||
self.target_path.value = ""
|
||||
|
||||
async def target_path_selected() -> None:
|
||||
self.build_command()
|
||||
|
||||
def build_command() -> None:
|
||||
try:
|
||||
prop_suffix = self.prop.value.split(":")[1]
|
||||
except IndexError:
|
||||
prop_suffix = ""
|
||||
base = ""
|
||||
for key, value in self.picked_options.items():
|
||||
base = base + f" --{key}{f' {value}' if value != '' else ''}"
|
||||
target_path = f"{f' {self.target_path.value}' if self.target_path.value != '' else ''}"
|
||||
base = base + f" {self.auto_name.value.lower()}" + target_path
|
||||
base = base + f" {prop_suffix}" + target_path
|
||||
self.command.value = base
|
||||
|
||||
def all_fs_to_lists():
|
||||
@@ -471,6 +468,17 @@ class Automation(Tab):
|
||||
self.children.update()
|
||||
self.exclude.update()
|
||||
|
||||
def validate_prop(value):
|
||||
parts = value.split(":")
|
||||
for part in parts:
|
||||
if part.find(" ") != -1:
|
||||
return False
|
||||
if len(part) < 1:
|
||||
return False
|
||||
if len(parts) != 2:
|
||||
return False
|
||||
return True
|
||||
|
||||
if name == "":
|
||||
self.default_options = {
|
||||
"verbose": "",
|
||||
@@ -491,9 +499,11 @@ class Automation(Tab):
|
||||
row.tailwind.width("[860px]").justify_content("center")
|
||||
with ui.column() as col:
|
||||
col.tailwind.height("full").width("[420px]")
|
||||
self.prop = el.DInput(label="Property", value="autobackup:{name}", on_change=build_command, validation=validate_prop)
|
||||
self.app_em.append(self.prop)
|
||||
self.target_host = el.DSelect(target_host, label="Target Host", on_change=target_host_selected)
|
||||
self.target_paths = [""]
|
||||
self.target_path = el.DSelect(self.target_paths, value="", label="Target Path", new_value_mode="add-unique", on_change=target_path_selected)
|
||||
self.target_path = el.DSelect(self.target_paths, value="", label="Target Path", new_value_mode="add-unique", on_change=build_command)
|
||||
self.hosts = el.DSelect(source_hosts, label="Source Host(s)", multiple=True, with_input=True)
|
||||
all_fs_to_lists()
|
||||
with ui.scroll_area().classes("col"):
|
||||
@@ -529,12 +539,15 @@ class Automation(Tab):
|
||||
col.tailwind.height("full").width("[420px]")
|
||||
options_controls()
|
||||
if name != "":
|
||||
self.prop.value = self.job_data.get("prop", "autobackup:{name}")
|
||||
self.target_host.value = self.job_data.get("target_host", "")
|
||||
target_path = self.job_data.get("target_path", "")
|
||||
tries = 0
|
||||
while target_path not in self.target_path.options and tries < 20:
|
||||
await asyncio.sleep(0.1)
|
||||
tries = tries + 1
|
||||
if target_path not in self.target_paths:
|
||||
self.target_paths.append(target_path)
|
||||
self.target_path.value = target_path
|
||||
self.parentchildren.value = self.fs["values"].get("parentchildren", None)
|
||||
self.parent.value = self.fs["values"].get("parent", None)
|
||||
@@ -751,7 +764,9 @@ class Automation(Tab):
|
||||
with el.WRow() as row:
|
||||
row.tailwind.height("[40px]")
|
||||
self.as_spinner = el.Spinner()
|
||||
self.app_em = el.ErrorAggregator()
|
||||
self.save = el.DButton("SAVE", on_click=lambda: automation_dialog.submit("save"))
|
||||
self.save.bind_enabled_from(self.app_em, "no_errors")
|
||||
el.Spinner(master=self.as_spinner)
|
||||
self.auto_name.value = name
|
||||
if name != "":
|
||||
@@ -772,11 +787,13 @@ class Automation(Tab):
|
||||
self.scheduler.scheduler.remove_job(job.id)
|
||||
for host in hosts:
|
||||
auto_id = f"{auto_name}@{host}"
|
||||
await self._remove_prop_from_all_fs(host=host, prop=auto_name)
|
||||
await self._add_prop_to_fs(host=host, prop=auto_name, value="true", filesystems=self.parentchildren.value)
|
||||
await self._add_prop_to_fs(host=host, prop=auto_name, value="parent", filesystems=self.parent.value)
|
||||
await self._add_prop_to_fs(host=host, prop=auto_name, value="child", filesystems=self.children.value)
|
||||
await self._add_prop_to_fs(host=host, prop=auto_name, value="false", filesystems=self.exclude.value)
|
||||
command = AutomationTemplate(self.prop.value)
|
||||
prop = command.safe_substitute(name=auto_name, host=host)
|
||||
await self._remove_prop_from_all_fs(host=host, prop=prop)
|
||||
await self._add_prop_to_fs(host=host, prop=prop, value="true", filesystems=self.parentchildren.value)
|
||||
await self._add_prop_to_fs(host=host, prop=prop, value="parent", filesystems=self.parent.value)
|
||||
await self._add_prop_to_fs(host=host, prop=prop, value="child", filesystems=self.children.value)
|
||||
await self._add_prop_to_fs(host=host, prop=prop, value="false", filesystems=self.exclude.value)
|
||||
self.fs["values"] = {}
|
||||
self.fs["values"]["parentchildren"] = self.parentchildren.value
|
||||
self.fs["values"]["parent"] = self.parent.value
|
||||
@@ -784,6 +801,7 @@ class Automation(Tab):
|
||||
self.fs["values"]["exclude"] = self.exclude.value
|
||||
auto = scheduler.Zfs_Autobackup(
|
||||
id=auto_id,
|
||||
name=auto_name,
|
||||
hosts=hosts,
|
||||
host=host,
|
||||
command="python -m zfs_autobackup.ZfsAutobackup" + self.command.value,
|
||||
@@ -796,6 +814,7 @@ class Automation(Tab):
|
||||
filesystems=self.fs,
|
||||
pipe_success=self.pipe_success.value,
|
||||
pipe_error=self.pipe_error.value,
|
||||
prop=self.prop.value,
|
||||
)
|
||||
self.scheduler.scheduler.add_job(
|
||||
automation_job,
|
||||
@@ -815,6 +834,7 @@ class Automation(Tab):
|
||||
auto_id = f"{auto_name}@{host}"
|
||||
auto = scheduler.Automation(
|
||||
id=auto_id,
|
||||
name=auto_name,
|
||||
app=self.app.value,
|
||||
hosts=hosts,
|
||||
host=host,
|
||||
@@ -837,6 +857,7 @@ class Automation(Tab):
|
||||
auto_id = f"{auto_name}@{self.host}"
|
||||
auto = scheduler.Automation(
|
||||
id=auto_id,
|
||||
name=auto_name,
|
||||
app=self.app.value,
|
||||
hosts=hosts,
|
||||
host=self.host,
|
||||
|
||||
+33
-6
@@ -1,7 +1,8 @@
|
||||
from datetime import datetime
|
||||
import json
|
||||
from . import SelectionConfirm, Tab
|
||||
from nicegui import ui, events
|
||||
import httpx
|
||||
from . import SelectionConfirm, Tab
|
||||
from bale import elements as el
|
||||
from bale.result import Result
|
||||
from bale.interfaces import zfs
|
||||
@@ -47,8 +48,16 @@ class History(Tab):
|
||||
"filter": "agTextColumnFilter",
|
||||
"flex": 1,
|
||||
},
|
||||
{"headerName": "Date", "field": "date", "filter": "agDateColumnFilter", "maxWidth": 100},
|
||||
{"headerName": "Time", "field": "time", "maxWidth": 100},
|
||||
{
|
||||
"headerName": "Timestamp",
|
||||
"field": "timestamp",
|
||||
"filter": "agTextColumnFilter",
|
||||
"maxWidth": 125,
|
||||
":cellRenderer": """(data) => {
|
||||
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
|
||||
return date;
|
||||
}""",
|
||||
},
|
||||
{
|
||||
"headerName": "Status",
|
||||
"field": "status",
|
||||
@@ -90,6 +99,20 @@ class History(Tab):
|
||||
http[status]["data"] = e.content["json"]["data"]
|
||||
http[status]["headers"] = e.content["json"]["headers"]
|
||||
|
||||
def test(status):
|
||||
try:
|
||||
url = http[status]["url"]
|
||||
data = self.process_pipe_data(result=Result(name=self.host, command="TEST COMMAND", status=status), data=http[status]["data"])
|
||||
headers = http[status]["headers"]
|
||||
post = httpx.post(url=url, json=data, headers=headers)
|
||||
print(post.status_code)
|
||||
if post.status_code == 200:
|
||||
el.notify("Test successful!", type="positive")
|
||||
else:
|
||||
el.notify(f"Test failed with status code {post.status_code}!", type="negative")
|
||||
except:
|
||||
el.notify("Test failed!", type="negative")
|
||||
|
||||
def show_controls(status):
|
||||
if status not in http:
|
||||
http[status] = {}
|
||||
@@ -107,7 +130,7 @@ class History(Tab):
|
||||
"topic": "mytopic",
|
||||
"tags": ["turtle"],
|
||||
"title": "Successful Automation Run for {name}",
|
||||
"message": "{stdout}",
|
||||
"message": "{command}",
|
||||
},
|
||||
)
|
||||
editor.properties["content"]["json"]["headers"] = self.get_pipe_status("http", status).get("headers", {"Authorization": "Bearer tk_..."})
|
||||
@@ -126,11 +149,15 @@ class History(Tab):
|
||||
with ui.step("On Success"):
|
||||
with el.WColumn().classes("col justify-start"):
|
||||
show_controls(status="success")
|
||||
el.LgButton("NEXT", on_click=lambda _: stepper.next())
|
||||
with el.WRow():
|
||||
el.LgButton("TEST", on_click=lambda _: test(status="success"))
|
||||
el.LgButton("NEXT", on_click=lambda _: stepper.next())
|
||||
with ui.step("On Error"):
|
||||
with el.WColumn().classes("col justify-start"):
|
||||
show_controls(status="error")
|
||||
el.DButton("SAVE", on_click=lambda: host_dialog.submit("save"))
|
||||
with el.WRow():
|
||||
el.LgButton("TEST", on_click=lambda _: test(status="error"))
|
||||
el.DButton("SAVE", on_click=lambda: host_dialog.submit("save"))
|
||||
|
||||
result = await host_dialog
|
||||
if result == "save":
|
||||
|
||||
+21
-10
@@ -1,6 +1,6 @@
|
||||
import asyncio
|
||||
from copy import deepcopy
|
||||
from nicegui import ui
|
||||
from nicegui import background_tasks, ui
|
||||
from . import SelectionConfirm, Tab, Task
|
||||
from bale.result import Result
|
||||
from bale import elements as el
|
||||
@@ -75,11 +75,19 @@ class Manage(Tab):
|
||||
"field": "used",
|
||||
"maxWidth": 100,
|
||||
":comparator": """(valueA, valueB, nodeA, nodeB, isInverted) => {
|
||||
return (nodeA.data.used_bytes > nodeB.data.used_bytes) ? -1 : 1;
|
||||
}""",
|
||||
return (nodeA.data.used_bytes > nodeB.data.used_bytes) ? -1 : 1;
|
||||
}""",
|
||||
},
|
||||
{
|
||||
"headerName": "Created",
|
||||
"field": "creation",
|
||||
"filter": "agTextColumnFilter",
|
||||
"maxWidth": 125,
|
||||
":cellRenderer": """(data) => {
|
||||
var date = new Date(data.value * 1000).toLocaleString(undefined, {dateStyle: 'short', timeStyle: 'short', hour12: false});;
|
||||
return date;
|
||||
}""",
|
||||
},
|
||||
{"headerName": "Creation Date", "field": "creation_date", "filter": "agDateColumnFilter", "maxWidth": 150},
|
||||
{"headerName": "Creation Time", "field": "creation_time", "maxWidth": 150},
|
||||
{"headerName": "Holds", "field": "userrefs", "filter": "agNumberColumnFilter", "maxWidth": 100},
|
||||
],
|
||||
"rowData": [],
|
||||
@@ -92,8 +100,8 @@ class Manage(Tab):
|
||||
self._spinner.visible = True
|
||||
self.zfs.invalidate_query()
|
||||
snapshots = await self.zfs.snapshots
|
||||
await self.zfs.filesystems
|
||||
await self.zfs.holds_for_snapshot()
|
||||
background_tasks.create(self.zfs.filesystems, name="zfs_filesystems")
|
||||
background_tasks.create(self.zfs.holds_for_snapshot(), name="zfs_holds")
|
||||
self._grid.options["rowData"] = list(snapshots.data.values())
|
||||
self._grid.update()
|
||||
self._spinner.visible = False
|
||||
@@ -103,9 +111,12 @@ class Manage(Tab):
|
||||
result = await SelectionConfirm(container=self._confirm, label=">BROWSE<")
|
||||
if result == "confirm":
|
||||
rows = await self._grid.get_selected_rows()
|
||||
filesystems = await self.zfs.filesystems
|
||||
mount_path = filesystems.data[rows[0]["filesystem"]]["mountpoint"]
|
||||
await sshdl.SshFileBrowse(zfs=self.zfs, path=f"{mount_path}/.zfs/snapshot/{rows[0]['name']}")
|
||||
try:
|
||||
filesystems = await self.zfs.filesystems
|
||||
mount_path = filesystems.data[rows[0]["filesystem"]]["mountpoint"]
|
||||
await sshdl.SshFileBrowse(zfs=self.zfs, path=f"{mount_path}/.zfs/snapshot/{rows[0]['name']}")
|
||||
except KeyError:
|
||||
el.notify(f"Unable to browse {rows[0]['filesystem']}", type="warning")
|
||||
self._set_selection()
|
||||
|
||||
async def _find(self) -> None:
|
||||
|
||||
Reference in New Issue
Block a user