Compare commits
49 Commits
dbd56c1b50
...
v0.0.1-alp
| Author | SHA1 | Date | |
|---|---|---|---|
| 8dfcbc5720 | |||
| 103ae77e9f | |||
| beeccc6e8d | |||
| 0880298cf5 | |||
| 34b0abac36 | |||
| 28c226ddbc | |||
| 42861cc69e | |||
| 5f3d683a13 | |||
| a17787e852 | |||
| 5865ac3b99 | |||
| 637de857f9 | |||
| 3ecf5fb916 | |||
| 92ba3ef512 | |||
| 7d6c2db89c | |||
| 74262beb65 | |||
| f3b8dd94e5 | |||
| 0059b9b850 | |||
| 1ad789b2b9 | |||
| 079478f932 | |||
| d6d5b451cd | |||
| 76747cf917 | |||
| 6e85991062 | |||
| 98e408cb85 | |||
| ed052dff3c | |||
| 8f59bba614 | |||
| fb2c5609aa | |||
| 17aed6cb89 | |||
| b02b93b83f | |||
| 9ceba8b5bb | |||
| 2c0dbf95c7 | |||
| 860207a60b | |||
| 5c6460012a | |||
| be1d4081e0 | |||
| 83a94cacf3 | |||
| 0ce3790675 | |||
| 5854889eb5 | |||
| 4caaf74569 | |||
| fe889ca757 | |||
| 699c124b0e | |||
| 7d55c5f431 | |||
| c4fd74fc93 | |||
| 3775760734 | |||
| 643d12ff18 | |||
| 82eaa23da7 | |||
| b18d1ced6d | |||
| 69c5cf87fd | |||
| 1fadf0ad25 | |||
| beae6eb648 | |||
| 82ab735982 |
@@ -1,8 +0,0 @@
|
|||||||
# Changesets
|
|
||||||
|
|
||||||
Hello and welcome! This folder has been automatically generated by `@changesets/cli`, a build tool that works
|
|
||||||
with multi-package repos, or single-package repos to help you version and publish your code. You can
|
|
||||||
find the full documentation for it [in our repository](https://github.com/changesets/changesets)
|
|
||||||
|
|
||||||
We have a quick list of common questions to get you started engaging with this project in
|
|
||||||
[our documentation](https://github.com/changesets/changesets/blob/main/docs/common-questions.md)
|
|
||||||
@@ -1,11 +0,0 @@
|
|||||||
{
|
|
||||||
"$schema": "https://unpkg.com/@changesets/config@3.1.2/schema.json",
|
|
||||||
"changelog": "@changesets/cli/changelog",
|
|
||||||
"commit": false,
|
|
||||||
"fixed": [],
|
|
||||||
"linked": [],
|
|
||||||
"access": "restricted",
|
|
||||||
"baseBranch": "main",
|
|
||||||
"updateInternalDependencies": "patch",
|
|
||||||
"ignore": []
|
|
||||||
}
|
|
||||||
@@ -6,4 +6,7 @@ Dockerfile
|
|||||||
docker-compose.yml
|
docker-compose.yml
|
||||||
npm-debug.log
|
npm-debug.log
|
||||||
builds
|
builds
|
||||||
testFiles
|
testFiles
|
||||||
|
nssm.exe
|
||||||
|
postgresql-17.9-2-windows-x64.exe
|
||||||
|
VSCodeUserSetup-x64-1.112.0.msi
|
||||||
31
.gitea/workflows/docker-build.yml
Normal file
31
.gitea/workflows/docker-build.yml
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
name: Build and Push LST Docker Image
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
docker:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout (local)
|
||||||
|
run: |
|
||||||
|
git clone https://git.tuffraid.net/cowch/lst_v3.git .
|
||||||
|
git checkout ${{ gitea.sha }}
|
||||||
|
|
||||||
|
- name: Login to registry
|
||||||
|
run: echo "${{ secrets.PASSWORD }}" | docker login git.tuffraid.net -u "cowch" --password-stdin
|
||||||
|
|
||||||
|
- name: Build image
|
||||||
|
run: |
|
||||||
|
docker build \
|
||||||
|
-t git.tuffraid.net/cowch/lst_v3:latest \
|
||||||
|
-t git.tuffraid.net/cowch/lst_v3:${{ gitea.sha }} \
|
||||||
|
.
|
||||||
|
|
||||||
|
- name: Push
|
||||||
|
run: |
|
||||||
|
docker push git.tuffraid.net/cowch/lst_v3:latest
|
||||||
|
docker push git.tuffraid.net/cowch/lst_v3:${{ gitea.sha }}
|
||||||
229
.gitea/workflows/release.yml
Normal file
229
.gitea/workflows/release.yml
Normal file
@@ -0,0 +1,229 @@
|
|||||||
|
name: Release and Build Image
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
tags:
|
||||||
|
- "v*"
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
release:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
env:
|
||||||
|
# Internal/origin Gitea URL. Do NOT use the Cloudflare fronted URL here.
|
||||||
|
# Examples:
|
||||||
|
# http://gitea.internal.lan:3000
|
||||||
|
# https://gitea-origin.yourdomain.local
|
||||||
|
GITEA_INTERNAL_URL: "https://git.tuffraid.net"
|
||||||
|
|
||||||
|
# Internal/origin registry host. Usually same host as above, but without protocol.
|
||||||
|
# Example:
|
||||||
|
# gitea.internal:3000
|
||||||
|
REGISTRY_HOST: "git.tuffraid.net"
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Check out repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Prepare release metadata
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
TAG="${GITHUB_REF_NAME:-${GITHUB_REF##refs/tags/}}"
|
||||||
|
VERSION="${TAG#v}"
|
||||||
|
IMAGE_NAME="${REGISTRY_HOST}/${{ gitea.repository }}"
|
||||||
|
|
||||||
|
echo "TAG=$TAG" >> "$GITHUB_ENV"
|
||||||
|
echo "VERSION=$VERSION" >> "$GITHUB_ENV"
|
||||||
|
echo "IMAGE_NAME=$IMAGE_NAME" >> "$GITHUB_ENV"
|
||||||
|
|
||||||
|
if [[ "$TAG" == *-* ]]; then
|
||||||
|
echo "PRERELEASE=true" >> "$GITHUB_ENV"
|
||||||
|
else
|
||||||
|
echo "PRERELEASE=false" >> "$GITHUB_ENV"
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Resolved TAG=$TAG"
|
||||||
|
echo "Resolved VERSION=$VERSION"
|
||||||
|
echo "Resolved IMAGE_NAME=$IMAGE_NAME"
|
||||||
|
|
||||||
|
- name: Log in to Gitea container registry
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
REGISTRY_USERNAME: ${{ secrets.REGISTRY_USERNAME }}
|
||||||
|
REGISTRY_TOKEN: ${{ secrets.RELEASE_TOKEN }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
echo "$REGISTRY_TOKEN" | docker login "$REGISTRY_HOST" -u "$REGISTRY_USERNAME" --password-stdin
|
||||||
|
|
||||||
|
- name: Build Docker image
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
docker build \
|
||||||
|
-t "$IMAGE_NAME:$TAG" \
|
||||||
|
-t "$IMAGE_NAME:latest" \
|
||||||
|
.
|
||||||
|
|
||||||
|
- name: Push version tag
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
docker push "$IMAGE_NAME:$TAG"
|
||||||
|
|
||||||
|
- name: Push latest tag
|
||||||
|
if: ${{ !contains(env.TAG, '-') }}
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
docker push "$IMAGE_NAME:latest"
|
||||||
|
|
||||||
|
- name: Push prerelease channel tag
|
||||||
|
if: ${{ contains(env.TAG, '-') }}
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
TAG: ${{ env.TAG }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
CHANNEL="${TAG#*-}"
|
||||||
|
CHANNEL="${CHANNEL%%.*}"
|
||||||
|
|
||||||
|
echo "Resolved prerelease channel: $CHANNEL"
|
||||||
|
|
||||||
|
docker tag "$IMAGE_NAME:$TAG" "$IMAGE_NAME:$CHANNEL"
|
||||||
|
docker push "$IMAGE_NAME:$CHANNEL"
|
||||||
|
|
||||||
|
- name: Extract matching CHANGELOG section
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
VERSION: ${{ env.VERSION }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
python3 - <<'PY'
|
||||||
|
import os
|
||||||
|
import re
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
version = os.environ["VERSION"]
|
||||||
|
changelog_path = Path("CHANGELOG.md")
|
||||||
|
|
||||||
|
if not changelog_path.exists():
|
||||||
|
Path("release_body.md").write_text(f"Release {version}\n", encoding="utf-8")
|
||||||
|
raise SystemExit(0)
|
||||||
|
|
||||||
|
text = changelog_path.read_text(encoding="utf-8")
|
||||||
|
|
||||||
|
# Matches headings like:
|
||||||
|
# ## [0.1.0]
|
||||||
|
# ## 0.1.0
|
||||||
|
# ## [0.1.0-alpha.1]
|
||||||
|
pattern = re.compile(
|
||||||
|
rf"^##\s+\[?{re.escape(version)}\]?[^\n]*\n(.*?)(?=^##\s+\[?[^\n]+|\Z)",
|
||||||
|
re.MULTILINE | re.DOTALL,
|
||||||
|
)
|
||||||
|
|
||||||
|
match = pattern.search(text)
|
||||||
|
if match:
|
||||||
|
body = match.group(1).strip()
|
||||||
|
else:
|
||||||
|
body = f"Release {version}"
|
||||||
|
|
||||||
|
if not body:
|
||||||
|
body = f"Release {version}"
|
||||||
|
|
||||||
|
Path("release_body.md").write_text(body + "\n", encoding="utf-8")
|
||||||
|
print("----- release_body.md -----")
|
||||||
|
print(body)
|
||||||
|
print("---------------------------")
|
||||||
|
PY
|
||||||
|
|
||||||
|
- name: Create Gitea release
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
RELEASE_TOKEN: ${{ secrets.RELEASE_TOKEN }}
|
||||||
|
GITEA_REPOSITORY: ${{ gitea.repository }}
|
||||||
|
GITEA_INTERNAL_URL: ${{ env.GITEA_INTERNAL_URL }}
|
||||||
|
TAG: ${{ env.TAG }}
|
||||||
|
PRERELEASE: ${{ env.PRERELEASE }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
python3 - <<'PY'
|
||||||
|
import json
|
||||||
|
import os
|
||||||
|
import urllib.request
|
||||||
|
import urllib.error
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
tag = os.environ["TAG"]
|
||||||
|
prerelease = os.environ["PRERELEASE"].lower() == "true"
|
||||||
|
server_url = os.environ["GITEA_INTERNAL_URL"].rstrip("/")
|
||||||
|
repo = os.environ["GITEA_REPOSITORY"]
|
||||||
|
token = os.environ["RELEASE_TOKEN"]
|
||||||
|
|
||||||
|
body = Path("release_body.md").read_text(encoding="utf-8").strip()
|
||||||
|
|
||||||
|
# Check if the release already exists for this tag
|
||||||
|
get_url = f"{server_url}/api/v1/repos/{repo}/releases/tags/{tag}"
|
||||||
|
get_req = urllib.request.Request(
|
||||||
|
get_url,
|
||||||
|
method="GET",
|
||||||
|
headers={
|
||||||
|
"Authorization": f"token {token}",
|
||||||
|
"Accept": "application/json",
|
||||||
|
"User-Agent": "lst-release-workflow/1.0",
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
existing_release = None
|
||||||
|
try:
|
||||||
|
with urllib.request.urlopen(get_req) as resp:
|
||||||
|
existing_release = json.loads(resp.read().decode("utf-8"))
|
||||||
|
except urllib.error.HTTPError as e:
|
||||||
|
if e.code != 404:
|
||||||
|
details = e.read().decode("utf-8", errors="replace")
|
||||||
|
print("Failed checking existing release:")
|
||||||
|
print(details)
|
||||||
|
raise
|
||||||
|
|
||||||
|
payload = {
|
||||||
|
"tag_name": tag,
|
||||||
|
"name": tag,
|
||||||
|
"body": body,
|
||||||
|
"draft": False,
|
||||||
|
"prerelease": prerelease,
|
||||||
|
}
|
||||||
|
|
||||||
|
data = json.dumps(payload).encode("utf-8")
|
||||||
|
|
||||||
|
if existing_release:
|
||||||
|
release_id = existing_release["id"]
|
||||||
|
url = f"{server_url}/api/v1/repos/{repo}/releases/{release_id}"
|
||||||
|
method = "PATCH"
|
||||||
|
print(f"Release already exists for tag {tag}, updating release id {release_id}")
|
||||||
|
else:
|
||||||
|
url = f"{server_url}/api/v1/repos/{repo}/releases"
|
||||||
|
method = "POST"
|
||||||
|
print(f"No release exists for tag {tag}, creating a new one")
|
||||||
|
|
||||||
|
req = urllib.request.Request(
|
||||||
|
url,
|
||||||
|
data=data,
|
||||||
|
method=method,
|
||||||
|
headers={
|
||||||
|
"Authorization": f"token {token}",
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
"Accept": "application/json",
|
||||||
|
"User-Agent": "lst-release-workflow/1.0",
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
try:
|
||||||
|
with urllib.request.urlopen(req) as resp:
|
||||||
|
print(resp.read().decode("utf-8"))
|
||||||
|
except urllib.error.HTTPError as e:
|
||||||
|
details = e.read().decode("utf-8", errors="replace")
|
||||||
|
print("Release create/update failed:")
|
||||||
|
print(details)
|
||||||
|
raise
|
||||||
|
PY
|
||||||
2
.gitignore
vendored
2
.gitignore
vendored
@@ -4,10 +4,12 @@ builds
|
|||||||
.includes
|
.includes
|
||||||
.buildNumber
|
.buildNumber
|
||||||
temp
|
temp
|
||||||
|
brunoApi
|
||||||
.scriptCreds
|
.scriptCreds
|
||||||
node-v24.14.0-x64.msi
|
node-v24.14.0-x64.msi
|
||||||
postgresql-17.9-2-windows-x64.exe
|
postgresql-17.9-2-windows-x64.exe
|
||||||
VSCodeUserSetup-x64-1.112.0.exe
|
VSCodeUserSetup-x64-1.112.0.exe
|
||||||
|
nssm.exe
|
||||||
|
|
||||||
# Logs
|
# Logs
|
||||||
logs
|
logs
|
||||||
|
|||||||
@@ -11,7 +11,7 @@
|
|||||||
{ "type": "ci", "hidden": false, "section": "📈 Project changes" },
|
{ "type": "ci", "hidden": false, "section": "📈 Project changes" },
|
||||||
{ "type": "build", "hidden": false, "section": "📈 Project Builds" }
|
{ "type": "build", "hidden": false, "section": "📈 Project Builds" }
|
||||||
],
|
],
|
||||||
"commitUrlFormat": "https://git.tuffraid.net/cowch/lst/commits/{{hash}}",
|
"commitUrlFormat": "https://git.tuffraid.net/cowch/lst_v3/commits/{{hash}}",
|
||||||
"compareUrlFormat": "https://git.tuffraid.net/cowch/lst/compare/{{previousTag}}...{{currentTag}}",
|
"compareUrlFormat": "https://git.tuffraid.net/cowch/lst_v3/compare/{{previousTag}}...{{currentTag}}",
|
||||||
"header": "# All Changes to LST can be found below.\n"
|
"header": "# All Changes to LST can be found below.\n"
|
||||||
}
|
}
|
||||||
2
.vscode/settings.json
vendored
2
.vscode/settings.json
vendored
@@ -54,8 +54,10 @@
|
|||||||
"alpla",
|
"alpla",
|
||||||
"alplamart",
|
"alplamart",
|
||||||
"alplaprod",
|
"alplaprod",
|
||||||
|
"alplapurchase",
|
||||||
"bookin",
|
"bookin",
|
||||||
"Datamart",
|
"Datamart",
|
||||||
|
"dotenvx",
|
||||||
"dyco",
|
"dyco",
|
||||||
"intiallally",
|
"intiallally",
|
||||||
"manadatory",
|
"manadatory",
|
||||||
|
|||||||
77
CHANGELOG.md
77
CHANGELOG.md
@@ -1,7 +1,76 @@
|
|||||||
# lst_v3
|
# All Changes to LST can be found below.
|
||||||
|
|
||||||
## 1.0.1
|
## [0.0.1-alpha.2](https://git.tuffraid.net/cowch/lst_v3/compare/v0.0.1-alpha.1...v0.0.1-alpha.2) (2026-04-08)
|
||||||
|
|
||||||
### Patch Changes
|
|
||||||
|
|
||||||
- cf18e94: core stuff
|
### 📈 Project Builds
|
||||||
|
|
||||||
|
* **release:** docker and release corrections ([103ae77](https://git.tuffraid.net/cowch/lst_v3/commits/103ae77e9f82fc008a8ae143b6feccc3ce802f8c))
|
||||||
|
|
||||||
|
## [0.0.1-alpha.1](https://git.tuffraid.net/cowch/lst_v3/compare/v0.0.1-alpha.0...v0.0.1-alpha.1) (2026-04-08)
|
||||||
|
|
||||||
|
|
||||||
|
* **notifcaion:** style changes to the notificaion card and started the table ([7d6c2db](https://git.tuffraid.net/cowch/lst_v3/commits/7d6c2db89cae1f137f126f5814dccd373f7ccb76))
|
||||||
|
|
||||||
|
|
||||||
|
### 🌟 Enhancements
|
||||||
|
|
||||||
|
* **notification:** base notifcaiton sub and admin compelted ([5865ac3](https://git.tuffraid.net/cowch/lst_v3/commits/5865ac3b99d60005c4245740369b0e0789c8fbbd))
|
||||||
|
* **notification:** reprint added ([a17787e](https://git.tuffraid.net/cowch/lst_v3/commits/a17787e85217f1fa4a5e5389e29c33ec09c286c5))
|
||||||
|
* **puchase history:** purhcase history changed to long running no notification ([34b0aba](https://git.tuffraid.net/cowch/lst_v3/commits/34b0abac36f645d0fe5f508881ddbef81ff04b7c))
|
||||||
|
* **purchase:** historical data capture for alpla purchase ([42861cc](https://git.tuffraid.net/cowch/lst_v3/commits/42861cc69e8d4aba5a9670aaed55417efda2b505))
|
||||||
|
* **user notifications:** added the ability for users to sub to notifications and add multi email ([637de85](https://git.tuffraid.net/cowch/lst_v3/commits/637de857f99499a41f7175181523f5d809d95d7e))
|
||||||
|
|
||||||
|
|
||||||
|
### 🐛 Bug fixes
|
||||||
|
|
||||||
|
* **build:** issue with how i wrote the release token ([fe889ca](https://git.tuffraid.net/cowch/lst_v3/commits/fe889ca75731af08c42ec714b7f2abf17cd1ee40))
|
||||||
|
* **build:** type in how we pushed the header over ([83a94ca](https://git.tuffraid.net/cowch/lst_v3/commits/83a94cacf3fc87287cdc0c0cc861b339e72e4b83))
|
||||||
|
* **build:** typo ([860207a](https://git.tuffraid.net/cowch/lst_v3/commits/860207a60b6e04b15736cba631be6c7eab74d020))
|
||||||
|
* **i suck:** more learning experance ([9ceba8b](https://git.tuffraid.net/cowch/lst_v3/commits/9ceba8b5bba17959f27b16b28f50a83c044863fb))
|
||||||
|
* **lala:** something here ([17aed6c](https://git.tuffraid.net/cowch/lst_v3/commits/17aed6cb89f8220570f6c66f78dba6bb202c1aaa))
|
||||||
|
* **release:** typo that caused errors ([76747cf](https://git.tuffraid.net/cowch/lst_v3/commits/76747cf91738bd0d0530afcf7b4f51f0db11ca98))
|
||||||
|
* **typo:** more dam typos ([079478f](https://git.tuffraid.net/cowch/lst_v3/commits/079478f93217dea31c9a1e8ffed85d2381a6977d))
|
||||||
|
* **wrelease:** forgot to save ([3775760](https://git.tuffraid.net/cowch/lst_v3/commits/377576073449e95d315defb913dc317759cc3f43))
|
||||||
|
|
||||||
|
|
||||||
|
### 📝 Chore
|
||||||
|
|
||||||
|
* **release:** 0.1.0-alpha.10 ([98e408c](https://git.tuffraid.net/cowch/lst_v3/commits/98e408cb8577da18e24821b55474198439434f3e))
|
||||||
|
* **release:** 0.1.0-alpha.11 ([d6d5b45](https://git.tuffraid.net/cowch/lst_v3/commits/d6d5b451cd9aeba642ef94654ca20f4acd0b827c))
|
||||||
|
* **release:** 0.1.0-alpha.12 ([1ad789b](https://git.tuffraid.net/cowch/lst_v3/commits/1ad789b2b91a20a2f5a8dc9e6f39af2e19ec9cdc))
|
||||||
|
* **release:** 0.1.0-alpha.9 ([8f59bba](https://git.tuffraid.net/cowch/lst_v3/commits/8f59bba614a8eaa3105bb56f0db36013d5e68485))
|
||||||
|
* **release:** version packages ([fb2c560](https://git.tuffraid.net/cowch/lst_v3/commits/fb2c5609aa12ea7823783c364d5bd029c48a64bd))
|
||||||
|
* **release:** version packages ([b02b93b](https://git.tuffraid.net/cowch/lst_v3/commits/b02b93b83f488fbcee6d24db080ad0d1fe1c5f59))
|
||||||
|
* **release:** version packages ([2c0dbf9](https://git.tuffraid.net/cowch/lst_v3/commits/2c0dbf95c7b8dfd2c98b476d3f44bc8929668c88))
|
||||||
|
* **release:** version packages ([5c64600](https://git.tuffraid.net/cowch/lst_v3/commits/5c6460012aa70d336fbc9702240b4f19262a6b41))
|
||||||
|
* **release:** version packages ([0ce3790](https://git.tuffraid.net/cowch/lst_v3/commits/0ce3790675bc408762eafe76cbd5ab496fd06e73))
|
||||||
|
* **release:** version packages ([4caaf74](https://git.tuffraid.net/cowch/lst_v3/commits/4caaf745693d4df847aefd3721ac5d0ae792114a))
|
||||||
|
* **release:** version packages ([699c124](https://git.tuffraid.net/cowch/lst_v3/commits/699c124b0efba8282e436210619504bda8878e90))
|
||||||
|
* **release:** version packages ([c4fd74f](https://git.tuffraid.net/cowch/lst_v3/commits/c4fd74fc93226cffd9e39602f507a05cd8ea628b))
|
||||||
|
|
||||||
|
|
||||||
|
### 📚 Documentation
|
||||||
|
|
||||||
|
* **readme:** updated progress data ([92ba3ef](https://git.tuffraid.net/cowch/lst_v3/commits/92ba3ef5121afd0d82d4f40a5a985e1fdc081011))
|
||||||
|
* **sop:** added more info ([be1d408](https://git.tuffraid.net/cowch/lst_v3/commits/be1d4081e07b0982b355a270b7850a852a4398f5))
|
||||||
|
|
||||||
|
|
||||||
|
### 🛠️ Code Refactor
|
||||||
|
|
||||||
|
* **build:** added in more info to the relase section ([5854889](https://git.tuffraid.net/cowch/lst_v3/commits/5854889eb5398feebda50a5d256ce7aec39ce112))
|
||||||
|
* **build:** changes to auto release when we cahnge version ([643d12f](https://git.tuffraid.net/cowch/lst_v3/commits/643d12ff182827e724e1569a583bd625a0d1dd0c))
|
||||||
|
* **build:** changes to the way we do release so it builds as well ([7d55c5f](https://git.tuffraid.net/cowch/lst_v3/commits/7d55c5f43173edb48d8709adcb972b7d8fbc3ebd))
|
||||||
|
* **changelog:** reverted back to commit-chagnelog, like more than changeset for solo dev ([ed052df](https://git.tuffraid.net/cowch/lst_v3/commits/ed052dff3c81a7064660a7d25685e0505065252c))
|
||||||
|
* **notification:** reprint - removed a console log as it shouldnt bc there ([5f3d683](https://git.tuffraid.net/cowch/lst_v3/commits/5f3d683a13c831229674166cced699e373131316))
|
||||||
|
* **notification:** select menu looks propper now ([74262be](https://git.tuffraid.net/cowch/lst_v3/commits/74262beb6596ddc971971cc9214a2688accf3a8e))
|
||||||
|
* **opendock refactor on how releases are posted:** this was a bug maybe just a better refactory ([0880298](https://git.tuffraid.net/cowch/lst_v3/commits/0880298cf53d83e487c706e73854e0874ae2d9da))
|
||||||
|
* **queries:** changed dev version to be 1500ms vs 5000ms ([f3b8dd9](https://git.tuffraid.net/cowch/lst_v3/commits/f3b8dd94e5ebae0cc4dd0a2689a19051942e94b8))
|
||||||
|
* **release:** changes to only have the changelog in the release ([6e85991](https://git.tuffraid.net/cowch/lst_v3/commits/6e8599106298ed13febd069d6fda8b354efb5b7b))
|
||||||
|
* **userprofile:** changes to have the table be blank and say nothing subscribed ([3ecf5fb](https://git.tuffraid.net/cowch/lst_v3/commits/3ecf5fb916d5dc1b1ffb224e2142d94f7a9cb126))
|
||||||
|
|
||||||
|
|
||||||
|
### 📈 Project Builds
|
||||||
|
|
||||||
|
* **agent:** added westbend into the flow ([28c226d](https://git.tuffraid.net/cowch/lst_v3/commits/28c226ddbc37ab85cd6a9a6aec091def3e5623d6))
|
||||||
|
* **changelog:** reset the change log after all crap testing ([0059b9b](https://git.tuffraid.net/cowch/lst_v3/commits/0059b9b850c9647695a3fecaf5927c2e3ee7b192))
|
||||||
|
|||||||
10
Dockerfile
10
Dockerfile
@@ -9,10 +9,13 @@ WORKDIR /app
|
|||||||
# Copy package files
|
# Copy package files
|
||||||
COPY . .
|
COPY . .
|
||||||
|
|
||||||
# Install production dependencies only
|
# build backend
|
||||||
RUN npm ci
|
RUN npm ci
|
||||||
|
RUN npm run build:docker
|
||||||
|
|
||||||
RUN npm run build
|
# build frontend
|
||||||
|
RUN npm --prefix frontend ci
|
||||||
|
RUN npm --prefix frontend run build
|
||||||
|
|
||||||
###########
|
###########
|
||||||
# Stage 2 #
|
# Stage 2 #
|
||||||
@@ -33,6 +36,9 @@ RUN npm ci --omit=dev
|
|||||||
|
|
||||||
|
|
||||||
COPY --from=build /app/dist ./dist
|
COPY --from=build /app/dist ./dist
|
||||||
|
COPY --from=build /app/frontend/dist ./frontend/dist
|
||||||
|
|
||||||
|
# TODO add in drizzle migrates
|
||||||
|
|
||||||
ENV RUNNING_IN_DOCKER=true
|
ENV RUNNING_IN_DOCKER=true
|
||||||
EXPOSE 3000
|
EXPOSE 3000
|
||||||
|
|||||||
10
README.md
10
README.md
@@ -7,7 +7,7 @@
|
|||||||
Quick summary of current rewrite/migration goal.
|
Quick summary of current rewrite/migration goal.
|
||||||
|
|
||||||
- **Phase:** Backend rewrite
|
- **Phase:** Backend rewrite
|
||||||
- **Last updated:** 2024-05-01
|
- **Last updated:** 2026-04-06
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
@@ -16,9 +16,9 @@ Quick summary of current rewrite/migration goal.
|
|||||||
| Feature | Description | Status |
|
| Feature | Description | Status |
|
||||||
|----------|--------------|--------|
|
|----------|--------------|--------|
|
||||||
| User Authentication | ~~Login~~, ~~Signup~~, API Key | 🟨 In Progress |
|
| User Authentication | ~~Login~~, ~~Signup~~, API Key | 🟨 In Progress |
|
||||||
| User Profile | Edit profile, upload avatar | ⏳ Not Started |
|
| User Profile | ~~Edit profile~~, upload avatar | 🟨 In Progress |
|
||||||
| User Admin | Edit user, create user, remove user, alplaprod user integration | ⏳ Not Started |
|
| User Admin | Edit user, create user, remove user, alplaprod user integration | ⏳ Not Started |
|
||||||
| Notifications | Subscribe, Create, Update, Remove, Manual Trigger | ⏳ Not Started |
|
| Notifications | ~~Subscribe~~, ~~Create~~, ~~Update~~, ~~~~Remove~~, Manual Trigger | 🟨 In Progress |
|
||||||
| Datamart | Create, Update, Run, Deactivate | 🔧 In Progress |
|
| Datamart | Create, Update, Run, Deactivate | 🔧 In Progress |
|
||||||
| Frontend | Analytics and charts | ⏳ Not Started |
|
| Frontend | Analytics and charts | ⏳ Not Started |
|
||||||
| Docs | Instructions and trouble shooting | ⏳ Not Started |
|
| Docs | Instructions and trouble shooting | ⏳ Not Started |
|
||||||
@@ -44,7 +44,7 @@ _Status legend:_
|
|||||||
How to run the current version of the app.
|
How to run the current version of the app.
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
git clone https://github.com/youruser/yourrepo.git
|
git clone https://git.tuffraid.net/cowch/lst_v3.git
|
||||||
cd yourrepo
|
cd lst_v3
|
||||||
npm install
|
npm install
|
||||||
npm run dev
|
npm run dev
|
||||||
@@ -26,7 +26,7 @@ const createApp = async () => {
|
|||||||
const __dirname = dirname(__filename);
|
const __dirname = dirname(__filename);
|
||||||
|
|
||||||
// well leave this active so we can monitor it to validate
|
// well leave this active so we can monitor it to validate
|
||||||
app.use(morgan("tiny"));
|
app.use(morgan("dev"));
|
||||||
app.set("trust proxy", true);
|
app.set("trust proxy", true);
|
||||||
app.use(lstCors());
|
app.use(lstCors());
|
||||||
app.all(`${baseUrl}/api/auth/*splat`, toNodeHandler(auth));
|
app.all(`${baseUrl}/api/auth/*splat`, toNodeHandler(auth));
|
||||||
@@ -34,11 +34,11 @@ const createApp = async () => {
|
|||||||
setupRoutes(baseUrl, app);
|
setupRoutes(baseUrl, app);
|
||||||
|
|
||||||
app.use(
|
app.use(
|
||||||
baseUrl + "/app",
|
`${baseUrl}/app`,
|
||||||
express.static(join(__dirname, "../frontend/dist")),
|
express.static(join(__dirname, "../frontend/dist")),
|
||||||
);
|
);
|
||||||
|
|
||||||
app.get(baseUrl + "/app/*splat", (_, res) => {
|
app.get(`${baseUrl}/app/*splat`, (_, res) => {
|
||||||
res.sendFile(join(__dirname, "../frontend/dist/index.html"));
|
res.sendFile(join(__dirname, "../frontend/dist/index.html"));
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
38
backend/db/schema/alplapurchase.schema.ts
Normal file
38
backend/db/schema/alplapurchase.schema.ts
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
import {
|
||||||
|
integer,
|
||||||
|
jsonb,
|
||||||
|
pgTable,
|
||||||
|
text,
|
||||||
|
timestamp,
|
||||||
|
uuid,
|
||||||
|
} from "drizzle-orm/pg-core";
|
||||||
|
import { createInsertSchema, createSelectSchema } from "drizzle-zod";
|
||||||
|
import type { z } from "zod";
|
||||||
|
|
||||||
|
export const alplaPurchaseHistory = pgTable("alpla_purchase_history", {
|
||||||
|
id: uuid("id").defaultRandom().primaryKey(),
|
||||||
|
apo: integer("apo"),
|
||||||
|
revision: integer("revision"),
|
||||||
|
confirmed: integer("confirmed"),
|
||||||
|
status: integer("status"),
|
||||||
|
statusText: text("status_text"),
|
||||||
|
journalNum: integer("journal_num"),
|
||||||
|
add_date: timestamp("add_date").defaultNow(),
|
||||||
|
add_user: text("add_user"),
|
||||||
|
upd_user: text("upd_user"),
|
||||||
|
upd_date: timestamp("upd_date").defaultNow(),
|
||||||
|
remark: text("remark"),
|
||||||
|
approvedStatus: text("approved_status").default("pending"),
|
||||||
|
position: jsonb("position").default([]),
|
||||||
|
createdAt: timestamp("created_at").defaultNow(),
|
||||||
|
});
|
||||||
|
|
||||||
|
export const alplaPurchaseHistorySchema =
|
||||||
|
createSelectSchema(alplaPurchaseHistory);
|
||||||
|
export const newAlplaPurchaseHistorySchema =
|
||||||
|
createInsertSchema(alplaPurchaseHistory);
|
||||||
|
|
||||||
|
export type AlplaPurchaseHistory = z.infer<typeof alplaPurchaseHistorySchema>;
|
||||||
|
export type NewAlplaPurchaseHistory = z.infer<
|
||||||
|
typeof newAlplaPurchaseHistorySchema
|
||||||
|
>;
|
||||||
@@ -1,4 +1,5 @@
|
|||||||
import {
|
import {
|
||||||
|
index,
|
||||||
integer,
|
integer,
|
||||||
jsonb,
|
jsonb,
|
||||||
pgTable,
|
pgTable,
|
||||||
@@ -9,14 +10,23 @@ import {
|
|||||||
import { createInsertSchema, createSelectSchema } from "drizzle-zod";
|
import { createInsertSchema, createSelectSchema } from "drizzle-zod";
|
||||||
import type { z } from "zod";
|
import type { z } from "zod";
|
||||||
|
|
||||||
export const opendockApt = pgTable("opendock_apt", {
|
export const opendockApt = pgTable(
|
||||||
id: uuid("id").defaultRandom().primaryKey(),
|
"opendock_apt",
|
||||||
release: integer("release").unique(),
|
{
|
||||||
openDockAptId: text("open_dock_apt_id").notNull(),
|
id: uuid("id").defaultRandom().primaryKey(),
|
||||||
appointment: jsonb("appointment").default([]),
|
release: integer("release").notNull().unique(),
|
||||||
upd_date: timestamp("upd_date").defaultNow(),
|
openDockAptId: text("open_dock_apt_id").notNull(),
|
||||||
createdAt: timestamp("created_at").defaultNow(),
|
appointment: jsonb("appointment").notNull().default([]),
|
||||||
});
|
upd_date: timestamp("upd_date").notNull().defaultNow(),
|
||||||
|
createdAt: timestamp("created_at").notNull().defaultNow(),
|
||||||
|
},
|
||||||
|
(table) => ({
|
||||||
|
releaseIdx: index("opendock_apt_release_idx").on(table.release),
|
||||||
|
openDockAptIdIdx: index("opendock_apt_opendock_id_idx").on(
|
||||||
|
table.openDockAptId,
|
||||||
|
),
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
|
||||||
export const opendockAptSchema = createSelectSchema(opendockApt);
|
export const opendockAptSchema = createSelectSchema(opendockApt);
|
||||||
export const newOpendockAptSchema = createInsertSchema(opendockApt);
|
export const newOpendockAptSchema = createInsertSchema(opendockApt);
|
||||||
|
|||||||
6
backend/db/schema/printerLogs.schema.ts
Normal file
6
backend/db/schema/printerLogs.schema.ts
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
import { integer, pgTable, text } from "drizzle-orm/pg-core";
|
||||||
|
|
||||||
|
export const opendockApt = pgTable("printer_log", {
|
||||||
|
id: integer().primaryKey().generatedAlwaysAsIdentity(),
|
||||||
|
name: text("name").notNull(),
|
||||||
|
});
|
||||||
113
backend/notification/notification.alplapurchase.ts
Normal file
113
backend/notification/notification.alplapurchase.ts
Normal file
@@ -0,0 +1,113 @@
|
|||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { notifications } from "../db/schema/notifications.schema.js";
|
||||||
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
|
import {
|
||||||
|
type SqlQuery,
|
||||||
|
sqlQuerySelector,
|
||||||
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { sendEmail } from "../utils/sendEmail.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
const func = async (data: any, emails: string) => {
|
||||||
|
// get the actual notification as items will be updated between intervals if no one touches
|
||||||
|
const { data: l, error: le } = (await tryCatch(
|
||||||
|
db.select().from(notifications).where(eq(notifications.id, data.id)),
|
||||||
|
)) as any;
|
||||||
|
|
||||||
|
if (le) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `${data.name} encountered an error while trying to get initial info`,
|
||||||
|
data: [le],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// search the query db for the query by name
|
||||||
|
const sqlQuery = sqlQuerySelector(`${data.name}`) as SqlQuery;
|
||||||
|
// create the ignore audit logs ids
|
||||||
|
const ignoreIds = l[0].options[0]?.auditId
|
||||||
|
? `${l[0].options[0]?.auditId}`
|
||||||
|
: "0";
|
||||||
|
|
||||||
|
// run the check
|
||||||
|
const { data: queryRun, error } = await tryCatch(
|
||||||
|
prodQuery(
|
||||||
|
sqlQuery.query
|
||||||
|
.replace("[intervalCheck]", l[0].interval)
|
||||||
|
.replace("[ignoreList]", ignoreIds),
|
||||||
|
`Running notification query: ${l[0].name}`,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
|
data: [error],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (queryRun.data.length > 0) {
|
||||||
|
// update the latest audit id
|
||||||
|
const { error: dbe } = await tryCatch(
|
||||||
|
db
|
||||||
|
.update(notifications)
|
||||||
|
.set({ options: [{ auditId: `${queryRun.data[0].id}` }] })
|
||||||
|
.where(eq(notifications.id, data.id)),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (dbe) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
|
data: [dbe],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// send the email
|
||||||
|
|
||||||
|
const sentEmail = await sendEmail({
|
||||||
|
email: emails,
|
||||||
|
subject: "Alert! Label Reprinted",
|
||||||
|
template: "reprintLabels",
|
||||||
|
context: {
|
||||||
|
items: queryRun.data,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!sentEmail?.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "email",
|
||||||
|
subModule: "notification",
|
||||||
|
message: `${l[0].name} failed to send the email`,
|
||||||
|
data: [sentEmail],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
console.log("doing nothing as there is nothing to do.");
|
||||||
|
}
|
||||||
|
// TODO send the error to systemAdmin users so they do not always need to be on the notifications.
|
||||||
|
// these errors are defined per notification.
|
||||||
|
};
|
||||||
|
|
||||||
|
export default func;
|
||||||
@@ -1,10 +1,106 @@
|
|||||||
const reprint = (data: any, emails: string) => {
|
import { eq } from "drizzle-orm";
|
||||||
// TODO: do the actual logic for the notification.
|
import { db } from "../db/db.controller.js";
|
||||||
console.log(data);
|
import { notifications } from "../db/schema/notifications.schema.js";
|
||||||
console.log(emails);
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
|
import {
|
||||||
|
type SqlQuery,
|
||||||
|
sqlQuerySelector,
|
||||||
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { sendEmail } from "../utils/sendEmail.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
// TODO send the error to systemAdmin users so they do not always need to be on the notifications.
|
const reprint = async (data: any, emails: string) => {
|
||||||
// these errors are defined per notification.
|
// TODO: do the actual logic for the notification.
|
||||||
|
const { data: l, error: le } = (await tryCatch(
|
||||||
|
db.select().from(notifications).where(eq(notifications.id, data.id)),
|
||||||
|
)) as any;
|
||||||
|
|
||||||
|
if (le) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `${data.name} encountered an error while trying to get initial info`,
|
||||||
|
data: [le],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// search the query db for the query by name
|
||||||
|
const sqlQuery = sqlQuerySelector(`${data.name}`) as SqlQuery;
|
||||||
|
// create the ignore audit logs ids
|
||||||
|
const ignoreIds = l[0].options[0]?.auditId
|
||||||
|
? `${l[0].options[0]?.auditId}`
|
||||||
|
: "0";
|
||||||
|
|
||||||
|
// run the check
|
||||||
|
const { data: queryRun, error } = await tryCatch(
|
||||||
|
prodQuery(
|
||||||
|
sqlQuery.query
|
||||||
|
.replace("[intervalCheck]", l[0].interval)
|
||||||
|
.replace("[ignoreList]", ignoreIds),
|
||||||
|
`Running notification query: ${l[0].name}`,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
|
data: [error],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (queryRun.data.length > 0) {
|
||||||
|
// update the latest audit id
|
||||||
|
const { error: dbe } = await tryCatch(
|
||||||
|
db
|
||||||
|
.update(notifications)
|
||||||
|
.set({ options: [{ auditId: `${queryRun.data[0].id}` }] })
|
||||||
|
.where(eq(notifications.id, data.id)),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (dbe) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
|
data: [dbe],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// send the email
|
||||||
|
|
||||||
|
const sentEmail = await sendEmail({
|
||||||
|
email: emails,
|
||||||
|
subject: "Alert! Label Reprinted",
|
||||||
|
template: "reprintLabels",
|
||||||
|
context: {
|
||||||
|
items: queryRun.data,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!sentEmail?.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "email",
|
||||||
|
subModule: "notification",
|
||||||
|
message: `${l[0].name} failed to send the email`,
|
||||||
|
data: [sentEmail],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
export default reprint;
|
export default reprint;
|
||||||
|
|||||||
@@ -3,12 +3,12 @@ import { type Response, Router } from "express";
|
|||||||
import z from "zod";
|
import z from "zod";
|
||||||
import { db } from "../db/db.controller.js";
|
import { db } from "../db/db.controller.js";
|
||||||
import { notificationSub } from "../db/schema/notifications.sub.schema.js";
|
import { notificationSub } from "../db/schema/notifications.sub.schema.js";
|
||||||
|
import { auth } from "../utils/auth.utils.js";
|
||||||
import { apiReturn } from "../utils/returnHelper.utils.js";
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
import { tryCatch } from "../utils/trycatch.utils.js";
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
import { modifiedNotification } from "./notification.controller.js";
|
import { modifiedNotification } from "./notification.controller.js";
|
||||||
|
|
||||||
const newSubscribe = z.object({
|
const newSubscribe = z.object({
|
||||||
emails: z.email().array().describe("An array of emails"),
|
|
||||||
userId: z.string().describe("User id."),
|
userId: z.string().describe("User id."),
|
||||||
notificationId: z.string().describe("Notification id"),
|
notificationId: z.string().describe("Notification id"),
|
||||||
});
|
});
|
||||||
@@ -16,14 +16,29 @@ const newSubscribe = z.object({
|
|||||||
const r = Router();
|
const r = Router();
|
||||||
|
|
||||||
r.delete("/", async (req, res: Response) => {
|
r.delete("/", async (req, res: Response) => {
|
||||||
|
const hasPermissions = await auth.api.userHasPermission({
|
||||||
|
body: {
|
||||||
|
//userId: req?.user?.id,
|
||||||
|
role: req.user?.roles as any,
|
||||||
|
permissions: {
|
||||||
|
notifications: ["readAll"], // This must match the structure in your access control
|
||||||
|
},
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
try {
|
try {
|
||||||
const validated = newSubscribe.parse(req.body);
|
const validated = newSubscribe.parse(req.body);
|
||||||
|
|
||||||
const { data, error } = await tryCatch(
|
const { data, error } = await tryCatch(
|
||||||
db
|
db
|
||||||
.delete(notificationSub)
|
.delete(notificationSub)
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(
|
||||||
eq(notificationSub.userId, validated.userId),
|
eq(
|
||||||
|
notificationSub.userId,
|
||||||
|
hasPermissions ? validated.userId : (req?.user?.id ?? ""),
|
||||||
|
), // allows the admin to delete this
|
||||||
|
//eq(notificationSub.userId, req?.user?.id ?? ""),
|
||||||
eq(notificationSub.notificationId, validated.notificationId),
|
eq(notificationSub.notificationId, validated.notificationId),
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
@@ -44,6 +59,18 @@ r.delete("/", async (req, res: Response) => {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (data.length <= 0) {
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: false,
|
||||||
|
level: "info",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "post",
|
||||||
|
message: `Subscription was not deleted invalid data sent over`,
|
||||||
|
data: data ?? [],
|
||||||
|
status: 200,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
return apiReturn(res, {
|
return apiReturn(res, {
|
||||||
success: true,
|
success: true,
|
||||||
level: "info",
|
level: "info",
|
||||||
|
|||||||
@@ -9,6 +9,8 @@ import { tryCatch } from "../utils/trycatch.utils.js";
|
|||||||
const r = Router();
|
const r = Router();
|
||||||
|
|
||||||
r.get("/", async (req, res: Response) => {
|
r.get("/", async (req, res: Response) => {
|
||||||
|
const { userId } = req.query;
|
||||||
|
|
||||||
const hasPermissions = await auth.api.userHasPermission({
|
const hasPermissions = await auth.api.userHasPermission({
|
||||||
body: {
|
body: {
|
||||||
//userId: req?.user?.id,
|
//userId: req?.user?.id,
|
||||||
@@ -19,6 +21,10 @@ r.get("/", async (req, res: Response) => {
|
|||||||
},
|
},
|
||||||
});
|
});
|
||||||
|
|
||||||
|
if (userId) {
|
||||||
|
hasPermissions.success = false;
|
||||||
|
}
|
||||||
|
|
||||||
const { data, error } = await tryCatch(
|
const { data, error } = await tryCatch(
|
||||||
db
|
db
|
||||||
.select()
|
.select()
|
||||||
@@ -47,7 +53,7 @@ r.get("/", async (req, res: Response) => {
|
|||||||
level: "info",
|
level: "info",
|
||||||
module: "notification",
|
module: "notification",
|
||||||
subModule: "post",
|
subModule: "post",
|
||||||
message: `Subscription deleted`,
|
message: `Subscriptions`,
|
||||||
data: data ?? [],
|
data: data ?? [],
|
||||||
status: 200,
|
status: 200,
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -25,8 +25,25 @@ r.post("/", async (req, res: Response) => {
|
|||||||
try {
|
try {
|
||||||
const validated = newSubscribe.parse(req.body);
|
const validated = newSubscribe.parse(req.body);
|
||||||
|
|
||||||
|
const emails = validated.emails
|
||||||
|
.map((e) => e.trim().toLowerCase())
|
||||||
|
.filter(Boolean);
|
||||||
|
|
||||||
|
const uniqueEmails = [...new Set(emails)];
|
||||||
|
|
||||||
const { data, error } = await tryCatch(
|
const { data, error } = await tryCatch(
|
||||||
db.insert(notificationSub).values(validated).returning(),
|
db
|
||||||
|
.insert(notificationSub)
|
||||||
|
.values({
|
||||||
|
userId: req?.user?.id ?? "",
|
||||||
|
notificationId: validated.notificationId,
|
||||||
|
emails: uniqueEmails,
|
||||||
|
})
|
||||||
|
.onConflictDoUpdate({
|
||||||
|
target: [notificationSub.userId, notificationSub.notificationId],
|
||||||
|
set: { emails: uniqueEmails },
|
||||||
|
})
|
||||||
|
.returning(),
|
||||||
);
|
);
|
||||||
|
|
||||||
await modifiedNotification(validated.notificationId);
|
await modifiedNotification(validated.notificationId);
|
||||||
|
|||||||
@@ -14,7 +14,27 @@ const note: NewNotification[] = [
|
|||||||
"Monitors the labels that are printed and returns a there data, if one falls withing the time frame.",
|
"Monitors the labels that are printed and returns a there data, if one falls withing the time frame.",
|
||||||
active: false,
|
active: false,
|
||||||
interval: "10",
|
interval: "10",
|
||||||
options: [{ prodID: 1 }],
|
options: [{ auditId: [0] }],
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "qualityBlocking",
|
||||||
|
description:
|
||||||
|
"Checks for new blocking orders that have been entered, recommend to get the most recent order in here before activating.",
|
||||||
|
active: false,
|
||||||
|
interval: "10",
|
||||||
|
options: [{ sentBlockingOrders: [{ timeStamp: "0", blockingOrder: 1 }] }],
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "alplaPurchaseHistory",
|
||||||
|
description:
|
||||||
|
"Will check the alpla purchase data for any changes, if the req has not been sent already then we will send this, for a po or fresh order we will ignore. ",
|
||||||
|
active: false,
|
||||||
|
interval: "5",
|
||||||
|
options: [
|
||||||
|
{ sentReqs: [{ timeStamp: "0", req: 1, approved: false }] },
|
||||||
|
{ sentAPOs: [{ timeStamp: "0", apo: 1 }] },
|
||||||
|
{ sentRCT: [{ timeStamp: "0", rct: 1 }] },
|
||||||
|
],
|
||||||
},
|
},
|
||||||
];
|
];
|
||||||
|
|
||||||
|
|||||||
36
backend/ocp/ocp.printer.listener.ts
Normal file
36
backend/ocp/ocp.printer.listener.ts
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
/**
|
||||||
|
* the route that listens for the printers post.
|
||||||
|
*
|
||||||
|
* and http-post alert should be setup on each printer pointing to at min you will want to make the alert for
|
||||||
|
* pause printer, you can have all on here as it will also monitor and do things on all messages
|
||||||
|
*
|
||||||
|
* http://{serverIP}:2222/lst/api/ocp/printer/listener/{printerName}
|
||||||
|
*
|
||||||
|
* the messages will be sent over to the db for logging as well as specific ones will do something
|
||||||
|
*
|
||||||
|
* pause will validate if can print
|
||||||
|
* close head will repause the printer so it wont print a label
|
||||||
|
* power up will just repause the printer so it wont print a label
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/printer/listener/:printer", async (req, res) => {
|
||||||
|
const { printer: printerName } = req.params;
|
||||||
|
console.log(req.body);
|
||||||
|
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "ocp",
|
||||||
|
subModule: "printing",
|
||||||
|
message: `${printerName} just passed over a message`,
|
||||||
|
data: req.body ?? [],
|
||||||
|
status: 200,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
export default r;
|
||||||
19
backend/ocp/ocp.printer.manage.ts
Normal file
19
backend/ocp/ocp.printer.manage.ts
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
/**
|
||||||
|
* this will do a prod sync, update or add alerts to the printer, validate the next pm intervale as well as head replacement.
|
||||||
|
*
|
||||||
|
* if a printer is upcoming on a pm or head replacement send to the plant to address.
|
||||||
|
*
|
||||||
|
* a trigger on the printer table will have the ability to run this as well
|
||||||
|
*
|
||||||
|
* heat beats on all assigned printers
|
||||||
|
*
|
||||||
|
* printer status will live here this will be how we manage all the levels of status like 3 paused, 1 printing, 8 error, 10 power up, etc...
|
||||||
|
*/
|
||||||
|
|
||||||
|
export const printerManager = async () => {};
|
||||||
|
|
||||||
|
export const printerHeartBeat = async () => {
|
||||||
|
// heat heats will be defaulted to 60 seconds no reason to allow anything else
|
||||||
|
};
|
||||||
|
|
||||||
|
//export const printerStatus = async (statusNr: number, printerId: number) => {};
|
||||||
22
backend/ocp/ocp.routes.ts
Normal file
22
backend/ocp/ocp.routes.ts
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
import { type Express, Router } from "express";
|
||||||
|
import { requireAuth } from "../middleware/auth.middleware.js";
|
||||||
|
import { featureCheck } from "../middleware/featureActive.middleware.js";
|
||||||
|
import listener from "./ocp.printer.listener.js";
|
||||||
|
|
||||||
|
export const setupOCPRoutes = (baseUrl: string, app: Express) => {
|
||||||
|
//setup all the routes
|
||||||
|
const router = Router();
|
||||||
|
|
||||||
|
// is the feature even on?
|
||||||
|
router.use(featureCheck("ocp"));
|
||||||
|
|
||||||
|
// non auth routes up here
|
||||||
|
router.use(listener);
|
||||||
|
|
||||||
|
// auth routes below here
|
||||||
|
router.use(requireAuth);
|
||||||
|
|
||||||
|
//router.use("");
|
||||||
|
|
||||||
|
app.use(`${baseUrl}/api/ocp`, router);
|
||||||
|
};
|
||||||
@@ -17,15 +17,6 @@ import { returnFunc } from "../utils/returnHelper.utils.js";
|
|||||||
import { tryCatch } from "../utils/trycatch.utils.js";
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
import { getToken, odToken } from "./opendock.utils.js";
|
import { getToken, odToken } from "./opendock.utils.js";
|
||||||
|
|
||||||
let lastCheck = formatInTimeZone(
|
|
||||||
new Date().toISOString(),
|
|
||||||
"America/New_York",
|
|
||||||
"yyyy-MM-dd HH:mm:ss",
|
|
||||||
);
|
|
||||||
|
|
||||||
//const queue: unknown[] = [];
|
|
||||||
//const isProcessing: boolean = false;
|
|
||||||
|
|
||||||
type Releases = {
|
type Releases = {
|
||||||
ReleaseNumber: number;
|
ReleaseNumber: number;
|
||||||
DeliveryState: number;
|
DeliveryState: number;
|
||||||
@@ -37,10 +28,38 @@ type Releases = {
|
|||||||
LineItemArticleWeight: number;
|
LineItemArticleWeight: number;
|
||||||
CustomerReleaseNumber: string;
|
CustomerReleaseNumber: string;
|
||||||
};
|
};
|
||||||
|
const timeZone = process.env.TIMEZONE as string;
|
||||||
const TWENTY_FOUR_HOURS = 24 * 60 * 60 * 1000;
|
const TWENTY_FOUR_HOURS = 24 * 60 * 60 * 1000;
|
||||||
const log = createLogger({ module: "opendock", subModule: "releaseMonitor" });
|
const log = createLogger({ module: "opendock", subModule: "releaseMonitor" });
|
||||||
|
|
||||||
|
// making the cron more safe when it comes to buffer stuff
|
||||||
|
let opendockSyncRunning = false;
|
||||||
|
|
||||||
|
let lastCheck = formatInTimeZone(
|
||||||
|
new Date().toISOString(),
|
||||||
|
timeZone,
|
||||||
|
"yyyy-MM-dd HH:mm:ss",
|
||||||
|
);
|
||||||
|
|
||||||
|
// const lastCheck = formatInTimeZone(
|
||||||
|
// new Date().toISOString(),
|
||||||
|
// `America/New_York`, //TODO: Pull timezone from the .env last as process.env.TIME_ZONE is not working so need to figure itout
|
||||||
|
// "yyyy-MM-dd HH:mm:ss",
|
||||||
|
// );
|
||||||
|
|
||||||
|
//const queue: unknown[] = [];
|
||||||
|
//const isProcessing: boolean = false;
|
||||||
|
|
||||||
|
// const parseDbDate = (value: string | Date) => {
|
||||||
|
// if (value instanceof Date) return value;
|
||||||
|
|
||||||
|
// // normalize "2026-04-08 13:10:43.280" -> "2026-04-08T13:10:43.280"
|
||||||
|
// const normalized = value.replace(" ", "T");
|
||||||
|
|
||||||
|
// // interpret that wall-clock time as America/New_York
|
||||||
|
// return fromZonedTime(normalized, timeZone);
|
||||||
|
// };
|
||||||
|
|
||||||
const postRelease = async (release: Releases) => {
|
const postRelease = async (release: Releases) => {
|
||||||
if (!odToken.odToken) {
|
if (!odToken.odToken) {
|
||||||
log.info({}, "Getting Auth Token");
|
log.info({}, "Getting Auth Token");
|
||||||
@@ -152,22 +171,25 @@ const postRelease = async (release: Releases) => {
|
|||||||
};
|
};
|
||||||
|
|
||||||
// TODO: pull the current added releases from the db and if one matches then we want to get its id and run the update vs create
|
// TODO: pull the current added releases from the db and if one matches then we want to get its id and run the update vs create
|
||||||
const { data: apt, error: aptError } = await tryCatch(
|
const { data: existingApt, error: aptError } = await tryCatch(
|
||||||
db.select().from(opendockApt),
|
db
|
||||||
|
.select()
|
||||||
|
.from(opendockApt)
|
||||||
|
.where(eq(opendockApt.release, release.ReleaseNumber))
|
||||||
|
.limit(1),
|
||||||
);
|
);
|
||||||
|
|
||||||
if (aptError) {
|
if (aptError) {
|
||||||
log.error({ error: aptError }, "Error getting apt data");
|
log.error({ error: aptError }, "Error getting apt data");
|
||||||
// TODO: send an error email on this one as it will cause issues
|
// TODO: send an error email on this one as it will cause issues
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const releaseCheck = apt.filter((r) => r.release === release.ReleaseNumber);
|
const existing = existingApt[0];
|
||||||
|
|
||||||
//console.log(releaseCheck);
|
//console.log(releaseCheck);
|
||||||
|
|
||||||
if (releaseCheck.length > 0) {
|
if (existing) {
|
||||||
const id = releaseCheck[0]?.openDockAptId;
|
const id = existing.openDockAptId;
|
||||||
try {
|
try {
|
||||||
const response = await axios.patch(
|
const response = await axios.patch(
|
||||||
`${process.env.OPENDOCK_URL}/appointment/${id}`,
|
`${process.env.OPENDOCK_URL}/appointment/${id}`,
|
||||||
@@ -196,7 +218,11 @@ const postRelease = async (release: Releases) => {
|
|||||||
})
|
})
|
||||||
.onConflictDoUpdate({
|
.onConflictDoUpdate({
|
||||||
target: opendockApt.release,
|
target: opendockApt.release,
|
||||||
set: { appointment: response.data.data, upd_date: sql`NOW()` },
|
set: {
|
||||||
|
openDockAptId: response.data.data.id,
|
||||||
|
appointment: response.data.data,
|
||||||
|
upd_date: sql`NOW()`,
|
||||||
|
},
|
||||||
})
|
})
|
||||||
.returning();
|
.returning();
|
||||||
|
|
||||||
@@ -250,8 +276,12 @@ const postRelease = async (release: Releases) => {
|
|||||||
appointment: response.data.data,
|
appointment: response.data.data,
|
||||||
})
|
})
|
||||||
.onConflictDoUpdate({
|
.onConflictDoUpdate({
|
||||||
target: opendockApt.id,
|
target: opendockApt.release,
|
||||||
set: { appointment: response.data.data, upd_date: sql`NOW()` },
|
set: {
|
||||||
|
openDockAptId: response.data.data.id,
|
||||||
|
appointment: response.data.data,
|
||||||
|
upd_date: sql`NOW()`,
|
||||||
|
},
|
||||||
})
|
})
|
||||||
.returning();
|
.returning();
|
||||||
|
|
||||||
@@ -270,7 +300,7 @@ const postRelease = async (release: Releases) => {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
await delay(500); // rate limit protection
|
await delay(750); // rate limit protection
|
||||||
};
|
};
|
||||||
|
|
||||||
export const monitorReleaseChanges = async () => {
|
export const monitorReleaseChanges = async () => {
|
||||||
@@ -298,184 +328,66 @@ export const monitorReleaseChanges = async () => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (openDockMonitor[0]?.active) {
|
if (openDockMonitor[0]?.active) {
|
||||||
createCronJob("opendock_sync", "*/15 * * * * *", async () => {
|
// const BUFFER_MS =
|
||||||
try {
|
// Math.floor(parseInt(openDockMonitor[0]?.value, 10) || 30) * 1.5 * 1000; // this should be >= to the interval we set in the cron TODO: should pull the buffer from the setting and give it an extra 10% then round to nearest int.
|
||||||
const result = await prodQuery(
|
|
||||||
sqlQuery.query.replace("[dateCheck]", `'${lastCheck}'`),
|
|
||||||
"Get release info",
|
|
||||||
);
|
|
||||||
|
|
||||||
if (result.data.length) {
|
createCronJob(
|
||||||
for (const release of result.data) {
|
"opendock_sync",
|
||||||
await postRelease(release);
|
`*/${parseInt(openDockMonitor[0]?.value, 10) || 30} * * * * *`,
|
||||||
|
async () => {
|
||||||
lastCheck = formatInTimeZone(
|
if (opendockSyncRunning) {
|
||||||
new Date(release.Upd_Date).toISOString(),
|
log.warn(
|
||||||
"UTC",
|
{},
|
||||||
"yyyy-MM-dd HH:mm:ss",
|
"Skipping opendock_sync because previous run is still active",
|
||||||
);
|
);
|
||||||
|
return;
|
||||||
await delay(500);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
} catch (e) {
|
|
||||||
console.error(
|
opendockSyncRunning = true;
|
||||||
{ error: e },
|
try {
|
||||||
"Error occurred while running the monitor job",
|
// set this to the latest time.
|
||||||
);
|
|
||||||
log.error({ error: e }, "Error occurred while running the monitor job");
|
const result = await prodQuery(
|
||||||
}
|
sqlQuery.query.replace("[dateCheck]", `'${lastCheck}'`),
|
||||||
});
|
"Get release info",
|
||||||
|
);
|
||||||
|
|
||||||
|
log.debug(
|
||||||
|
{ lastCheck },
|
||||||
|
`${result.data.length} Changes to a release have been made`,
|
||||||
|
);
|
||||||
|
|
||||||
|
if (result.data.length) {
|
||||||
|
for (const release of result.data) {
|
||||||
|
await postRelease(release);
|
||||||
|
|
||||||
|
// add a 2 seconds to account for a massive influx of orders and when we dont finish in 1 go it wont try to grab the same amount again
|
||||||
|
const nDate = new Date(release.Upd_Date);
|
||||||
|
nDate.setSeconds(nDate.getSeconds() + 2);
|
||||||
|
|
||||||
|
lastCheck = formatInTimeZone(
|
||||||
|
nDate.toISOString(),
|
||||||
|
"UTC",
|
||||||
|
"yyyy-MM-dd HH:mm:ss",
|
||||||
|
);
|
||||||
|
log.debug({ lastCheck }, "Changes to a release have been made");
|
||||||
|
await delay(500);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} catch (e) {
|
||||||
|
console.error(
|
||||||
|
{ error: e },
|
||||||
|
"Error occurred while running the monitor job",
|
||||||
|
);
|
||||||
|
log.error(
|
||||||
|
{ error: e },
|
||||||
|
"Error occurred while running the monitor job",
|
||||||
|
);
|
||||||
|
} finally {
|
||||||
|
opendockSyncRunning = false;
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"monitorReleaseChanges",
|
||||||
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// run the main game loop
|
|
||||||
// while (openDockSetting) {
|
|
||||||
// try {
|
|
||||||
// const result = await prodQuery(
|
|
||||||
// sqlQuery.query.replace("[dateCheck]", `'${lastCheck}'`),
|
|
||||||
// "Get release info",
|
|
||||||
// );
|
|
||||||
|
|
||||||
// if (result.data.length) {
|
|
||||||
// for (const release of result.data) {
|
|
||||||
// // potentially move this to a buffer table to easy up on memory
|
|
||||||
// await postRelease(release);
|
|
||||||
|
|
||||||
// // Move checkpoint AFTER successful post
|
|
||||||
// lastCheck = formatInTimeZone(
|
|
||||||
// new Date(release.Upd_Date).toISOString(),
|
|
||||||
// "UTC",
|
|
||||||
// "yyyy-MM-dd HH:mm:ss",
|
|
||||||
// );
|
|
||||||
|
|
||||||
// await delay(500);
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// } catch (e) {
|
|
||||||
// console.error("Monitor error:", e);
|
|
||||||
// }
|
|
||||||
|
|
||||||
// await delay(15 * 1000); // making this 15 seconds as we would really only see issues if we have a mass burst.
|
|
||||||
// }
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// export const monitorReleaseChanges = async () => {
|
|
||||||
// console.log("Starting release monitor", lastCheck);
|
|
||||||
// setInterval(async () => {
|
|
||||||
// try {
|
|
||||||
// const result = await prodQuery(
|
|
||||||
// releaseQuery.replace("[dateCheck]", `'${lastCheck}'`),
|
|
||||||
// "get last release change",
|
|
||||||
// );
|
|
||||||
|
|
||||||
// //console.log(releaseQuery.replace("[dateCheck]", `'${lastCheck}'`));
|
|
||||||
// if (result.data.length > 0) {
|
|
||||||
// console.log(
|
|
||||||
// formatInTimeZone(
|
|
||||||
// result.data[result.data.length - 1].Upd_Date,
|
|
||||||
// "UTC",
|
|
||||||
// "yyyy-MM-dd HH:mm:ss",
|
|
||||||
// ),
|
|
||||||
// lastCheck,
|
|
||||||
// );
|
|
||||||
// lastCheck = formatInTimeZone(
|
|
||||||
// result.data[result.data.length - 1].Upd_Date,
|
|
||||||
// "UTC",
|
|
||||||
// "yyyy-MM-dd HH:mm:ss",
|
|
||||||
// );
|
|
||||||
// const releases = result.data;
|
|
||||||
// for (let i = 0; i < releases.length; i++) {
|
|
||||||
// const newDockApt = {
|
|
||||||
// status: "Scheduled",
|
|
||||||
// userId: "ee956455-e193-47fc-b53b-dff30fabdf4b", // this should be the carrierid
|
|
||||||
// loadTypeId: "0aa7988e-b17b-4f10-acdd-3d029b44a773", // well get this and make it a default one
|
|
||||||
// dockId: "00ba4386-ce5a-4dd1-9356-6e6d10a24609", // this the warehouse we want it in to start out
|
|
||||||
// refNumbers: [releases[i].ReleaseNumber],
|
|
||||||
// refNumber: releases[i].ReleaseNumber,
|
|
||||||
// start: releases[i].DeliveryDate,
|
|
||||||
// end: addHours(releases[i].DeliveryDate, 1),
|
|
||||||
// notes: "",
|
|
||||||
// ccEmails: [""],
|
|
||||||
// muteNotifications: true,
|
|
||||||
// metadata: {
|
|
||||||
// externalValidationFailed: false,
|
|
||||||
// externalValidationErrorMessage: null,
|
|
||||||
// },
|
|
||||||
// units: null,
|
|
||||||
// customFields: [
|
|
||||||
// {
|
|
||||||
// name: "strArticle",
|
|
||||||
// type: "str",
|
|
||||||
// label: "Article",
|
|
||||||
// value: `${releases[i].LineItemHumanReadableId} - ${releases[i].ArticleAlias}`,
|
|
||||||
// description: "What bottle are we sending ",
|
|
||||||
// placeholder: "",
|
|
||||||
// dropDownValues: [],
|
|
||||||
// minLengthOrValue: 1,
|
|
||||||
// hiddenFromCarrier: false,
|
|
||||||
// requiredForCarrier: false,
|
|
||||||
// requiredForWarehouse: false,
|
|
||||||
// },
|
|
||||||
// {
|
|
||||||
// name: "intPallet Count",
|
|
||||||
// type: "int",
|
|
||||||
// label: "Pallet Count",
|
|
||||||
// value: parseInt(releases[i].LoadingUnits, 10),
|
|
||||||
// description: "How many pallets",
|
|
||||||
// placeholder: "22",
|
|
||||||
// dropDownValues: [],
|
|
||||||
// minLengthOrValue: 1,
|
|
||||||
// hiddenFromCarrier: false,
|
|
||||||
// requiredForCarrier: false,
|
|
||||||
// requiredForWarehouse: false,
|
|
||||||
// },
|
|
||||||
// {
|
|
||||||
// name: "strTotal Weight",
|
|
||||||
// type: "str",
|
|
||||||
// label: "Total Weight",
|
|
||||||
// value: `${(((releases[i].Quantity * releases[i].LineItemArticleWeight) / 1000) * 2.20462).toFixed(2)}`,
|
|
||||||
// description: "What is the total weight of the load",
|
|
||||||
// placeholder: "",
|
|
||||||
// dropDownValues: [],
|
|
||||||
// minLengthOrValue: 1,
|
|
||||||
// hiddenFromCarrier: false,
|
|
||||||
// requiredForCarrier: false,
|
|
||||||
// requiredForWarehouse: false,
|
|
||||||
// },
|
|
||||||
// {
|
|
||||||
// name: "strCustomer ReleaseNumber",
|
|
||||||
// type: "str",
|
|
||||||
// label: "Customer Release Number",
|
|
||||||
// value: `${releases[i].CustomerReleaseNumber}`,
|
|
||||||
// description: "What is the customer release number",
|
|
||||||
// placeholder: "",
|
|
||||||
// dropDownValues: [],
|
|
||||||
// minLengthOrValue: 1,
|
|
||||||
// hiddenFromCarrier: false,
|
|
||||||
// requiredForCarrier: false,
|
|
||||||
// requiredForWarehouse: false,
|
|
||||||
// },
|
|
||||||
// ],
|
|
||||||
// };
|
|
||||||
|
|
||||||
// //console.log(newDockApt);
|
|
||||||
|
|
||||||
// const newDockResult = await axios.post(
|
|
||||||
// "https://neutron.staging.opendock.com/appointment",
|
|
||||||
// newDockApt,
|
|
||||||
// {
|
|
||||||
// headers: {
|
|
||||||
// "content-type": "application/json; charset=utf-8",
|
|
||||||
// },
|
|
||||||
// },
|
|
||||||
// );
|
|
||||||
|
|
||||||
// console.log(newDockResult.statusText);
|
|
||||||
// await delay(500);
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// } catch (e) {
|
|
||||||
// console.log(e);
|
|
||||||
// }
|
|
||||||
// }, 5 * 1000);
|
|
||||||
// };
|
|
||||||
|
|||||||
@@ -28,7 +28,7 @@ export const getToken = async () => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
odToken = { odToken: data.access_token, tokenDate: new Date() };
|
odToken = { odToken: data.access_token, tokenDate: new Date() };
|
||||||
log.info({}, "Token added");
|
log.info({ odToken }, "Token added");
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
log.error({ error: e }, "Error getting/refreshing token");
|
log.error({ error: e }, "Error getting/refreshing token");
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -36,12 +36,12 @@ export const opendockSocketMonitor = async () => {
|
|||||||
// console.log(data);
|
// console.log(data);
|
||||||
// });
|
// });
|
||||||
|
|
||||||
socket.on("create-Appointment", (data) => {
|
socket.on("create-Appointment", () => {
|
||||||
console.log("appt create:", data);
|
//console.log("appt create:", data);
|
||||||
});
|
});
|
||||||
|
|
||||||
socket.on("update-Appointment", (data) => {
|
socket.on("update-Appointment", () => {
|
||||||
console.log("appt update:", data);
|
//console.log("appt update:", data);
|
||||||
});
|
});
|
||||||
|
|
||||||
socket.on("error", (data) => {
|
socket.on("error", (data) => {
|
||||||
|
|||||||
63
backend/prodSql/queries/alplapurchase.sql
Normal file
63
backend/prodSql/queries/alplapurchase.sql
Normal file
@@ -0,0 +1,63 @@
|
|||||||
|
use AlplaPROD_test1
|
||||||
|
declare @intervalCheck as int = '[interval]'
|
||||||
|
|
||||||
|
/*
|
||||||
|
Monitors alpla purchase for thing new. this will not update unless the order status is updated.
|
||||||
|
this means if a user just reopens the order it will update but everything changed in the position will not be updated until the user reorders or cancels the po
|
||||||
|
*/
|
||||||
|
|
||||||
|
select
|
||||||
|
IdBestellung as apo
|
||||||
|
,po.revision as revision
|
||||||
|
,po.Bestaetigt as confirmed
|
||||||
|
,po.status
|
||||||
|
,case po.Status
|
||||||
|
when 1 then 'Created'
|
||||||
|
when 2 then 'Ordered'
|
||||||
|
when 22 then 'Reopened'
|
||||||
|
when 11 then 'Reopened'
|
||||||
|
when 4 then 'Planned'
|
||||||
|
when 5 then 'Partly Delivered'
|
||||||
|
when 6 then 'Delivered'
|
||||||
|
when 7 then 'Canceled'
|
||||||
|
when 8 then 'Closed'
|
||||||
|
else 'Unknown' end as statusText
|
||||||
|
,po.IdJournal as journalNum -- use this to validate if we used it already.
|
||||||
|
,po.Add_User as add_user
|
||||||
|
,po.Add_Date as add_date
|
||||||
|
,po.Upd_User as upd_user
|
||||||
|
,po.Upd_Date as upd_Date
|
||||||
|
,po.Bemerkung as remark
|
||||||
|
,po.IdJournal as journal -- use this to validate if we used it already.
|
||||||
|
,isnull((
|
||||||
|
select
|
||||||
|
o.IdArtikelVarianten as av
|
||||||
|
,a.Bezeichnung as alias
|
||||||
|
,Lieferdatum as deliveryDate
|
||||||
|
,cast(BestellMenge as decimal(18,2)) as qty
|
||||||
|
,cast(BestellMengeVPK as decimal(18,0)) as pkg
|
||||||
|
,cast(PreisProEinheit as decimal(18,0)) as price
|
||||||
|
,PositionsStatus
|
||||||
|
,case PositionsStatus
|
||||||
|
when 1 then 'Created'
|
||||||
|
when 2 then 'Ordered'
|
||||||
|
when 22 then 'Reopened'
|
||||||
|
when 4 then 'Planned'
|
||||||
|
when 5 then 'Partly Delivered'
|
||||||
|
when 6 then 'Delivered'
|
||||||
|
when 7 then 'Canceled'
|
||||||
|
when 8 then 'Closed'
|
||||||
|
else 'Unknown' end as statusText
|
||||||
|
,o.upd_user
|
||||||
|
,o.upd_date
|
||||||
|
from T_Bestellpositionen (nolock) as o
|
||||||
|
|
||||||
|
left join
|
||||||
|
T_Artikelvarianten as a on
|
||||||
|
a.IdArtikelvarianten = o.IdArtikelVarianten
|
||||||
|
where o.IdBestellung = po.IdBestellung
|
||||||
|
for json path
|
||||||
|
), '[]') as position
|
||||||
|
--,*
|
||||||
|
from T_Bestellungen (nolock) as po
|
||||||
|
where po.Upd_Date > dateadd(MINUTE, -@intervalCheck, getdate())
|
||||||
28
backend/prodSql/queries/reprintLabels.sql
Normal file
28
backend/prodSql/queries/reprintLabels.sql
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
SELECT
|
||||||
|
--JSON_VALUE(content, '$.EntityId') as labelId
|
||||||
|
a.id
|
||||||
|
,ActorName
|
||||||
|
,FORMAT(PrintDate, 'yyyy-MM-dd HH:mm') as printDate
|
||||||
|
,FORMAT(CreatedDateTime, 'yyyy-MM-dd HH:mm') createdDateTime
|
||||||
|
,l.ArticleHumanReadableId as av
|
||||||
|
,l.ArticleDescription as alias
|
||||||
|
,PrintedCopies
|
||||||
|
,p.name as printerName
|
||||||
|
,RunningNumber
|
||||||
|
--,*
|
||||||
|
FROM [support].[AuditLog] (nolock) as a
|
||||||
|
|
||||||
|
left join
|
||||||
|
[labelling].[InternalLabel] (nolock) as l on
|
||||||
|
l.id = JSON_VALUE(content, '$.EntityId')
|
||||||
|
|
||||||
|
left join
|
||||||
|
[masterData].[printer] (nolock) as p on
|
||||||
|
p.id = l.PrinterId
|
||||||
|
|
||||||
|
where message like '%reprint%'
|
||||||
|
and CreatedDateTime > DATEADD(minute, -[intervalCheck], SYSDATETIMEOFFSET())
|
||||||
|
and a.id > [ignoreList]
|
||||||
|
order by CreatedDateTime desc
|
||||||
97
backend/purchase/purchase.controller.ts
Normal file
97
backend/purchase/purchase.controller.ts
Normal file
@@ -0,0 +1,97 @@
|
|||||||
|
/**
|
||||||
|
* This will monitor alpla purchase
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import {
|
||||||
|
alplaPurchaseHistory,
|
||||||
|
type NewAlplaPurchaseHistory,
|
||||||
|
} from "../db/schema/alplapurchase.schema.js";
|
||||||
|
import { settings } from "../db/schema/settings.schema.js";
|
||||||
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
|
import {
|
||||||
|
type SqlQuery,
|
||||||
|
sqlQuerySelector,
|
||||||
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import { createCronJob } from "../utils/croner.utils.js";
|
||||||
|
import { delay } from "../utils/delay.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
|
const log = createLogger({ module: "purchase", subModule: "purchaseMonitor" });
|
||||||
|
|
||||||
|
export const monitorAlplaPurchase = async () => {
|
||||||
|
const purchaseMonitor = await db
|
||||||
|
.select()
|
||||||
|
.from(settings)
|
||||||
|
.where(eq(settings.name, "purchaseMonitor"));
|
||||||
|
|
||||||
|
const sqlQuery = sqlQuerySelector(`alplapurchase`) as SqlQuery;
|
||||||
|
|
||||||
|
if (!sqlQuery.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "purchase",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Error getting alpla purchase info`,
|
||||||
|
data: [sqlQuery.message],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (purchaseMonitor[0]?.active) {
|
||||||
|
createCronJob("purchaseMonitor", "0 */5 * * * *", async () => {
|
||||||
|
try {
|
||||||
|
const result = await prodQuery(
|
||||||
|
sqlQuery.query.replace(
|
||||||
|
"[interval]",
|
||||||
|
`${purchaseMonitor[0]?.value || "5"}`,
|
||||||
|
),
|
||||||
|
"Get release info",
|
||||||
|
);
|
||||||
|
|
||||||
|
log.debug(
|
||||||
|
{},
|
||||||
|
`There are ${result.data.length} pending to be updated from the last ${purchaseMonitor[0]?.value}`,
|
||||||
|
);
|
||||||
|
if (result.data.length) {
|
||||||
|
const convertedData = result.data.map((i) => ({
|
||||||
|
...i,
|
||||||
|
position: JSON.parse(i.position),
|
||||||
|
})) as NewAlplaPurchaseHistory;
|
||||||
|
|
||||||
|
const { data, error } = await tryCatch(
|
||||||
|
db.insert(alplaPurchaseHistory).values(convertedData).returning(),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (data) {
|
||||||
|
log.debug(
|
||||||
|
{ data },
|
||||||
|
"New data was just added to alpla purchase history",
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
log.error(
|
||||||
|
{ error },
|
||||||
|
"There was an error adding alpla purchase history",
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
await delay(500);
|
||||||
|
}
|
||||||
|
} catch (e) {
|
||||||
|
console.error(
|
||||||
|
{ error: e },
|
||||||
|
"Error occurred while running the monitor job",
|
||||||
|
);
|
||||||
|
log.error({ error: e }, "Error occurred while running the monitor job");
|
||||||
|
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
@@ -5,6 +5,7 @@ import { setupAuthRoutes } from "./auth/auth.routes.js";
|
|||||||
import { setupApiDocsRoutes } from "./configs/scaler.config.js";
|
import { setupApiDocsRoutes } from "./configs/scaler.config.js";
|
||||||
import { setupDatamartRoutes } from "./datamart/datamart.routes.js";
|
import { setupDatamartRoutes } from "./datamart/datamart.routes.js";
|
||||||
import { setupNotificationRoutes } from "./notification/notification.routes.js";
|
import { setupNotificationRoutes } from "./notification/notification.routes.js";
|
||||||
|
import { setupOCPRoutes } from "./ocp/ocp.routes.js";
|
||||||
import { setupOpendockRoutes } from "./opendock/opendock.routes.js";
|
import { setupOpendockRoutes } from "./opendock/opendock.routes.js";
|
||||||
import { setupProdSqlRoutes } from "./prodSql/prodSql.routes.js";
|
import { setupProdSqlRoutes } from "./prodSql/prodSql.routes.js";
|
||||||
import { setupSystemRoutes } from "./system/system.routes.js";
|
import { setupSystemRoutes } from "./system/system.routes.js";
|
||||||
@@ -20,4 +21,5 @@ export const setupRoutes = (baseUrl: string, app: Express) => {
|
|||||||
setupUtilsRoutes(baseUrl, app);
|
setupUtilsRoutes(baseUrl, app);
|
||||||
setupOpendockRoutes(baseUrl, app);
|
setupOpendockRoutes(baseUrl, app);
|
||||||
setupNotificationRoutes(baseUrl, app);
|
setupNotificationRoutes(baseUrl, app);
|
||||||
|
setupOCPRoutes(baseUrl, app);
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ import { createNotifications } from "./notification/notifications.master.js";
|
|||||||
import { monitorReleaseChanges } from "./opendock/openDockRreleaseMonitor.utils.js";
|
import { monitorReleaseChanges } from "./opendock/openDockRreleaseMonitor.utils.js";
|
||||||
import { opendockSocketMonitor } from "./opendock/opendockSocketMonitor.utils.js";
|
import { opendockSocketMonitor } from "./opendock/opendockSocketMonitor.utils.js";
|
||||||
import { connectProdSql } from "./prodSql/prodSqlConnection.controller.js";
|
import { connectProdSql } from "./prodSql/prodSqlConnection.controller.js";
|
||||||
|
import { monitorAlplaPurchase } from "./purchase/purchase.controller.js";
|
||||||
import { setupSocketIORoutes } from "./socket.io/serverSetup.js";
|
import { setupSocketIORoutes } from "./socket.io/serverSetup.js";
|
||||||
import { baseSettingValidationCheck } from "./system/settingsBase.controller.js";
|
import { baseSettingValidationCheck } from "./system/settingsBase.controller.js";
|
||||||
import { createCronJob } from "./utils/croner.utils.js";
|
import { createCronJob } from "./utils/croner.utils.js";
|
||||||
@@ -36,7 +37,7 @@ const start = async () => {
|
|||||||
// also we always want to have long lived processes inside a setting check.
|
// also we always want to have long lived processes inside a setting check.
|
||||||
setTimeout(() => {
|
setTimeout(() => {
|
||||||
if (systemSettings.filter((n) => n.name === "opendock_sync")[0]?.active) {
|
if (systemSettings.filter((n) => n.name === "opendock_sync")[0]?.active) {
|
||||||
log.info({}, "Opendock is not active");
|
log.info({}, "Opendock is active");
|
||||||
monitorReleaseChanges(); // this is od monitoring the db for all new releases
|
monitorReleaseChanges(); // this is od monitoring the db for all new releases
|
||||||
opendockSocketMonitor();
|
opendockSocketMonitor();
|
||||||
createCronJob("opendockAptCleanup", "0 30 5 * * *", () =>
|
createCronJob("opendockAptCleanup", "0 30 5 * * *", () =>
|
||||||
@@ -44,6 +45,10 @@ const start = async () => {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (systemSettings.filter((n) => n.name === "purchaseMonitor")[0]?.active) {
|
||||||
|
monitorAlplaPurchase();
|
||||||
|
}
|
||||||
|
|
||||||
// these jobs below are system jobs and should run no matter what.
|
// these jobs below are system jobs and should run no matter what.
|
||||||
createCronJob("JobAuditLogCleanUp", "0 0 5 * * *", () =>
|
createCronJob("JobAuditLogCleanUp", "0 0 5 * * *", () =>
|
||||||
dbCleanup("jobs", 30),
|
dbCleanup("jobs", 30),
|
||||||
|
|||||||
@@ -8,8 +8,8 @@ const newSettings: NewSetting[] = [
|
|||||||
// feature settings
|
// feature settings
|
||||||
{
|
{
|
||||||
name: "opendock_sync",
|
name: "opendock_sync",
|
||||||
value: "0",
|
value: "15",
|
||||||
active: true,
|
active: false,
|
||||||
description: "Dock Scheduling system",
|
description: "Dock Scheduling system",
|
||||||
moduleName: "opendock",
|
moduleName: "opendock",
|
||||||
settingType: "feature",
|
settingType: "feature",
|
||||||
@@ -19,7 +19,7 @@ const newSettings: NewSetting[] = [
|
|||||||
{
|
{
|
||||||
name: "ocp",
|
name: "ocp",
|
||||||
value: "1",
|
value: "1",
|
||||||
active: true,
|
active: false,
|
||||||
description: "One click print",
|
description: "One click print",
|
||||||
moduleName: "ocp",
|
moduleName: "ocp",
|
||||||
settingType: "feature",
|
settingType: "feature",
|
||||||
@@ -29,7 +29,7 @@ const newSettings: NewSetting[] = [
|
|||||||
{
|
{
|
||||||
name: "ocme",
|
name: "ocme",
|
||||||
value: "0",
|
value: "0",
|
||||||
active: true,
|
active: false,
|
||||||
description: "Dayton Agv system",
|
description: "Dayton Agv system",
|
||||||
moduleName: "ocme",
|
moduleName: "ocme",
|
||||||
settingType: "feature",
|
settingType: "feature",
|
||||||
@@ -39,7 +39,7 @@ const newSettings: NewSetting[] = [
|
|||||||
{
|
{
|
||||||
name: "demandManagement",
|
name: "demandManagement",
|
||||||
value: "1",
|
value: "1",
|
||||||
active: true,
|
active: false,
|
||||||
description: "Fake EDI System",
|
description: "Fake EDI System",
|
||||||
moduleName: "demandManagement",
|
moduleName: "demandManagement",
|
||||||
settingType: "feature",
|
settingType: "feature",
|
||||||
@@ -49,7 +49,7 @@ const newSettings: NewSetting[] = [
|
|||||||
{
|
{
|
||||||
name: "qualityRequest",
|
name: "qualityRequest",
|
||||||
value: "0",
|
value: "0",
|
||||||
active: true,
|
active: false,
|
||||||
description: "Quality System",
|
description: "Quality System",
|
||||||
moduleName: "qualityRequest",
|
moduleName: "qualityRequest",
|
||||||
settingType: "feature",
|
settingType: "feature",
|
||||||
@@ -59,13 +59,23 @@ const newSettings: NewSetting[] = [
|
|||||||
{
|
{
|
||||||
name: "tms",
|
name: "tms",
|
||||||
value: "0",
|
value: "0",
|
||||||
active: true,
|
active: false,
|
||||||
description: "Transport system integration",
|
description: "Transport system integration",
|
||||||
moduleName: "tms",
|
moduleName: "tms",
|
||||||
settingType: "feature",
|
settingType: "feature",
|
||||||
roles: ["admin"],
|
roles: ["admin"],
|
||||||
seedVersion: 1,
|
seedVersion: 1,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
name: "purchaseMonitor",
|
||||||
|
value: "5",
|
||||||
|
active: true,
|
||||||
|
description: "Monitors alpla purchase fo all changes",
|
||||||
|
moduleName: "purchase",
|
||||||
|
settingType: "feature",
|
||||||
|
roles: ["admin"],
|
||||||
|
seedVersion: 1,
|
||||||
|
},
|
||||||
|
|
||||||
// standard settings
|
// standard settings
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ import {
|
|||||||
killOpendockSocket,
|
killOpendockSocket,
|
||||||
opendockSocketMonitor,
|
opendockSocketMonitor,
|
||||||
} from "../opendock/opendockSocketMonitor.utils.js";
|
} from "../opendock/opendockSocketMonitor.utils.js";
|
||||||
|
import { monitorAlplaPurchase } from "../purchase/purchase.controller.js";
|
||||||
import {
|
import {
|
||||||
createCronJob,
|
createCronJob,
|
||||||
resumeCronJob,
|
resumeCronJob,
|
||||||
@@ -31,8 +32,24 @@ export const featureControl = async (data: Setting) => {
|
|||||||
createCronJob("opendockAptCleanup", "0 30 5 * * *", () =>
|
createCronJob("opendockAptCleanup", "0 30 5 * * *", () =>
|
||||||
dbCleanup("opendockApt", 90),
|
dbCleanup("opendockApt", 90),
|
||||||
);
|
);
|
||||||
} else {
|
}
|
||||||
|
|
||||||
|
if (data.name === "opendock_sync" && !data.active) {
|
||||||
killOpendockSocket();
|
killOpendockSocket();
|
||||||
stopCronJob("opendockAptCleanup");
|
stopCronJob("opendockAptCleanup");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// purchase stuff
|
||||||
|
if (data.name === "purchaseMonitor" && data.active) {
|
||||||
|
monitorAlplaPurchase();
|
||||||
|
}
|
||||||
|
|
||||||
|
if (data.name === "purchaseMonitor" && !data.active) {
|
||||||
|
stopCronJob("purchaseMonitor");
|
||||||
|
}
|
||||||
|
|
||||||
|
// this means the data time has changed
|
||||||
|
if (data.name === "purchaseMonitor" && data.value) {
|
||||||
|
monitorAlplaPurchase();
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -15,6 +15,7 @@ export const allowedOrigins = [
|
|||||||
`http://${process.env.PROD_SERVER}:3000`,
|
`http://${process.env.PROD_SERVER}:3000`,
|
||||||
`http://${process.env.PROD_SERVER}:3100`, // temp
|
`http://${process.env.PROD_SERVER}:3100`, // temp
|
||||||
`http://usmcd1olp082:3000`,
|
`http://usmcd1olp082:3000`,
|
||||||
|
`${process.env.EXTERNAL_URL}`, // internal docker
|
||||||
];
|
];
|
||||||
export const lstCors = () => {
|
export const lstCors = () => {
|
||||||
return cors({
|
return cors({
|
||||||
|
|||||||
@@ -18,7 +18,9 @@ export interface JobInfo {
|
|||||||
|
|
||||||
// Store running cronjobs
|
// Store running cronjobs
|
||||||
export const runningCrons: Record<string, Cron> = {};
|
export const runningCrons: Record<string, Cron> = {};
|
||||||
|
const activeRuns = new Set<string>();
|
||||||
const log = createLogger({ module: "system", subModule: "croner" });
|
const log = createLogger({ module: "system", subModule: "croner" });
|
||||||
|
const cronStats: Record<string, { created: number; replaced: number }> = {};
|
||||||
|
|
||||||
// how to se the times
|
// how to se the times
|
||||||
// * ┌──────────────── (optional) second (0 - 59)
|
// * ┌──────────────── (optional) second (0 - 59)
|
||||||
@@ -38,17 +40,36 @@ const log = createLogger({ module: "system", subModule: "croner" });
|
|||||||
* @param name Name of the job we want to run
|
* @param name Name of the job we want to run
|
||||||
* @param schedule Cron expression (example: `*\/5 * * * * *`)
|
* @param schedule Cron expression (example: `*\/5 * * * * *`)
|
||||||
* @param task Async function that will run
|
* @param task Async function that will run
|
||||||
|
* @param source we can add where it came from to assist in getting this tracked down, more for debugging
|
||||||
*/
|
*/
|
||||||
export const createCronJob = async (
|
export const createCronJob = async (
|
||||||
name: string,
|
name: string,
|
||||||
schedule: string, // cron string with 8 8 IE: */5 * * * * * every 5th second
|
schedule: string, // cron string with 8 8 IE: */5 * * * * * every 5th second
|
||||||
task: () => Promise<void>, // what function are we passing over
|
task: () => Promise<void>, // what function are we passing over
|
||||||
|
source = "unknown",
|
||||||
) => {
|
) => {
|
||||||
// get the timezone based on the os timezone set
|
// get the timezone based on the os timezone set
|
||||||
const timeZone = Intl.DateTimeFormat().resolvedOptions().timeZone;
|
const timeZone = Intl.DateTimeFormat().resolvedOptions().timeZone;
|
||||||
|
|
||||||
|
// initial go so just store it this is more for debugging to see if something crazy keeps happening
|
||||||
|
if (!cronStats[name]) {
|
||||||
|
cronStats[name] = { created: 0, replaced: 0 };
|
||||||
|
}
|
||||||
|
|
||||||
// Destroy existing job if it exist
|
// Destroy existing job if it exist
|
||||||
if (runningCrons[name]) {
|
if (runningCrons[name]) {
|
||||||
|
cronStats[name].replaced += 1;
|
||||||
|
log.warn(
|
||||||
|
{
|
||||||
|
job: name,
|
||||||
|
source,
|
||||||
|
oldSchedule: runningCrons[name].getPattern?.(),
|
||||||
|
newSchedule: schedule,
|
||||||
|
replaceCount: cronStats[name].replaced,
|
||||||
|
},
|
||||||
|
`Cron job "${name}" already existed and is being replaced`,
|
||||||
|
);
|
||||||
|
|
||||||
runningCrons[name].stop();
|
runningCrons[name].stop();
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -61,6 +82,13 @@ export const createCronJob = async (
|
|||||||
name: name,
|
name: name,
|
||||||
},
|
},
|
||||||
async () => {
|
async () => {
|
||||||
|
if (activeRuns.has(name)) {
|
||||||
|
log.warn({ jobName: name }, "Skipping overlapping cron execution");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
activeRuns.add(name);
|
||||||
|
|
||||||
const startedAt = new Date();
|
const startedAt = new Date();
|
||||||
const start = Date.now();
|
const start = Date.now();
|
||||||
|
|
||||||
@@ -91,14 +119,19 @@ export const createCronJob = async (
|
|||||||
.where(eq(jobAuditLog.id, executionId));
|
.where(eq(jobAuditLog.id, executionId));
|
||||||
} catch (e: any) {
|
} catch (e: any) {
|
||||||
if (executionId) {
|
if (executionId) {
|
||||||
await db.update(jobAuditLog).set({
|
await db
|
||||||
finishedAt: new Date(),
|
.update(jobAuditLog)
|
||||||
durationMs: Date.now() - start,
|
.set({
|
||||||
status: "error",
|
finishedAt: new Date(),
|
||||||
errorMessage: e.message,
|
durationMs: Date.now() - start,
|
||||||
errorStack: e.stack,
|
status: "error",
|
||||||
});
|
errorMessage: e.message,
|
||||||
|
errorStack: e.stack,
|
||||||
|
})
|
||||||
|
.where(eq(jobAuditLog.id, executionId));
|
||||||
}
|
}
|
||||||
|
} finally {
|
||||||
|
activeRuns.delete(name);
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
47
backend/utils/mailViews/reprintLabels.hbs
Normal file
47
backend/utils/mailViews/reprintLabels.hbs
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
<!DOCTYPE html>
|
||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8">
|
||||||
|
<meta name="viewport" content="width=device-width, initial-scale=1.0">
|
||||||
|
{{!-- <link rel="stylesheet" href="styles/styles.css" /> --}}
|
||||||
|
{{> styles}}
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<p>All,</p>
|
||||||
|
<p>The below labels have been reprinted.</p>
|
||||||
|
<table >
|
||||||
|
<thead>
|
||||||
|
<tr>
|
||||||
|
<th>AV</th>
|
||||||
|
<th>Description</th>
|
||||||
|
<th>Label Number</th>
|
||||||
|
<th>Date Added</th>
|
||||||
|
<th>Date Reprinted</th>
|
||||||
|
<th>Who printed/Updated</th>
|
||||||
|
<th>What printer it came from</th>
|
||||||
|
|
||||||
|
</tr>
|
||||||
|
</thead>
|
||||||
|
|
||||||
|
<tbody>
|
||||||
|
{{#each items}}
|
||||||
|
<tr>
|
||||||
|
<td>{{av}}</td>
|
||||||
|
<td>{{alias}}</td>
|
||||||
|
<td>{{RunningNumber}}</td>
|
||||||
|
<td>{{printDate}}</td>
|
||||||
|
<td>{{createdDateTime}}</td>
|
||||||
|
<td>{{ActorName}}</td>
|
||||||
|
<td>{{printerName}}</td>
|
||||||
|
|
||||||
|
</tr>
|
||||||
|
{{/each}}
|
||||||
|
</tbody>
|
||||||
|
</table>
|
||||||
|
|
||||||
|
<div>
|
||||||
|
<p>Thank you,</p>
|
||||||
|
<p>LST Team</p>
|
||||||
|
</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
@@ -10,7 +10,9 @@ interface Data<T = unknown[]> {
|
|||||||
| "datamart"
|
| "datamart"
|
||||||
| "utils"
|
| "utils"
|
||||||
| "opendock"
|
| "opendock"
|
||||||
| "notification";
|
| "notification"
|
||||||
|
| "email"
|
||||||
|
| "purchase";
|
||||||
subModule:
|
subModule:
|
||||||
| "db"
|
| "db"
|
||||||
| "labeling"
|
| "labeling"
|
||||||
@@ -28,7 +30,8 @@ interface Data<T = unknown[]> {
|
|||||||
| "delete"
|
| "delete"
|
||||||
| "post"
|
| "post"
|
||||||
| "notification"
|
| "notification"
|
||||||
| "delete";
|
| "delete"
|
||||||
|
| "printing";
|
||||||
level: "info" | "error" | "debug" | "fatal";
|
level: "info" | "error" | "debug" | "fatal";
|
||||||
message: string;
|
message: string;
|
||||||
room?: string;
|
room?: string;
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
vars {
|
vars {
|
||||||
url: http://localhost:3000/lst
|
url: http://localhost:3600/lst
|
||||||
readerIp: 10.44.14.215
|
readerIp: 10.44.14.215
|
||||||
}
|
}
|
||||||
vars:secret [
|
vars:secret [
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ body:json {
|
|||||||
{
|
{
|
||||||
"userId":"m6AbQXFwOXoX3YKLfwWgq2LIdDqS5jqv",
|
"userId":"m6AbQXFwOXoX3YKLfwWgq2LIdDqS5jqv",
|
||||||
"notificationId": "0399eb2a-39df-48b7-9f1c-d233cec94d2e",
|
"notificationId": "0399eb2a-39df-48b7-9f1c-d233cec94d2e",
|
||||||
"emails": ["blake.mattes@alpla.com"]
|
"emails": ["blake.mattes@alpla.com","cowchmonkey@gmail.com"]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
22
brunoApi/ocp/Printer Listenter.bru
Normal file
22
brunoApi/ocp/Printer Listenter.bru
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
meta {
|
||||||
|
name: Printer Listenter
|
||||||
|
type: http
|
||||||
|
seq: 1
|
||||||
|
}
|
||||||
|
|
||||||
|
post {
|
||||||
|
url: {{url}}/api/ocp/printer/listener/line_1
|
||||||
|
body: json
|
||||||
|
auth: inherit
|
||||||
|
}
|
||||||
|
|
||||||
|
body:json {
|
||||||
|
{
|
||||||
|
"message":"xnvjdhhgsdfr"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
settings {
|
||||||
|
encodeUrl: true
|
||||||
|
timeout: 0
|
||||||
|
}
|
||||||
8
brunoApi/ocp/folder.bru
Normal file
8
brunoApi/ocp/folder.bru
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
meta {
|
||||||
|
name: ocp
|
||||||
|
seq: 9
|
||||||
|
}
|
||||||
|
|
||||||
|
auth {
|
||||||
|
mode: inherit
|
||||||
|
}
|
||||||
@@ -1,16 +1,26 @@
|
|||||||
services:
|
services:
|
||||||
app:
|
lst:
|
||||||
build:
|
image: git.tuffraid.net/cowch/lst_v3:latest
|
||||||
context: .
|
|
||||||
dockerfile: Dockerfile
|
|
||||||
container_name: lst_app
|
container_name: lst_app
|
||||||
|
restart: unless-stopped
|
||||||
|
# app:
|
||||||
|
# build:
|
||||||
|
# context: .
|
||||||
|
# dockerfile: Dockerfile
|
||||||
|
# container_name: lst_app
|
||||||
ports:
|
ports:
|
||||||
#- "${VITE_PORT:-4200}:4200"
|
#- "${VITE_PORT:-4200}:4200"
|
||||||
- "3600:3000"
|
- "3600:3000"
|
||||||
|
dns:
|
||||||
|
- 10.193.9.250
|
||||||
|
- 10.193.9.251 # your internal DNS server
|
||||||
|
dns_search:
|
||||||
|
- alpla.net # or your internal search suffix
|
||||||
environment:
|
environment:
|
||||||
- NODE_ENV=production
|
- NODE_ENV=production
|
||||||
- LOG_LEVEL=info
|
- LOG_LEVEL=info
|
||||||
- DATABASE_HOST=host.docker.internal
|
- EXTERNAL_URL=http://192.168.8.222:3600
|
||||||
|
- DATABASE_HOST=host.docker.internal # if running on the same docker then do this
|
||||||
- DATABASE_PORT=5433
|
- DATABASE_PORT=5433
|
||||||
- DATABASE_USER=${DATABASE_USER}
|
- DATABASE_USER=${DATABASE_USER}
|
||||||
- DATABASE_PASSWORD=${DATABASE_PASSWORD}
|
- DATABASE_PASSWORD=${DATABASE_PASSWORD}
|
||||||
@@ -21,7 +31,6 @@ services:
|
|||||||
- PROD_PASSWORD=${PROD_PASSWORD}
|
- PROD_PASSWORD=${PROD_PASSWORD}
|
||||||
- BETTER_AUTH_SECRET=${BETTER_AUTH_SECRET}
|
- BETTER_AUTH_SECRET=${BETTER_AUTH_SECRET}
|
||||||
- BETTER_AUTH_URL=${URL}
|
- BETTER_AUTH_URL=${URL}
|
||||||
restart: unless-stopped
|
|
||||||
# for all host including prod servers, plc's, printers, or other de
|
# for all host including prod servers, plc's, printers, or other de
|
||||||
# extra_hosts:
|
# extra_hosts:
|
||||||
# - "${PROD_SERVER}:${PROD_IP}"
|
# - "${PROD_SERVER}:${PROD_IP}"
|
||||||
|
|||||||
@@ -1,69 +0,0 @@
|
|||||||
# React + TypeScript + Vite
|
|
||||||
|
|
||||||
This template provides a minimal setup to get React working in Vite with HMR and some ESLint rules.
|
|
||||||
|
|
||||||
Currently, two official plugins are available:
|
|
||||||
|
|
||||||
- [@vitejs/plugin-react](https://github.com/vitejs/vite-plugin-react/blob/main/packages/plugin-react) uses [Babel](https://babeljs.io/) for Fast Refresh
|
|
||||||
- [@vitejs/plugin-react-swc](https://github.com/vitejs/vite-plugin-react/blob/main/packages/plugin-react-swc) uses [SWC](https://swc.rs/) for Fast Refresh
|
|
||||||
|
|
||||||
## Expanding the ESLint configuration
|
|
||||||
|
|
||||||
If you are developing a production application, we recommend updating the configuration to enable type-aware lint rules:
|
|
||||||
|
|
||||||
```js
|
|
||||||
export default defineConfig([
|
|
||||||
globalIgnores(['dist']),
|
|
||||||
{
|
|
||||||
files: ['**/*.{ts,tsx}'],
|
|
||||||
extends: [
|
|
||||||
// Other configs...
|
|
||||||
|
|
||||||
// Remove tseslint.configs.recommended and replace with this
|
|
||||||
tseslint.configs.recommendedTypeChecked,
|
|
||||||
// Alternatively, use this for stricter rules
|
|
||||||
tseslint.configs.strictTypeChecked,
|
|
||||||
// Optionally, add this for stylistic rules
|
|
||||||
tseslint.configs.stylisticTypeChecked,
|
|
||||||
|
|
||||||
// Other configs...
|
|
||||||
],
|
|
||||||
languageOptions: {
|
|
||||||
parserOptions: {
|
|
||||||
project: ['./tsconfig.node.json', './tsconfig.app.json'],
|
|
||||||
tsconfigRootDir: import.meta.dirname,
|
|
||||||
},
|
|
||||||
// other options...
|
|
||||||
},
|
|
||||||
},
|
|
||||||
])
|
|
||||||
```
|
|
||||||
|
|
||||||
You can also install [eslint-plugin-react-x](https://github.com/Rel1cx/eslint-react/tree/main/packages/plugins/eslint-plugin-react-x) and [eslint-plugin-react-dom](https://github.com/Rel1cx/eslint-react/tree/main/packages/plugins/eslint-plugin-react-dom) for React-specific lint rules:
|
|
||||||
|
|
||||||
```js
|
|
||||||
// eslint.config.js
|
|
||||||
import reactX from 'eslint-plugin-react-x'
|
|
||||||
import reactDom from 'eslint-plugin-react-dom'
|
|
||||||
|
|
||||||
export default defineConfig([
|
|
||||||
globalIgnores(['dist']),
|
|
||||||
{
|
|
||||||
files: ['**/*.{ts,tsx}'],
|
|
||||||
extends: [
|
|
||||||
// Other configs...
|
|
||||||
// Enable lint rules for React
|
|
||||||
reactX.configs['recommended-typescript'],
|
|
||||||
// Enable lint rules for React DOM
|
|
||||||
reactDom.configs.recommended,
|
|
||||||
],
|
|
||||||
languageOptions: {
|
|
||||||
parserOptions: {
|
|
||||||
project: ['./tsconfig.node.json', './tsconfig.app.json'],
|
|
||||||
tsconfigRootDir: import.meta.dirname,
|
|
||||||
},
|
|
||||||
// other options...
|
|
||||||
},
|
|
||||||
},
|
|
||||||
])
|
|
||||||
```
|
|
||||||
@@ -1,5 +1,5 @@
|
|||||||
import { Link } from "@tanstack/react-router";
|
import { Link } from "@tanstack/react-router";
|
||||||
import { Logs } from "lucide-react";
|
import { Bell, Logs, Settings } from "lucide-react";
|
||||||
|
|
||||||
import {
|
import {
|
||||||
SidebarGroup,
|
SidebarGroup,
|
||||||
@@ -11,23 +11,31 @@ import {
|
|||||||
useSidebar,
|
useSidebar,
|
||||||
} from "../ui/sidebar";
|
} from "../ui/sidebar";
|
||||||
|
|
||||||
type AdminSidebarProps = {
|
// type AdminSidebarProps = {
|
||||||
session: {
|
// session: {
|
||||||
user: {
|
// user: {
|
||||||
name?: string | null;
|
// name?: string | null;
|
||||||
email?: string | null;
|
// email?: string | null;
|
||||||
role?: string | string[];
|
// role?: string | string[];
|
||||||
};
|
// };
|
||||||
} | null;
|
// } | null;
|
||||||
};
|
//};
|
||||||
|
|
||||||
export default function AdminSidebar({ session }: any) {
|
export default function AdminSidebar({ session }: any) {
|
||||||
const { setOpen } = useSidebar();
|
const { setOpen } = useSidebar();
|
||||||
const items = [
|
const items = [
|
||||||
|
{
|
||||||
|
title: "Notifications",
|
||||||
|
url: "/admin/notifications",
|
||||||
|
icon: Bell,
|
||||||
|
role: ["systemAdmin", "admin"],
|
||||||
|
module: "admin",
|
||||||
|
active: true,
|
||||||
|
},
|
||||||
{
|
{
|
||||||
title: "Settings",
|
title: "Settings",
|
||||||
url: "/admin/settings",
|
url: "/admin/settings",
|
||||||
icon: Logs,
|
icon: Settings,
|
||||||
role: ["systemAdmin"],
|
role: ["systemAdmin"],
|
||||||
module: "admin",
|
module: "admin",
|
||||||
active: true,
|
active: true,
|
||||||
|
|||||||
190
frontend/src/components/ui/select.tsx
Normal file
190
frontend/src/components/ui/select.tsx
Normal file
@@ -0,0 +1,190 @@
|
|||||||
|
import * as React from "react"
|
||||||
|
import { Select as SelectPrimitive } from "radix-ui"
|
||||||
|
|
||||||
|
import { cn } from "@/lib/utils"
|
||||||
|
import { ChevronDownIcon, CheckIcon, ChevronUpIcon } from "lucide-react"
|
||||||
|
|
||||||
|
function Select({
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Root>) {
|
||||||
|
return <SelectPrimitive.Root data-slot="select" {...props} />
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectGroup({
|
||||||
|
className,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Group>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.Group
|
||||||
|
data-slot="select-group"
|
||||||
|
className={cn("scroll-my-1 p-1", className)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectValue({
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Value>) {
|
||||||
|
return <SelectPrimitive.Value data-slot="select-value" {...props} />
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectTrigger({
|
||||||
|
className,
|
||||||
|
size = "default",
|
||||||
|
children,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Trigger> & {
|
||||||
|
size?: "sm" | "default"
|
||||||
|
}) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.Trigger
|
||||||
|
data-slot="select-trigger"
|
||||||
|
data-size={size}
|
||||||
|
className={cn(
|
||||||
|
"flex w-fit items-center justify-between gap-1.5 rounded-lg border border-input bg-transparent py-2 pr-2 pl-2.5 text-sm whitespace-nowrap transition-colors outline-none select-none focus-visible:border-ring focus-visible:ring-3 focus-visible:ring-ring/50 disabled:cursor-not-allowed disabled:opacity-50 aria-invalid:border-destructive aria-invalid:ring-3 aria-invalid:ring-destructive/20 data-placeholder:text-muted-foreground data-[size=default]:h-8 data-[size=sm]:h-7 data-[size=sm]:rounded-[min(var(--radius-md),10px)] *:data-[slot=select-value]:line-clamp-1 *:data-[slot=select-value]:flex *:data-[slot=select-value]:items-center *:data-[slot=select-value]:gap-1.5 dark:bg-input/30 dark:hover:bg-input/50 dark:aria-invalid:border-destructive/50 dark:aria-invalid:ring-destructive/40 [&_svg]:pointer-events-none [&_svg]:shrink-0 [&_svg:not([class*='size-'])]:size-4",
|
||||||
|
className
|
||||||
|
)}
|
||||||
|
{...props}
|
||||||
|
>
|
||||||
|
{children}
|
||||||
|
<SelectPrimitive.Icon asChild>
|
||||||
|
<ChevronDownIcon className="pointer-events-none size-4 text-muted-foreground" />
|
||||||
|
</SelectPrimitive.Icon>
|
||||||
|
</SelectPrimitive.Trigger>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectContent({
|
||||||
|
className,
|
||||||
|
children,
|
||||||
|
position = "item-aligned",
|
||||||
|
align = "center",
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Content>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.Portal>
|
||||||
|
<SelectPrimitive.Content
|
||||||
|
data-slot="select-content"
|
||||||
|
data-align-trigger={position === "item-aligned"}
|
||||||
|
className={cn("relative z-50 max-h-(--radix-select-content-available-height) min-w-36 origin-(--radix-select-content-transform-origin) overflow-x-hidden overflow-y-auto rounded-lg bg-popover text-popover-foreground shadow-md ring-1 ring-foreground/10 duration-100 data-[align-trigger=true]:animate-none data-[side=bottom]:slide-in-from-top-2 data-[side=left]:slide-in-from-right-2 data-[side=right]:slide-in-from-left-2 data-[side=top]:slide-in-from-bottom-2 data-open:animate-in data-open:fade-in-0 data-open:zoom-in-95 data-closed:animate-out data-closed:fade-out-0 data-closed:zoom-out-95", position ==="popper"&&"data-[side=bottom]:translate-y-1 data-[side=left]:-translate-x-1 data-[side=right]:translate-x-1 data-[side=top]:-translate-y-1", className )}
|
||||||
|
position={position}
|
||||||
|
align={align}
|
||||||
|
{...props}
|
||||||
|
>
|
||||||
|
<SelectScrollUpButton />
|
||||||
|
<SelectPrimitive.Viewport
|
||||||
|
data-position={position}
|
||||||
|
className={cn(
|
||||||
|
"data-[position=popper]:h-(--radix-select-trigger-height) data-[position=popper]:w-full data-[position=popper]:min-w-(--radix-select-trigger-width)",
|
||||||
|
position === "popper" && ""
|
||||||
|
)}
|
||||||
|
>
|
||||||
|
{children}
|
||||||
|
</SelectPrimitive.Viewport>
|
||||||
|
<SelectScrollDownButton />
|
||||||
|
</SelectPrimitive.Content>
|
||||||
|
</SelectPrimitive.Portal>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectLabel({
|
||||||
|
className,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Label>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.Label
|
||||||
|
data-slot="select-label"
|
||||||
|
className={cn("px-1.5 py-1 text-xs text-muted-foreground", className)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectItem({
|
||||||
|
className,
|
||||||
|
children,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Item>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.Item
|
||||||
|
data-slot="select-item"
|
||||||
|
className={cn(
|
||||||
|
"relative flex w-full cursor-default items-center gap-1.5 rounded-md py-1 pr-8 pl-1.5 text-sm outline-hidden select-none focus:bg-accent focus:text-accent-foreground not-data-[variant=destructive]:focus:**:text-accent-foreground data-disabled:pointer-events-none data-disabled:opacity-50 [&_svg]:pointer-events-none [&_svg]:shrink-0 [&_svg:not([class*='size-'])]:size-4 *:[span]:last:flex *:[span]:last:items-center *:[span]:last:gap-2",
|
||||||
|
className
|
||||||
|
)}
|
||||||
|
{...props}
|
||||||
|
>
|
||||||
|
<span className="pointer-events-none absolute right-2 flex size-4 items-center justify-center">
|
||||||
|
<SelectPrimitive.ItemIndicator>
|
||||||
|
<CheckIcon className="pointer-events-none" />
|
||||||
|
</SelectPrimitive.ItemIndicator>
|
||||||
|
</span>
|
||||||
|
<SelectPrimitive.ItemText>{children}</SelectPrimitive.ItemText>
|
||||||
|
</SelectPrimitive.Item>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectSeparator({
|
||||||
|
className,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.Separator>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.Separator
|
||||||
|
data-slot="select-separator"
|
||||||
|
className={cn("pointer-events-none -mx-1 my-1 h-px bg-border", className)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectScrollUpButton({
|
||||||
|
className,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.ScrollUpButton>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.ScrollUpButton
|
||||||
|
data-slot="select-scroll-up-button"
|
||||||
|
className={cn(
|
||||||
|
"z-10 flex cursor-default items-center justify-center bg-popover py-1 [&_svg:not([class*='size-'])]:size-4",
|
||||||
|
className
|
||||||
|
)}
|
||||||
|
{...props}
|
||||||
|
>
|
||||||
|
<ChevronUpIcon
|
||||||
|
/>
|
||||||
|
</SelectPrimitive.ScrollUpButton>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function SelectScrollDownButton({
|
||||||
|
className,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof SelectPrimitive.ScrollDownButton>) {
|
||||||
|
return (
|
||||||
|
<SelectPrimitive.ScrollDownButton
|
||||||
|
data-slot="select-scroll-down-button"
|
||||||
|
className={cn(
|
||||||
|
"z-10 flex cursor-default items-center justify-center bg-popover py-1 [&_svg:not([class*='size-'])]:size-4",
|
||||||
|
className
|
||||||
|
)}
|
||||||
|
{...props}
|
||||||
|
>
|
||||||
|
<ChevronDownIcon
|
||||||
|
/>
|
||||||
|
</SelectPrimitive.ScrollDownButton>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
export {
|
||||||
|
Select,
|
||||||
|
SelectContent,
|
||||||
|
SelectGroup,
|
||||||
|
SelectItem,
|
||||||
|
SelectLabel,
|
||||||
|
SelectScrollDownButton,
|
||||||
|
SelectScrollUpButton,
|
||||||
|
SelectSeparator,
|
||||||
|
SelectTrigger,
|
||||||
|
SelectValue,
|
||||||
|
}
|
||||||
87
frontend/src/lib/formSutff/DynamicInput.Field.tsx
Normal file
87
frontend/src/lib/formSutff/DynamicInput.Field.tsx
Normal file
@@ -0,0 +1,87 @@
|
|||||||
|
import { Trash2 } from "lucide-react";
|
||||||
|
import { Input } from "@/components/ui/input";
|
||||||
|
import { Label } from "@/components/ui/label";
|
||||||
|
import { Button } from "../../components/ui/button";
|
||||||
|
import { useFieldContext } from ".";
|
||||||
|
import { FieldErrors } from "./Errors.Field";
|
||||||
|
|
||||||
|
type DynamicInputField = {
|
||||||
|
name?: string;
|
||||||
|
label: string;
|
||||||
|
inputType: "text" | "email" | "password" | "number" | "username";
|
||||||
|
required?: boolean;
|
||||||
|
description?: string;
|
||||||
|
addLabel?: string;
|
||||||
|
placeholder?: string;
|
||||||
|
disabled?: boolean;
|
||||||
|
};
|
||||||
|
|
||||||
|
const autoCompleteMap: Record<string, string> = {
|
||||||
|
email: "email",
|
||||||
|
password: "current-password",
|
||||||
|
text: "off",
|
||||||
|
username: "username",
|
||||||
|
};
|
||||||
|
|
||||||
|
export const DynamicInputField = ({
|
||||||
|
label,
|
||||||
|
inputType = "text",
|
||||||
|
required = false,
|
||||||
|
description,
|
||||||
|
addLabel,
|
||||||
|
}: DynamicInputField) => {
|
||||||
|
const field = useFieldContext<any>();
|
||||||
|
const values = Array.isArray(field.state.value) ? field.state.value : [];
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div className="grid gap-3 mt-2">
|
||||||
|
<div className="flex items-start justify-between gap-4">
|
||||||
|
<div className="space-y-1">
|
||||||
|
<Label>{label}</Label>
|
||||||
|
{description ? (
|
||||||
|
<p className="text-sm text-muted-foreground">{description}</p>
|
||||||
|
) : null}
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<Button
|
||||||
|
type="button"
|
||||||
|
variant="secondary"
|
||||||
|
onClick={() => {
|
||||||
|
field.pushValue("");
|
||||||
|
}}
|
||||||
|
>
|
||||||
|
{addLabel}
|
||||||
|
</Button>
|
||||||
|
</div>
|
||||||
|
<div className="grid gap-3">
|
||||||
|
{values.map((_: string, index: number) => (
|
||||||
|
<div key={`${field.name}-${index}`} className="grid gap-2">
|
||||||
|
<div className="flex items-center gap-2">
|
||||||
|
<Label htmlFor={field.name}>{label}</Label>
|
||||||
|
<Input
|
||||||
|
id={field.name}
|
||||||
|
autoComplete={autoCompleteMap[inputType] ?? "off"}
|
||||||
|
value={field.state.value?.[index] ?? ""}
|
||||||
|
onChange={(e) => field.replaceValue(index, e.target.value)}
|
||||||
|
onBlur={field.handleBlur}
|
||||||
|
type={inputType}
|
||||||
|
required={required}
|
||||||
|
/>
|
||||||
|
{values.length > 1 ? (
|
||||||
|
<Button
|
||||||
|
type="button"
|
||||||
|
size={"icon"}
|
||||||
|
variant="destructive"
|
||||||
|
onClick={() => field.removeValue(index)}
|
||||||
|
>
|
||||||
|
<Trash2 className="w-32 h-32" />
|
||||||
|
</Button>
|
||||||
|
) : null}
|
||||||
|
<FieldErrors meta={field.state.meta} />
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
))}
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
};
|
||||||
57
frontend/src/lib/formSutff/Select.Field.tsx
Normal file
57
frontend/src/lib/formSutff/Select.Field.tsx
Normal file
@@ -0,0 +1,57 @@
|
|||||||
|
import { Label } from "../../components/ui/label";
|
||||||
|
import {
|
||||||
|
Select,
|
||||||
|
SelectContent,
|
||||||
|
SelectItem,
|
||||||
|
SelectTrigger,
|
||||||
|
SelectValue,
|
||||||
|
} from "../../components/ui/select";
|
||||||
|
import { useFieldContext } from ".";
|
||||||
|
import { FieldErrors } from "./Errors.Field";
|
||||||
|
|
||||||
|
type SelectOption = {
|
||||||
|
value: string;
|
||||||
|
label: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
type SelectFieldProps = {
|
||||||
|
label: string;
|
||||||
|
options: SelectOption[];
|
||||||
|
placeholder?: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
export const SelectField = ({
|
||||||
|
label,
|
||||||
|
options,
|
||||||
|
placeholder,
|
||||||
|
}: SelectFieldProps) => {
|
||||||
|
const field = useFieldContext<string>();
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div className="grid gap-3">
|
||||||
|
<div className="grid gap-3">
|
||||||
|
<Label htmlFor={field.name}>{label}</Label>
|
||||||
|
<Select
|
||||||
|
value={field.state.value}
|
||||||
|
onValueChange={(value) => field.handleChange(value)}
|
||||||
|
>
|
||||||
|
<SelectTrigger
|
||||||
|
id={field.name}
|
||||||
|
onBlur={field.handleBlur}
|
||||||
|
className="w-min-2/3 w-max-fit"
|
||||||
|
>
|
||||||
|
<SelectValue placeholder={placeholder} />
|
||||||
|
</SelectTrigger>
|
||||||
|
<SelectContent position={"popper"}>
|
||||||
|
{options.map((option) => (
|
||||||
|
<SelectItem key={option.value} value={option.value}>
|
||||||
|
{option.label}
|
||||||
|
</SelectItem>
|
||||||
|
))}
|
||||||
|
</SelectContent>
|
||||||
|
</Select>
|
||||||
|
</div>
|
||||||
|
<FieldErrors meta={field.state.meta} />
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
};
|
||||||
@@ -1,7 +1,9 @@
|
|||||||
import { createFormHook, createFormHookContexts } from "@tanstack/react-form";
|
import { createFormHook, createFormHookContexts } from "@tanstack/react-form";
|
||||||
import { CheckboxField } from "./CheckBox.Field";
|
import { CheckboxField } from "./CheckBox.Field";
|
||||||
|
import { DynamicInputField } from "./DynamicInput.Field";
|
||||||
import { InputField } from "./Input.Field";
|
import { InputField } from "./Input.Field";
|
||||||
import { InputPasswordField } from "./InputPassword.Field";
|
import { InputPasswordField } from "./InputPassword.Field";
|
||||||
|
import { SelectField } from "./Select.Field";
|
||||||
import { SubmitButton } from "./SubmitButton";
|
import { SubmitButton } from "./SubmitButton";
|
||||||
import { SwitchField } from "./Switch.Field";
|
import { SwitchField } from "./Switch.Field";
|
||||||
|
|
||||||
@@ -12,12 +14,13 @@ export const { useAppForm } = createFormHook({
|
|||||||
fieldComponents: {
|
fieldComponents: {
|
||||||
InputField,
|
InputField,
|
||||||
InputPasswordField,
|
InputPasswordField,
|
||||||
//SelectField,
|
SelectField,
|
||||||
CheckboxField,
|
CheckboxField,
|
||||||
//DateField,
|
//DateField,
|
||||||
//TextArea,
|
//TextArea,
|
||||||
//Searchable,
|
//Searchable,
|
||||||
SwitchField,
|
SwitchField,
|
||||||
|
DynamicInputField,
|
||||||
},
|
},
|
||||||
formComponents: { SubmitButton },
|
formComponents: { SubmitButton },
|
||||||
fieldContext,
|
fieldContext,
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ export function getSettings() {
|
|||||||
|
|
||||||
const fetch = async () => {
|
const fetch = async () => {
|
||||||
if (window.location.hostname === "localhost") {
|
if (window.location.hostname === "localhost") {
|
||||||
await new Promise((res) => setTimeout(res, 5000));
|
await new Promise((res) => setTimeout(res, 1500));
|
||||||
}
|
}
|
||||||
|
|
||||||
const { data } = await axios.get("/lst/api/settings");
|
const { data } = await axios.get("/lst/api/settings");
|
||||||
|
|||||||
24
frontend/src/lib/queries/notificationSubs.ts
Normal file
24
frontend/src/lib/queries/notificationSubs.ts
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
import { keepPreviousData, queryOptions } from "@tanstack/react-query";
|
||||||
|
import axios from "axios";
|
||||||
|
|
||||||
|
export function notificationSubs(userId?: string) {
|
||||||
|
return queryOptions({
|
||||||
|
queryKey: ["notificationSubs"],
|
||||||
|
queryFn: () => fetch(userId),
|
||||||
|
staleTime: 5000,
|
||||||
|
refetchOnWindowFocus: true,
|
||||||
|
placeholderData: keepPreviousData,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const fetch = async (userId?: string) => {
|
||||||
|
if (window.location.hostname === "localhost") {
|
||||||
|
await new Promise((res) => setTimeout(res, 1500));
|
||||||
|
}
|
||||||
|
|
||||||
|
const { data } = await axios.get(
|
||||||
|
`/lst/api/notification/sub${userId ? `?userId=${userId}` : ""}`,
|
||||||
|
);
|
||||||
|
|
||||||
|
return data.data;
|
||||||
|
};
|
||||||
22
frontend/src/lib/queries/notifications.ts
Normal file
22
frontend/src/lib/queries/notifications.ts
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
import { keepPreviousData, queryOptions } from "@tanstack/react-query";
|
||||||
|
import axios from "axios";
|
||||||
|
|
||||||
|
export function notifications() {
|
||||||
|
return queryOptions({
|
||||||
|
queryKey: ["notifications"],
|
||||||
|
queryFn: () => fetch(),
|
||||||
|
staleTime: 5000,
|
||||||
|
refetchOnWindowFocus: true,
|
||||||
|
placeholderData: keepPreviousData,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const fetch = async () => {
|
||||||
|
if (window.location.hostname === "localhost") {
|
||||||
|
await new Promise((res) => setTimeout(res, 1500));
|
||||||
|
}
|
||||||
|
|
||||||
|
const { data } = await axios.get("/lst/api/notification");
|
||||||
|
|
||||||
|
return data.data;
|
||||||
|
};
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
import type { Column } from "@tanstack/react-table";
|
import type { Column } from "@tanstack/react-table";
|
||||||
import { ArrowDown, ArrowUp, Search } from "lucide-react";
|
import { ArrowDown, ArrowUp, Search } from "lucide-react";
|
||||||
import React, { useState } from "react";
|
import { useState } from "react";
|
||||||
import { Button } from "../../components/ui/button";
|
import { Button } from "../../components/ui/button";
|
||||||
import {
|
import {
|
||||||
DropdownMenu,
|
DropdownMenu,
|
||||||
|
|||||||
@@ -12,6 +12,7 @@ import { Route as rootRouteImport } from './routes/__root'
|
|||||||
import { Route as AboutRouteImport } from './routes/about'
|
import { Route as AboutRouteImport } from './routes/about'
|
||||||
import { Route as IndexRouteImport } from './routes/index'
|
import { Route as IndexRouteImport } from './routes/index'
|
||||||
import { Route as AdminSettingsRouteImport } from './routes/admin/settings'
|
import { Route as AdminSettingsRouteImport } from './routes/admin/settings'
|
||||||
|
import { Route as AdminNotificationsRouteImport } from './routes/admin/notifications'
|
||||||
import { Route as AdminLogsRouteImport } from './routes/admin/logs'
|
import { Route as AdminLogsRouteImport } from './routes/admin/logs'
|
||||||
import { Route as authLoginRouteImport } from './routes/(auth)/login'
|
import { Route as authLoginRouteImport } from './routes/(auth)/login'
|
||||||
import { Route as authUserSignupRouteImport } from './routes/(auth)/user.signup'
|
import { Route as authUserSignupRouteImport } from './routes/(auth)/user.signup'
|
||||||
@@ -33,6 +34,11 @@ const AdminSettingsRoute = AdminSettingsRouteImport.update({
|
|||||||
path: '/admin/settings',
|
path: '/admin/settings',
|
||||||
getParentRoute: () => rootRouteImport,
|
getParentRoute: () => rootRouteImport,
|
||||||
} as any)
|
} as any)
|
||||||
|
const AdminNotificationsRoute = AdminNotificationsRouteImport.update({
|
||||||
|
id: '/admin/notifications',
|
||||||
|
path: '/admin/notifications',
|
||||||
|
getParentRoute: () => rootRouteImport,
|
||||||
|
} as any)
|
||||||
const AdminLogsRoute = AdminLogsRouteImport.update({
|
const AdminLogsRoute = AdminLogsRouteImport.update({
|
||||||
id: '/admin/logs',
|
id: '/admin/logs',
|
||||||
path: '/admin/logs',
|
path: '/admin/logs',
|
||||||
@@ -64,6 +70,7 @@ export interface FileRoutesByFullPath {
|
|||||||
'/about': typeof AboutRoute
|
'/about': typeof AboutRoute
|
||||||
'/login': typeof authLoginRoute
|
'/login': typeof authLoginRoute
|
||||||
'/admin/logs': typeof AdminLogsRoute
|
'/admin/logs': typeof AdminLogsRoute
|
||||||
|
'/admin/notifications': typeof AdminNotificationsRoute
|
||||||
'/admin/settings': typeof AdminSettingsRoute
|
'/admin/settings': typeof AdminSettingsRoute
|
||||||
'/user/profile': typeof authUserProfileRoute
|
'/user/profile': typeof authUserProfileRoute
|
||||||
'/user/resetpassword': typeof authUserResetpasswordRoute
|
'/user/resetpassword': typeof authUserResetpasswordRoute
|
||||||
@@ -74,6 +81,7 @@ export interface FileRoutesByTo {
|
|||||||
'/about': typeof AboutRoute
|
'/about': typeof AboutRoute
|
||||||
'/login': typeof authLoginRoute
|
'/login': typeof authLoginRoute
|
||||||
'/admin/logs': typeof AdminLogsRoute
|
'/admin/logs': typeof AdminLogsRoute
|
||||||
|
'/admin/notifications': typeof AdminNotificationsRoute
|
||||||
'/admin/settings': typeof AdminSettingsRoute
|
'/admin/settings': typeof AdminSettingsRoute
|
||||||
'/user/profile': typeof authUserProfileRoute
|
'/user/profile': typeof authUserProfileRoute
|
||||||
'/user/resetpassword': typeof authUserResetpasswordRoute
|
'/user/resetpassword': typeof authUserResetpasswordRoute
|
||||||
@@ -85,6 +93,7 @@ export interface FileRoutesById {
|
|||||||
'/about': typeof AboutRoute
|
'/about': typeof AboutRoute
|
||||||
'/(auth)/login': typeof authLoginRoute
|
'/(auth)/login': typeof authLoginRoute
|
||||||
'/admin/logs': typeof AdminLogsRoute
|
'/admin/logs': typeof AdminLogsRoute
|
||||||
|
'/admin/notifications': typeof AdminNotificationsRoute
|
||||||
'/admin/settings': typeof AdminSettingsRoute
|
'/admin/settings': typeof AdminSettingsRoute
|
||||||
'/(auth)/user/profile': typeof authUserProfileRoute
|
'/(auth)/user/profile': typeof authUserProfileRoute
|
||||||
'/(auth)/user/resetpassword': typeof authUserResetpasswordRoute
|
'/(auth)/user/resetpassword': typeof authUserResetpasswordRoute
|
||||||
@@ -97,6 +106,7 @@ export interface FileRouteTypes {
|
|||||||
| '/about'
|
| '/about'
|
||||||
| '/login'
|
| '/login'
|
||||||
| '/admin/logs'
|
| '/admin/logs'
|
||||||
|
| '/admin/notifications'
|
||||||
| '/admin/settings'
|
| '/admin/settings'
|
||||||
| '/user/profile'
|
| '/user/profile'
|
||||||
| '/user/resetpassword'
|
| '/user/resetpassword'
|
||||||
@@ -107,6 +117,7 @@ export interface FileRouteTypes {
|
|||||||
| '/about'
|
| '/about'
|
||||||
| '/login'
|
| '/login'
|
||||||
| '/admin/logs'
|
| '/admin/logs'
|
||||||
|
| '/admin/notifications'
|
||||||
| '/admin/settings'
|
| '/admin/settings'
|
||||||
| '/user/profile'
|
| '/user/profile'
|
||||||
| '/user/resetpassword'
|
| '/user/resetpassword'
|
||||||
@@ -117,6 +128,7 @@ export interface FileRouteTypes {
|
|||||||
| '/about'
|
| '/about'
|
||||||
| '/(auth)/login'
|
| '/(auth)/login'
|
||||||
| '/admin/logs'
|
| '/admin/logs'
|
||||||
|
| '/admin/notifications'
|
||||||
| '/admin/settings'
|
| '/admin/settings'
|
||||||
| '/(auth)/user/profile'
|
| '/(auth)/user/profile'
|
||||||
| '/(auth)/user/resetpassword'
|
| '/(auth)/user/resetpassword'
|
||||||
@@ -128,6 +140,7 @@ export interface RootRouteChildren {
|
|||||||
AboutRoute: typeof AboutRoute
|
AboutRoute: typeof AboutRoute
|
||||||
authLoginRoute: typeof authLoginRoute
|
authLoginRoute: typeof authLoginRoute
|
||||||
AdminLogsRoute: typeof AdminLogsRoute
|
AdminLogsRoute: typeof AdminLogsRoute
|
||||||
|
AdminNotificationsRoute: typeof AdminNotificationsRoute
|
||||||
AdminSettingsRoute: typeof AdminSettingsRoute
|
AdminSettingsRoute: typeof AdminSettingsRoute
|
||||||
authUserProfileRoute: typeof authUserProfileRoute
|
authUserProfileRoute: typeof authUserProfileRoute
|
||||||
authUserResetpasswordRoute: typeof authUserResetpasswordRoute
|
authUserResetpasswordRoute: typeof authUserResetpasswordRoute
|
||||||
@@ -157,6 +170,13 @@ declare module '@tanstack/react-router' {
|
|||||||
preLoaderRoute: typeof AdminSettingsRouteImport
|
preLoaderRoute: typeof AdminSettingsRouteImport
|
||||||
parentRoute: typeof rootRouteImport
|
parentRoute: typeof rootRouteImport
|
||||||
}
|
}
|
||||||
|
'/admin/notifications': {
|
||||||
|
id: '/admin/notifications'
|
||||||
|
path: '/admin/notifications'
|
||||||
|
fullPath: '/admin/notifications'
|
||||||
|
preLoaderRoute: typeof AdminNotificationsRouteImport
|
||||||
|
parentRoute: typeof rootRouteImport
|
||||||
|
}
|
||||||
'/admin/logs': {
|
'/admin/logs': {
|
||||||
id: '/admin/logs'
|
id: '/admin/logs'
|
||||||
path: '/admin/logs'
|
path: '/admin/logs'
|
||||||
@@ -200,6 +220,7 @@ const rootRouteChildren: RootRouteChildren = {
|
|||||||
AboutRoute: AboutRoute,
|
AboutRoute: AboutRoute,
|
||||||
authLoginRoute: authLoginRoute,
|
authLoginRoute: authLoginRoute,
|
||||||
AdminLogsRoute: AdminLogsRoute,
|
AdminLogsRoute: AdminLogsRoute,
|
||||||
|
AdminNotificationsRoute: AdminNotificationsRoute,
|
||||||
AdminSettingsRoute: AdminSettingsRoute,
|
AdminSettingsRoute: AdminSettingsRoute,
|
||||||
authUserProfileRoute: authUserProfileRoute,
|
authUserProfileRoute: authUserProfileRoute,
|
||||||
authUserResetpasswordRoute: authUserResetpasswordRoute,
|
authUserResetpasswordRoute: authUserResetpasswordRoute,
|
||||||
|
|||||||
104
frontend/src/routes/(auth)/-components/NotificationsSubCard.tsx
Normal file
104
frontend/src/routes/(auth)/-components/NotificationsSubCard.tsx
Normal file
@@ -0,0 +1,104 @@
|
|||||||
|
import { useSuspenseQuery } from "@tanstack/react-query";
|
||||||
|
import axios from "axios";
|
||||||
|
import { toast } from "sonner";
|
||||||
|
import {
|
||||||
|
Card,
|
||||||
|
CardContent,
|
||||||
|
CardDescription,
|
||||||
|
CardHeader,
|
||||||
|
CardTitle,
|
||||||
|
} from "../../../components/ui/card";
|
||||||
|
import { useAppForm } from "../../../lib/formSutff";
|
||||||
|
import { notificationSubs } from "../../../lib/queries/notificationSubs";
|
||||||
|
import { notifications } from "../../../lib/queries/notifications";
|
||||||
|
|
||||||
|
export default function NotificationsSubCard({ user }: any) {
|
||||||
|
const { data } = useSuspenseQuery(notifications());
|
||||||
|
const { refetch } = useSuspenseQuery(notificationSubs(user.id));
|
||||||
|
const form = useAppForm({
|
||||||
|
defaultValues: {
|
||||||
|
notificationId: "",
|
||||||
|
emails: [user.email],
|
||||||
|
},
|
||||||
|
onSubmit: async ({ value }) => {
|
||||||
|
if (value.notificationId === "") {
|
||||||
|
toast.error("Please select a notification before trying to subscribe.");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
const postD = { ...value, userId: user.id };
|
||||||
|
|
||||||
|
try {
|
||||||
|
const res = await axios.post("/lst/api/notification/sub", postD, {
|
||||||
|
withCredentials: true,
|
||||||
|
});
|
||||||
|
|
||||||
|
if (res.status === 200) {
|
||||||
|
toast.success("Notification Subbed");
|
||||||
|
refetch();
|
||||||
|
form.reset();
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error(error);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
let n: any = [];
|
||||||
|
if (data) {
|
||||||
|
n = data.map((i: any) => ({
|
||||||
|
label: i.name,
|
||||||
|
value: i.id,
|
||||||
|
}));
|
||||||
|
}
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div>
|
||||||
|
<Card className="p-3 w-lg">
|
||||||
|
<CardHeader>
|
||||||
|
<CardTitle>Notifications</CardTitle>
|
||||||
|
<CardDescription>
|
||||||
|
All currently active notifications you can subscribe to. selecting a
|
||||||
|
notification will give you a brief description on how it works
|
||||||
|
</CardDescription>
|
||||||
|
</CardHeader>
|
||||||
|
<CardContent>
|
||||||
|
<form
|
||||||
|
onSubmit={(e) => {
|
||||||
|
e.preventDefault();
|
||||||
|
form.handleSubmit();
|
||||||
|
}}
|
||||||
|
>
|
||||||
|
<div>
|
||||||
|
<form.AppField name="notificationId">
|
||||||
|
{(field) => (
|
||||||
|
<field.SelectField
|
||||||
|
label="Notifications"
|
||||||
|
placeholder="Select Notification"
|
||||||
|
options={n}
|
||||||
|
/>
|
||||||
|
)}
|
||||||
|
</form.AppField>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<form.AppField name="emails" mode="array">
|
||||||
|
{(field) => (
|
||||||
|
<field.DynamicInputField
|
||||||
|
label="Notification Emails"
|
||||||
|
description="Add more email addresses for notification delivery."
|
||||||
|
inputType="email"
|
||||||
|
addLabel="Add Email"
|
||||||
|
//initialValue={session?.user.email}
|
||||||
|
/>
|
||||||
|
)}
|
||||||
|
</form.AppField>
|
||||||
|
<div className="flex justify-end mt-6">
|
||||||
|
<form.AppForm>
|
||||||
|
<form.SubmitButton>Subscribe</form.SubmitButton>
|
||||||
|
</form.AppForm>
|
||||||
|
</div>
|
||||||
|
</form>
|
||||||
|
</CardContent>
|
||||||
|
</Card>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
114
frontend/src/routes/(auth)/-components/NotificationsTable.tsx
Normal file
114
frontend/src/routes/(auth)/-components/NotificationsTable.tsx
Normal file
@@ -0,0 +1,114 @@
|
|||||||
|
import { useSuspenseQuery } from "@tanstack/react-query";
|
||||||
|
import { createColumnHelper } from "@tanstack/react-table";
|
||||||
|
import axios from "axios";
|
||||||
|
import { Trash } from "lucide-react";
|
||||||
|
import { toast } from "sonner";
|
||||||
|
import type { Notifications } from "../../../../types/notifications";
|
||||||
|
import { Button } from "../../../components/ui/button";
|
||||||
|
import { Card, CardContent, CardHeader } from "../../../components/ui/card";
|
||||||
|
import {
|
||||||
|
Tooltip,
|
||||||
|
TooltipContent,
|
||||||
|
TooltipTrigger,
|
||||||
|
} from "../../../components/ui/tooltip";
|
||||||
|
import { notificationSubs } from "../../../lib/queries/notificationSubs";
|
||||||
|
import { notifications } from "../../../lib/queries/notifications";
|
||||||
|
import LstTable from "../../../lib/tableStuff/LstTable";
|
||||||
|
import SearchableHeader from "../../../lib/tableStuff/SearchableHeader";
|
||||||
|
|
||||||
|
export default function NotificationsTable({ userId }: any) {
|
||||||
|
const { data: subs, refetch } = useSuspenseQuery(notificationSubs(userId));
|
||||||
|
const { data: note } = useSuspenseQuery(notifications());
|
||||||
|
const columnHelper = createColumnHelper<Notifications>();
|
||||||
|
|
||||||
|
// filter out the current
|
||||||
|
const notificationMap = Object.fromEntries(note.map((n: any) => [n.id, n]));
|
||||||
|
|
||||||
|
const data = subs.map((sub: any) => ({
|
||||||
|
...sub,
|
||||||
|
name: notificationMap[sub.notificationId].name || null,
|
||||||
|
description: notificationMap[sub.notificationId].description || null,
|
||||||
|
emails: sub.emails ? sub.emails.join(",") : null,
|
||||||
|
}));
|
||||||
|
|
||||||
|
const removeNotification = async (ns: any) => {
|
||||||
|
try {
|
||||||
|
const res = await axios.delete(`/lst/api/notification/sub`, {
|
||||||
|
withCredentials: true,
|
||||||
|
data: {
|
||||||
|
userId: ns.userId,
|
||||||
|
notificationId: ns.notificationId,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
if (res.data.success) {
|
||||||
|
toast.success(`Subscription removed`);
|
||||||
|
refetch();
|
||||||
|
} else {
|
||||||
|
console.info(res);
|
||||||
|
toast.error(res.data.message);
|
||||||
|
}
|
||||||
|
} catch {
|
||||||
|
toast.error(`There was an error removing subscription.`);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
const column = [
|
||||||
|
columnHelper.accessor("name", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Name" searchable={true} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => i.getValue(),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("description", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Description" />
|
||||||
|
),
|
||||||
|
cell: (i) => (
|
||||||
|
<Tooltip>
|
||||||
|
<TooltipTrigger>
|
||||||
|
{i.getValue()?.length > 25 ? (
|
||||||
|
<span>{i.getValue().slice(0, 25)}...</span>
|
||||||
|
) : (
|
||||||
|
<span>{i.getValue()}</span>
|
||||||
|
)}
|
||||||
|
</TooltipTrigger>
|
||||||
|
<TooltipContent>{i.getValue()}</TooltipContent>
|
||||||
|
</Tooltip>
|
||||||
|
),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("emails", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Emails" searchable={true} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => i.getValue(),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("remove", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Remove" searchable={false} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => {
|
||||||
|
return (
|
||||||
|
<Button
|
||||||
|
size="icon"
|
||||||
|
variant={"destructive"}
|
||||||
|
onClick={() => removeNotification(i.row.original)}
|
||||||
|
>
|
||||||
|
<Trash />
|
||||||
|
</Button>
|
||||||
|
);
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
];
|
||||||
|
return (
|
||||||
|
<Card>
|
||||||
|
<CardHeader className="text-center">Subscriptions</CardHeader>
|
||||||
|
<CardContent>
|
||||||
|
<LstTable data={data} columns={column} />
|
||||||
|
</CardContent>
|
||||||
|
</Card>
|
||||||
|
);
|
||||||
|
}
|
||||||
@@ -1,4 +1,5 @@
|
|||||||
import { createFileRoute, redirect } from "@tanstack/react-router";
|
import { createFileRoute, redirect } from "@tanstack/react-router";
|
||||||
|
import { Suspense } from "react";
|
||||||
import { toast } from "sonner";
|
import { toast } from "sonner";
|
||||||
import {
|
import {
|
||||||
Card,
|
Card,
|
||||||
@@ -9,7 +10,10 @@ import {
|
|||||||
} from "@/components/ui/card";
|
} from "@/components/ui/card";
|
||||||
import { authClient, useSession } from "@/lib/auth-client";
|
import { authClient, useSession } from "@/lib/auth-client";
|
||||||
import { useAppForm } from "@/lib/formSutff";
|
import { useAppForm } from "@/lib/formSutff";
|
||||||
|
import { Spinner } from "../../components/ui/spinner";
|
||||||
import ChangePassword from "./-components/ChangePassword";
|
import ChangePassword from "./-components/ChangePassword";
|
||||||
|
import NotificationsSubCard from "./-components/NotificationsSubCard";
|
||||||
|
import NotificationsTable from "./-components/NotificationsTable";
|
||||||
|
|
||||||
export const Route = createFileRoute("/(auth)/user/profile")({
|
export const Route = createFileRoute("/(auth)/user/profile")({
|
||||||
beforeLoad: async () => {
|
beforeLoad: async () => {
|
||||||
@@ -54,44 +58,86 @@ function RouteComponent() {
|
|||||||
},
|
},
|
||||||
});
|
});
|
||||||
return (
|
return (
|
||||||
<div className="flex justify-center flex-col pt-4 gap-2 lg:flex-row">
|
<div className="flex justify-center flex-col pt-4 gap-2">
|
||||||
<div>
|
<div className="flex justify-center flex-col pt-4 gap-2 lg:flex-row">
|
||||||
<Card className="p-6 w-96">
|
<div>
|
||||||
<CardHeader>
|
<Card className="p-6 w-96">
|
||||||
<CardTitle>Profile</CardTitle>
|
<CardHeader>
|
||||||
<CardDescription>
|
<CardTitle>Profile</CardTitle>
|
||||||
Change your profile and password below
|
<CardDescription>
|
||||||
</CardDescription>
|
Change your profile and password below
|
||||||
</CardHeader>
|
</CardDescription>
|
||||||
|
</CardHeader>
|
||||||
|
|
||||||
<CardContent>
|
<CardContent>
|
||||||
<form
|
<form
|
||||||
onSubmit={(e) => {
|
onSubmit={(e) => {
|
||||||
e.preventDefault();
|
e.preventDefault();
|
||||||
form.handleSubmit();
|
form.handleSubmit();
|
||||||
}}
|
}}
|
||||||
>
|
>
|
||||||
<form.AppField name="name">
|
<form.AppField name="name">
|
||||||
{(field) => (
|
{(field) => (
|
||||||
<field.InputField
|
<field.InputField
|
||||||
label="Name"
|
label="Name"
|
||||||
inputType="string"
|
inputType="string"
|
||||||
required={true}
|
required={true}
|
||||||
/>
|
/>
|
||||||
)}
|
)}
|
||||||
</form.AppField>
|
</form.AppField>
|
||||||
|
|
||||||
<div className="flex justify-end mt-6">
|
<div className="flex justify-end mt-6">
|
||||||
<form.AppForm>
|
<form.AppForm>
|
||||||
<form.SubmitButton>Update Profile</form.SubmitButton>
|
<form.SubmitButton>Update Profile</form.SubmitButton>
|
||||||
</form.AppForm>
|
</form.AppForm>
|
||||||
</div>
|
</div>
|
||||||
</form>
|
</form>
|
||||||
</CardContent>
|
</CardContent>
|
||||||
</Card>
|
</Card>
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<ChangePassword />
|
||||||
|
</div>
|
||||||
|
<div>
|
||||||
|
<Suspense
|
||||||
|
fallback={
|
||||||
|
<Card className="p-3 w-lg">
|
||||||
|
<CardHeader>
|
||||||
|
<CardTitle>Notifications</CardTitle>
|
||||||
|
</CardHeader>
|
||||||
|
<CardContent>
|
||||||
|
<div className="flex justify-center m-auto">
|
||||||
|
<div>
|
||||||
|
<Spinner className="size-32" />
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</CardContent>
|
||||||
|
</Card>
|
||||||
|
}
|
||||||
|
>
|
||||||
|
{session && <NotificationsSubCard user={session.user} />}
|
||||||
|
</Suspense>
|
||||||
|
</div>
|
||||||
</div>
|
</div>
|
||||||
<div>
|
<div className="w-fill">
|
||||||
<ChangePassword />
|
<Suspense
|
||||||
|
fallback={
|
||||||
|
<Card className="p-3">
|
||||||
|
<CardHeader>
|
||||||
|
<CardTitle className="text-center">Subscriptions</CardTitle>
|
||||||
|
</CardHeader>
|
||||||
|
<CardContent>
|
||||||
|
<div className="flex justify-center m-auto">
|
||||||
|
<div>
|
||||||
|
<Spinner className="size-32" />
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
</CardContent>
|
||||||
|
</Card>
|
||||||
|
}
|
||||||
|
>
|
||||||
|
{session && <NotificationsTable userId={`${session.user.id}`} />}
|
||||||
|
</Suspense>
|
||||||
</div>
|
</div>
|
||||||
</div>
|
</div>
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ import { TanStackRouterDevtools } from "@tanstack/react-router-devtools";
|
|||||||
import { Toaster } from "sonner";
|
import { Toaster } from "sonner";
|
||||||
import Header from "@/components/Header";
|
import Header from "@/components/Header";
|
||||||
import { AppSidebar } from "@/components/Sidebar/sidebar";
|
import { AppSidebar } from "@/components/Sidebar/sidebar";
|
||||||
import { SidebarInset, SidebarProvider } from "@/components/ui/sidebar";
|
import { SidebarProvider } from "@/components/ui/sidebar";
|
||||||
import { ThemeProvider } from "@/lib/theme-provider";
|
import { ThemeProvider } from "@/lib/theme-provider";
|
||||||
|
|
||||||
const RootLayout = () => (
|
const RootLayout = () => (
|
||||||
|
|||||||
@@ -1,6 +1,5 @@
|
|||||||
import { useSuspenseQuery } from "@tanstack/react-query";
|
import { useSuspenseQuery } from "@tanstack/react-query";
|
||||||
import axios from "axios";
|
import axios from "axios";
|
||||||
import React from "react";
|
|
||||||
import { toast } from "sonner";
|
import { toast } from "sonner";
|
||||||
import { Card, CardDescription, CardHeader } from "../../../components/ui/card";
|
import { Card, CardDescription, CardHeader } from "../../../components/ui/card";
|
||||||
import { useAppForm } from "../../../lib/formSutff";
|
import { useAppForm } from "../../../lib/formSutff";
|
||||||
@@ -33,7 +32,9 @@ export default function FeatureCard({ item }: { item: Setting }) {
|
|||||||
const { data } = await axios.patch(`/lst/api/settings/${item.name}`, {
|
const { data } = await axios.patch(`/lst/api/settings/${item.name}`, {
|
||||||
value: value.value,
|
value: value.value,
|
||||||
active: value.active ? "true" : "false",
|
active: value.active ? "true" : "false",
|
||||||
});
|
}, {
|
||||||
|
withCredentials: true,
|
||||||
|
});
|
||||||
|
|
||||||
refetch();
|
refetch();
|
||||||
toast.success(
|
toast.success(
|
||||||
|
|||||||
316
frontend/src/routes/admin/notifications.tsx
Normal file
316
frontend/src/routes/admin/notifications.tsx
Normal file
@@ -0,0 +1,316 @@
|
|||||||
|
import { useMutation, useSuspenseQuery } from "@tanstack/react-query";
|
||||||
|
import { createFileRoute, redirect } from "@tanstack/react-router";
|
||||||
|
import { createColumnHelper } from "@tanstack/react-table";
|
||||||
|
import axios from "axios";
|
||||||
|
import { Trash } from "lucide-react";
|
||||||
|
import { Suspense, useState } from "react";
|
||||||
|
import { toast } from "sonner";
|
||||||
|
import type { Notifications } from "../../../types/notifications";
|
||||||
|
import { Button } from "../../components/ui/button";
|
||||||
|
import { Card, CardContent } from "../../components/ui/card";
|
||||||
|
import { Label } from "../../components/ui/label";
|
||||||
|
import { Switch } from "../../components/ui/switch";
|
||||||
|
import {
|
||||||
|
Tabs,
|
||||||
|
TabsContent,
|
||||||
|
TabsList,
|
||||||
|
TabsTrigger,
|
||||||
|
} from "../../components/ui/tabs";
|
||||||
|
import {
|
||||||
|
Tooltip,
|
||||||
|
TooltipContent,
|
||||||
|
TooltipTrigger,
|
||||||
|
} from "../../components/ui/tooltip";
|
||||||
|
import { authClient } from "../../lib/auth-client";
|
||||||
|
import { notificationSubs } from "../../lib/queries/notificationSubs";
|
||||||
|
import { notifications } from "../../lib/queries/notifications";
|
||||||
|
import EditableCellInput from "../../lib/tableStuff/EditableCellInput";
|
||||||
|
import LstTable from "../../lib/tableStuff/LstTable";
|
||||||
|
import SearchableHeader from "../../lib/tableStuff/SearchableHeader";
|
||||||
|
import SkellyTable from "../../lib/tableStuff/SkellyTable";
|
||||||
|
|
||||||
|
const updateNotifications = async (
|
||||||
|
id: string,
|
||||||
|
data: Record<string, string | number | boolean | null>,
|
||||||
|
) => {
|
||||||
|
//console.log(id, data);
|
||||||
|
try {
|
||||||
|
const res = await axios.patch(
|
||||||
|
`/lst/api/notification/${id}`,
|
||||||
|
{ interval: data.interval },
|
||||||
|
{
|
||||||
|
withCredentials: true,
|
||||||
|
},
|
||||||
|
);
|
||||||
|
toast.success(`Notification was just updated`);
|
||||||
|
return res;
|
||||||
|
} catch (err) {
|
||||||
|
toast.error("Error in updating the settings");
|
||||||
|
return err;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
export const Route = createFileRoute("/admin/notifications")({
|
||||||
|
beforeLoad: async ({ location }) => {
|
||||||
|
const { data: session } = await authClient.getSession();
|
||||||
|
const allowedRole = ["systemAdmin", "admin"];
|
||||||
|
|
||||||
|
if (!session?.user) {
|
||||||
|
throw redirect({
|
||||||
|
to: "/",
|
||||||
|
search: {
|
||||||
|
redirect: location.href,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!allowedRole.includes(session.user.role as string)) {
|
||||||
|
throw redirect({
|
||||||
|
to: "/",
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
return { user: session.user };
|
||||||
|
},
|
||||||
|
component: RouteComponent,
|
||||||
|
});
|
||||||
|
|
||||||
|
const NotificationTable = () => {
|
||||||
|
const { data, refetch } = useSuspenseQuery(notifications());
|
||||||
|
const { data: subs, refetch: subRefetch } = useSuspenseQuery(
|
||||||
|
notificationSubs(),
|
||||||
|
);
|
||||||
|
const columnHelper = createColumnHelper<Notifications>();
|
||||||
|
|
||||||
|
const notificationMap = Object.fromEntries(data.map((n: any) => [n.id, n]));
|
||||||
|
|
||||||
|
const subData = subs.map((sub: any) => ({
|
||||||
|
...sub,
|
||||||
|
name: notificationMap[sub.notificationId].name || null,
|
||||||
|
description: notificationMap[sub.notificationId].description || null,
|
||||||
|
emails: sub.emails ? sub.emails.join(",") : null,
|
||||||
|
}));
|
||||||
|
|
||||||
|
const updateNotification = useMutation({
|
||||||
|
mutationFn: ({
|
||||||
|
id,
|
||||||
|
field,
|
||||||
|
value,
|
||||||
|
}: {
|
||||||
|
id: string;
|
||||||
|
field: string;
|
||||||
|
value: string | number | boolean | null;
|
||||||
|
}) => updateNotifications(id, { [field]: value }),
|
||||||
|
|
||||||
|
onSuccess: () => {
|
||||||
|
// refetch or update cache
|
||||||
|
refetch();
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
const removeNotification = async (ns: any) => {
|
||||||
|
try {
|
||||||
|
const res = await axios.delete(`/lst/api/notification/sub`, {
|
||||||
|
withCredentials: true,
|
||||||
|
data: {
|
||||||
|
userId: ns.userId,
|
||||||
|
notificationId: ns.notificationId,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
if (res.data.success) {
|
||||||
|
toast.success(`Subscription removed`);
|
||||||
|
subRefetch();
|
||||||
|
} else {
|
||||||
|
console.info(res);
|
||||||
|
toast.error(res.data.message);
|
||||||
|
}
|
||||||
|
} catch {
|
||||||
|
toast.error(`There was an error removing subscription.`);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
const column = [
|
||||||
|
columnHelper.accessor("name", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Name" searchable={true} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => i.getValue(),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("description", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Description" />
|
||||||
|
),
|
||||||
|
cell: (i) => (
|
||||||
|
<Tooltip>
|
||||||
|
<TooltipTrigger>
|
||||||
|
{i.getValue()?.length > 25 ? (
|
||||||
|
<span>{i.getValue().slice(0, 25)}...</span>
|
||||||
|
) : (
|
||||||
|
<span>{i.getValue()}</span>
|
||||||
|
)}
|
||||||
|
</TooltipTrigger>
|
||||||
|
<TooltipContent>{i.getValue()}</TooltipContent>
|
||||||
|
</Tooltip>
|
||||||
|
),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("active", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Active" searchable={false} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => {
|
||||||
|
// biome-ignore lint: just removing the lint for now to get this going will maybe fix later
|
||||||
|
const [activeToggle, setActiveToggle] = useState(i.getValue());
|
||||||
|
|
||||||
|
const onToggle = async (e: boolean) => {
|
||||||
|
setActiveToggle(e);
|
||||||
|
|
||||||
|
try {
|
||||||
|
const res = await axios.patch(
|
||||||
|
`/lst/api/notification/${i.row.original.id}`,
|
||||||
|
{
|
||||||
|
active: !activeToggle,
|
||||||
|
},
|
||||||
|
{ withCredentials: true },
|
||||||
|
);
|
||||||
|
|
||||||
|
if (res.data.success) {
|
||||||
|
toast.success(
|
||||||
|
`${i.row.original.name} was set to ${activeToggle ? "Inactive" : "Active"}`,
|
||||||
|
);
|
||||||
|
refetch();
|
||||||
|
}
|
||||||
|
} catch (error) {
|
||||||
|
console.error(error);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div className="w-48">
|
||||||
|
<div className="flex items-center space-x-2">
|
||||||
|
<Switch
|
||||||
|
id={i.row.original.id}
|
||||||
|
checked={activeToggle}
|
||||||
|
onCheckedChange={(e) => onToggle(e)}
|
||||||
|
//onBlur={field.handleBlur}
|
||||||
|
/>
|
||||||
|
<Label htmlFor={i.row.original.id}>
|
||||||
|
{activeToggle ? "Active" : "Deactivated"}
|
||||||
|
</Label>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("interval", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Interval" />
|
||||||
|
),
|
||||||
|
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: ({ row, getValue }) => {
|
||||||
|
return (
|
||||||
|
<EditableCellInput
|
||||||
|
value={getValue()}
|
||||||
|
id={row.original.id}
|
||||||
|
field="interval"
|
||||||
|
onSubmit={({ id, field, value }) => {
|
||||||
|
updateNotification.mutate({ id, field, value });
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
);
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
];
|
||||||
|
|
||||||
|
const subsColumn = [
|
||||||
|
columnHelper.accessor("name", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Name" searchable={true} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => i.getValue(),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("description", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Description" />
|
||||||
|
),
|
||||||
|
cell: (i) => (
|
||||||
|
<Tooltip>
|
||||||
|
<TooltipTrigger>
|
||||||
|
{i.getValue()?.length > 25 ? (
|
||||||
|
<span>{i.getValue().slice(0, 25)}...</span>
|
||||||
|
) : (
|
||||||
|
<span>{i.getValue()}</span>
|
||||||
|
)}
|
||||||
|
</TooltipTrigger>
|
||||||
|
<TooltipContent>{i.getValue()}</TooltipContent>
|
||||||
|
</Tooltip>
|
||||||
|
),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("emails", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Emails" searchable={true} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => i.getValue(),
|
||||||
|
}),
|
||||||
|
columnHelper.accessor("remove", {
|
||||||
|
header: ({ column }) => (
|
||||||
|
<SearchableHeader column={column} title="Remove" searchable={false} />
|
||||||
|
),
|
||||||
|
filterFn: "includesString",
|
||||||
|
cell: (i) => {
|
||||||
|
return (
|
||||||
|
<Button
|
||||||
|
size="icon"
|
||||||
|
variant={"destructive"}
|
||||||
|
onClick={() => removeNotification(i.row.original)}
|
||||||
|
>
|
||||||
|
<Trash />
|
||||||
|
</Button>
|
||||||
|
);
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
];
|
||||||
|
|
||||||
|
return (
|
||||||
|
<>
|
||||||
|
<TabsContent value="notifications">
|
||||||
|
<LstTable data={data} columns={column} />
|
||||||
|
</TabsContent>
|
||||||
|
<TabsContent value="subscriptions">
|
||||||
|
<LstTable data={subData} columns={subsColumn} />
|
||||||
|
</TabsContent>
|
||||||
|
</>
|
||||||
|
);
|
||||||
|
};
|
||||||
|
|
||||||
|
function RouteComponent() {
|
||||||
|
return (
|
||||||
|
<div className="space-y-6">
|
||||||
|
<div className="space-y-2">
|
||||||
|
<h1 className="text-2xl font-semibold">Notifications</h1>
|
||||||
|
<p className="text-sm text-muted-foreground">
|
||||||
|
Manage all notification settings and user subs.
|
||||||
|
</p>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<Card>
|
||||||
|
<CardContent>
|
||||||
|
<Tabs defaultValue="notifications" className="w-full">
|
||||||
|
<TabsList>
|
||||||
|
<TabsTrigger value="notifications">Notifications</TabsTrigger>
|
||||||
|
<TabsTrigger value="subscriptions">Subscriptions</TabsTrigger>
|
||||||
|
</TabsList>
|
||||||
|
<Suspense fallback={<SkellyTable />}>
|
||||||
|
<NotificationTable />
|
||||||
|
</Suspense>
|
||||||
|
</Tabs>
|
||||||
|
</CardContent>
|
||||||
|
</Card>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
@@ -46,7 +46,7 @@ const updateSettings = async (
|
|||||||
id: string,
|
id: string,
|
||||||
data: Record<string, string | number | boolean | null>,
|
data: Record<string, string | number | boolean | null>,
|
||||||
) => {
|
) => {
|
||||||
console.log(id, data);
|
//console.log(id, data);
|
||||||
try {
|
try {
|
||||||
const res = await axios.patch(`/lst/api/settings/${id}`, data, {
|
const res = await axios.patch(`/lst/api/settings/${id}`, data, {
|
||||||
withCredentials: true,
|
withCredentials: true,
|
||||||
|
|||||||
@@ -34,7 +34,7 @@ function Index() {
|
|||||||
<p>
|
<p>
|
||||||
This is active in your plant today due to having warehousing activated
|
This is active in your plant today due to having warehousing activated
|
||||||
and new functions needed to be introduced, you should be still using LST
|
and new functions needed to be introduced, you should be still using LST
|
||||||
as you were before
|
as you were before.
|
||||||
</p>
|
</p>
|
||||||
<br></br>
|
<br></br>
|
||||||
<p>
|
<p>
|
||||||
|
|||||||
10
frontend/types/notifications.ts
Normal file
10
frontend/types/notifications.ts
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
export type Notifications = {
|
||||||
|
id: string;
|
||||||
|
name: string;
|
||||||
|
emails: string;
|
||||||
|
description: string;
|
||||||
|
remove?: unknown;
|
||||||
|
active?: boolean;
|
||||||
|
interval: number;
|
||||||
|
options: unknown[];
|
||||||
|
};
|
||||||
4
migrations/0019_large_thunderbird.sql
Normal file
4
migrations/0019_large_thunderbird.sql
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
CREATE TABLE "printer_log" (
|
||||||
|
"id" integer PRIMARY KEY GENERATED ALWAYS AS IDENTITY (sequence name "printer_log_id_seq" INCREMENT BY 1 MINVALUE 1 MAXVALUE 2147483647 START WITH 1 CACHE 1),
|
||||||
|
"name" text NOT NULL
|
||||||
|
);
|
||||||
15
migrations/0020_stale_ma_gnuci.sql
Normal file
15
migrations/0020_stale_ma_gnuci.sql
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
CREATE TABLE "alpla_purchase_history" (
|
||||||
|
"id" uuid PRIMARY KEY DEFAULT gen_random_uuid() NOT NULL,
|
||||||
|
"apo" integer,
|
||||||
|
"revision" integer,
|
||||||
|
"confirmed" integer,
|
||||||
|
"status" integer,
|
||||||
|
"status_text" integer,
|
||||||
|
"add_date" timestamp DEFAULT now(),
|
||||||
|
"upd_date" timestamp DEFAULT now(),
|
||||||
|
"add_user" text,
|
||||||
|
"upd_user" text,
|
||||||
|
"remark" text,
|
||||||
|
"position" jsonb DEFAULT '[]'::jsonb,
|
||||||
|
"created_at" timestamp DEFAULT now()
|
||||||
|
);
|
||||||
2
migrations/0021_slimy_master_mold.sql
Normal file
2
migrations/0021_slimy_master_mold.sql
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
ALTER TABLE "alpla_purchase_history" ADD COLUMN "journal_num" integer;--> statement-breakpoint
|
||||||
|
ALTER TABLE "alpla_purchase_history" ADD COLUMN "approved_status" text;
|
||||||
1
migrations/0022_large_sumo.sql
Normal file
1
migrations/0022_large_sumo.sql
Normal file
@@ -0,0 +1 @@
|
|||||||
|
ALTER TABLE "alpla_purchase_history" ALTER COLUMN "approved_status" SET DEFAULT 'pending';
|
||||||
1
migrations/0023_normal_hellion.sql
Normal file
1
migrations/0023_normal_hellion.sql
Normal file
@@ -0,0 +1 @@
|
|||||||
|
ALTER TABLE "alpla_purchase_history" ALTER COLUMN "status_text" SET DATA TYPE text;
|
||||||
6
migrations/0024_absent_barracuda.sql
Normal file
6
migrations/0024_absent_barracuda.sql
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
ALTER TABLE "opendock_apt" ALTER COLUMN "release" SET NOT NULL;--> statement-breakpoint
|
||||||
|
ALTER TABLE "opendock_apt" ALTER COLUMN "appointment" SET NOT NULL;--> statement-breakpoint
|
||||||
|
ALTER TABLE "opendock_apt" ALTER COLUMN "upd_date" SET NOT NULL;--> statement-breakpoint
|
||||||
|
ALTER TABLE "opendock_apt" ALTER COLUMN "created_at" SET NOT NULL;--> statement-breakpoint
|
||||||
|
CREATE INDEX "opendock_apt_release_idx" ON "opendock_apt" USING btree ("release");--> statement-breakpoint
|
||||||
|
CREATE INDEX "opendock_apt_opendock_id_idx" ON "opendock_apt" USING btree ("open_dock_apt_id");
|
||||||
1327
migrations/meta/0019_snapshot.json
Normal file
1327
migrations/meta/0019_snapshot.json
Normal file
File diff suppressed because it is too large
Load Diff
1423
migrations/meta/0020_snapshot.json
Normal file
1423
migrations/meta/0020_snapshot.json
Normal file
File diff suppressed because it is too large
Load Diff
1435
migrations/meta/0021_snapshot.json
Normal file
1435
migrations/meta/0021_snapshot.json
Normal file
File diff suppressed because it is too large
Load Diff
1436
migrations/meta/0022_snapshot.json
Normal file
1436
migrations/meta/0022_snapshot.json
Normal file
File diff suppressed because it is too large
Load Diff
1436
migrations/meta/0023_snapshot.json
Normal file
1436
migrations/meta/0023_snapshot.json
Normal file
File diff suppressed because it is too large
Load Diff
1467
migrations/meta/0024_snapshot.json
Normal file
1467
migrations/meta/0024_snapshot.json
Normal file
File diff suppressed because it is too large
Load Diff
@@ -134,6 +134,48 @@
|
|||||||
"when": 1774032587305,
|
"when": 1774032587305,
|
||||||
"tag": "0018_lowly_wallow",
|
"tag": "0018_lowly_wallow",
|
||||||
"breakpoints": true
|
"breakpoints": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"idx": 19,
|
||||||
|
"version": "7",
|
||||||
|
"when": 1775159956510,
|
||||||
|
"tag": "0019_large_thunderbird",
|
||||||
|
"breakpoints": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"idx": 20,
|
||||||
|
"version": "7",
|
||||||
|
"when": 1775566910220,
|
||||||
|
"tag": "0020_stale_ma_gnuci",
|
||||||
|
"breakpoints": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"idx": 21,
|
||||||
|
"version": "7",
|
||||||
|
"when": 1775647109925,
|
||||||
|
"tag": "0021_slimy_master_mold",
|
||||||
|
"breakpoints": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"idx": 22,
|
||||||
|
"version": "7",
|
||||||
|
"when": 1775649219780,
|
||||||
|
"tag": "0022_large_sumo",
|
||||||
|
"breakpoints": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"idx": 23,
|
||||||
|
"version": "7",
|
||||||
|
"when": 1775650901523,
|
||||||
|
"tag": "0023_normal_hellion",
|
||||||
|
"breakpoints": true
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"idx": 24,
|
||||||
|
"version": "7",
|
||||||
|
"when": 1775661516749,
|
||||||
|
"tag": "0024_absent_barracuda",
|
||||||
|
"breakpoints": true
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
3122
package-lock.json
generated
3122
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
21
package.json
21
package.json
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "lst_v3",
|
"name": "lst_v3",
|
||||||
"version": "1.0.1",
|
"version": "0.0.1-alpha.2",
|
||||||
"description": "The tool that supports us in our everyday alplaprod",
|
"description": "The tool that supports us in our everyday alplaprod",
|
||||||
"main": "index.js",
|
"main": "index.js",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
@@ -13,23 +13,17 @@
|
|||||||
"build": "rimraf dist && npm run dev:db:generate && npm run dev:db:migrate && npm run build:app && npm run build:copySql && cd frontend && npm run build",
|
"build": "rimraf dist && npm run dev:db:generate && npm run dev:db:migrate && npm run build:app && npm run build:copySql && cd frontend && npm run build",
|
||||||
"build:app": "tsc",
|
"build:app": "tsc",
|
||||||
"agent": "powershell -ExecutionPolicy Bypass -File scripts/agentController.ps1",
|
"agent": "powershell -ExecutionPolicy Bypass -File scripts/agentController.ps1",
|
||||||
"build:docker": "docker compose up --force-recreate --build -d",
|
"build:docker": "rimraf dist && npm run build:app && npm run build:copySql",
|
||||||
"build:copySql": "xcopy backend\\prodSql\\queries dist\\prodSql\\queries\\ /E /I /Y ",
|
"build:copySql": "cpy \"backend/prodSql/queries/**/*\" dist/prodSql/queries --parents",
|
||||||
"lint": "tsc && biome lint",
|
"lint": "tsc && biome lint",
|
||||||
"start": "npm run start:server",
|
"start": "npm run start:server",
|
||||||
"start:server": "dotenvx run -f .env -- node dist/server.js",
|
"start:server": "dotenvx run -f .env -- node dist/server.js",
|
||||||
"start:docker": "node dist/server.js",
|
"start:docker": "node dist/server.js",
|
||||||
"commit": "cz",
|
|
||||||
"changeset": "changeset",
|
|
||||||
"version": "changeset version",
|
"version": "changeset version",
|
||||||
"release": "dotenvx run -f .env -- npm run version && git push --follow-tags && node scripts/create-release.js",
|
"specCheck": "node scripts/check-route-specs.mjs",
|
||||||
"specCheck": "node scripts/check-route-specs.mjs"
|
"commit": "cz",
|
||||||
|
"release": "commit-and-tag-version"
|
||||||
},
|
},
|
||||||
"workspaces": [
|
|
||||||
"backend",
|
|
||||||
"agent",
|
|
||||||
"shared"
|
|
||||||
],
|
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://git.tuffraid.net/cowch/lst_v3.git"
|
"url": "https://git.tuffraid.net/cowch/lst_v3.git"
|
||||||
@@ -40,7 +34,6 @@
|
|||||||
"type": "module",
|
"type": "module",
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@biomejs/biome": "2.4.8",
|
"@biomejs/biome": "2.4.8",
|
||||||
"@changesets/cli": "^2.30.0",
|
|
||||||
"@commitlint/cli": "^20.5.0",
|
"@commitlint/cli": "^20.5.0",
|
||||||
"@commitlint/config-conventional": "^20.5.0",
|
"@commitlint/config-conventional": "^20.5.0",
|
||||||
"@types/cors": "^2.8.19",
|
"@types/cors": "^2.8.19",
|
||||||
@@ -56,7 +49,9 @@
|
|||||||
"@types/supertest": "^7.2.0",
|
"@types/supertest": "^7.2.0",
|
||||||
"@types/swagger-jsdoc": "^6.0.4",
|
"@types/swagger-jsdoc": "^6.0.4",
|
||||||
"@types/swagger-ui-express": "^4.1.8",
|
"@types/swagger-ui-express": "^4.1.8",
|
||||||
|
"commit-and-tag-version": "^12.7.1",
|
||||||
"commitizen": "^4.3.1",
|
"commitizen": "^4.3.1",
|
||||||
|
"cpy-cli": "^7.0.0",
|
||||||
"cz-conventional-changelog": "^3.3.0",
|
"cz-conventional-changelog": "^3.3.0",
|
||||||
"npm-check-updates": "^19.6.5",
|
"npm-check-updates": "^19.6.5",
|
||||||
"openapi-types": "^12.1.3",
|
"openapi-types": "^12.1.3",
|
||||||
|
|||||||
@@ -17,6 +17,11 @@ $Servers = @(
|
|||||||
token = "uslim1"
|
token = "uslim1"
|
||||||
loc = "D$\LST_V3"
|
loc = "D$\LST_V3"
|
||||||
},
|
},
|
||||||
|
[PSCustomObject]@{
|
||||||
|
server = "ushou1vms006"
|
||||||
|
token = "ushou1"
|
||||||
|
loc = "D$\LST_V3"
|
||||||
|
},
|
||||||
[PSCustomObject]@{
|
[PSCustomObject]@{
|
||||||
server = "usday1vms006"
|
server = "usday1vms006"
|
||||||
token = "usday1"
|
token = "usday1"
|
||||||
@@ -32,12 +37,15 @@ $Servers = @(
|
|||||||
token = "test2"
|
token = "test2"
|
||||||
loc = "E$\LST_V3"
|
loc = "E$\LST_V3"
|
||||||
}
|
}
|
||||||
#@{ server = "usmcd1vms036"; token = "test1"; loc = "E$\LST\lst_backend"; }
|
,
|
||||||
#@{ server = "usiow1vms036"; token = "test1"; loc = "E$\LST\lst_backend"; }
|
[PSCustomObject]@{
|
||||||
|
server = "usweb1vms006"
|
||||||
|
token = "usweb1"
|
||||||
|
loc = "D$\LST_V3"
|
||||||
|
}
|
||||||
#@{ server = "usbet1vms006"; token = "usbet1";loc = "C$\Users\adm_matthes01\Desktop\lst_backend"; }
|
#@{ server = "usbet1vms006"; token = "usbet1";loc = "C$\Users\adm_matthes01\Desktop\lst_backend"; }
|
||||||
#@{ server = "usbow1vms006"; token = "usbow1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
#@{ server = "usbow1vms006"; token = "usbow1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
||||||
#@{ server = "usbow2vms006"; token = "usbow2"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
#@{ server = "usbow2vms006"; token = "usbow2"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
||||||
#@{ server = "usday1vms006"; token = "usday1"; loc = "E$\LST\lst_backend" ; }
|
|
||||||
#@{ server = "usflo1vms006"; token = "usflo1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
#@{ server = "usflo1vms006"; token = "usflo1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
||||||
#@{ server = "ushou1vms006"; token = "ushou1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
#@{ server = "ushou1vms006"; token = "ushou1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
||||||
#@{ server = "usiow1vms006"; token = "usiow1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
#@{ server = "usiow1vms006"; token = "usiow1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ; }
|
||||||
@@ -48,7 +56,6 @@ $Servers = @(
|
|||||||
#@{ server = "usshe1vms006"; token = "usshe1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
#@{ server = "usshe1vms006"; token = "usshe1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
||||||
#@{ server = "usslc1vms006"; token = "usslc1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
#@{ server = "usslc1vms006"; token = "usslc1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
||||||
#@{ server = "usstp1vms006"; token = "usstp1"; loc = "E$\LST\lst_backend" ; }
|
#@{ server = "usstp1vms006"; token = "usstp1"; loc = "E$\LST\lst_backend" ; }
|
||||||
#@{ server = "usweb1vms006"; token = "usweb1"; loc = "C$\Users\adm_matthes01\Desktop\lst_backend" ;}
|
|
||||||
#@{ server = "usmar1vms006"; token = "test1"; loc = "E$\LST\lst_backend"; }
|
#@{ server = "usmar1vms006"; token = "test1"; loc = "E$\LST\lst_backend"; }
|
||||||
|
|
||||||
)
|
)
|
||||||
@@ -334,7 +341,7 @@ function Update-Server {
|
|||||||
# do the install/update
|
# do the install/update
|
||||||
Push-Location $LocalPath
|
Push-Location $LocalPath
|
||||||
Write-Host "Running install/update in: $LocalPath"
|
Write-Host "Running install/update in: $LocalPath"
|
||||||
npm install
|
npm install --omit=dev
|
||||||
Start-Sleep -Seconds 3
|
Start-Sleep -Seconds 3
|
||||||
Write-Host "Install/update completed."
|
Write-Host "Install/update completed."
|
||||||
# do the migrations
|
# do the migrations
|
||||||
|
|||||||
132
scripts/dockerscripts.md
Normal file
132
scripts/dockerscripts.md
Normal file
@@ -0,0 +1,132 @@
|
|||||||
|
|
||||||
|
|
||||||
|
docker build -t git.tuffraid.net/cowch/lst_v3:latest .
|
||||||
|
docker push git.tuffraid.net/cowch/lst_v3:latest
|
||||||
|
|
||||||
|
docker compose pull && docker compose up -d --force-recreate
|
||||||
|
|
||||||
|
How to choose the bump
|
||||||
|
|
||||||
|
Use this rule:
|
||||||
|
|
||||||
|
patch = bug fix, small safe improvement
|
||||||
|
minor = new feature, backward compatible
|
||||||
|
major = breaking change
|
||||||
|
|
||||||
|
Changesets uses semver bump ty
|
||||||
|
|
||||||
|
|
||||||
|
### daily process
|
||||||
|
npm commit
|
||||||
|
|
||||||
|
- when closing a issue at the end add
|
||||||
|
Use one of these in the commit body or PR description:
|
||||||
|
|
||||||
|
- - Closes #123
|
||||||
|
- - Fixes #123
|
||||||
|
- - Resolves #123
|
||||||
|
|
||||||
|
Common ones:
|
||||||
|
|
||||||
|
- - Closes #123
|
||||||
|
- - Fixes #123
|
||||||
|
- - Resolves #123
|
||||||
|
Reference an issue without closing it
|
||||||
|
|
||||||
|
Use:
|
||||||
|
|
||||||
|
- - Refs #123
|
||||||
|
- - Related to #123
|
||||||
|
- - See #123
|
||||||
|
|
||||||
|
Good safe one:
|
||||||
|
|
||||||
|
- - Refs #123
|
||||||
|
Good example commit
|
||||||
|
|
||||||
|
Subject:
|
||||||
|
|
||||||
|
- - fix(cors): normalize external url origin
|
||||||
|
|
||||||
|
Body:
|
||||||
|
|
||||||
|
- - Refs #42
|
||||||
|
|
||||||
|
Or if this should close it:
|
||||||
|
|
||||||
|
- - Closes #42
|
||||||
|
|
||||||
|
# Release flow
|
||||||
|
npm run changeset:add
|
||||||
|
|
||||||
|
Pick one:
|
||||||
|
|
||||||
|
- patch = bug fix
|
||||||
|
- minor = new feature, non-breaking
|
||||||
|
- major = breaking change
|
||||||
|
|
||||||
|
Edit the generated .md file in .changeset it will be randomly named and add anything else in here from all the commits that are new to this release
|
||||||
|
|
||||||
|
Recommended release command
|
||||||
|
npm run changeset:version
|
||||||
|
|
||||||
|
stage the change log file
|
||||||
|
|
||||||
|
git commit -m "chore(release): version packages"
|
||||||
|
|
||||||
|
git tag v0.0.1-alpha.0 change this to the same version thats in the pkg.json
|
||||||
|
|
||||||
|
then push it
|
||||||
|
|
||||||
|
git push
|
||||||
|
git push --tags
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
### release type
|
||||||
|
|
||||||
|
when we want to go from alpha to normal well do
|
||||||
|
npx changeset pre enter alpha
|
||||||
|
npx changeset pre enter rc
|
||||||
|
|
||||||
|
go to full production
|
||||||
|
npx changeset pre exit
|
||||||
|
npx changeset version
|
||||||
|
|
||||||
|
### Steps will make it cleaner later
|
||||||
|
Daily work
|
||||||
|
1. Stage files
|
||||||
|
2. npm run commit
|
||||||
|
3. Add issue keyword if needed
|
||||||
|
4. git push when ready
|
||||||
|
|
||||||
|
Release flow
|
||||||
|
1. npx changeset
|
||||||
|
2. pick patch/minor/major
|
||||||
|
3. edit the generated md file with better notes
|
||||||
|
4. npx changeset version
|
||||||
|
5. git add .
|
||||||
|
6. git commit -m "chore(release): version packages"
|
||||||
|
7. git tag vX.X.X
|
||||||
|
8. git push
|
||||||
|
9. git push --tags
|
||||||
|
|
||||||
|
|
||||||
|
# normal work
|
||||||
|
stage files
|
||||||
|
npm run commit
|
||||||
|
|
||||||
|
# if releasing
|
||||||
|
npm run commit
|
||||||
|
npm run release -- --prerelease alpha
|
||||||
|
git push
|
||||||
|
git push --tags
|
||||||
|
|
||||||
|
|
||||||
|
git add .
|
||||||
|
git commit -m "chore(release): version packages"
|
||||||
|
git tag v0.0.1-alpha.0
|
||||||
|
git push
|
||||||
|
git push --tags
|
||||||
|
|
||||||
Reference in New Issue
Block a user