Compare commits
34 Commits
v0.0.1-alp
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
| 3734d9daac | |||
| a1eeadeec4 | |||
| 3639c1b77c | |||
| cfbc156517 | |||
| fb3cd85b41 | |||
| 5b1c88546f | |||
| ba3227545d | |||
| 84909bfcf8 | |||
| e0d0ac2077 | |||
| 52a6c821f4 | |||
| eccaf17332 | |||
| 6307037985 | |||
| 4b6061c478 | |||
| fc6dc82d84 | |||
| 6ba905a887 | |||
| f33587a3d9 | |||
| 80189baf90 | |||
| 87f738702a | |||
| 38a0b65e94 | |||
| 9a0ef8e51a | |||
| dcb3f2dd13 | |||
| e47ea9ec52 | |||
| ca3425d327 | |||
| 3bf024cfc9 | |||
| 9d39c13510 | |||
| c9eb59e2ad | |||
| b0e5fd7999 | |||
| 07ebf88806 | |||
| 79e653efa3 | |||
| d05a0ce930 | |||
| 995b1dda7c | |||
| 97f93a1830 | |||
| 635635b356 | |||
| a691dc276e |
58
.env-example
58
.env-example
@@ -1,32 +1,52 @@
|
|||||||
|
NODE_ENV=development
|
||||||
# Server
|
# Server
|
||||||
PORT=3000
|
PORT=3000
|
||||||
URL=http://localhost:3000
|
URL=http://localhost:3000
|
||||||
|
SERVER_IP=10.75.2.38
|
||||||
|
TIMEZONE=America/New_York
|
||||||
|
TCP_PORT=2222
|
||||||
|
|
||||||
# authentication
|
# Better auth Secret
|
||||||
BETTER_AUTH_SECRET=""
|
BETTER_AUTH_SECRET=
|
||||||
RESET_EXPIRY_SECONDS=3600
|
RESET_EXPIRY_SECONDS=3600
|
||||||
|
|
||||||
# logging
|
# logging
|
||||||
LOG_LEVEL=debug
|
LOG_LEVEL=
|
||||||
|
|
||||||
# prodServer
|
# SMTP password
|
||||||
PROD_SERVER=usmcd1vms036
|
SMTP_PASSWORD=
|
||||||
PROD_PLANT_TOKEN=test3
|
|
||||||
PROD_USER=alplaprod
|
# opendock
|
||||||
PROD_PASSWORD=password
|
OPENDOCK_URL=https://neutron.opendock.com
|
||||||
|
OPENDOCK_PASSWORD=
|
||||||
|
DEFAULT_DOCK=
|
||||||
|
DEFAULT_LOAD_TYPE=
|
||||||
|
DEFAULT_CARRIER=
|
||||||
|
|
||||||
|
# prodServer when ruining on an actual prod server use localhost this way we don't go out and back in.
|
||||||
|
PROD_SERVER=
|
||||||
|
PROD_PLANT_TOKEN=
|
||||||
|
PROD_USER=
|
||||||
|
PROD_PASSWORD=
|
||||||
|
|
||||||
|
# Tech user for alplaprod api
|
||||||
|
TEC_API_KEY=
|
||||||
|
|
||||||
|
# AD STUFF
|
||||||
|
# this is mainly used for purchase stuff to reference reqs
|
||||||
|
LDAP_URL=
|
||||||
|
|
||||||
# postgres connection
|
# postgres connection
|
||||||
DATABASE_HOST=localhost
|
DATABASE_HOST=localhost
|
||||||
DATABASE_PORT=5433
|
DATABASE_PORT=5432
|
||||||
DATABASE_USER=user
|
DATABASE_USER=
|
||||||
DATABASE_PASSWORD=password
|
DATABASE_PASSWORD=
|
||||||
DATABASE_DB=lst_dev
|
DATABASE_DB=
|
||||||
|
|
||||||
# how is the app running server or client when in client mode you must provide the server
|
# Gp connection
|
||||||
APP_RUNNING_IN=server
|
GP_USER=
|
||||||
SERVER_NAME=localhost
|
GP_PASSWORD=
|
||||||
|
|
||||||
#dev stuff
|
# how often to check for new/updated queries in min
|
||||||
GITEA_TOKEN=""
|
QUERY_TIME_TYPE=m #valid options are m, h
|
||||||
EMAIL_USER=""
|
QUERY_CHECK=1
|
||||||
EMAIL_PASSWORD=""
|
|
||||||
|
|||||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -5,6 +5,7 @@ builds
|
|||||||
.buildNumber
|
.buildNumber
|
||||||
temp
|
temp
|
||||||
brunoApi
|
brunoApi
|
||||||
|
downloads
|
||||||
.scriptCreds
|
.scriptCreds
|
||||||
node-v24.14.0-x64.msi
|
node-v24.14.0-x64.msi
|
||||||
postgresql-17.9-2-windows-x64.exe
|
postgresql-17.9-2-windows-x64.exe
|
||||||
|
|||||||
4
.vscode/settings.json
vendored
4
.vscode/settings.json
vendored
@@ -65,12 +65,14 @@
|
|||||||
"onnotice",
|
"onnotice",
|
||||||
"opendock",
|
"opendock",
|
||||||
"opendocks",
|
"opendocks",
|
||||||
|
"palletizer",
|
||||||
"ppoo",
|
"ppoo",
|
||||||
"preseed",
|
"preseed",
|
||||||
"prodlabels",
|
"prodlabels",
|
||||||
"prolink",
|
"prolink",
|
||||||
"Skelly",
|
"Skelly",
|
||||||
"trycatch"
|
"trycatch",
|
||||||
|
"whse"
|
||||||
],
|
],
|
||||||
"gitea.token": "8456def90e1c651a761a8711763d6ef225d6b2db",
|
"gitea.token": "8456def90e1c651a761a8711763d6ef225d6b2db",
|
||||||
"gitea.instanceURL": "https://git.tuffraid.net",
|
"gitea.instanceURL": "https://git.tuffraid.net",
|
||||||
|
|||||||
58
CHANGELOG.md
58
CHANGELOG.md
@@ -1,5 +1,63 @@
|
|||||||
# All Changes to LST can be found below.
|
# All Changes to LST can be found below.
|
||||||
|
|
||||||
|
## [0.0.1-alpha.4](https://git.tuffraid.net/cowch/lst_v3/compare/v0.0.1-alpha.3...v0.0.1-alpha.4) (2026-04-15)
|
||||||
|
|
||||||
|
|
||||||
|
### 🌟 Enhancements
|
||||||
|
|
||||||
|
* **datamart:** migrations completed remaining is the deactivation that will be ran by anylitics ([eccaf17](https://git.tuffraid.net/cowch/lst_v3/commits/eccaf17332fb1c63b8d6bbea6f668c3bb42d44b7))
|
||||||
|
* **datamart:** psi data has been added :D ([e0d0ac2](https://git.tuffraid.net/cowch/lst_v3/commits/e0d0ac20773159373495d65023587b76b47df34f))
|
||||||
|
* **migrate:** quality alert migrated ([b0e5fd7](https://git.tuffraid.net/cowch/lst_v3/commits/b0e5fd79998d551d4f155d58416157a324498fbd))
|
||||||
|
* **ocp:** printer sync and logging logic added ([80189ba](https://git.tuffraid.net/cowch/lst_v3/commits/80189baf906224da43ec1b9b7521153d2a49e059))
|
||||||
|
* **tcp crud:** tcp server start, stop, restart endpoints + status check ([6307037](https://git.tuffraid.net/cowch/lst_v3/commits/6307037985162bc6b49f9f711132853296f43eee))
|
||||||
|
|
||||||
|
|
||||||
|
### 🐛 Bug fixes
|
||||||
|
|
||||||
|
* **datamart:** error when running build and crashed everything ([52a6c82](https://git.tuffraid.net/cowch/lst_v3/commits/52a6c821f4632e4b5b51e0528a0d620e2e0deffc))
|
||||||
|
|
||||||
|
|
||||||
|
### 📚 Documentation
|
||||||
|
|
||||||
|
* **docs:** removed docusorus as all docs will be inside lst now to better assist users ([6ba905a](https://git.tuffraid.net/cowch/lst_v3/commits/6ba905a887dbd8f306d71fed75bb34c71fee74c9))
|
||||||
|
* **env example:** updated the file ([ca3425d](https://git.tuffraid.net/cowch/lst_v3/commits/ca3425d327757120c2cc876fff28e8668c76838d))
|
||||||
|
* **notifcations:** docs for intro, notifcations, reprint added ([87f7387](https://git.tuffraid.net/cowch/lst_v3/commits/87f738702a935279a248d471541cdd9d49330565))
|
||||||
|
|
||||||
|
|
||||||
|
### 🛠️ Code Refactor
|
||||||
|
|
||||||
|
* **agent:** changed to have the test servers on there own push for better testing ([3bf024c](https://git.tuffraid.net/cowch/lst_v3/commits/3bf024cfc97d2841130d54d1a7c5cb5f09f0f598))
|
||||||
|
* **connection:** corrected the connection to the old system ([38a0b65](https://git.tuffraid.net/cowch/lst_v3/commits/38a0b65e9450c65b8300a10058a8f0357400f4e6))
|
||||||
|
* **logging:** when notify is true send the error to systemAdmins ([79e653e](https://git.tuffraid.net/cowch/lst_v3/commits/79e653efa3bcb2941ccee06b28378e709e085ec0))
|
||||||
|
* **notification:** blocking added ([9a0ef8e](https://git.tuffraid.net/cowch/lst_v3/commits/9a0ef8e51a36e3ab45b601b977f1b5cf35d56947))
|
||||||
|
* **puchase:** changes how the error handling works so a better email can be sent ([9d39c13](https://git.tuffraid.net/cowch/lst_v3/commits/9d39c13510974b5ada2a6f6c2448da3f1b755a5c))
|
||||||
|
* **reprint:** new query added to deactivate the old notifcation so no chance of duplicates ([c9eb59e](https://git.tuffraid.net/cowch/lst_v3/commits/c9eb59e2ad9847418ac55cb8a4a91c013f6c97bb))
|
||||||
|
* **server:** added in serverCrash email ([dcb3f2d](https://git.tuffraid.net/cowch/lst_v3/commits/dcb3f2dd1382986639b722778fad113392533b28))
|
||||||
|
* **services:** added in examples for migration stuff ([fc6dc82](https://git.tuffraid.net/cowch/lst_v3/commits/fc6dc82d8458a9928050dd3770778d6a6e1eea7f))
|
||||||
|
* **sql:** corrections to the way we reconnect so the app can error out and be reactivated later ([f33587a](https://git.tuffraid.net/cowch/lst_v3/commits/f33587a3d9a72ca72806635fac9d1214bb1452f1))
|
||||||
|
* **templates:** corrections for new notify process on critcal errors ([07ebf88](https://git.tuffraid.net/cowch/lst_v3/commits/07ebf88806b93b9320f8f9d36b867572dd9a9580))
|
||||||
|
|
||||||
|
|
||||||
|
### 📈 Project changes
|
||||||
|
|
||||||
|
* **agent:** added in jeff city ([e47ea9e](https://git.tuffraid.net/cowch/lst_v3/commits/e47ea9ec52a6ebaf5a8f67a7e8bd2c73da6186fb))
|
||||||
|
* **agent:** added in sherman ([4b6061c](https://git.tuffraid.net/cowch/lst_v3/commits/4b6061c478cbeba7c845dc1c8a015b9998721456))
|
||||||
|
* **service:** changes to the script to allow running the powershell on execution palicy restrictions ([84909bf](https://git.tuffraid.net/cowch/lst_v3/commits/84909bfcf85b91d085ea9dca78be00482b7fd231))
|
||||||
|
|
||||||
|
## [0.0.1-alpha.3](https://git.tuffraid.net/cowch/lst_v3/compare/v0.0.1-alpha.2...v0.0.1-alpha.3) (2026-04-10)
|
||||||
|
|
||||||
|
|
||||||
|
### 🌟 Enhancements
|
||||||
|
|
||||||
|
* **puchase hist:** finished up purhcase historical / gp updates ([a691dc2](https://git.tuffraid.net/cowch/lst_v3/commits/a691dc276e8650c669409241f73d7b2d7a1f9176))
|
||||||
|
|
||||||
|
|
||||||
|
### 🛠️ Code Refactor
|
||||||
|
|
||||||
|
* **gp connect:** gp connect as was added to long live services ([635635b](https://git.tuffraid.net/cowch/lst_v3/commits/635635b356e1262e1c0b063408fe2209e6a8d4ec))
|
||||||
|
* **reprints:** changes the module and submodule around to be more accurate ([97f93a1](https://git.tuffraid.net/cowch/lst_v3/commits/97f93a1830761437118863372108df810ce9977a))
|
||||||
|
* **send email:** changes the error message to show the true message in the error ([995b1dd](https://git.tuffraid.net/cowch/lst_v3/commits/995b1dda7cdfebf4367d301ccac38fd339fab6dd))
|
||||||
|
|
||||||
## [0.0.1-alpha.2](https://git.tuffraid.net/cowch/lst_v3/compare/v0.0.1-alpha.1...v0.0.1-alpha.2) (2026-04-08)
|
## [0.0.1-alpha.2](https://git.tuffraid.net/cowch/lst_v3/compare/v0.0.1-alpha.1...v0.0.1-alpha.2) (2026-04-08)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ Quick summary of current rewrite/migration goal.
|
|||||||
| User Profile | ~~Edit profile~~, upload avatar | 🟨 In Progress |
|
| User Profile | ~~Edit profile~~, upload avatar | 🟨 In Progress |
|
||||||
| User Admin | Edit user, create user, remove user, alplaprod user integration | ⏳ Not Started |
|
| User Admin | Edit user, create user, remove user, alplaprod user integration | ⏳ Not Started |
|
||||||
| Notifications | ~~Subscribe~~, ~~Create~~, ~~Update~~, ~~~~Remove~~, Manual Trigger | 🟨 In Progress |
|
| Notifications | ~~Subscribe~~, ~~Create~~, ~~Update~~, ~~~~Remove~~, Manual Trigger | 🟨 In Progress |
|
||||||
| Datamart | Create, Update, Run, Deactivate | 🔧 In Progress |
|
| Datamart | ~~Create~~, ~~Update~~, ~~Run~~, Deactivate | 🟨 In Progress |
|
||||||
| Frontend | Analytics and charts | ⏳ Not Started |
|
| Frontend | Analytics and charts | ⏳ Not Started |
|
||||||
| Docs | Instructions and trouble shooting | ⏳ Not Started |
|
| Docs | Instructions and trouble shooting | ⏳ Not Started |
|
||||||
| One Click Print | Get printers, monitor printers, label process, material process, Special processes | ⏳ Not Started |
|
| One Click Print | Get printers, monitor printers, label process, material process, Special processes | ⏳ Not Started |
|
||||||
|
|||||||
23
backend/configs/gpSql.config.ts
Normal file
23
backend/configs/gpSql.config.ts
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
import type sql from "mssql";
|
||||||
|
|
||||||
|
const username = "gpviewer";
|
||||||
|
const password = "gp$$ViewOnly!";
|
||||||
|
|
||||||
|
export const gpSqlConfig: sql.config = {
|
||||||
|
server: `USMCD1VMS011`,
|
||||||
|
database: `ALPLA`,
|
||||||
|
user: username,
|
||||||
|
password: password,
|
||||||
|
options: {
|
||||||
|
encrypt: true,
|
||||||
|
trustServerCertificate: true,
|
||||||
|
},
|
||||||
|
requestTimeout: 90000, // how long until we kill the query and fail it
|
||||||
|
pool: {
|
||||||
|
max: 20, // Maximum number of connections in the pool
|
||||||
|
min: 0, // Minimum number of connections in the pool
|
||||||
|
idleTimeoutMillis: 10000, // How long a connection is allowed to be idle before being released
|
||||||
|
reapIntervalMillis: 1000, // how often to check for idle resources to destroy
|
||||||
|
acquireTimeoutMillis: 100000, // How long until a complete timeout happens
|
||||||
|
},
|
||||||
|
};
|
||||||
@@ -13,6 +13,10 @@
|
|||||||
*
|
*
|
||||||
* when a criteria is password over we will handle it by counting how many were passed up to 3 then deal with each one respectively
|
* when a criteria is password over we will handle it by counting how many were passed up to 3 then deal with each one respectively
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
import { and, between, inArray, notInArray } from "drizzle-orm";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { invHistoricalData } from "../db/schema/historicalInv.schema.js";
|
||||||
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
import {
|
import {
|
||||||
type SqlQuery,
|
type SqlQuery,
|
||||||
@@ -22,37 +26,118 @@ import { returnFunc } from "../utils/returnHelper.utils.js";
|
|||||||
import { tryCatch } from "../utils/trycatch.utils.js";
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
import { datamartData } from "./datamartData.utlis.js";
|
import { datamartData } from "./datamartData.utlis.js";
|
||||||
|
|
||||||
type Options = {
|
|
||||||
name: string;
|
|
||||||
value: string;
|
|
||||||
};
|
|
||||||
type Data = {
|
type Data = {
|
||||||
name: string;
|
name: string;
|
||||||
options: Options;
|
options: any;
|
||||||
optionsRequired?: boolean;
|
optionsRequired?: boolean;
|
||||||
howManyOptionsRequired?: number;
|
howManyOptionsRequired?: number;
|
||||||
};
|
};
|
||||||
|
|
||||||
|
const lstDbRun = async (data: Data) => {
|
||||||
|
if (data.options) {
|
||||||
|
if (data.name === "psiInventory") {
|
||||||
|
const ids = data.options.articles.split(",").map((id: any) => id.trim());
|
||||||
|
const whse = data.options.whseToInclude
|
||||||
|
? data.options.whseToInclude
|
||||||
|
.split(",")
|
||||||
|
.map((w: any) => w.trim())
|
||||||
|
.filter(Boolean)
|
||||||
|
: [];
|
||||||
|
|
||||||
|
const locations = data.options.exludeLanes
|
||||||
|
? data.options.exludeLanes
|
||||||
|
.split(",")
|
||||||
|
.map((l: any) => l.trim())
|
||||||
|
.filter(Boolean)
|
||||||
|
: [];
|
||||||
|
|
||||||
|
const conditions = [
|
||||||
|
inArray(invHistoricalData.article, ids),
|
||||||
|
between(
|
||||||
|
invHistoricalData.histDate,
|
||||||
|
data.options.startDate,
|
||||||
|
data.options.endDate,
|
||||||
|
),
|
||||||
|
];
|
||||||
|
|
||||||
|
// only add the warehouse condition if there are any whse values
|
||||||
|
if (whse.length > 0) {
|
||||||
|
conditions.push(inArray(invHistoricalData.whseId, whse));
|
||||||
|
}
|
||||||
|
|
||||||
|
// locations we dont want in the system
|
||||||
|
if (locations.length > 0) {
|
||||||
|
conditions.push(notInArray(invHistoricalData.location, locations));
|
||||||
|
}
|
||||||
|
|
||||||
|
return await db
|
||||||
|
.select()
|
||||||
|
.from(invHistoricalData)
|
||||||
|
.where(and(...conditions));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return [];
|
||||||
|
};
|
||||||
export const runDatamartQuery = async (data: Data) => {
|
export const runDatamartQuery = async (data: Data) => {
|
||||||
// search the query db for the query by name
|
// search the query db for the query by name
|
||||||
const sqlQuery = sqlQuerySelector(`${data.name}`) as SqlQuery;
|
const considerLstDBRuns = ["psiInventory"];
|
||||||
|
|
||||||
|
if (considerLstDBRuns.includes(data.name)) {
|
||||||
|
const lstDB = await lstDbRun(data);
|
||||||
|
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "datamart",
|
||||||
|
subModule: "lstDBrn",
|
||||||
|
message: `Data for: ${data.name}`,
|
||||||
|
data: lstDB,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const featureQ = sqlQuerySelector(`featureCheck`) as SqlQuery;
|
||||||
|
|
||||||
|
const { data: fd, error: fe } = await tryCatch(
|
||||||
|
prodQuery(featureQ.query, `Running feature check`),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (fe) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "datamart",
|
||||||
|
subModule: "query",
|
||||||
|
message: `feature check failed`,
|
||||||
|
data: fe as any,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// for queries that will need to be ran on legacy until we get the plant updated need to go in here
|
||||||
|
const doubleQueries = ["inventory"];
|
||||||
|
const sqlQuery = sqlQuerySelector(
|
||||||
|
`datamart.${fd.data[0].activated > 0 && !doubleQueries.includes(data.name) ? data.name : `legacy.${data.name}`}`,
|
||||||
|
) as SqlQuery;
|
||||||
|
|
||||||
|
// checking if warehousing is as it will start to effect a lot of queries for plants that are not on 2.
|
||||||
|
|
||||||
const getDataMartInfo = datamartData.filter((x) => x.endpoint === data.name);
|
const getDataMartInfo = datamartData.filter((x) => x.endpoint === data.name);
|
||||||
|
|
||||||
// const optionsMissing =
|
// const optionsMissing =
|
||||||
// !data.options || Object.keys(data.options).length === 0;
|
// !data.options || Object.keys(data.options).length === 0;
|
||||||
|
|
||||||
const optionCount =
|
const isValid =
|
||||||
Object.keys(data.options).length ===
|
Object.keys(data.options ?? {}).length >=
|
||||||
getDataMartInfo[0]?.howManyOptionsRequired;
|
(getDataMartInfo[0]?.howManyOptionsRequired ?? 0);
|
||||||
|
|
||||||
if (getDataMartInfo[0]?.optionsRequired && !optionCount) {
|
if (getDataMartInfo[0]?.optionsRequired && !isValid) {
|
||||||
return returnFunc({
|
return returnFunc({
|
||||||
success: false,
|
success: false,
|
||||||
level: "error",
|
level: "error",
|
||||||
module: "datamart",
|
module: "datamart",
|
||||||
subModule: "query",
|
subModule: "query",
|
||||||
message: `This query is required to have the ${getDataMartInfo[0]?.howManyOptionsRequired} options set in order use it.`,
|
message: `This query is required to have ${getDataMartInfo[0]?.howManyOptionsRequired} option(s) set in order use it, please add in your option(s) data and try again.`,
|
||||||
data: [getDataMartInfo[0].options],
|
data: [getDataMartInfo[0].options],
|
||||||
notify: false,
|
notify: false,
|
||||||
});
|
});
|
||||||
@@ -75,11 +160,121 @@ export const runDatamartQuery = async (data: Data) => {
|
|||||||
|
|
||||||
// split the criteria by "," then and then update the query
|
// split the criteria by "," then and then update the query
|
||||||
if (data.options) {
|
if (data.options) {
|
||||||
Object.entries(data.options ?? {}).forEach(([key, value]) => {
|
switch (data.name) {
|
||||||
const pattern = new RegExp(`\\[${key.trim()}\\]`, "g");
|
case "activeArticles":
|
||||||
datamartQuery = datamartQuery.replace(pattern, String(value).trim());
|
break;
|
||||||
|
case "deliveryByDateRange":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replace("[startDate]", `${data.options.startDate}`)
|
||||||
|
.replace("[endDate]", `${data.options.endDate}`)
|
||||||
|
.replace(
|
||||||
|
"--and r.ArticleHumanReadableId in ([articles]) ",
|
||||||
|
data.options.articles
|
||||||
|
? `and r.ArticleHumanReadableId in (${data.options.articles})`
|
||||||
|
: "--and r.ArticleHumanReadableId in ([articles]) ",
|
||||||
|
);
|
||||||
|
|
||||||
|
break;
|
||||||
|
case "customerInventory":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replace(
|
||||||
|
"--and IdAdressen",
|
||||||
|
`and IdAdressen in (${data.options.customer})`,
|
||||||
|
)
|
||||||
|
.replace(
|
||||||
|
"--and x.IdWarenlager in (0)",
|
||||||
|
`${data.options.whseToInclude ? `and x.IdWarenlager in (${data.options.whseToInclude})` : `--and x.IdWarenlager in (0)`}`,
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
case "openOrders":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replace("[startDay]", `${data.options.startDay}`)
|
||||||
|
.replace("[endDay]", `${data.options.endDay}`);
|
||||||
|
break;
|
||||||
|
case "inventory":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replaceAll(
|
||||||
|
"--,l.RunningNumber",
|
||||||
|
`${data.options.includeRunningNumbers ? `,l.RunningNumber` : `--,l.RunningNumber`}`,
|
||||||
|
)
|
||||||
|
.replaceAll(
|
||||||
|
"--,l.MachineLocation,l.MachineName,l.ProductionLotRunningNumber as lot",
|
||||||
|
`${data.options.lots ? `,l.MachineLocation,l.MachineName,l.ProductionLotRunningNumber as lot` : `--,l.MachineLocation,l.MachineName,l.ProductionLotRunningNumber as lot`}`,
|
||||||
|
)
|
||||||
|
.replaceAll(
|
||||||
|
"--,l.WarehouseDescription,l.LaneDescription",
|
||||||
|
`${data.options.locations ? `,l.WarehouseDescription,l.LaneDescription` : `--,l.WarehouseDescription,l.LaneDescription`}`,
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
case "fakeEDIUpdate":
|
||||||
|
datamartQuery = datamartQuery.replace(
|
||||||
|
"--AND h.CustomerHumanReadableId in (0)",
|
||||||
|
`${data.options.address ? `AND h.CustomerHumanReadableId in (${data.options.address})` : `--AND h.CustomerHumanReadableId in (0)`}`,
|
||||||
|
);
|
||||||
|
|
||||||
|
break;
|
||||||
|
case "forecast":
|
||||||
|
datamartQuery = datamartQuery.replace(
|
||||||
|
"where DeliveryAddressHumanReadableId in ([customers])",
|
||||||
|
data.options.customers
|
||||||
|
? `where DeliveryAddressHumanReadableId in (${data.options.customers})`
|
||||||
|
: "--where DeliveryAddressHumanReadableId in ([customers])",
|
||||||
|
);
|
||||||
|
|
||||||
|
break;
|
||||||
|
case "activeArticles2":
|
||||||
|
datamartQuery = datamartQuery.replace(
|
||||||
|
"and a.HumanReadableId in ([articles])",
|
||||||
|
data.options.articles
|
||||||
|
? `and a.HumanReadableId in (${data.options.articles})`
|
||||||
|
: "--and a.HumanReadableId in ([articles])",
|
||||||
|
);
|
||||||
|
|
||||||
|
break;
|
||||||
|
|
||||||
|
case "psiDeliveryData":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replace("[startDate]", `${data.options.startDate}`)
|
||||||
|
.replace("[endDate]", `${data.options.endDate}`)
|
||||||
|
.replace(
|
||||||
|
"[articles]",
|
||||||
|
data.options.articles ? `${data.options.articles}` : "[articles]",
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
case "productionData":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replace("[startDate]", `${data.options.startDate}`)
|
||||||
|
.replace("[endDate]", `${data.options.endDate}`)
|
||||||
|
.replace(
|
||||||
|
"and ArticleHumanReadableId in ([articles])",
|
||||||
|
data.options.articles
|
||||||
|
? `and ArticleHumanReadableId in (${data.options.articles})`
|
||||||
|
: "--and ArticleHumanReadableId in ([articles])",
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
case "psiPlanningData":
|
||||||
|
datamartQuery = datamartQuery
|
||||||
|
.replace("[startDate]", `${data.options.startDate}`)
|
||||||
|
.replace("[endDate]", `${data.options.endDate}`)
|
||||||
|
.replace(
|
||||||
|
"and p.IdArtikelvarianten in ([articles])",
|
||||||
|
data.options.articles
|
||||||
|
? `and p.IdArtikelvarianten in (${data.options.articles})`
|
||||||
|
: "--and p.IdArtikelvarianten in ([articles])",
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
default:
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "datamart",
|
||||||
|
subModule: "query",
|
||||||
|
message: `${data.name} encountered an error as it might not exist in LST please contact support if this continues to happen`,
|
||||||
|
data: [sqlQuery.message],
|
||||||
|
notify: true,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
const { data: queryRun, error } = await tryCatch(
|
const { data: queryRun, error } = await tryCatch(
|
||||||
prodQuery(datamartQuery, `Running datamart query: ${data.name}`),
|
prodQuery(datamartQuery, `Running datamart query: ${data.name}`),
|
||||||
|
|||||||
@@ -10,14 +10,50 @@ export const datamartData = [
|
|||||||
name: "Active articles",
|
name: "Active articles",
|
||||||
endpoint: "activeArticles",
|
endpoint: "activeArticles",
|
||||||
description: "returns all active articles for the server with custom data",
|
description: "returns all active articles for the server with custom data",
|
||||||
options: "", // set as a string and each item will be seperated by a , this way we can split it later in the excel file.
|
options: "",
|
||||||
optionsRequired: false,
|
optionsRequired: false,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "Delivery by date range",
|
name: "Delivery by date range",
|
||||||
endpoint: "deliveryByDateRange",
|
endpoint: "deliveryByDateRange",
|
||||||
description: `Returns all Deliverys in selected date range IE: 1/1/${new Date(Date.now()).getFullYear()} to 1/31/${new Date(Date.now()).getFullYear()}`,
|
description: `Returns all Deliveries in selected date range IE: 1/1/${new Date(Date.now()).getFullYear()} to 1/31/${new Date(Date.now()).getFullYear()}`,
|
||||||
options: "startDate,endDate", // set as a string and each item will be seperated by a , this way we can split it later in the excel file.
|
options: "startDate,endDate",
|
||||||
|
optionsRequired: true,
|
||||||
|
howManyOptionsRequired: 2,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Get Customer Inventory",
|
||||||
|
endpoint: "customerInventory",
|
||||||
|
description: `Returns specific customer inventory based on there address ID, IE: 8,12,145. \nWith option to include specific warehousesIds, IE 36,41,5. \nNOTES: *leaving warehouse blank will just pull everything for the customer, Inventory dose not include PPOO or INV`,
|
||||||
|
options: "customer,whseToInclude",
|
||||||
|
optionsRequired: true,
|
||||||
|
howManyOptionsRequired: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Get open order",
|
||||||
|
endpoint: "openOrders",
|
||||||
|
description: `Returns open orders based on day count sent over, IE: startDay 15 days in the past endDay 5 days in the future, can be left empty for this default days`,
|
||||||
|
options: "startDay,endDay",
|
||||||
|
optionsRequired: true,
|
||||||
|
howManyOptionsRequired: 2,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Get inventory",
|
||||||
|
endpoint: "inventory",
|
||||||
|
description: `Returns all inventory, excludes inv location. adding an x in one of the options will enable it.`,
|
||||||
|
options: "includeRunningNumbers,locations,lots",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Fake EDI Update",
|
||||||
|
endpoint: "fakeEDIUpdate",
|
||||||
|
description: `Returns all open orders to correct and resubmit via lst demand mgt, leaving blank will get everything putting an address only returns the specified address. \nNOTE: only orders that were created via edi will populate here.`,
|
||||||
|
options: "address",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "Production Data",
|
||||||
|
endpoint: "productionData",
|
||||||
|
description: `Returns all production data from the date range with the option to have 1 to many avs to search by.`,
|
||||||
|
options: "startDate,endDate,articles",
|
||||||
optionsRequired: true,
|
optionsRequired: true,
|
||||||
howManyOptionsRequired: 2,
|
howManyOptionsRequired: 2,
|
||||||
},
|
},
|
||||||
|
|||||||
@@ -22,9 +22,10 @@ export const alplaPurchaseHistory = pgTable("alpla_purchase_history", {
|
|||||||
upd_user: text("upd_user"),
|
upd_user: text("upd_user"),
|
||||||
upd_date: timestamp("upd_date").defaultNow(),
|
upd_date: timestamp("upd_date").defaultNow(),
|
||||||
remark: text("remark"),
|
remark: text("remark"),
|
||||||
approvedStatus: text("approved_status").default("pending"),
|
approvedStatus: text("approved_status").default("new"),
|
||||||
position: jsonb("position").default([]),
|
position: jsonb("position").default([]),
|
||||||
createdAt: timestamp("created_at").defaultNow(),
|
createdAt: timestamp("created_at").defaultNow(),
|
||||||
|
updatedAt: timestamp("updated_at").defaultNow(),
|
||||||
});
|
});
|
||||||
|
|
||||||
export const alplaPurchaseHistorySchema =
|
export const alplaPurchaseHistorySchema =
|
||||||
|
|||||||
30
backend/db/schema/historicalInv.schema.ts
Normal file
30
backend/db/schema/historicalInv.schema.ts
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
import { date, pgTable, text, timestamp, uuid } from "drizzle-orm/pg-core";
|
||||||
|
import { createInsertSchema, createSelectSchema } from "drizzle-zod";
|
||||||
|
import type z from "zod";
|
||||||
|
|
||||||
|
export const invHistoricalData = pgTable("inv_historical_data", {
|
||||||
|
inv: uuid("id").defaultRandom().primaryKey(),
|
||||||
|
histDate: date("hist_date").notNull(), // this date should always be yesterday when we post it.
|
||||||
|
plantToken: text("plant_token"),
|
||||||
|
article: text("article").notNull(),
|
||||||
|
articleDescription: text("article_description").notNull(),
|
||||||
|
materialType: text("material_type"),
|
||||||
|
total_QTY: text("total_QTY"),
|
||||||
|
available_QTY: text("available_QTY"),
|
||||||
|
coa_QTY: text("coa_QTY"),
|
||||||
|
held_QTY: text("held_QTY"),
|
||||||
|
consignment_QTY: text("consignment_qty"),
|
||||||
|
lot_Number: text("lot_number"),
|
||||||
|
locationId: text("location_id"),
|
||||||
|
location: text("location"),
|
||||||
|
whseId: text("whse_id").default(""),
|
||||||
|
whseName: text("whse_name").default("missing whseName"),
|
||||||
|
upd_user: text("upd_user").default("lst-system"),
|
||||||
|
upd_date: timestamp("upd_date").defaultNow(),
|
||||||
|
});
|
||||||
|
|
||||||
|
export const invHistoricalDataSchema = createSelectSchema(invHistoricalData);
|
||||||
|
export const newInvHistoricalDataSchema = createInsertSchema(invHistoricalData);
|
||||||
|
|
||||||
|
export type InvHistoricalData = z.infer<typeof invHistoricalDataSchema>;
|
||||||
|
export type NewInvHistoricalData = z.infer<typeof newInvHistoricalDataSchema>;
|
||||||
@@ -1,6 +1,11 @@
|
|||||||
import { integer, pgTable, text } from "drizzle-orm/pg-core";
|
import { integer, pgTable, text, timestamp } from "drizzle-orm/pg-core";
|
||||||
|
|
||||||
export const opendockApt = pgTable("printer_log", {
|
export const printerLog = pgTable("printer_log", {
|
||||||
id: integer().primaryKey().generatedAlwaysAsIdentity(),
|
id: integer().primaryKey().generatedAlwaysAsIdentity(),
|
||||||
name: text("name").notNull(),
|
name: text("name"),
|
||||||
|
ip: text("ip"),
|
||||||
|
printerSN: text("printer_sn"),
|
||||||
|
condition: text("condition").notNull(),
|
||||||
|
message: text("message"),
|
||||||
|
createdAt: timestamp("created_at").defaultNow(),
|
||||||
});
|
});
|
||||||
|
|||||||
44
backend/db/schema/printers.schema.ts
Normal file
44
backend/db/schema/printers.schema.ts
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
import {
|
||||||
|
boolean,
|
||||||
|
integer,
|
||||||
|
jsonb,
|
||||||
|
pgTable,
|
||||||
|
text,
|
||||||
|
timestamp,
|
||||||
|
uniqueIndex,
|
||||||
|
uuid,
|
||||||
|
} from "drizzle-orm/pg-core";
|
||||||
|
import { createInsertSchema, createSelectSchema } from "drizzle-zod";
|
||||||
|
import type z from "zod";
|
||||||
|
|
||||||
|
export const printerData = pgTable(
|
||||||
|
"printer_data",
|
||||||
|
{
|
||||||
|
id: uuid("id").defaultRandom().primaryKey(),
|
||||||
|
humanReadableId: text("humanReadable_id").unique().notNull(),
|
||||||
|
name: text("name").notNull(),
|
||||||
|
ipAddress: text("ipAddress"),
|
||||||
|
port: integer("port"),
|
||||||
|
status: text("status"),
|
||||||
|
statusText: text("statusText"),
|
||||||
|
printerSN: text("printer_sn"),
|
||||||
|
lastTimePrinted: timestamp("last_time_printed").notNull().defaultNow(),
|
||||||
|
assigned: boolean("assigned").default(false),
|
||||||
|
remark: text("remark"),
|
||||||
|
printDelay: integer("printDelay").default(90),
|
||||||
|
processes: jsonb("processes").default([]),
|
||||||
|
printDelayOverride: boolean("print_delay_override").default(false), // this will be more for if we have the lot time active but want to over ride this single line for some reason
|
||||||
|
add_Date: timestamp("add_Date").defaultNow(),
|
||||||
|
upd_date: timestamp("upd_date").defaultNow(),
|
||||||
|
},
|
||||||
|
(table) => [
|
||||||
|
//uniqueIndex("emailUniqueIndex").on(sql`lower(${table.email})`),
|
||||||
|
uniqueIndex("printer_id").on(table.humanReadableId),
|
||||||
|
],
|
||||||
|
);
|
||||||
|
|
||||||
|
export const printerSchema = createSelectSchema(printerData);
|
||||||
|
export const newPrinterSchema = createInsertSchema(printerData);
|
||||||
|
|
||||||
|
export type Printer = z.infer<typeof printerSchema>;
|
||||||
|
export type NewPrinter = z.infer<typeof newPrinterSchema>;
|
||||||
17
backend/gpSql/gpSql.routes.ts
Normal file
17
backend/gpSql/gpSql.routes.ts
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
import { type Express, Router } from "express";
|
||||||
|
import { requireAuth } from "../middleware/auth.middleware.js";
|
||||||
|
import restart from "./gpSqlRestart.route.js";
|
||||||
|
import start from "./gpSqlStart.route.js";
|
||||||
|
import stop from "./gpSqlStop.route.js";
|
||||||
|
export const setupGPSqlRoutes = (baseUrl: string, app: Express) => {
|
||||||
|
//setup all the routes
|
||||||
|
// Apply auth to entire router
|
||||||
|
const router = Router();
|
||||||
|
router.use(requireAuth);
|
||||||
|
|
||||||
|
router.use(start);
|
||||||
|
router.use(stop);
|
||||||
|
router.use(restart);
|
||||||
|
|
||||||
|
app.use(`${baseUrl}/api/system/gpSql`, router);
|
||||||
|
};
|
||||||
148
backend/gpSql/gpSqlConnection.controller.ts
Normal file
148
backend/gpSql/gpSqlConnection.controller.ts
Normal file
@@ -0,0 +1,148 @@
|
|||||||
|
import sql from "mssql";
|
||||||
|
import { gpSqlConfig } from "../configs/gpSql.config.js";
|
||||||
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import { checkHostnamePort } from "../utils/checkHost.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
|
||||||
|
export let pool2: sql.ConnectionPool;
|
||||||
|
export let connected: boolean = false;
|
||||||
|
export let reconnecting = false;
|
||||||
|
// start the delay out as 2 seconds
|
||||||
|
let delayStart = 2000;
|
||||||
|
let attempt = 0;
|
||||||
|
const maxAttempts = 10;
|
||||||
|
|
||||||
|
export const connectGPSql = async () => {
|
||||||
|
const serverUp = await checkHostnamePort(`USMCD1VMS011:1433`);
|
||||||
|
if (!serverUp) {
|
||||||
|
// we will try to reconnect
|
||||||
|
connected = false;
|
||||||
|
reconnectToSql;
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: "GP server is offline or unreachable.",
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// if we are trying to click restart from the api for some reason we want to kick back and say no
|
||||||
|
if (connected) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: "The Sql server is already connected.",
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// try to connect to the sql server
|
||||||
|
try {
|
||||||
|
pool2 = new sql.ConnectionPool(gpSqlConfig);
|
||||||
|
await pool2.connect();
|
||||||
|
connected = true;
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: `${gpSqlConfig.server} is connected to ${gpSqlConfig.database}`,
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
} catch (error) {
|
||||||
|
reconnectToSql;
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: "Failed to connect to the prod sql server.",
|
||||||
|
data: [error],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
export const closePool = async () => {
|
||||||
|
if (!connected) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: "There is no connection to the prod server currently.",
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
await pool2.close();
|
||||||
|
connected = false;
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: "The sql connection has been closed.",
|
||||||
|
});
|
||||||
|
} catch (error) {
|
||||||
|
connected = false;
|
||||||
|
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
message: "There was an error closing the sql connection",
|
||||||
|
data: [error],
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
|
export const reconnectToSql = async () => {
|
||||||
|
const log = createLogger({
|
||||||
|
module: "system",
|
||||||
|
subModule: "db",
|
||||||
|
});
|
||||||
|
if (reconnecting) return;
|
||||||
|
|
||||||
|
//set reconnecting to true while we try to reconnect
|
||||||
|
reconnecting = true;
|
||||||
|
|
||||||
|
while (!connected && attempt < maxAttempts) {
|
||||||
|
attempt++;
|
||||||
|
log.info(
|
||||||
|
`Reconnect attempt ${attempt}/${maxAttempts} in ${delayStart / 1000}s ...`,
|
||||||
|
);
|
||||||
|
|
||||||
|
await new Promise((res) => setTimeout(res, delayStart));
|
||||||
|
|
||||||
|
const serverUp = await checkHostnamePort(`${process.env.PROD_SERVER}:1433`);
|
||||||
|
|
||||||
|
if (!serverUp) {
|
||||||
|
delayStart = Math.min(delayStart * 2, 30000); // exponential backoff until up to 30000
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
pool2 = await sql.connect(gpSqlConfig);
|
||||||
|
reconnecting = false;
|
||||||
|
connected = true;
|
||||||
|
log.info(`${gpSqlConfig.server} is connected to ${gpSqlConfig.database}`);
|
||||||
|
} catch (error) {
|
||||||
|
delayStart = Math.min(delayStart * 2, 30000);
|
||||||
|
log.error({ error }, "Failed to reconnect to the prod sql server.");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!connected && attempt >= maxAttempts) {
|
||||||
|
log.error(
|
||||||
|
{ notify: true },
|
||||||
|
"Max reconnect attempts reached on the prodSql server. Stopping retries.",
|
||||||
|
);
|
||||||
|
|
||||||
|
reconnecting = false;
|
||||||
|
// TODO: exit alert someone here
|
||||||
|
}
|
||||||
|
};
|
||||||
78
backend/gpSql/gpSqlQuery.controller.ts
Normal file
78
backend/gpSql/gpSqlQuery.controller.ts
Normal file
@@ -0,0 +1,78 @@
|
|||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { connected, pool2 } from "./gpSqlConnection.controller.js";
|
||||||
|
|
||||||
|
interface SqlError extends Error {
|
||||||
|
code?: string;
|
||||||
|
originalError?: {
|
||||||
|
info?: { message?: string };
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Run a prod query
|
||||||
|
* just pass over the query as a string and the name of the query.
|
||||||
|
* Query should be like below.
|
||||||
|
* * select * from AlplaPROD_test1.dbo.table
|
||||||
|
* You must use test1 always as it will be changed via query
|
||||||
|
*/
|
||||||
|
export const gpQuery = async (queryToRun: string, name: string) => {
|
||||||
|
if (!connected) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "gpSql",
|
||||||
|
message: `${process.env.PROD_PLANT_TOKEN} is offline or attempting to reconnect`,
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
//change to the correct server
|
||||||
|
const query = queryToRun.replaceAll(
|
||||||
|
"test1",
|
||||||
|
`${process.env.PROD_PLANT_TOKEN}`,
|
||||||
|
);
|
||||||
|
|
||||||
|
try {
|
||||||
|
const result = await pool2.request().query(query);
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
message: `Query results for: ${name}`,
|
||||||
|
data: result.recordset ?? [],
|
||||||
|
};
|
||||||
|
} catch (error: unknown) {
|
||||||
|
const err = error as SqlError;
|
||||||
|
if (err.code === "ETIMEOUT") {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
module: "system",
|
||||||
|
subModule: "gpSql",
|
||||||
|
level: "error",
|
||||||
|
message: `${name} did not run due to a timeout.`,
|
||||||
|
notify: false,
|
||||||
|
data: [],
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (err.code === "EREQUEST") {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
module: "system",
|
||||||
|
subModule: "gpSql",
|
||||||
|
level: "error",
|
||||||
|
message: `${name} encountered an error ${err.originalError?.info?.message || "undefined error"}`,
|
||||||
|
data: [],
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
module: "system",
|
||||||
|
subModule: "gpSql",
|
||||||
|
level: "error",
|
||||||
|
message: `${name} encountered an unknown error.`,
|
||||||
|
data: [],
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
29
backend/gpSql/gpSqlQuerySelector.utils.ts
Normal file
29
backend/gpSql/gpSqlQuerySelector.utils.ts
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
import { readFileSync } from "node:fs";
|
||||||
|
|
||||||
|
export type SqlGPQuery = {
|
||||||
|
query: string;
|
||||||
|
success: boolean;
|
||||||
|
message: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
export const sqlGpQuerySelector = (name: string) => {
|
||||||
|
try {
|
||||||
|
const queryFile = readFileSync(
|
||||||
|
new URL(`../gpSql/queries/${name}.sql`, import.meta.url),
|
||||||
|
"utf8",
|
||||||
|
);
|
||||||
|
|
||||||
|
return {
|
||||||
|
success: true,
|
||||||
|
message: `Query for: ${name}`,
|
||||||
|
query: queryFile,
|
||||||
|
};
|
||||||
|
} catch (e) {
|
||||||
|
console.error(e);
|
||||||
|
return {
|
||||||
|
success: false,
|
||||||
|
message:
|
||||||
|
"Error getting the query file, please make sure you have the correct name.",
|
||||||
|
};
|
||||||
|
}
|
||||||
|
};
|
||||||
23
backend/gpSql/gpSqlRestart.route.ts
Normal file
23
backend/gpSql/gpSqlRestart.route.ts
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { closePool, connectGPSql } from "./gpSqlConnection.controller.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/restart", async (_, res) => {
|
||||||
|
await closePool();
|
||||||
|
|
||||||
|
await new Promise((r) => setTimeout(r, 2000));
|
||||||
|
|
||||||
|
const connect = await connectGPSql();
|
||||||
|
apiReturn(res, {
|
||||||
|
success: connect.success,
|
||||||
|
level: connect.success ? "info" : "error",
|
||||||
|
module: "routes",
|
||||||
|
subModule: "prodSql",
|
||||||
|
message: "Sql Server has been restarted",
|
||||||
|
data: connect.data,
|
||||||
|
status: connect.success ? 200 : 400,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
export default r;
|
||||||
20
backend/gpSql/gpSqlStart.route.ts
Normal file
20
backend/gpSql/gpSqlStart.route.ts
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { connectGPSql } from "./gpSqlConnection.controller.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/start", async (_, res) => {
|
||||||
|
const connect = await connectGPSql();
|
||||||
|
apiReturn(res, {
|
||||||
|
success: connect.success,
|
||||||
|
level: connect.success ? "info" : "error",
|
||||||
|
module: "routes",
|
||||||
|
subModule: "prodSql",
|
||||||
|
message: connect.message,
|
||||||
|
data: connect.data,
|
||||||
|
status: connect.success ? 200 : 400,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
export default r;
|
||||||
20
backend/gpSql/gpSqlStop.route.ts
Normal file
20
backend/gpSql/gpSqlStop.route.ts
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { closePool } from "./gpSqlConnection.controller.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/stop", async (_, res) => {
|
||||||
|
const connect = await closePool();
|
||||||
|
apiReturn(res, {
|
||||||
|
success: connect.success,
|
||||||
|
level: connect.success ? "info" : "error",
|
||||||
|
module: "routes",
|
||||||
|
subModule: "prodSql",
|
||||||
|
message: connect.message,
|
||||||
|
data: connect.data,
|
||||||
|
status: connect.success ? 200 : 400,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
export default r;
|
||||||
39
backend/gpSql/queries/reqCheck.sql
Normal file
39
backend/gpSql/queries/reqCheck.sql
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
USE [ALPLA]
|
||||||
|
|
||||||
|
SELECT Distinct r.[POPRequisitionNumber] as req,
|
||||||
|
r.[ApprovalStatus] as approvalStatus,
|
||||||
|
r.[Requested By] requestedBy,
|
||||||
|
format(t.[Created Date], 'yyyy-MM-dd') as createdAt,
|
||||||
|
format(r.[Requisition Date], 'MM/dd/yyyy') as expectedDate,
|
||||||
|
r.[Requisition Amount] as glAccount,
|
||||||
|
case when r.[Account Segment 2] is null or r.[Account Segment 2] = '' then '999' else cast(r.[Account Segment 2] as varchar) end as plant
|
||||||
|
,t.Status as status
|
||||||
|
,t.[Document Status] as docStatus
|
||||||
|
,t.[Workflow Status] as reqState
|
||||||
|
,CASE
|
||||||
|
WHEN [Workflow Status] = 'Completed'
|
||||||
|
THEN 'Pending APO convertion'
|
||||||
|
WHEN [Workflow Status] = 'Pending User Action'
|
||||||
|
AND r.[ApprovalStatus] = 'Pending Approval'
|
||||||
|
THEN 'Pending plant approver'
|
||||||
|
WHEN [Workflow Status] = ''
|
||||||
|
AND r.[ApprovalStatus] = 'Not Submitted'
|
||||||
|
THEN 'Req not submited'
|
||||||
|
ELSE 'Unknown reason'
|
||||||
|
END AS approvedStatus
|
||||||
|
|
||||||
|
FROM [dbo].[PORequisitions] r (nolock)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
left join
|
||||||
|
[dbo].[PurchaseRequisitions] as t (nolock) on
|
||||||
|
t.[Requisition Number] = r.[POPRequisitionNumber]
|
||||||
|
|
||||||
|
|
||||||
|
--where ApprovalStatus = 'Pending Approval'
|
||||||
|
--and [Account Segment 2] = 80
|
||||||
|
|
||||||
|
where r.POPRequisitionNumber in ([reqsToCheck])
|
||||||
|
|
||||||
|
Order By r.POPRequisitionNumber
|
||||||
@@ -5,6 +5,7 @@ import { db } from "../db/db.controller.js";
|
|||||||
import { logs } from "../db/schema/logs.schema.js";
|
import { logs } from "../db/schema/logs.schema.js";
|
||||||
import { emitToRoom } from "../socket.io/roomEmitter.socket.js";
|
import { emitToRoom } from "../socket.io/roomEmitter.socket.js";
|
||||||
import { tryCatch } from "../utils/trycatch.utils.js";
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
import { notifySystemIssue } from "./logger.notify.js";
|
||||||
//import build from "pino-abstract-transport";
|
//import build from "pino-abstract-transport";
|
||||||
|
|
||||||
export const logLevel = process.env.LOG_LEVEL || "info";
|
export const logLevel = process.env.LOG_LEVEL || "info";
|
||||||
@@ -45,6 +46,10 @@ const dbStream = new Writable({
|
|||||||
console.error(res.error);
|
console.error(res.error);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (obj.notify) {
|
||||||
|
notifySystemIssue(obj);
|
||||||
|
}
|
||||||
|
|
||||||
if (obj.room) {
|
if (obj.room) {
|
||||||
emitToRoom(obj.room, res.data ? res.data[0] : obj);
|
emitToRoom(obj.room, res.data ? res.data[0] : obj);
|
||||||
}
|
}
|
||||||
|
|||||||
44
backend/logger/logger.notify.ts
Normal file
44
backend/logger/logger.notify.ts
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
/**
|
||||||
|
* For all logging that has notify set to true well send an email to the system admins, if we have a discord webhook set well send it there as well
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { user } from "../db/schema/auth.schema.js";
|
||||||
|
import { sendEmail } from "../utils/sendEmail.utils.js";
|
||||||
|
|
||||||
|
type NotifyData = {
|
||||||
|
module: string;
|
||||||
|
submodule: string;
|
||||||
|
hostname: string;
|
||||||
|
msg: string;
|
||||||
|
stack: unknown[];
|
||||||
|
};
|
||||||
|
|
||||||
|
export const notifySystemIssue = async (data: NotifyData) => {
|
||||||
|
// build the email out
|
||||||
|
|
||||||
|
const formattedError = Array.isArray(data.stack)
|
||||||
|
? data.stack.map((e: any) => e.error || e)
|
||||||
|
: data.stack;
|
||||||
|
|
||||||
|
const sysAdmin = await db
|
||||||
|
.select()
|
||||||
|
.from(user)
|
||||||
|
.where(eq(user.role, "systemAdmin"));
|
||||||
|
|
||||||
|
await sendEmail({
|
||||||
|
email: sysAdmin.map((r) => r.email).join("; ") ?? "cowchmonkey@gmail.com", // change to pull in system admin emails
|
||||||
|
subject: `${data.hostname} has encountered a critical issue.`,
|
||||||
|
template: "serverCritialIssue",
|
||||||
|
context: {
|
||||||
|
plant: data.hostname,
|
||||||
|
module: data.module,
|
||||||
|
subModule: data.submodule,
|
||||||
|
message: data.msg,
|
||||||
|
error: JSON.stringify(formattedError, null, 2),
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
// TODO: add discord
|
||||||
|
};
|
||||||
223
backend/logistics/logistics.historicalInv.ts
Normal file
223
backend/logistics/logistics.historicalInv.ts
Normal file
@@ -0,0 +1,223 @@
|
|||||||
|
import { format } from "date-fns";
|
||||||
|
import { eq, sql } from "drizzle-orm";
|
||||||
|
import { runDatamartQuery } from "../datamart/datamart.controller.js";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { invHistoricalData } from "../db/schema/historicalInv.schema.js";
|
||||||
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
|
import {
|
||||||
|
type SqlQuery,
|
||||||
|
sqlQuerySelector,
|
||||||
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import { createCronJob } from "../utils/croner.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
|
type Inventory = {
|
||||||
|
article: string;
|
||||||
|
alias: string;
|
||||||
|
materialType: string;
|
||||||
|
total_palletQTY: string;
|
||||||
|
available_QTY: string;
|
||||||
|
coa_QTY: string;
|
||||||
|
held_QTY: string;
|
||||||
|
consignment_qty: string;
|
||||||
|
lot: string;
|
||||||
|
locationId: string;
|
||||||
|
laneDescription: string;
|
||||||
|
warehouseId: string;
|
||||||
|
warehouseDescription: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
const historicalInvImport = async () => {
|
||||||
|
const today = new Date();
|
||||||
|
const { data, error } = await tryCatch(
|
||||||
|
db
|
||||||
|
.select()
|
||||||
|
.from(invHistoricalData)
|
||||||
|
.where(eq(invHistoricalData.histDate, format(today, "yyyy-MM-dd"))),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "system",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Error getting historical inv info`,
|
||||||
|
data: error as any,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (data?.length === 0) {
|
||||||
|
const avSQLQuery = sqlQuerySelector(`datamart.activeArticles`) as SqlQuery;
|
||||||
|
|
||||||
|
if (!avSQLQuery.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Error getting Article info`,
|
||||||
|
data: [avSQLQuery.message],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const { data: inv, error: invError } = await tryCatch(
|
||||||
|
//prodQuery(sqlQuery.query, "Inventory data"),
|
||||||
|
runDatamartQuery({
|
||||||
|
name: "inventory",
|
||||||
|
options: { lots: "x", locations: "x" },
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
|
||||||
|
const { data: av, error: avError } = (await tryCatch(
|
||||||
|
runDatamartQuery({ name: "activeArticles", options: {} }),
|
||||||
|
)) as any;
|
||||||
|
|
||||||
|
if (invError) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Error getting inventory info from prod query`,
|
||||||
|
data: invError as any,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (avError) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Error getting article info from prod query`,
|
||||||
|
data: invError as any,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// shape the data to go into our table
|
||||||
|
const plantToken = process.env.PROD_PLANT_TOKEN ?? "test1";
|
||||||
|
const importInv = (inv.data ? inv.data : []) as Inventory[];
|
||||||
|
const importData = importInv.map((i) => {
|
||||||
|
return {
|
||||||
|
histDate: sql`(NOW())::date`,
|
||||||
|
plantToken: plantToken,
|
||||||
|
article: i.article,
|
||||||
|
articleDescription: i.alias,
|
||||||
|
materialType:
|
||||||
|
av.data.filter((a: any) => a.article === i.article).length > 0
|
||||||
|
? av.data.filter((a: any) => a.article === i.article)[0]
|
||||||
|
?.TypeOfMaterial
|
||||||
|
: "Item not defined",
|
||||||
|
total_QTY: i.total_palletQTY ?? "0.00",
|
||||||
|
available_QTY: i.available_QTY ?? "0.00",
|
||||||
|
coa_QTY: i.coa_QTY ?? "0.00",
|
||||||
|
held_QTY: i.held_QTY ?? "0.00",
|
||||||
|
consignment_QTY: i.consignment_qty ?? "0.00",
|
||||||
|
lot_Number: i.lot ?? "0",
|
||||||
|
locationId: i.locationId ?? "0",
|
||||||
|
location: i.laneDescription ?? "Missing lane",
|
||||||
|
whseId: i.warehouseId ?? "0",
|
||||||
|
whseName: i.warehouseDescription ?? "Missing warehouse",
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
|
const { data: dataImport, error: errorImport } = await tryCatch(
|
||||||
|
db.insert(invHistoricalData).values(importData),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (errorImport) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Error adding historical data to lst db`,
|
||||||
|
data: errorImport as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (dataImport) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "info",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Historical data was added to lst :D`,
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "info",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Historical Data for: ${format(today, "yyyy-MM-dd")}, is already added and nothing to do.`,
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "info",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "inv",
|
||||||
|
message: `Some weird crazy error just happened and didnt get captured during the historical inv check.`,
|
||||||
|
data: [],
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
export const historicalSchedule = async () => {
|
||||||
|
// running the history in case my silly ass dose an update around the shift change time lol, this will prevent loss data. it might be off a little but no one cares
|
||||||
|
historicalInvImport();
|
||||||
|
|
||||||
|
const sqlQuery = sqlQuerySelector(`shiftChange`) as SqlQuery;
|
||||||
|
|
||||||
|
if (!sqlQuery.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Error getting shiftChange sql file`,
|
||||||
|
data: [sqlQuery.message],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const { data, error } = await tryCatch(
|
||||||
|
prodQuery(sqlQuery.query, "Shift Change data"),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "logistics",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Error getting shiftChange info`,
|
||||||
|
data: error as any,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
// shift split
|
||||||
|
const shiftTimeSplit = data?.data[0]?.shiftChange.split(":");
|
||||||
|
|
||||||
|
const cronSetup = `0 ${
|
||||||
|
shiftTimeSplit?.length > 0 ? `${parseInt(shiftTimeSplit[1])}` : "0"
|
||||||
|
} ${
|
||||||
|
shiftTimeSplit?.length > 0 ? `${parseInt(shiftTimeSplit[0])}` : "7"
|
||||||
|
} * * *`;
|
||||||
|
|
||||||
|
createCronJob("historicalInv", cronSetup, () => historicalInvImport());
|
||||||
|
};
|
||||||
96
backend/notification/notification.manualTrigger.ts
Normal file
96
backend/notification/notification.manualTrigger.ts
Normal file
@@ -0,0 +1,96 @@
|
|||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import { type Response, Router } from "express";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { notifications } from "../db/schema/notifications.schema.js";
|
||||||
|
import { auth } from "../utils/auth.utils.js";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/", async (req, res: Response) => {
|
||||||
|
const hasPermissions = await auth.api.userHasPermission({
|
||||||
|
body: {
|
||||||
|
//userId: req?.user?.id,
|
||||||
|
role: req.user?.roles as any,
|
||||||
|
permissions: {
|
||||||
|
notifications: ["readAll"], // This must match the structure in your access control
|
||||||
|
},
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!hasPermissions) {
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "post",
|
||||||
|
message: `You do not have permissions to be here`,
|
||||||
|
data: [],
|
||||||
|
status: 400,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const { data: nName, error: nError } = await tryCatch(
|
||||||
|
db
|
||||||
|
.select()
|
||||||
|
.from(notifications)
|
||||||
|
.where(eq(notifications.name, req.body.name)),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (nError) {
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "get",
|
||||||
|
message: `There was an error getting the notifications `,
|
||||||
|
data: [nError],
|
||||||
|
status: 400,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const { data: sub, error: sError } = await tryCatch(
|
||||||
|
db
|
||||||
|
.select()
|
||||||
|
.from(notifications)
|
||||||
|
.where(eq(notifications.name, req.body.name)),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (sError) {
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "get",
|
||||||
|
message: `There was an error getting the subs `,
|
||||||
|
data: [sError],
|
||||||
|
status: 400,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
const emailString = [
|
||||||
|
...new Set(
|
||||||
|
sub.flatMap((e: any) =>
|
||||||
|
e.emails?.map((email: any) => email.trim().toLowerCase()),
|
||||||
|
),
|
||||||
|
),
|
||||||
|
].join(";");
|
||||||
|
|
||||||
|
console.log(emailString);
|
||||||
|
const { default: runFun } = await import(
|
||||||
|
`./notification.${req.body.name.trim()}.js`
|
||||||
|
);
|
||||||
|
const manual = await runFun(nName[0], "blake.matthes@alpla.com");
|
||||||
|
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "post",
|
||||||
|
message: `Manual Trigger ran`,
|
||||||
|
data: manual ?? [],
|
||||||
|
status: 200,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
export default r;
|
||||||
114
backend/notification/notification.qualityBlocking.ts
Normal file
114
backend/notification/notification.qualityBlocking.ts
Normal file
@@ -0,0 +1,114 @@
|
|||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { notifications } from "../db/schema/notifications.schema.js";
|
||||||
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
|
import {
|
||||||
|
type SqlQuery,
|
||||||
|
sqlQuerySelector,
|
||||||
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import { delay } from "../utils/delay.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { sendEmail } from "../utils/sendEmail.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
import { v2QueryRun } from "../utils/pgConnectToLst.utils.js";
|
||||||
|
|
||||||
|
let shutoffv1 = false
|
||||||
|
const func = async (data: any, emails: string) => {
|
||||||
|
// TODO: remove this disable once all 17 plants are on this new lst
|
||||||
|
if (!shutoffv1){
|
||||||
|
v2QueryRun(`update public.notifications set active = false where name = '${data.name}'`)
|
||||||
|
shutoffv1 = true
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
const { data: l, error: le } = (await tryCatch(
|
||||||
|
db.select().from(notifications).where(eq(notifications.id, data.id)),
|
||||||
|
)) as any;
|
||||||
|
|
||||||
|
if (le) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `${data.name} encountered an error while trying to get initial info`,
|
||||||
|
data: le as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
// search the query db for the query by name
|
||||||
|
const sqlQuery = sqlQuerySelector(`${data.name}`) as SqlQuery;
|
||||||
|
// create the ignore audit logs ids
|
||||||
|
|
||||||
|
// get get the latest blocking order id that was sent
|
||||||
|
const blockingOrderId = l[0].options[0].lastBlockingOrderIdSent ?? 69;
|
||||||
|
|
||||||
|
// run the check
|
||||||
|
const { data: queryRun, error } = await tryCatch(
|
||||||
|
prodQuery(
|
||||||
|
sqlQuery.query.replace("[lastBlocking]", blockingOrderId),
|
||||||
|
`Running notification query: ${l[0].name}`,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
|
data: error as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (queryRun.data.length > 0) {
|
||||||
|
for (const bo of queryRun.data) {
|
||||||
|
const sentEmail = await sendEmail({
|
||||||
|
email: emails,
|
||||||
|
subject: bo.subject,
|
||||||
|
template: "qualityBlocking",
|
||||||
|
context: {
|
||||||
|
items: bo,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!sentEmail?.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "email",
|
||||||
|
message: `${l[0].name} failed to send the email`,
|
||||||
|
data: sentEmail?.data as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
await delay(1500);
|
||||||
|
|
||||||
|
const { error: dbe } = await tryCatch(
|
||||||
|
db
|
||||||
|
.update(notifications)
|
||||||
|
.set({ options: [{ lastBlockingOrderIdSent: bo.blockingNumber }] })
|
||||||
|
.where(eq(notifications.id, data.id)),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (dbe) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "notification",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
|
data: dbe as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
export default func;
|
||||||
@@ -9,9 +9,16 @@ import {
|
|||||||
import { returnFunc } from "../utils/returnHelper.utils.js";
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
import { sendEmail } from "../utils/sendEmail.utils.js";
|
import { sendEmail } from "../utils/sendEmail.utils.js";
|
||||||
import { tryCatch } from "../utils/trycatch.utils.js";
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
import { v2QueryRun } from "../utils/pgConnectToLst.utils.js";
|
||||||
|
|
||||||
|
let shutoffv1 = false
|
||||||
|
const func = async (data: any, emails: string) => {
|
||||||
|
// TODO: remove this disable once all 17 plants are on this new lst
|
||||||
|
if (!shutoffv1){
|
||||||
|
v2QueryRun(`update public.notifications set active = false where name = '${data.name}'`)
|
||||||
|
shutoffv1 = true
|
||||||
|
}
|
||||||
|
|
||||||
const reprint = async (data: any, emails: string) => {
|
|
||||||
// TODO: do the actual logic for the notification.
|
|
||||||
const { data: l, error: le } = (await tryCatch(
|
const { data: l, error: le } = (await tryCatch(
|
||||||
db.select().from(notifications).where(eq(notifications.id, data.id)),
|
db.select().from(notifications).where(eq(notifications.id, data.id)),
|
||||||
)) as any;
|
)) as any;
|
||||||
@@ -23,7 +30,7 @@ const reprint = async (data: any, emails: string) => {
|
|||||||
module: "notification",
|
module: "notification",
|
||||||
subModule: "query",
|
subModule: "query",
|
||||||
message: `${data.name} encountered an error while trying to get initial info`,
|
message: `${data.name} encountered an error while trying to get initial info`,
|
||||||
data: [le],
|
data: le as any,
|
||||||
notify: true,
|
notify: true,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
@@ -52,7 +59,7 @@ const reprint = async (data: any, emails: string) => {
|
|||||||
module: "notification",
|
module: "notification",
|
||||||
subModule: "query",
|
subModule: "query",
|
||||||
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
data: [error],
|
data: error as any,
|
||||||
notify: true,
|
notify: true,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
@@ -73,7 +80,7 @@ const reprint = async (data: any, emails: string) => {
|
|||||||
module: "notification",
|
module: "notification",
|
||||||
subModule: "query",
|
subModule: "query",
|
||||||
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
message: `Data for: ${l[0].name} encountered an error while trying to get it`,
|
||||||
data: [dbe],
|
data: dbe as any,
|
||||||
notify: true,
|
notify: true,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
@@ -93,14 +100,14 @@ const reprint = async (data: any, emails: string) => {
|
|||||||
return returnFunc({
|
return returnFunc({
|
||||||
success: false,
|
success: false,
|
||||||
level: "error",
|
level: "error",
|
||||||
module: "email",
|
module: "notification",
|
||||||
subModule: "notification",
|
subModule: "email",
|
||||||
message: `${l[0].name} failed to send the email`,
|
message: `${l[0].name} failed to send the email`,
|
||||||
data: [sentEmail],
|
data: sentEmail?.data as any,
|
||||||
notify: true,
|
notify: true,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
export default reprint;
|
export default func;
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
import type { Express } from "express";
|
import type { Express } from "express";
|
||||||
import { requireAuth } from "../middleware/auth.middleware.js";
|
import { requireAuth } from "../middleware/auth.middleware.js";
|
||||||
|
import manual from "./notification.manualTrigger.js";
|
||||||
import getNotifications from "./notification.route.js";
|
import getNotifications from "./notification.route.js";
|
||||||
import updateNote from "./notification.update.route.js";
|
import updateNote from "./notification.update.route.js";
|
||||||
import deleteSub from "./notificationSub.delete.route.js";
|
import deleteSub from "./notificationSub.delete.route.js";
|
||||||
@@ -11,6 +12,7 @@ export const setupNotificationRoutes = (baseUrl: string, app: Express) => {
|
|||||||
//stats will be like this as we dont need to change this
|
//stats will be like this as we dont need to change this
|
||||||
app.use(`${baseUrl}/api/notification`, requireAuth, getNotifications);
|
app.use(`${baseUrl}/api/notification`, requireAuth, getNotifications);
|
||||||
app.use(`${baseUrl}/api/notification`, requireAuth, updateNote);
|
app.use(`${baseUrl}/api/notification`, requireAuth, updateNote);
|
||||||
|
app.use(`${baseUrl}/api/notification/manual`, requireAuth, manual);
|
||||||
app.use(`${baseUrl}/api/notification/sub`, requireAuth, subs);
|
app.use(`${baseUrl}/api/notification/sub`, requireAuth, subs);
|
||||||
app.use(`${baseUrl}/api/notification/sub`, requireAuth, newSub);
|
app.use(`${baseUrl}/api/notification/sub`, requireAuth, newSub);
|
||||||
app.use(`${baseUrl}/api/notification/sub`, requireAuth, updateSub);
|
app.use(`${baseUrl}/api/notification/sub`, requireAuth, updateSub);
|
||||||
|
|||||||
@@ -22,7 +22,7 @@ const note: NewNotification[] = [
|
|||||||
"Checks for new blocking orders that have been entered, recommend to get the most recent order in here before activating.",
|
"Checks for new blocking orders that have been entered, recommend to get the most recent order in here before activating.",
|
||||||
active: false,
|
active: false,
|
||||||
interval: "10",
|
interval: "10",
|
||||||
options: [{ sentBlockingOrders: [{ timeStamp: "0", blockingOrder: 1 }] }],
|
options: [{ lastBlockingOrderIdSent: 1 }],
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
name: "alplaPurchaseHistory",
|
name: "alplaPurchaseHistory",
|
||||||
|
|||||||
@@ -14,20 +14,82 @@
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
import { Router } from "express";
|
import { Router } from "express";
|
||||||
|
import multer from "multer";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { printerLog } from "../db/schema/printerLogs.schema.js";
|
||||||
import { apiReturn } from "../utils/returnHelper.utils.js";
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
|
type PrinterEvent = {
|
||||||
|
name: string;
|
||||||
|
condition: string;
|
||||||
|
message: string;
|
||||||
|
};
|
||||||
const r = Router();
|
const r = Router();
|
||||||
|
const upload = multer();
|
||||||
|
|
||||||
r.post("/printer/listener/:printer", async (req, res) => {
|
const parseZebraAlert = (body: any): PrinterEvent => {
|
||||||
|
const name = body.uniqueId || "unknown";
|
||||||
|
const decoded = decodeURIComponent(body.alertMsg || "");
|
||||||
|
|
||||||
|
const [conditionRaw, ...rest] = decoded.split(":");
|
||||||
|
const condition = conditionRaw?.toLowerCase()?.trim() || "unknown";
|
||||||
|
const message = rest.join(":").trim();
|
||||||
|
|
||||||
|
return {
|
||||||
|
name,
|
||||||
|
condition,
|
||||||
|
message,
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|
||||||
|
r.post("/printer/listener/:printer", upload.any(), async (req, res) => {
|
||||||
const { printer: printerName } = req.params;
|
const { printer: printerName } = req.params;
|
||||||
console.log(req.body);
|
const event: PrinterEvent = parseZebraAlert(req.body);
|
||||||
|
|
||||||
|
const rawIp =
|
||||||
|
req.headers["x-forwarded-for"]?.toString().split(",")[0]?.trim() ||
|
||||||
|
req.socket.remoteAddress ||
|
||||||
|
req.ip;
|
||||||
|
|
||||||
|
const ip = rawIp?.replace("::ffff:", "");
|
||||||
|
|
||||||
|
// post the new message
|
||||||
|
const { data, error } = await tryCatch(
|
||||||
|
db
|
||||||
|
.insert(printerLog)
|
||||||
|
.values({
|
||||||
|
ip: ip?.replace("::ffff:", ""),
|
||||||
|
name: printerName,
|
||||||
|
printerSN: event.name,
|
||||||
|
condition: event.condition,
|
||||||
|
message: event.message,
|
||||||
|
})
|
||||||
|
.returning(),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: false,
|
||||||
|
level: "info",
|
||||||
|
module: "ocp",
|
||||||
|
subModule: "printing",
|
||||||
|
message: `${printerName} encountered an error posting the log`,
|
||||||
|
data: error as any,
|
||||||
|
status: 400,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (data) {
|
||||||
|
// TODO: send message over to the controller to decide what to do next with it
|
||||||
|
}
|
||||||
|
|
||||||
return apiReturn(res, {
|
return apiReturn(res, {
|
||||||
success: true,
|
success: true,
|
||||||
level: "info",
|
level: "info",
|
||||||
module: "ocp",
|
module: "ocp",
|
||||||
subModule: "printing",
|
subModule: "printing",
|
||||||
message: `${printerName} just passed over a message`,
|
message: `${printerName} just sent a message`,
|
||||||
data: req.body ?? [],
|
data: req.body ?? [],
|
||||||
status: 200,
|
status: 200,
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -10,10 +10,323 @@
|
|||||||
* printer status will live here this will be how we manage all the levels of status like 3 paused, 1 printing, 8 error, 10 power up, etc...
|
* printer status will live here this will be how we manage all the levels of status like 3 paused, 1 printing, 8 error, 10 power up, etc...
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import net from "net";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { printerData } from "../db/schema/printers.schema.js";
|
||||||
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import { delay } from "../utils/delay.utils.js";
|
||||||
|
import { runProdApi } from "../utils/prodEndpoint.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
|
||||||
|
type Printer = {
|
||||||
|
name: string;
|
||||||
|
humanReadableId: string;
|
||||||
|
type: number;
|
||||||
|
ipAddress: string;
|
||||||
|
port: number;
|
||||||
|
default: boolean;
|
||||||
|
labelInstanceIpAddress: string;
|
||||||
|
labelInstancePort: number;
|
||||||
|
active: boolean;
|
||||||
|
remark: string;
|
||||||
|
processes: number[];
|
||||||
|
};
|
||||||
|
|
||||||
|
const log = createLogger({ module: "ocp", subModule: "printers" });
|
||||||
|
|
||||||
export const printerManager = async () => {};
|
export const printerManager = async () => {};
|
||||||
|
|
||||||
export const printerHeartBeat = async () => {
|
export const printerHeartBeat = async () => {
|
||||||
// heat heats will be defaulted to 60 seconds no reason to allow anything else
|
// heat heats will be defaulted to 60 seconds no reason to allow anything else, and heart beats will only go to assigned printers no need to be monitoring non labeling printers
|
||||||
};
|
};
|
||||||
|
|
||||||
//export const printerStatus = async (statusNr: number, printerId: number) => {};
|
//export const printerStatus = async (statusNr: number, printerId: number) => {};
|
||||||
|
export const printerSync = async () => {
|
||||||
|
// pull the printers from alpla prod and update them in lst
|
||||||
|
|
||||||
|
const printers = await runProdApi({
|
||||||
|
method: "get",
|
||||||
|
endpoint: "/public/v1.0/Administration/Printers",
|
||||||
|
});
|
||||||
|
|
||||||
|
if (!printers?.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "ocp",
|
||||||
|
subModule: "printer",
|
||||||
|
message: printers?.message ?? "",
|
||||||
|
data: printers?.data ?? [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (printers?.success && Array.isArray(printers.data)) {
|
||||||
|
const ignorePrinters = ["pdf24", "standard"];
|
||||||
|
|
||||||
|
const validPrinters =
|
||||||
|
printers.data.filter(
|
||||||
|
(n: any) =>
|
||||||
|
!ignorePrinters.includes(n.name.toLowerCase()) && n.ipAddress,
|
||||||
|
) ?? [];
|
||||||
|
if (validPrinters.length) {
|
||||||
|
for (const printer of validPrinters as Printer[]) {
|
||||||
|
// run an update for each printer, do on conflicts based on the printer id
|
||||||
|
log.debug({}, `Add/Updating ${printer.name}`);
|
||||||
|
|
||||||
|
if (printer.active) {
|
||||||
|
await db
|
||||||
|
.insert(printerData)
|
||||||
|
.values({
|
||||||
|
name: printer.name,
|
||||||
|
humanReadableId: printer.humanReadableId,
|
||||||
|
ipAddress: printer.ipAddress,
|
||||||
|
port: printer.port,
|
||||||
|
remark: printer.remark,
|
||||||
|
processes: printer.processes,
|
||||||
|
})
|
||||||
|
.onConflictDoUpdate({
|
||||||
|
target: printerData.humanReadableId,
|
||||||
|
set: {
|
||||||
|
name: printer.name,
|
||||||
|
humanReadableId: printer.humanReadableId,
|
||||||
|
ipAddress: printer.ipAddress,
|
||||||
|
port: printer.port,
|
||||||
|
remark: printer.remark,
|
||||||
|
processes: printer.processes,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
.returning();
|
||||||
|
await tcpPrinter(printer);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!printer.active) {
|
||||||
|
log.warn({}, `${printer.name} is not active so removing from lst.`);
|
||||||
|
await db
|
||||||
|
.delete(printerData)
|
||||||
|
.where(eq(printerData.humanReadableId, printer.humanReadableId));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "ocp",
|
||||||
|
subModule: "printer",
|
||||||
|
message: `${printers.data.length} printers were just synced, this includes new and old printers`,
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "ocp",
|
||||||
|
subModule: "printer",
|
||||||
|
message: `No printers to update`,
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
const tcpPrinter = (printer: Printer) => {
|
||||||
|
return new Promise<void>((resolve) => {
|
||||||
|
const socket = new net.Socket();
|
||||||
|
const timeoutMs = 15 * 1000;
|
||||||
|
|
||||||
|
const commands = [
|
||||||
|
{
|
||||||
|
key: "clearAlerts",
|
||||||
|
command: '! U1 setvar "alerts.configured" ""\r\n',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: "addAlert",
|
||||||
|
command: `! U1 setvar "alerts.add" "ALL MESSAGES,HTTP-POST,Y,Y,http://${process.env.SERVER_IP}:${process.env.PORT}/lst/api/ocp/printer/listener/${printer.name},0,N,printer"\r\n`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: "setFriendlyName",
|
||||||
|
command: `! U1 setvar "device.friendly_name" "${printer.name}"\r\n`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: "getUniqueId",
|
||||||
|
command: '! U1 getvar "device.unique_id"\r\n',
|
||||||
|
},
|
||||||
|
] as const;
|
||||||
|
|
||||||
|
let currentCommandIndex = 0;
|
||||||
|
let awaitingSerial = false;
|
||||||
|
let settled = false;
|
||||||
|
|
||||||
|
const cleanup = () => {
|
||||||
|
socket.removeAllListeners();
|
||||||
|
socket.destroy();
|
||||||
|
};
|
||||||
|
|
||||||
|
const finish = (err?: unknown) => {
|
||||||
|
if (settled) return;
|
||||||
|
settled = true;
|
||||||
|
clearTimeout(timeout);
|
||||||
|
cleanup();
|
||||||
|
|
||||||
|
if (err) {
|
||||||
|
log.error(
|
||||||
|
{ err, printer: printer.name },
|
||||||
|
`Printer update failed for ${printer.name}: doing the name and alert add directly on the printer.`,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
resolve();
|
||||||
|
};
|
||||||
|
|
||||||
|
const timeout = setTimeout(() => {
|
||||||
|
finish(`${printer.name} timed out while updating printer config`);
|
||||||
|
}, timeoutMs);
|
||||||
|
|
||||||
|
const sendNext = async () => {
|
||||||
|
if (currentCommandIndex >= commands.length) {
|
||||||
|
socket.end();
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
const current = commands[currentCommandIndex];
|
||||||
|
|
||||||
|
if (!current) {
|
||||||
|
socket.end();
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
awaitingSerial = current.key === "getUniqueId";
|
||||||
|
|
||||||
|
log.info(
|
||||||
|
{ printer: printer.name, command: current.key },
|
||||||
|
`Sending command to ${printer.name}`,
|
||||||
|
);
|
||||||
|
|
||||||
|
socket.write(current.command);
|
||||||
|
|
||||||
|
currentCommandIndex++;
|
||||||
|
|
||||||
|
// Small pause between commands so the printer has breathing room
|
||||||
|
if (currentCommandIndex < commands.length) {
|
||||||
|
await delay(1500);
|
||||||
|
await sendNext();
|
||||||
|
} else {
|
||||||
|
// last command was sent, now wait for final data/close
|
||||||
|
await delay(1500);
|
||||||
|
socket.end();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
socket.connect(printer.port, printer.ipAddress, async () => {
|
||||||
|
log.info({}, `Connected to ${printer.name}`);
|
||||||
|
|
||||||
|
try {
|
||||||
|
await sendNext();
|
||||||
|
} catch (error) {
|
||||||
|
finish(
|
||||||
|
error instanceof Error
|
||||||
|
? error
|
||||||
|
: new Error(
|
||||||
|
`Unknown error while sending commands to ${printer.name}`,
|
||||||
|
),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
socket.on("data", async (data) => {
|
||||||
|
const response = data.toString().trim().replaceAll('"', "");
|
||||||
|
|
||||||
|
log.info(
|
||||||
|
{ printer: printer.name, response },
|
||||||
|
`Received printer response from ${printer.name}`,
|
||||||
|
);
|
||||||
|
|
||||||
|
if (!awaitingSerial) return;
|
||||||
|
|
||||||
|
awaitingSerial = false;
|
||||||
|
|
||||||
|
try {
|
||||||
|
await db
|
||||||
|
.update(printerData)
|
||||||
|
.set({ printerSN: response })
|
||||||
|
.where(eq(printerData.humanReadableId, printer.humanReadableId));
|
||||||
|
} catch (error) {
|
||||||
|
finish(
|
||||||
|
error instanceof Error
|
||||||
|
? error
|
||||||
|
: new Error(`Failed to update printer SN for ${printer.name}`),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
socket.on("close", () => {
|
||||||
|
log.info({}, `Closed connection to ${printer.name}`);
|
||||||
|
finish();
|
||||||
|
});
|
||||||
|
|
||||||
|
socket.on("error", (err) => {
|
||||||
|
finish(err);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
// const tcpPrinter = async (printer: Printer) => {
|
||||||
|
// const p = new net.Socket();
|
||||||
|
// const commands = [
|
||||||
|
// '! U1 setvar "alerts.configured" ""\r\n', // clean install just remove all alerts
|
||||||
|
// `! U1 setvar "alerts.add" "ALL MESSAGES,HTTP-POST,Y,Y,http://${process.env.SERVER_IP}:${process.env.PORT}/lst/api/ocp/printer/listener/${printer.name},0,N,printer"\r\n`, // add in the all alert
|
||||||
|
// `! U1 setvar "device.friendly_name" "${printer.name}"\r\n`, // change the name to match the alplaprod name
|
||||||
|
// `! U1 getvar "device.unique_id"\r\n`, // this will get mapped into the printer as this is the one we will link to in the db.
|
||||||
|
// //'! U1 getvar "alerts.configured" ""\r\n',
|
||||||
|
// ];
|
||||||
|
|
||||||
|
// let index = 0;
|
||||||
|
// const sendNext = async () => {
|
||||||
|
// if (index >= commands.length) {
|
||||||
|
// p.end();
|
||||||
|
// return;
|
||||||
|
// }
|
||||||
|
|
||||||
|
// const cmd = commands[index] as string;
|
||||||
|
// p.write(cmd);
|
||||||
|
// return;
|
||||||
|
// };
|
||||||
|
|
||||||
|
// p.connect(printer.port, printer.ipAddress, async () => {
|
||||||
|
// log.info({}, `Connected to ${printer.name}`);
|
||||||
|
// while (index < commands.length) {
|
||||||
|
// await sendNext();
|
||||||
|
// await delay(2000);
|
||||||
|
// index++;
|
||||||
|
// }
|
||||||
|
// });
|
||||||
|
|
||||||
|
// p.on("data", async (data) => {
|
||||||
|
// // this is just the sn that comes over so we will update this printer.
|
||||||
|
// await db
|
||||||
|
// .update(printerData)
|
||||||
|
// .set({ printerSN: data.toString().trim().replaceAll('"', "") })
|
||||||
|
// .where(eq(printerData.humanReadableId, printer.humanReadableId));
|
||||||
|
|
||||||
|
// // get the name
|
||||||
|
// // p.write('! U1 getvar "device.friendly_name"\r\n');
|
||||||
|
// // p.write('! U1 getvar "device.unique_id"\r\n');
|
||||||
|
// // p.write('! U1 getvar "alerts.configured"\r\n');
|
||||||
|
// });
|
||||||
|
|
||||||
|
// p.on("close", () => {
|
||||||
|
// log.info({}, `Closed connection to ${printer.name}`);
|
||||||
|
// p.destroy();
|
||||||
|
// return;
|
||||||
|
// });
|
||||||
|
|
||||||
|
// p.on("error", (err) => {
|
||||||
|
// log.info(
|
||||||
|
// { stack: err },
|
||||||
|
// `${printer.name} encountered an error while trying to update`,
|
||||||
|
// );
|
||||||
|
// return;
|
||||||
|
// });
|
||||||
|
// };
|
||||||
|
|||||||
38
backend/ocp/ocp.printer.update.ts
Normal file
38
backend/ocp/ocp.printer.update.ts
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
/**
|
||||||
|
* the route that listens for the printers post.
|
||||||
|
*
|
||||||
|
* and http-post alert should be setup on each printer pointing to at min you will want to make the alert for
|
||||||
|
* pause printer, you can have all on here as it will also monitor and do things on all messages
|
||||||
|
*
|
||||||
|
* http://{serverIP}:2222/lst/api/ocp/printer/listener/{printerName}
|
||||||
|
*
|
||||||
|
* the messages will be sent over to the db for logging as well as specific ones will do something
|
||||||
|
*
|
||||||
|
* pause will validate if can print
|
||||||
|
* close head will repause the printer so it wont print a label
|
||||||
|
* power up will just repause the printer so it wont print a label
|
||||||
|
*/
|
||||||
|
|
||||||
|
import { Router } from "express";
|
||||||
|
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
//import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
import { printerSync } from "./ocp.printer.manage.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/printer/update", async (_, res) => {
|
||||||
|
printerSync();
|
||||||
|
return apiReturn(res, {
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "ocp",
|
||||||
|
subModule: "printing",
|
||||||
|
message:
|
||||||
|
"Printer update has been triggered to monitor progress please head to the logs.",
|
||||||
|
data: [],
|
||||||
|
status: 200,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
export default r;
|
||||||
@@ -2,6 +2,7 @@ import { type Express, Router } from "express";
|
|||||||
import { requireAuth } from "../middleware/auth.middleware.js";
|
import { requireAuth } from "../middleware/auth.middleware.js";
|
||||||
import { featureCheck } from "../middleware/featureActive.middleware.js";
|
import { featureCheck } from "../middleware/featureActive.middleware.js";
|
||||||
import listener from "./ocp.printer.listener.js";
|
import listener from "./ocp.printer.listener.js";
|
||||||
|
import update from "./ocp.printer.update.js";
|
||||||
|
|
||||||
export const setupOCPRoutes = (baseUrl: string, app: Express) => {
|
export const setupOCPRoutes = (baseUrl: string, app: Express) => {
|
||||||
//setup all the routes
|
//setup all the routes
|
||||||
@@ -16,6 +17,7 @@ export const setupOCPRoutes = (baseUrl: string, app: Express) => {
|
|||||||
// auth routes below here
|
// auth routes below here
|
||||||
router.use(requireAuth);
|
router.use(requireAuth);
|
||||||
|
|
||||||
|
router.use(update);
|
||||||
//router.use("");
|
//router.use("");
|
||||||
|
|
||||||
app.use(`${baseUrl}/api/ocp`, router);
|
app.use(`${baseUrl}/api/ocp`, router);
|
||||||
|
|||||||
@@ -7,12 +7,17 @@ import { returnFunc } from "../utils/returnHelper.utils.js";
|
|||||||
export let pool: sql.ConnectionPool;
|
export let pool: sql.ConnectionPool;
|
||||||
export let connected: boolean = false;
|
export let connected: boolean = false;
|
||||||
export let reconnecting = false;
|
export let reconnecting = false;
|
||||||
|
// start the delay out as 2 seconds
|
||||||
|
let delayStart = 2000;
|
||||||
|
let attempt = 0;
|
||||||
|
const maxAttempts = 10;
|
||||||
|
|
||||||
export const connectProdSql = async () => {
|
export const connectProdSql = async () => {
|
||||||
const serverUp = await checkHostnamePort(`${process.env.PROD_SERVER}:1433`);
|
const serverUp = await checkHostnamePort(`${process.env.PROD_SERVER}:1433`);
|
||||||
if (!serverUp) {
|
if (!serverUp) {
|
||||||
// we will try to reconnect
|
// we will try to reconnect
|
||||||
connected = false;
|
connected = false;
|
||||||
|
reconnectToSql();
|
||||||
return returnFunc({
|
return returnFunc({
|
||||||
success: false,
|
success: false,
|
||||||
level: "error",
|
level: "error",
|
||||||
@@ -48,6 +53,7 @@ export const connectProdSql = async () => {
|
|||||||
notify: false,
|
notify: false,
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
reconnectToSql();
|
||||||
return returnFunc({
|
return returnFunc({
|
||||||
success: false,
|
success: false,
|
||||||
level: "error",
|
level: "error",
|
||||||
@@ -104,11 +110,6 @@ export const reconnectToSql = async () => {
|
|||||||
//set reconnecting to true while we try to reconnect
|
//set reconnecting to true while we try to reconnect
|
||||||
reconnecting = true;
|
reconnecting = true;
|
||||||
|
|
||||||
// start the delay out as 2 seconds
|
|
||||||
let delayStart = 2000;
|
|
||||||
let attempt = 0;
|
|
||||||
const maxAttempts = 10;
|
|
||||||
|
|
||||||
while (!connected && attempt < maxAttempts) {
|
while (!connected && attempt < maxAttempts) {
|
||||||
attempt++;
|
attempt++;
|
||||||
log.info(
|
log.info(
|
||||||
@@ -121,7 +122,7 @@ export const reconnectToSql = async () => {
|
|||||||
|
|
||||||
if (!serverUp) {
|
if (!serverUp) {
|
||||||
delayStart = Math.min(delayStart * 2, 30000); // exponential backoff until up to 30000
|
delayStart = Math.min(delayStart * 2, 30000); // exponential backoff until up to 30000
|
||||||
return;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
@@ -133,19 +134,12 @@ export const reconnectToSql = async () => {
|
|||||||
);
|
);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
delayStart = Math.min(delayStart * 2, 30000);
|
delayStart = Math.min(delayStart * 2, 30000);
|
||||||
return returnFunc({
|
delayStart = Math.min(delayStart * 2, 30000);
|
||||||
success: false,
|
log.error({ error }, "Failed to reconnect to the prod sql server.");
|
||||||
level: "error",
|
|
||||||
module: "system",
|
|
||||||
subModule: "db",
|
|
||||||
message: "Failed to reconnect to the prod sql server.",
|
|
||||||
data: [error],
|
|
||||||
notify: false,
|
|
||||||
});
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!connected) {
|
if (!connected && attempt >= maxAttempts) {
|
||||||
log.error(
|
log.error(
|
||||||
{ notify: true },
|
{ notify: true },
|
||||||
"Max reconnect attempts reached on the prodSql server. Stopping retries.",
|
"Max reconnect attempts reached on the prodSql server. Stopping retries.",
|
||||||
|
|||||||
@@ -1,10 +1,5 @@
|
|||||||
import { returnFunc } from "../utils/returnHelper.utils.js";
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
import {
|
import { connected, pool } from "./prodSqlConnection.controller.js";
|
||||||
connected,
|
|
||||||
pool,
|
|
||||||
reconnecting,
|
|
||||||
reconnectToSql,
|
|
||||||
} from "./prodSqlConnection.controller.js";
|
|
||||||
|
|
||||||
interface SqlError extends Error {
|
interface SqlError extends Error {
|
||||||
code?: string;
|
code?: string;
|
||||||
@@ -22,29 +17,15 @@ interface SqlError extends Error {
|
|||||||
*/
|
*/
|
||||||
export const prodQuery = async (queryToRun: string, name: string) => {
|
export const prodQuery = async (queryToRun: string, name: string) => {
|
||||||
if (!connected) {
|
if (!connected) {
|
||||||
reconnectToSql();
|
|
||||||
|
|
||||||
if (reconnecting) {
|
|
||||||
return returnFunc({
|
return returnFunc({
|
||||||
success: false,
|
success: false,
|
||||||
level: "error",
|
level: "error",
|
||||||
module: "system",
|
module: "system",
|
||||||
subModule: "prodSql",
|
subModule: "prodSql",
|
||||||
message: `The sql ${process.env.PROD_PLANT_TOKEN} is trying to reconnect already`,
|
message: `${process.env.PROD_PLANT_TOKEN} is offline or attempting to reconnect`,
|
||||||
data: [],
|
data: [],
|
||||||
notify: false,
|
notify: false,
|
||||||
});
|
});
|
||||||
} else {
|
|
||||||
return returnFunc({
|
|
||||||
success: false,
|
|
||||||
level: "error",
|
|
||||||
module: "system",
|
|
||||||
subModule: "prodSql",
|
|
||||||
message: `${process.env.PROD_PLANT_TOKEN} is not connected, and failed to connect.`,
|
|
||||||
data: [],
|
|
||||||
notify: true,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
//change to the correct server
|
//change to the correct server
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
use AlplaPROD_test1
|
use AlplaPROD_test1
|
||||||
|
|
||||||
SELECT V_Artikel.IdArtikelvarianten,
|
SELECT V_Artikel.IdArtikelvarianten as article,
|
||||||
V_Artikel.Bezeichnung,
|
V_Artikel.Bezeichnung,
|
||||||
V_Artikel.ArtikelvariantenTypBez,
|
V_Artikel.ArtikelvariantenTypBez,
|
||||||
V_Artikel.PreisEinheitBez,
|
V_Artikel.PreisEinheitBez,
|
||||||
43
backend/prodSql/queries/datamart.activeArticles2.sql
Normal file
43
backend/prodSql/queries/datamart.activeArticles2.sql
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
/**
|
||||||
|
This will be replacing activeArticles once all data is remapped into this query.
|
||||||
|
make a note in the docs this activeArticles will go stale sooner or later.
|
||||||
|
**/
|
||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
select a.Id,
|
||||||
|
a.HumanReadableId as av,
|
||||||
|
a.Alias as alias,
|
||||||
|
p.LoadingUnitsPerTruck as loadingUnitsPerTruck,
|
||||||
|
p.LoadingUnitsPerTruck * p.LoadingUnitPieces as qtyPerTruck,
|
||||||
|
p.LoadingUnitPieces,
|
||||||
|
case when i.MinQuantity IS NOT NULL then round(cast(i.MinQuantity as float), 2) else 0 end as min,
|
||||||
|
case when i.MaxQuantity IS NOT NULL then round(cast(i.MaxQuantity as float),2) else 0 end as max
|
||||||
|
from masterData.Article (nolock) as a
|
||||||
|
|
||||||
|
/* sales price */
|
||||||
|
left join
|
||||||
|
(select *
|
||||||
|
from (select
|
||||||
|
id,
|
||||||
|
PackagingId,
|
||||||
|
ArticleId,
|
||||||
|
DefaultCustomer,
|
||||||
|
ROW_NUMBER() OVER (PARTITION BY ArticleId ORDER BY ValidAfter DESC) AS RowNum
|
||||||
|
from masterData.SalesPrice (nolock)
|
||||||
|
where DefaultCustomer = 1) as x
|
||||||
|
where RowNum = 1
|
||||||
|
) as s
|
||||||
|
on a.id = s.ArticleId
|
||||||
|
|
||||||
|
/* pkg instructions */
|
||||||
|
left join
|
||||||
|
masterData.PackagingInstruction (nolock) as p
|
||||||
|
on s.PackagingId = p.id
|
||||||
|
|
||||||
|
/* stock limits */
|
||||||
|
left join
|
||||||
|
masterData.StockLimit (nolock) as i
|
||||||
|
on a.id = i.ArticleId
|
||||||
|
|
||||||
|
where a.active = 1
|
||||||
|
and a.HumanReadableId in ([articles])
|
||||||
45
backend/prodSql/queries/datamart.customerInventory.sql
Normal file
45
backend/prodSql/queries/datamart.customerInventory.sql
Normal file
@@ -0,0 +1,45 @@
|
|||||||
|
select x.idartikelVarianten as av
|
||||||
|
,ArtikelVariantenAlias as Alias
|
||||||
|
--x.Lfdnr as RunningNumber,
|
||||||
|
--,round(sum(EinlagerungsMengeVPKSum),0) as Total_Pallets
|
||||||
|
--,sum(EinlagerungsMengeSum) as Total_PalletQTY
|
||||||
|
,round(sum(VerfuegbareMengeVPKSum),0) as Avalible_Pallets
|
||||||
|
,sum(VerfuegbareMengeSum) as Avaliable_PalletQTY
|
||||||
|
,sum(case when c.Description LIKE '%COA%' then GesperrteMengeVPKSum else 0 end) as COA_Pallets
|
||||||
|
,sum(case when c.Description LIKE '%COA%' then GesperrteMengeSum else 0 end) as COA_QTY
|
||||||
|
--,sum(case when c.Description NOT LIKE '%COA%' then GesperrteMengeVPKSum else 0 end) as Held_Pallets
|
||||||
|
--,sum(case when c.Description NOT LIKE '%COA%' then GesperrteMengeSum else 0 end) as Held_QTY
|
||||||
|
,IdProdPlanung as Lot
|
||||||
|
--,IdAdressen
|
||||||
|
--,x.AdressBez
|
||||||
|
--,*
|
||||||
|
from [AlplaPROD_test1].dbo.[V_LagerPositionenBarcodes] (nolock) x
|
||||||
|
|
||||||
|
left join
|
||||||
|
[AlplaPROD_test1].dbo.T_EtikettenGedruckt (nolock) on
|
||||||
|
x.Lfdnr = T_EtikettenGedruckt.Lfdnr AND T_EtikettenGedruckt.Lfdnr > 1
|
||||||
|
|
||||||
|
left join
|
||||||
|
|
||||||
|
(SELECT *
|
||||||
|
FROM [AlplaPROD_test1].[dbo].[T_BlockingDefects] (nolock) where Active = 1) as c
|
||||||
|
on x.IdMainDefect = c.IdBlockingDefect
|
||||||
|
/*
|
||||||
|
The data below will be controlled by the user in excell by default everything will be passed over
|
||||||
|
IdAdressen = 3
|
||||||
|
*/
|
||||||
|
where
|
||||||
|
--IdArtikelTyp = 1
|
||||||
|
x.IdWarenlager not in (6, 1)
|
||||||
|
--and IdAdressen
|
||||||
|
--and x.IdWarenlager in (0)
|
||||||
|
|
||||||
|
|
||||||
|
group by x.IdArtikelVarianten
|
||||||
|
,ArtikelVariantenAlias
|
||||||
|
,IdProdPlanung
|
||||||
|
--,c.Description
|
||||||
|
,IdAdressen
|
||||||
|
,x.AdressBez
|
||||||
|
--, x.Lfdnr
|
||||||
|
order by x.IdArtikelVarianten
|
||||||
74
backend/prodSql/queries/datamart.deliveryByDateRange.sql
Normal file
74
backend/prodSql/queries/datamart.deliveryByDateRange.sql
Normal file
@@ -0,0 +1,74 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
DECLARE @StartDate DATE = '[startDate]' -- 2025-1-1
|
||||||
|
DECLARE @EndDate DATE = '[endDate]' -- 2025-1-31
|
||||||
|
SELECT
|
||||||
|
r.[ArticleHumanReadableId]
|
||||||
|
,[ReleaseNumber]
|
||||||
|
,h.CustomerOrderNumber
|
||||||
|
,x.CustomerLineItemNumber
|
||||||
|
,[CustomerReleaseNumber]
|
||||||
|
,[ReleaseState]
|
||||||
|
,[DeliveryState]
|
||||||
|
,ea.JournalNummer as BOL_Number
|
||||||
|
,[ReleaseConfirmationState]
|
||||||
|
,[PlanningState]
|
||||||
|
,format(r.[OrderDate], 'yyyy-MM-dd HH:mm') as OrderDate
|
||||||
|
--,r.[OrderDate]
|
||||||
|
,FORMAT(r.[DeliveryDate], 'yyyy-MM-dd HH:mm') as DeliveryDate
|
||||||
|
--,r.[DeliveryDate]
|
||||||
|
,FORMAT(r.[LoadingDate], 'yyyy-MM-dd HH:mm') as LoadingDate
|
||||||
|
--,r.[LoadingDate]
|
||||||
|
,[Quantity]
|
||||||
|
,[DeliveredQuantity]
|
||||||
|
,r.[AdditionalInformation1]
|
||||||
|
,r.[AdditionalInformation2]
|
||||||
|
,[TradeUnits]
|
||||||
|
,[LoadingUnits]
|
||||||
|
,[Trucks]
|
||||||
|
,[LoadingToleranceType]
|
||||||
|
,[SalesPrice]
|
||||||
|
,[Currency]
|
||||||
|
,[QuantityUnit]
|
||||||
|
,[SalesPriceRemark]
|
||||||
|
,r.[Remark]
|
||||||
|
,[Irradiated]
|
||||||
|
,r.[CreatedByEdi]
|
||||||
|
,[DeliveryAddressHumanReadableId]
|
||||||
|
,DeliveryAddressDescription
|
||||||
|
,[CustomerArtNo]
|
||||||
|
,[TotalPrice]
|
||||||
|
,r.[ArticleAlias]
|
||||||
|
|
||||||
|
FROM [order].[Release] (nolock) as r
|
||||||
|
|
||||||
|
left join
|
||||||
|
[order].LineItem as x on
|
||||||
|
|
||||||
|
r.LineItemId = x.id
|
||||||
|
|
||||||
|
left join
|
||||||
|
[order].Header as h on
|
||||||
|
x.HeaderId = h.id
|
||||||
|
|
||||||
|
--bol stuff
|
||||||
|
left join
|
||||||
|
AlplaPROD_test1.dbo.V_LadePlanungenLadeAuftragAbruf (nolock) as zz
|
||||||
|
on zz.AbrufIdAuftragsAbruf = r.ReleaseNumber
|
||||||
|
|
||||||
|
left join
|
||||||
|
(select * from (SELECT
|
||||||
|
ROW_NUMBER() OVER (PARTITION BY IdJournal ORDER BY add_date DESC) AS RowNum
|
||||||
|
,*
|
||||||
|
FROM [AlplaPROD_test1].[dbo].[T_Lieferungen] (nolock)) x
|
||||||
|
|
||||||
|
where RowNum = 1) as ea on
|
||||||
|
zz.IdLieferschein = ea.IdJournal
|
||||||
|
|
||||||
|
where
|
||||||
|
--r.ReleaseNumber = 1452
|
||||||
|
|
||||||
|
r.DeliveryDate between @StartDate AND @EndDate
|
||||||
|
and DeliveredQuantity > 0
|
||||||
|
--and r.ArticleHumanReadableId in ([articles])
|
||||||
|
--and Journalnummer = 169386
|
||||||
29
backend/prodSql/queries/datamart.fakeEDIUpdate.sql
Normal file
29
backend/prodSql/queries/datamart.fakeEDIUpdate.sql
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
select
|
||||||
|
customerartno as CustomerArticleNumber
|
||||||
|
,h.CustomerOrderNumber as CustomerOrderNumber
|
||||||
|
,l.CustomerLineItemNumber as CustomerLineNumber
|
||||||
|
,r.CustomerReleaseNumber as CustomerRealeaseNumber
|
||||||
|
,r.Quantity
|
||||||
|
,format(r.DeliveryDate, 'MM/dd/yyyy HH:mm') as DeliveryDate
|
||||||
|
,h.CustomerHumanReadableId as CustomerID
|
||||||
|
,r.Remark
|
||||||
|
--,*
|
||||||
|
from [order].[Release] as r (nolock)
|
||||||
|
|
||||||
|
left join
|
||||||
|
[order].LineItem as l (nolock) on
|
||||||
|
l.id = r.LineItemId
|
||||||
|
|
||||||
|
left join
|
||||||
|
[order].Header as h (nolock) on
|
||||||
|
h.id = l.HeaderId
|
||||||
|
|
||||||
|
WHERE releaseState not in (1, 2, 3, 4)
|
||||||
|
AND h.CreatedByEdi = 1
|
||||||
|
AND r.deliveryDate < getdate() + 1
|
||||||
|
--AND h.CustomerHumanReadableId in (0)
|
||||||
|
|
||||||
|
|
||||||
|
order by r.deliveryDate
|
||||||
8
backend/prodSql/queries/datamart.forecast.sql
Normal file
8
backend/prodSql/queries/datamart.forecast.sql
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
SELECT format(RequirementDate, 'yyyy-MM-dd') as requirementDate
|
||||||
|
,ArticleHumanReadableId
|
||||||
|
,CustomerArticleNumber
|
||||||
|
,ArticleDescription
|
||||||
|
,Quantity
|
||||||
|
FROM [test1_AlplaPROD2.0_Read].[forecast].[Forecast]
|
||||||
|
where DeliveryAddressHumanReadableId in ([customers])
|
||||||
|
order by RequirementDate
|
||||||
58
backend/prodSql/queries/datamart.inventory.sql
Normal file
58
backend/prodSql/queries/datamart.inventory.sql
Normal file
@@ -0,0 +1,58 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
select
|
||||||
|
ArticleHumanReadableId as article
|
||||||
|
,ArticleAlias as alias
|
||||||
|
,round(sum(QuantityLoadingUnits),2) total_pallets
|
||||||
|
,round(sum(Quantity),2) as total_palletQTY
|
||||||
|
,round(sum(case when State = 0 then QuantityLoadingUnits else 0 end),2) available_Pallets
|
||||||
|
,round(sum(case when State = 0 then Quantity else 0 end),2) available_QTY
|
||||||
|
,round(sum(case when b.HumanReadableId = 864 then QuantityLoadingUnits else 0 end),2) as coa_Pallets
|
||||||
|
,round(sum(case when b.HumanReadableId = 864 then Quantity else 0 end),2) as coa_QTY
|
||||||
|
,round(sum(case when b.HumanReadableId <> 864 then QuantityLoadingUnits else 0 end),2) as held_Pallets
|
||||||
|
,round(sum(case when b.HumanReadableId <> 864 then Quantity else 0 end),2) as held_QTY
|
||||||
|
,round(sum(case when w.type = 7 then QuantityLoadingUnits else 0 end),2) as consignment_Pallets
|
||||||
|
,round(sum(case when w.type = 7 then Quantity else 0 end),2) as consignment_qty
|
||||||
|
--,l.RunningNumber
|
||||||
|
|
||||||
|
/** datamart include lot number **/
|
||||||
|
--,l.MachineLocation,l.MachineName,l.ProductionLotRunningNumber as lot
|
||||||
|
|
||||||
|
/** data mart include location data **/
|
||||||
|
--,l.WarehouseDescription,l.LaneDescription
|
||||||
|
|
||||||
|
,articleTypeName
|
||||||
|
|
||||||
|
FROM [warehousing].[WarehouseUnit] as l (nolock)
|
||||||
|
left join
|
||||||
|
(
|
||||||
|
SELECT [Id]
|
||||||
|
,[HumanReadableId]
|
||||||
|
,d.[Description]
|
||||||
|
,[DefectGroupId]
|
||||||
|
,[IsActive]
|
||||||
|
FROM [blocking].[BlockingDefect] as g (nolock)
|
||||||
|
|
||||||
|
left join
|
||||||
|
[AlplaPROD_test1].dbo.[T_BlockingDefects] as d (nolock) on
|
||||||
|
d.IdGlobalBlockingDefect = g.HumanReadableId
|
||||||
|
) as b on
|
||||||
|
b.id = l.MainDefectId
|
||||||
|
|
||||||
|
left join
|
||||||
|
[warehousing].[warehouse] as w (nolock) on
|
||||||
|
w.id = l.warehouseid
|
||||||
|
|
||||||
|
where LaneHumanReadableId not in (20000,21000)
|
||||||
|
group by ArticleHumanReadableId,
|
||||||
|
ArticleAlias,
|
||||||
|
ArticleTypeName
|
||||||
|
--,l.RunningNumber
|
||||||
|
|
||||||
|
/** datamart include lot number **/
|
||||||
|
--,l.MachineLocation,l.MachineName,l.ProductionLotRunningNumber
|
||||||
|
|
||||||
|
/** data mart include location data **/
|
||||||
|
--,l.WarehouseDescription,l.LaneDescription
|
||||||
|
|
||||||
|
order by ArticleHumanReadableId
|
||||||
48
backend/prodSql/queries/datamart.legacy.inventory.sql
Normal file
48
backend/prodSql/queries/datamart.legacy.inventory.sql
Normal file
@@ -0,0 +1,48 @@
|
|||||||
|
select
|
||||||
|
x.idartikelVarianten as article,
|
||||||
|
x.ArtikelVariantenAlias as alias
|
||||||
|
--x.Lfdnr as RunningNumber,
|
||||||
|
,round(sum(EinlagerungsMengeVPKSum),2) as total_pallets
|
||||||
|
,sum(EinlagerungsMengeSum) as total_palletQTY
|
||||||
|
,round(sum(VerfuegbareMengeVPKSum),0) as available_Pallets
|
||||||
|
,sum(VerfuegbareMengeSum) as available_QTY
|
||||||
|
,sum(case when c.Description LIKE '%COA%' then GesperrteMengeVPKSum else 0 end) as coa_Pallets
|
||||||
|
,sum(case when c.Description LIKE '%COA%' then GesperrteMengeSum else 0 end) as coa_QTY
|
||||||
|
,sum(case when c.Description NOT LIKE '%COA%' or x.IdMainDefect = -1 then GesperrteMengeVPKSum else 0 end) as held_Pallets
|
||||||
|
,sum(case when c.Description NOT LIKE '%COA%' or x.IdMainDefect = -1 then GesperrteMengeSum else 0 end) as held_QTY
|
||||||
|
,sum(case when x.WarenLagerLagerTyp = 8 then VerfuegbareMengeSum else 0 end) as consignment_qty
|
||||||
|
,IdProdPlanung as lot
|
||||||
|
----,IdAdressen,
|
||||||
|
,x.AdressBez
|
||||||
|
,x.IdLagerAbteilung as locationId
|
||||||
|
,x.LagerAbteilungKurzBez as laneDescription
|
||||||
|
,x.IdWarenlager as warehouseId
|
||||||
|
,x.WarenLagerKurzBez as warehouseDescription
|
||||||
|
--,*
|
||||||
|
from [AlplaPROD_test1].dbo.[V_LagerPositionenBarcodes] (nolock) x
|
||||||
|
|
||||||
|
left join
|
||||||
|
[AlplaPROD_test1].dbo.T_EtikettenGedruckt as l(nolock) on
|
||||||
|
x.Lfdnr = l.Lfdnr AND l.Lfdnr > 1
|
||||||
|
|
||||||
|
left join
|
||||||
|
|
||||||
|
(SELECT *
|
||||||
|
FROM [AlplaPROD_test1].[dbo].[T_BlockingDefects] where Active = 1) as c
|
||||||
|
on x.IdMainDefect = c.IdBlockingDefect
|
||||||
|
/*
|
||||||
|
The data below will be controlled by the user in excell by default everything will be passed over
|
||||||
|
IdAdressen = 3
|
||||||
|
*/
|
||||||
|
where /*IdArtikelTyp = 1 and */x.IdWarenlager not in (6, 1)
|
||||||
|
|
||||||
|
group by x.idartikelVarianten, ArtikelVariantenAlias, c.Description
|
||||||
|
--,IdAdressen
|
||||||
|
,x.AdressBez
|
||||||
|
,IdProdPlanung
|
||||||
|
,x.IdLagerAbteilung
|
||||||
|
,x.LagerAbteilungKurzBez
|
||||||
|
,x.IdWarenlager
|
||||||
|
,x.WarenLagerKurzBez
|
||||||
|
--, x.Lfdnr
|
||||||
|
order by x.IdArtikelVarianten
|
||||||
33
backend/prodSql/queries/datamart.openOrders.sql
Normal file
33
backend/prodSql/queries/datamart.openOrders.sql
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
select
|
||||||
|
customerartno
|
||||||
|
,r.ArticleHumanReadableId as article
|
||||||
|
,r.ArticleAlias as articleAlias
|
||||||
|
,ReleaseNumber
|
||||||
|
,h.CustomerOrderNumber as header
|
||||||
|
,l.CustomerLineItemNumber as lineItem
|
||||||
|
,r.CustomerReleaseNumber as releaseNumber
|
||||||
|
,r.LoadingUnits
|
||||||
|
,r.Quantity
|
||||||
|
,r.TradeUnits
|
||||||
|
,h.CustomerHumanReadableId
|
||||||
|
,r.DeliveryAddressDescription
|
||||||
|
,format(r.LoadingDate, 'MM/dd/yyyy HH:mm') as loadingDate
|
||||||
|
,format(r.DeliveryDate, 'MM/dd/yyyy HH:mm') as deliveryDate
|
||||||
|
,r.Remark
|
||||||
|
--,*
|
||||||
|
from [order].[Release] as r (nolock)
|
||||||
|
|
||||||
|
left join
|
||||||
|
[order].LineItem as l (nolock) on
|
||||||
|
l.id = r.LineItemId
|
||||||
|
|
||||||
|
left join
|
||||||
|
[order].Header as h (nolock) on
|
||||||
|
h.id = l.HeaderId
|
||||||
|
|
||||||
|
WHERE releasestate not in (1, 2, 4)
|
||||||
|
AND r.deliverydate between getDate() + -[startDay] and getdate() + [endDay]
|
||||||
|
|
||||||
|
order by r.deliverydate
|
||||||
19
backend/prodSql/queries/datamart.productionData.sql
Normal file
19
backend/prodSql/queries/datamart.productionData.sql
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Reporting]
|
||||||
|
|
||||||
|
declare @startDate nvarchar(30) = '[startDate]' --'2024-12-30'
|
||||||
|
declare @endDate nvarchar(30) = '[endDate]' --'2025-08-09'
|
||||||
|
|
||||||
|
select MachineLocation,
|
||||||
|
ArticleHumanReadableId as article,
|
||||||
|
sum(Quantity) as Produced,
|
||||||
|
count(Quantity) as palletsProdued,
|
||||||
|
FORMAT(convert(date, ProductionDay), 'M/d/yyyy') as ProductionDay,
|
||||||
|
ProductionLotHumanReadableId as productionLot
|
||||||
|
|
||||||
|
from [reporting_productionControlling].[ScannedUnit] (nolock)
|
||||||
|
|
||||||
|
where convert(date, ProductionDay) between @startDate and @endDate
|
||||||
|
and ArticleHumanReadableId in ([articles])
|
||||||
|
and BookedOut is null
|
||||||
|
|
||||||
|
group by MachineLocation, ArticleHumanReadableId,ProductionDay, ProductionLotHumanReadableId
|
||||||
@@ -1,5 +1,10 @@
|
|||||||
use [test1_AlplaPROD2.0_Read]
|
use AlplaPROD_test1
|
||||||
|
/**
|
||||||
|
|
||||||
|
move this over to the delivery date range query once we have the shift data mapped over correctly.
|
||||||
|
|
||||||
|
update the psi stuff on this as well.
|
||||||
|
**/
|
||||||
DECLARE @StartDate DATE = '[startDate]' -- 2025-1-1
|
DECLARE @StartDate DATE = '[startDate]' -- 2025-1-1
|
||||||
DECLARE @EndDate DATE = '[endDate]' -- 2025-1-31
|
DECLARE @EndDate DATE = '[endDate]' -- 2025-1-31
|
||||||
SELECT
|
SELECT
|
||||||
@@ -66,9 +71,9 @@ ROW_NUMBER() OVER (PARTITION BY IdJournal ORDER BY add_date DESC) AS RowNum
|
|||||||
zz.IdLieferschein = ea.IdJournal
|
zz.IdLieferschein = ea.IdJournal
|
||||||
|
|
||||||
where
|
where
|
||||||
--r.ArticleHumanReadableId in ([articles])
|
r.ArticleHumanReadableId in ([articles])
|
||||||
--r.ReleaseNumber = 1452
|
--r.ReleaseNumber = 1452
|
||||||
|
|
||||||
r.DeliveryDate between @StartDate AND @EndDate
|
and r.DeliveryDate between @StartDate AND @EndDate
|
||||||
and DeliveredQuantity > 0
|
--and DeliveredQuantity > 0
|
||||||
--and Journalnummer = 169386
|
--and Journalnummer = 169386
|
||||||
32
backend/prodSql/queries/datamart.psiPlanningData.sql
Normal file
32
backend/prodSql/queries/datamart.psiPlanningData.sql
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
use AlplaPROD_test1
|
||||||
|
declare @start_date nvarchar(30) = '[startDate]' --'2025-01-01'
|
||||||
|
declare @end_date nvarchar(30) = '[endDate]' --'2025-08-09'
|
||||||
|
/*
|
||||||
|
articles will need to be passed over as well as the date structure we want to see
|
||||||
|
*/
|
||||||
|
|
||||||
|
select x.IdArtikelvarianten As Article,
|
||||||
|
ProduktionAlias as Description,
|
||||||
|
standort as MachineId,
|
||||||
|
MaschinenBezeichnung as MachineName,
|
||||||
|
--MaschZyklus as PlanningCycleTime,
|
||||||
|
x.IdProdPlanung as LotNumber,
|
||||||
|
FORMAT(ProdTag, 'MM/dd/yyyy') as ProductionDay,
|
||||||
|
x.planMenge as TotalPlanned,
|
||||||
|
ProduktionMenge as QTYPerDay,
|
||||||
|
round(ProduktionMengeVPK, 2) PalDay,
|
||||||
|
Status as finished
|
||||||
|
--MaschStdAuslastung as nee
|
||||||
|
|
||||||
|
from dbo.V_ProdLosProduktionJeProdTag_PLANNING (nolock) as x
|
||||||
|
|
||||||
|
left join
|
||||||
|
dbo.V_ProdPlanung (nolock) as p on
|
||||||
|
x.IdProdPlanung = p.IdProdPlanung
|
||||||
|
|
||||||
|
where ProdTag between @start_date and @end_date
|
||||||
|
and p.IdArtikelvarianten in ([articles])
|
||||||
|
--and V_ProdLosProduktionJeProdTag_PLANNING.IdKunde = 10
|
||||||
|
--and IdProdPlanung = 18442
|
||||||
|
|
||||||
|
order by ProdTag desc
|
||||||
11
backend/prodSql/queries/featureCheck.sql
Normal file
11
backend/prodSql/queries/featureCheck.sql
Normal file
@@ -0,0 +1,11 @@
|
|||||||
|
SELECT count(*) as activated
|
||||||
|
FROM [test1_AlplaPROD2.0_Read].[support].[FeatureActivation]
|
||||||
|
|
||||||
|
where feature in (108,7)
|
||||||
|
|
||||||
|
|
||||||
|
/*
|
||||||
|
as more features get activated and need to have this checked to include the new endpoints add here so we can check this.
|
||||||
|
108 = waste
|
||||||
|
7 = warehousing
|
||||||
|
*/
|
||||||
44
backend/prodSql/queries/qualityBlocking.sql
Normal file
44
backend/prodSql/queries/qualityBlocking.sql
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
use [test1_AlplaPROD2.0_Read]
|
||||||
|
|
||||||
|
SELECT
|
||||||
|
'Alert! new blocking order: #' + cast(bo.HumanReadableId as varchar) + ' - ' + bo.ArticleVariantDescription as subject
|
||||||
|
,cast(bo.[HumanReadableId] as varchar) as blockingNumber
|
||||||
|
,bo.[ArticleVariantDescription] as article
|
||||||
|
,cast(bo.[CustomerHumanReadableId] as varchar) + ' - ' + bo.[CustomerDescription] as customer
|
||||||
|
,convert(varchar(10), bo.[BlockingDate], 101) + ' ' + convert(varchar(5), bo.[BlockingDate], 108) as blockingDate
|
||||||
|
,cast(ArticleVariantHumanReadableId as varchar) + ' - ' + ArticleVariantDescription as av
|
||||||
|
,case when bo.Remark = '' or bo.Remark is NULL then 'Please reach out to quality for the reason this was placed on hold as a remark was not entered during the blocking processs' else bo.Remark end as remark
|
||||||
|
,cast(FORMAT(TotalAmountOfPieces, '###,###') as varchar) + ' / ' + cast(LoadingUnit as varchar) as peicesAndLoadingUnits
|
||||||
|
,bo.ProductionLotHumanReadableId as lotNumber
|
||||||
|
,cast(osd.IdBlockingDefectsGroup as varchar) + ' - ' + osd.Description as mainDefectGroup
|
||||||
|
,cast(df.HumanReadableId as varchar) + ' - ' + os.Description as mainDefect
|
||||||
|
,lot.MachineLocation as line
|
||||||
|
--,*
|
||||||
|
FROM [blocking].[BlockingOrder] (nolock) as bo
|
||||||
|
|
||||||
|
|
||||||
|
/*** get the defect details ***/
|
||||||
|
join
|
||||||
|
[blocking].[BlockingDefect] (nolock) AS df
|
||||||
|
on df.id = bo.MainDefectId
|
||||||
|
|
||||||
|
/*** pull description from 1.0 ***/
|
||||||
|
left join
|
||||||
|
[AlplaPROD_test1].[dbo].[T_BlockingDefects] (nolock) as os
|
||||||
|
on os.IdGlobalBlockingDefect = df.HumanReadableId
|
||||||
|
|
||||||
|
/*** join in 1.0 defect group ***/
|
||||||
|
left join
|
||||||
|
[AlplaPROD_test1].[dbo].[T_BlockingDefectsGroups] (nolock) as osd
|
||||||
|
on osd.IdBlockingDefectsGroup = os.IdBlockingDefectsGroup
|
||||||
|
|
||||||
|
left join
|
||||||
|
[productionControlling].[ProducedLot] (nolock) as lot
|
||||||
|
on lot.id = bo.ProductionLotId
|
||||||
|
|
||||||
|
|
||||||
|
where
|
||||||
|
bo.[BlockingDate] between getdate() - 2 and getdate() + 3 and
|
||||||
|
bo.BlockingTrigger = 1 -- so we only get the ir blocking and not coa
|
||||||
|
--and HumanReadableId NOT IN ([sentBlockingOrders])
|
||||||
|
and bo.HumanReadableId > [lastBlocking]
|
||||||
4
backend/prodSql/queries/shiftChange.sql
Normal file
4
backend/prodSql/queries/shiftChange.sql
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
select top(1) convert(varchar(8) ,
|
||||||
|
convert(time,startdate), 108) as shiftChange
|
||||||
|
from [test1_AlplaPROD2.0_Read].[masterData].[ShiftDefinition]
|
||||||
|
where teamNumber = 1
|
||||||
125
backend/purchase/puchase.gpCheck.ts
Normal file
125
backend/purchase/puchase.gpCheck.ts
Normal file
@@ -0,0 +1,125 @@
|
|||||||
|
import { gpQuery } from "../gpSql/gpSqlQuery.controller.js";
|
||||||
|
import {
|
||||||
|
type SqlGPQuery,
|
||||||
|
sqlGpQuerySelector,
|
||||||
|
} from "../gpSql/gpSqlQuerySelector.utils.js";
|
||||||
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import type { GpStatus } from "../types/purhcaseTypes.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
|
||||||
|
const log = createLogger({ module: "purchase", subModule: "gp" });
|
||||||
|
|
||||||
|
export const gpReqCheck = async (data: GpStatus[]) => {
|
||||||
|
const gpReqCheck = sqlGpQuerySelector("reqCheck") as SqlGPQuery;
|
||||||
|
const reqs = data.map((r) => r.req.trim());
|
||||||
|
|
||||||
|
if (!gpReqCheck.success) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "purchase",
|
||||||
|
subModule: "query",
|
||||||
|
message: `Error getting alpla purchase info`,
|
||||||
|
data: gpReqCheck.message as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
// check the initial req table
|
||||||
|
const result = await gpQuery(
|
||||||
|
gpReqCheck.query.replace(
|
||||||
|
"[reqsToCheck]",
|
||||||
|
data.map((r) => `'${r.req}'`).join(", ") ?? "xo",
|
||||||
|
),
|
||||||
|
"Get req info",
|
||||||
|
);
|
||||||
|
|
||||||
|
log.debug(
|
||||||
|
{},
|
||||||
|
`There are ${result.data.length} reqs that need to be updated with there current status`,
|
||||||
|
);
|
||||||
|
|
||||||
|
const firstFound = result.data.map((r) => ({
|
||||||
|
req: r.req.trim(),
|
||||||
|
approvedStatus: r.approvedStatus,
|
||||||
|
}));
|
||||||
|
|
||||||
|
const firstFoundSet = new Set(result.data.map((r) => r.req.trim()));
|
||||||
|
|
||||||
|
const missing1Reqs = reqs.filter((req) => !firstFoundSet.has(req));
|
||||||
|
|
||||||
|
//check if we have a recall on our req
|
||||||
|
const reqCheck = await gpQuery(
|
||||||
|
`select
|
||||||
|
[Requisition Number] as req
|
||||||
|
,case when [Workflow Status] = 'recall' then 'returned' else [Workflow Status] end as approvedStatus
|
||||||
|
--,*
|
||||||
|
from [dbo].[PurchaseRequisitions] where [Requisition Number] in (${missing1Reqs.map((r) => `'${r}'`).join(", ") ?? "xo"})`,
|
||||||
|
"validate req is not in recall",
|
||||||
|
);
|
||||||
|
|
||||||
|
const secondFound = reqCheck.data.map((r) => ({
|
||||||
|
req: r.req.trim(),
|
||||||
|
approvedStatus: r.approvedStatus,
|
||||||
|
}));
|
||||||
|
|
||||||
|
const secondFoundSet =
|
||||||
|
new Set(reqCheck.data.map((r) => r.req.trim())) ?? [];
|
||||||
|
|
||||||
|
const missing2Reqs = missing1Reqs.filter((req) => !secondFoundSet.has(req));
|
||||||
|
|
||||||
|
// check if we have a po already
|
||||||
|
const apoCheck = await gpQuery(
|
||||||
|
`select
|
||||||
|
SOPNUMBE
|
||||||
|
,PONUMBER
|
||||||
|
,reqStatus='converted'
|
||||||
|
,*
|
||||||
|
from alpla.dbo.sop60100 (nolock) where sopnumbe in (${missing2Reqs.map((r) => `'${r}'`).join(", ") ?? "xo"})`,
|
||||||
|
"Get release info",
|
||||||
|
);
|
||||||
|
|
||||||
|
const thirdRound = apoCheck.data.map((r) => ({
|
||||||
|
req: r.req.trim(),
|
||||||
|
approvedStatus: r.approvedStatus,
|
||||||
|
}));
|
||||||
|
|
||||||
|
const missing3Reqs = missing2Reqs.filter((req) => !secondFoundSet.has(req));
|
||||||
|
|
||||||
|
// remaining just got canceled or no longer exist
|
||||||
|
const remaining = missing3Reqs.map((m) => ({
|
||||||
|
req: m,
|
||||||
|
approvedStatus: "canceled",
|
||||||
|
}));
|
||||||
|
|
||||||
|
const allFound = [
|
||||||
|
...firstFound,
|
||||||
|
...secondFound,
|
||||||
|
...thirdRound,
|
||||||
|
...remaining,
|
||||||
|
];
|
||||||
|
|
||||||
|
const statusMap = new Map(
|
||||||
|
allFound.map((r: any) => [r.req, r.approvedStatus]),
|
||||||
|
);
|
||||||
|
|
||||||
|
const updateData = data.map((row) => ({
|
||||||
|
id: row.id,
|
||||||
|
//req: row.req,
|
||||||
|
approvedStatus: statusMap.get(row.req.trim()) ?? null,
|
||||||
|
}));
|
||||||
|
|
||||||
|
return updateData;
|
||||||
|
} catch (error: any) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "purchase",
|
||||||
|
subModule: "gpChecks",
|
||||||
|
message: error.message,
|
||||||
|
data: error.stack as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
@@ -2,7 +2,7 @@
|
|||||||
* This will monitor alpla purchase
|
* This will monitor alpla purchase
|
||||||
*/
|
*/
|
||||||
|
|
||||||
import { eq } from "drizzle-orm";
|
import { eq, sql } from "drizzle-orm";
|
||||||
import { db } from "../db/db.controller.js";
|
import { db } from "../db/db.controller.js";
|
||||||
import {
|
import {
|
||||||
alplaPurchaseHistory,
|
alplaPurchaseHistory,
|
||||||
@@ -15,10 +15,12 @@ import {
|
|||||||
type SqlQuery,
|
type SqlQuery,
|
||||||
sqlQuerySelector,
|
sqlQuerySelector,
|
||||||
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import type { GpStatus, StatusUpdate } from "../types/purhcaseTypes.js";
|
||||||
import { createCronJob } from "../utils/croner.utils.js";
|
import { createCronJob } from "../utils/croner.utils.js";
|
||||||
import { delay } from "../utils/delay.utils.js";
|
import { delay } from "../utils/delay.utils.js";
|
||||||
import { returnFunc } from "../utils/returnHelper.utils.js";
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
import { tryCatch } from "../utils/trycatch.utils.js";
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
import { gpReqCheck } from "./puchase.gpCheck.js";
|
||||||
|
|
||||||
const log = createLogger({ module: "purchase", subModule: "purchaseMonitor" });
|
const log = createLogger({ module: "purchase", subModule: "purchaseMonitor" });
|
||||||
|
|
||||||
@@ -37,13 +39,13 @@ export const monitorAlplaPurchase = async () => {
|
|||||||
module: "purchase",
|
module: "purchase",
|
||||||
subModule: "query",
|
subModule: "query",
|
||||||
message: `Error getting alpla purchase info`,
|
message: `Error getting alpla purchase info`,
|
||||||
data: [sqlQuery.message],
|
data: sqlQuery.message as any,
|
||||||
notify: false,
|
notify: true,
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
if (purchaseMonitor[0]?.active) {
|
if (purchaseMonitor[0]?.active) {
|
||||||
createCronJob("purchaseMonitor", "0 */5 * * * *", async () => {
|
createCronJob("purchaseMonitor", "0 5 * * * *", async () => {
|
||||||
try {
|
try {
|
||||||
const result = await prodQuery(
|
const result = await prodQuery(
|
||||||
sqlQuery.query.replace(
|
sqlQuery.query.replace(
|
||||||
@@ -76,7 +78,7 @@ export const monitorAlplaPurchase = async () => {
|
|||||||
|
|
||||||
if (error) {
|
if (error) {
|
||||||
log.error(
|
log.error(
|
||||||
{ error },
|
{ error, notify: true },
|
||||||
"There was an error adding alpla purchase history",
|
"There was an error adding alpla purchase history",
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
@@ -84,14 +86,147 @@ export const monitorAlplaPurchase = async () => {
|
|||||||
await delay(500);
|
await delay(500);
|
||||||
}
|
}
|
||||||
} catch (e) {
|
} catch (e) {
|
||||||
console.error(
|
log.error(
|
||||||
{ error: e },
|
{ error: e, notify: true },
|
||||||
"Error occurred while running the monitor job",
|
"Error occurred while running the monitor job",
|
||||||
);
|
);
|
||||||
log.error({ error: e }, "Error occurred while running the monitor job");
|
|
||||||
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// re-pull re-pull everything that has approvedStatus is pending
|
||||||
|
|
||||||
|
const { data: allReq, error: errorReq } = await tryCatch(
|
||||||
|
db
|
||||||
|
.select()
|
||||||
|
.from(alplaPurchaseHistory)
|
||||||
|
.where(eq(alplaPurchaseHistory.approvedStatus, "new")),
|
||||||
|
);
|
||||||
|
|
||||||
|
// if theres no reqs just end meow
|
||||||
|
if (errorReq) {
|
||||||
|
log.error(
|
||||||
|
{ stack: errorReq, notify: true },
|
||||||
|
"There was an error getting history data",
|
||||||
|
);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
log.debug({}, `There are ${allReq.length} pending reqs to be updated`);
|
||||||
|
|
||||||
|
if (!allReq.length) {
|
||||||
|
log.debug({}, "There are not reqs to be processed");
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* approvedStatus
|
||||||
|
* remark = '' then pending req/manual po
|
||||||
|
* pending = pending
|
||||||
|
* approved = approved
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
|
||||||
|
// the flow for all the fun stuff
|
||||||
|
|
||||||
|
const needsGpLookup: GpStatus[] = [];
|
||||||
|
const updates: StatusUpdate[] = [];
|
||||||
|
|
||||||
|
for (const row of allReq ?? []) {
|
||||||
|
const remark = row.remark?.toLowerCase() ?? "";
|
||||||
|
|
||||||
|
if (remark === "") {
|
||||||
|
updates.push({ id: row.id, approvedStatus: "initial" });
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (remark.includes("rct")) {
|
||||||
|
updates.push({ id: row.id, approvedStatus: "received" });
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (remark.includes("apo")) {
|
||||||
|
updates.push({ id: row.id, approvedStatus: "approved" });
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
// not handled locally, defer to GP lookup
|
||||||
|
needsGpLookup.push({ id: row.id, req: row.remark?.trim() ?? "" });
|
||||||
|
}
|
||||||
|
|
||||||
|
const gpSmash = (await gpReqCheck(needsGpLookup)) as StatusUpdate[];
|
||||||
|
|
||||||
|
const merge = [...updates, ...gpSmash];
|
||||||
|
|
||||||
|
if (merge.length > 0) {
|
||||||
|
await db.execute(sql`
|
||||||
|
UPDATE ${alplaPurchaseHistory}
|
||||||
|
SET approved_status = CASE
|
||||||
|
${sql.join(
|
||||||
|
merge.map(
|
||||||
|
(row) =>
|
||||||
|
sql`WHEN ${alplaPurchaseHistory.id} = ${row.id} THEN ${row.approvedStatus}`,
|
||||||
|
),
|
||||||
|
sql` `,
|
||||||
|
)}
|
||||||
|
ELSE approved_status
|
||||||
|
END,
|
||||||
|
updated_at = NOW()
|
||||||
|
WHERE ${alplaPurchaseHistory.id} IN (
|
||||||
|
${sql.join(
|
||||||
|
merge.map((row) => sql`${row.id}`),
|
||||||
|
sql`, `,
|
||||||
|
)}
|
||||||
|
)
|
||||||
|
`);
|
||||||
|
log.info(
|
||||||
|
{},
|
||||||
|
"All alpla purchase orders have been processed and updated",
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
// for reqs, create a string of reqs then run them through the gp req table to see there status. then update in lst ass see fit.
|
||||||
|
|
||||||
|
// then double check if we have all reqs covered, for the reqs missing from above restring them and check the po table
|
||||||
|
|
||||||
|
// these ones will be called to as converted to po
|
||||||
|
|
||||||
|
// for the remaining reqs from above check the actual req table to see the status of it if the workflow is set at Recall this means a change was requested from purchasing team and needs to be re approved
|
||||||
|
|
||||||
|
// for all remaining reqs we change them to replace/canceled
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
// const updates = (allReq ?? [])
|
||||||
|
// .map((row) => {
|
||||||
|
// const remark = row.remark?.toLowerCase() ?? "";
|
||||||
|
|
||||||
|
// let approvedStatus: string | null = null;
|
||||||
|
|
||||||
|
// // priority order matters here
|
||||||
|
// if (remark === "") {
|
||||||
|
// approvedStatus = "initial";
|
||||||
|
// } else if (remark.includes("rct")) {
|
||||||
|
// approvedStatus = "received";
|
||||||
|
// } else if (remark.includes("apo")) {
|
||||||
|
// approvedStatus = "approved";
|
||||||
|
// }
|
||||||
|
|
||||||
|
// // add your next 4 checks here
|
||||||
|
// // else if (...) approvedStatus = "somethingElse";
|
||||||
|
|
||||||
|
// if (!approvedStatus) return null;
|
||||||
|
|
||||||
|
// return {
|
||||||
|
// id: row.id,
|
||||||
|
// approvedStatus,
|
||||||
|
// };
|
||||||
|
// })
|
||||||
|
// .filter(
|
||||||
|
// (
|
||||||
|
// row,
|
||||||
|
// ): row is {
|
||||||
|
// id: string;
|
||||||
|
// approvedStatus: string;
|
||||||
|
// } => row !== null,
|
||||||
|
// );
|
||||||
|
|||||||
@@ -4,11 +4,13 @@ import { setupAuthRoutes } from "./auth/auth.routes.js";
|
|||||||
// import the routes and route setups
|
// import the routes and route setups
|
||||||
import { setupApiDocsRoutes } from "./configs/scaler.config.js";
|
import { setupApiDocsRoutes } from "./configs/scaler.config.js";
|
||||||
import { setupDatamartRoutes } from "./datamart/datamart.routes.js";
|
import { setupDatamartRoutes } from "./datamart/datamart.routes.js";
|
||||||
|
import { setupGPSqlRoutes } from "./gpSql/gpSql.routes.js";
|
||||||
import { setupNotificationRoutes } from "./notification/notification.routes.js";
|
import { setupNotificationRoutes } from "./notification/notification.routes.js";
|
||||||
import { setupOCPRoutes } from "./ocp/ocp.routes.js";
|
import { setupOCPRoutes } from "./ocp/ocp.routes.js";
|
||||||
import { setupOpendockRoutes } from "./opendock/opendock.routes.js";
|
import { setupOpendockRoutes } from "./opendock/opendock.routes.js";
|
||||||
import { setupProdSqlRoutes } from "./prodSql/prodSql.routes.js";
|
import { setupProdSqlRoutes } from "./prodSql/prodSql.routes.js";
|
||||||
import { setupSystemRoutes } from "./system/system.routes.js";
|
import { setupSystemRoutes } from "./system/system.routes.js";
|
||||||
|
import { setupTCPRoutes } from "./tcpServer/tcp.routes.js";
|
||||||
import { setupUtilsRoutes } from "./utils/utils.routes.js";
|
import { setupUtilsRoutes } from "./utils/utils.routes.js";
|
||||||
|
|
||||||
export const setupRoutes = (baseUrl: string, app: Express) => {
|
export const setupRoutes = (baseUrl: string, app: Express) => {
|
||||||
@@ -16,10 +18,12 @@ export const setupRoutes = (baseUrl: string, app: Express) => {
|
|||||||
setupSystemRoutes(baseUrl, app);
|
setupSystemRoutes(baseUrl, app);
|
||||||
setupApiDocsRoutes(baseUrl, app);
|
setupApiDocsRoutes(baseUrl, app);
|
||||||
setupProdSqlRoutes(baseUrl, app);
|
setupProdSqlRoutes(baseUrl, app);
|
||||||
|
setupGPSqlRoutes(baseUrl, app);
|
||||||
setupDatamartRoutes(baseUrl, app);
|
setupDatamartRoutes(baseUrl, app);
|
||||||
setupAuthRoutes(baseUrl, app);
|
setupAuthRoutes(baseUrl, app);
|
||||||
setupUtilsRoutes(baseUrl, app);
|
setupUtilsRoutes(baseUrl, app);
|
||||||
setupOpendockRoutes(baseUrl, app);
|
setupOpendockRoutes(baseUrl, app);
|
||||||
setupNotificationRoutes(baseUrl, app);
|
setupNotificationRoutes(baseUrl, app);
|
||||||
setupOCPRoutes(baseUrl, app);
|
setupOCPRoutes(baseUrl, app);
|
||||||
|
setupTCPRoutes(baseUrl, app);
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -4,16 +4,21 @@ import createApp from "./app.js";
|
|||||||
import { db } from "./db/db.controller.js";
|
import { db } from "./db/db.controller.js";
|
||||||
import { dbCleanup } from "./db/dbCleanup.controller.js";
|
import { dbCleanup } from "./db/dbCleanup.controller.js";
|
||||||
import { type Setting, settings } from "./db/schema/settings.schema.js";
|
import { type Setting, settings } from "./db/schema/settings.schema.js";
|
||||||
|
import { connectGPSql } from "./gpSql/gpSqlConnection.controller.js";
|
||||||
import { createLogger } from "./logger/logger.controller.js";
|
import { createLogger } from "./logger/logger.controller.js";
|
||||||
|
import { historicalSchedule } from "./logistics/logistics.historicalInv.js";
|
||||||
import { startNotifications } from "./notification/notification.controller.js";
|
import { startNotifications } from "./notification/notification.controller.js";
|
||||||
import { createNotifications } from "./notification/notifications.master.js";
|
import { createNotifications } from "./notification/notifications.master.js";
|
||||||
|
import { printerSync } from "./ocp/ocp.printer.manage.js";
|
||||||
import { monitorReleaseChanges } from "./opendock/openDockRreleaseMonitor.utils.js";
|
import { monitorReleaseChanges } from "./opendock/openDockRreleaseMonitor.utils.js";
|
||||||
import { opendockSocketMonitor } from "./opendock/opendockSocketMonitor.utils.js";
|
import { opendockSocketMonitor } from "./opendock/opendockSocketMonitor.utils.js";
|
||||||
import { connectProdSql } from "./prodSql/prodSqlConnection.controller.js";
|
import { connectProdSql } from "./prodSql/prodSqlConnection.controller.js";
|
||||||
import { monitorAlplaPurchase } from "./purchase/purchase.controller.js";
|
import { monitorAlplaPurchase } from "./purchase/purchase.controller.js";
|
||||||
import { setupSocketIORoutes } from "./socket.io/serverSetup.js";
|
import { setupSocketIORoutes } from "./socket.io/serverSetup.js";
|
||||||
import { baseSettingValidationCheck } from "./system/settingsBase.controller.js";
|
import { baseSettingValidationCheck } from "./system/settingsBase.controller.js";
|
||||||
|
import { startTCPServer } from "./tcpServer/tcp.server.js";
|
||||||
import { createCronJob } from "./utils/croner.utils.js";
|
import { createCronJob } from "./utils/croner.utils.js";
|
||||||
|
import { sendEmail } from "./utils/sendEmail.utils.js";
|
||||||
|
|
||||||
const port = Number(process.env.PORT) || 3000;
|
const port = Number(process.env.PORT) || 3000;
|
||||||
export let systemSettings: Setting[] = [];
|
export let systemSettings: Setting[] = [];
|
||||||
@@ -27,7 +32,9 @@ const start = async () => {
|
|||||||
const log = createLogger({ module: "system", subModule: "main start" });
|
const log = createLogger({ module: "system", subModule: "main start" });
|
||||||
|
|
||||||
// triggering long lived processes
|
// triggering long lived processes
|
||||||
|
startTCPServer();
|
||||||
connectProdSql();
|
connectProdSql();
|
||||||
|
connectGPSql();
|
||||||
|
|
||||||
// trigger startup processes these must run before anything else can run
|
// trigger startup processes these must run before anything else can run
|
||||||
await baseSettingValidationCheck();
|
await baseSettingValidationCheck();
|
||||||
@@ -49,17 +56,39 @@ const start = async () => {
|
|||||||
monitorAlplaPurchase();
|
monitorAlplaPurchase();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (systemSettings.filter((n) => n.name === "ocp")[0]?.active) {
|
||||||
|
printerSync();
|
||||||
|
}
|
||||||
|
|
||||||
// these jobs below are system jobs and should run no matter what.
|
// these jobs below are system jobs and should run no matter what.
|
||||||
createCronJob("JobAuditLogCleanUp", "0 0 5 * * *", () =>
|
createCronJob("JobAuditLogCleanUp", "0 0 5 * * *", () =>
|
||||||
dbCleanup("jobs", 30),
|
dbCleanup("jobs", 30),
|
||||||
);
|
);
|
||||||
createCronJob("logsCleanup", "0 15 5 * * *", () => dbCleanup("logs", 120));
|
createCronJob("logsCleanup", "0 15 5 * * *", () => dbCleanup("logs", 120));
|
||||||
|
historicalSchedule();
|
||||||
|
|
||||||
// one shots only needed to run on server startups
|
// one shots only needed to run on server startups
|
||||||
createNotifications();
|
createNotifications();
|
||||||
startNotifications();
|
startNotifications();
|
||||||
}, 5 * 1000);
|
}, 5 * 1000);
|
||||||
|
|
||||||
|
process.on("uncaughtException", async (err) => {
|
||||||
|
console.error("Uncaught Exception:", err);
|
||||||
|
//await closePool();
|
||||||
|
const emailData = {
|
||||||
|
email: "blake.matthes@alpla.com", // should be moved to the db so it can be reused.
|
||||||
|
subject: `${os.hostname()} has just encountered a crash.`,
|
||||||
|
template: "serverCrash",
|
||||||
|
context: {
|
||||||
|
error: err,
|
||||||
|
plant: `${os.hostname()}`,
|
||||||
|
},
|
||||||
|
};
|
||||||
|
|
||||||
|
await sendEmail(emailData);
|
||||||
|
//process.exit(1);
|
||||||
|
});
|
||||||
|
|
||||||
server.listen(port, async () => {
|
server.listen(port, async () => {
|
||||||
log.info(
|
log.info(
|
||||||
`Listening on http://${os.hostname()}:${port}${baseUrl}, logging in ${process.env.LOG_LEVEL}, current ENV ${process.env.NODE_ENV ? process.env.NODE_ENV : "development"}`,
|
`Listening on http://${os.hostname()}:${port}${baseUrl}, logging in ${process.env.LOG_LEVEL}, current ENV ${process.env.NODE_ENV ? process.env.NODE_ENV : "development"}`,
|
||||||
|
|||||||
@@ -1,9 +1,12 @@
|
|||||||
import { Router } from "express";
|
import { Router } from "express";
|
||||||
|
import { connected as gpSql } from "../gpSql/gpSqlConnection.controller.js";
|
||||||
|
import { connected as prodSql } from "../prodSql/prodSqlConnection.controller.js";
|
||||||
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
import { prodQuery } from "../prodSql/prodSqlQuery.controller.js";
|
||||||
import {
|
import {
|
||||||
type SqlQuery,
|
type SqlQuery,
|
||||||
sqlQuerySelector,
|
sqlQuerySelector,
|
||||||
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
} from "../prodSql/prodSqlQuerySelector.utils.js";
|
||||||
|
import { isServerRunning } from "../tcpServer/tcp.server.js";
|
||||||
|
|
||||||
const router = Router();
|
const router = Router();
|
||||||
|
|
||||||
@@ -25,6 +28,9 @@ router.get("/", async (_, res) => {
|
|||||||
: [],
|
: [],
|
||||||
eomFGPkgSheetVersion: 1, // this is the excel file version when we have a change to the macro we want to grab this
|
eomFGPkgSheetVersion: 1, // this is the excel file version when we have a change to the macro we want to grab this
|
||||||
masterMacroFile: 1,
|
masterMacroFile: 1,
|
||||||
|
tcpServerOnline: isServerRunning,
|
||||||
|
sqlServerConnected: prodSql,
|
||||||
|
gpServerConnected: gpSql,
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|||||||
49
backend/system/system.mobileApp.ts
Normal file
49
backend/system/system.mobileApp.ts
Normal file
@@ -0,0 +1,49 @@
|
|||||||
|
import fs from "node:fs";
|
||||||
|
import { Router } from "express";
|
||||||
|
import path from "path";
|
||||||
|
import { fileURLToPath } from "url";
|
||||||
|
|
||||||
|
const router = Router();
|
||||||
|
|
||||||
|
const __filename = fileURLToPath(import.meta.url);
|
||||||
|
const __dirname = path.dirname(__filename);
|
||||||
|
|
||||||
|
const downloadDir = path.resolve(__dirname, "../../downloads/mobile");
|
||||||
|
|
||||||
|
const currentApk = {
|
||||||
|
packageName: "net.alpla.lst.mobile",
|
||||||
|
versionName: "0.0.1-alpha",
|
||||||
|
versionCode: 1,
|
||||||
|
minSupportedVersionCode: 1,
|
||||||
|
fileName: "lst-mobile.apk",
|
||||||
|
};
|
||||||
|
|
||||||
|
router.get("/version", async (req, res) => {
|
||||||
|
const baseUrl = `${req.protocol}://${req.get("host")}`;
|
||||||
|
|
||||||
|
res.json({
|
||||||
|
packageName: currentApk.packageName,
|
||||||
|
versionName: currentApk.versionName,
|
||||||
|
versionCode: currentApk.versionCode,
|
||||||
|
minSupportedVersionCode: currentApk.minSupportedVersionCode,
|
||||||
|
downloadUrl: `${baseUrl}/lst/api/mobile/apk/latest`,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
router.get("/apk/latest", (_, res) => {
|
||||||
|
const apkPath = path.join(downloadDir, currentApk.fileName);
|
||||||
|
|
||||||
|
if (!fs.existsSync(apkPath)) {
|
||||||
|
return res.status(404).json({ success: false, message: "APK not found" });
|
||||||
|
}
|
||||||
|
|
||||||
|
res.setHeader("Content-Type", "application/vnd.android.package-archive");
|
||||||
|
res.setHeader(
|
||||||
|
"Content-Disposition",
|
||||||
|
`attachment; filename="${currentApk.fileName}"`,
|
||||||
|
);
|
||||||
|
|
||||||
|
return res.sendFile(apkPath);
|
||||||
|
});
|
||||||
|
|
||||||
|
export default router;
|
||||||
@@ -3,10 +3,12 @@ import { requireAuth } from "../middleware/auth.middleware.js";
|
|||||||
import getSettings from "./settings.route.js";
|
import getSettings from "./settings.route.js";
|
||||||
import updSetting from "./settingsUpdate.route.js";
|
import updSetting from "./settingsUpdate.route.js";
|
||||||
import stats from "./stats.route.js";
|
import stats from "./stats.route.js";
|
||||||
|
import mobile from "./system.mobileApp.js";
|
||||||
|
|
||||||
export const setupSystemRoutes = (baseUrl: string, app: Express) => {
|
export const setupSystemRoutes = (baseUrl: string, app: Express) => {
|
||||||
//stats will be like this as we dont need to change this
|
//stats will be like this as we dont need to change this
|
||||||
app.use(`${baseUrl}/api/stats`, stats);
|
app.use(`${baseUrl}/api/stats`, stats);
|
||||||
|
app.use(`${baseUrl}/api/mobile`, mobile);
|
||||||
app.use(`${baseUrl}/api/settings`, getSettings);
|
app.use(`${baseUrl}/api/settings`, getSettings);
|
||||||
app.use(`${baseUrl}/api/settings`, requireAuth, updSetting);
|
app.use(`${baseUrl}/api/settings`, requireAuth, updSetting);
|
||||||
|
|
||||||
|
|||||||
51
backend/tcpServer/tcp.printerListener.ts
Normal file
51
backend/tcpServer/tcp.printerListener.ts
Normal file
@@ -0,0 +1,51 @@
|
|||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { printerLog } from "../db/schema/printerLogs.schema.js";
|
||||||
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
|
||||||
|
export type PrinterData = {
|
||||||
|
ip: string;
|
||||||
|
name: string;
|
||||||
|
condition: string;
|
||||||
|
message: string;
|
||||||
|
date?: string;
|
||||||
|
printerSN: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
const log = createLogger({ module: "tcp", submodule: "create_server" });
|
||||||
|
|
||||||
|
export const printerListen = async (tcpData: PrinterData) => {
|
||||||
|
const ip = tcpData.ip?.replace("::ffff:", "");
|
||||||
|
|
||||||
|
// post the new message
|
||||||
|
const { data, error } = await tryCatch(
|
||||||
|
db
|
||||||
|
.insert(printerLog)
|
||||||
|
.values({
|
||||||
|
ip,
|
||||||
|
name: tcpData.name,
|
||||||
|
condition: tcpData.condition,
|
||||||
|
message: tcpData.message,
|
||||||
|
printerSN: tcpData.printerSN,
|
||||||
|
})
|
||||||
|
.returning(),
|
||||||
|
);
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "tcp",
|
||||||
|
subModule: "post",
|
||||||
|
message: "Failed to post tcp printer data.",
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (data) {
|
||||||
|
log.info({}, `${tcpData.name} sent a message over`);
|
||||||
|
// TODO: send message over to the controller to decide what to do next with it
|
||||||
|
}
|
||||||
|
};
|
||||||
14
backend/tcpServer/tcp.routes.ts
Normal file
14
backend/tcpServer/tcp.routes.ts
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
import type { Express } from "express";
|
||||||
|
import { requireAuth } from "../middleware/auth.middleware.js";
|
||||||
|
import restart from "./tcpRestart.route.js";
|
||||||
|
import start from "./tcpStart.route.js";
|
||||||
|
import stop from "./tcpStop.route.js";
|
||||||
|
|
||||||
|
export const setupTCPRoutes = (baseUrl: string, app: Express) => {
|
||||||
|
//stats will be like this as we dont need to change this
|
||||||
|
app.use(`${baseUrl}/api/tcp/start`, requireAuth, start);
|
||||||
|
app.use(`${baseUrl}/api/tcp/stop`, requireAuth, stop);
|
||||||
|
app.use(`${baseUrl}/api/tcp/restart`, requireAuth, restart);
|
||||||
|
|
||||||
|
// all other system should be under /api/system/*
|
||||||
|
};
|
||||||
180
backend/tcpServer/tcp.server.ts
Normal file
180
backend/tcpServer/tcp.server.ts
Normal file
@@ -0,0 +1,180 @@
|
|||||||
|
import net from "node:net";
|
||||||
|
import { eq } from "drizzle-orm";
|
||||||
|
import { db } from "../db/db.controller.js";
|
||||||
|
import { printerData } from "../db/schema/printers.schema.js";
|
||||||
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import { delay } from "../utils/delay.utils.js";
|
||||||
|
import { returnFunc } from "../utils/returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "../utils/trycatch.utils.js";
|
||||||
|
import { type PrinterData, printerListen } from "./tcp.printerListener.js";
|
||||||
|
|
||||||
|
let tcpServer: net.Server;
|
||||||
|
const tcpSockets: Set<net.Socket> = new Set();
|
||||||
|
export let isServerRunning = false;
|
||||||
|
|
||||||
|
const port = parseInt(process.env.TCP_PORT ?? "2222", 10);
|
||||||
|
|
||||||
|
const parseTcpAlert = (input: string) => {
|
||||||
|
// guard
|
||||||
|
const colonIndex = input.indexOf(":");
|
||||||
|
if (colonIndex === -1) return null;
|
||||||
|
|
||||||
|
const condition = input.slice(0, colonIndex).trim();
|
||||||
|
const rest = input.slice(colonIndex + 1).trim();
|
||||||
|
|
||||||
|
// extract all [ ... ] blocks from rest
|
||||||
|
const matches = [...rest.matchAll(/\[(.*?)\]/g)];
|
||||||
|
|
||||||
|
const date = matches[0]?.[1] ?? "";
|
||||||
|
const name = matches[1]?.[1] ?? "";
|
||||||
|
|
||||||
|
// message = everything before first "["
|
||||||
|
const bracketIndex = rest.indexOf("[");
|
||||||
|
const message =
|
||||||
|
bracketIndex !== -1 ? rest.slice(0, bracketIndex).trim() : rest;
|
||||||
|
|
||||||
|
return {
|
||||||
|
condition,
|
||||||
|
message,
|
||||||
|
date,
|
||||||
|
name,
|
||||||
|
};
|
||||||
|
};
|
||||||
|
const log = createLogger({ module: "tcp", submodule: "create_server" });
|
||||||
|
export const startTCPServer = async () => {
|
||||||
|
tcpServer = net.createServer(async (socket) => {
|
||||||
|
tcpSockets.add(socket);
|
||||||
|
socket.on("data", async (data: Buffer) => {
|
||||||
|
const parseData = data.toString("utf-8").trimEnd();
|
||||||
|
|
||||||
|
// check where the data came from then we do something.
|
||||||
|
|
||||||
|
const ip = socket.remoteAddress ?? "127.0.0.1";
|
||||||
|
const { data: printer, error: pError } = await tryCatch(
|
||||||
|
db
|
||||||
|
.select()
|
||||||
|
.from(printerData)
|
||||||
|
.where(eq(printerData.ipAddress, ip.replace("::ffff:", ""))),
|
||||||
|
);
|
||||||
|
if (pError) {
|
||||||
|
log.error(
|
||||||
|
{ stack: pError },
|
||||||
|
"There was an error getting printer data for tcp check",
|
||||||
|
);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (printer?.length) {
|
||||||
|
const printerData = {
|
||||||
|
...parseTcpAlert(parseData),
|
||||||
|
ip,
|
||||||
|
printerSN: printer[0]?.printerSN,
|
||||||
|
name: printer[0]?.name,
|
||||||
|
};
|
||||||
|
|
||||||
|
printerListen(printerData as PrinterData);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
socket.on("end", () => {
|
||||||
|
log.debug({}, "Client disconnected");
|
||||||
|
// just in case we dont fully disconnect
|
||||||
|
setTimeout(() => {
|
||||||
|
if (!socket.destroyed) {
|
||||||
|
socket.destroy();
|
||||||
|
}
|
||||||
|
}, 1000);
|
||||||
|
tcpSockets.delete(socket);
|
||||||
|
});
|
||||||
|
|
||||||
|
socket.on("error", (err: Error) => {
|
||||||
|
log.error({ stack: err }, `Socket error:", ${err}`);
|
||||||
|
// just in case we dont fully disconnect
|
||||||
|
setTimeout(() => {
|
||||||
|
if (!socket.destroyed) {
|
||||||
|
socket.destroy();
|
||||||
|
}
|
||||||
|
}, 1000);
|
||||||
|
tcpSockets.delete(socket);
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
tcpServer.listen(port, () => {
|
||||||
|
log.info({}, `TCP Server listening on port ${port}`);
|
||||||
|
});
|
||||||
|
|
||||||
|
isServerRunning = true;
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "tcp",
|
||||||
|
subModule: "create_server",
|
||||||
|
message: "TCP server started.",
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
room: "",
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
export const stopTCPServer = async () => {
|
||||||
|
if (!isServerRunning)
|
||||||
|
return { success: false, message: "Server is not running" };
|
||||||
|
for (const socket of tcpSockets) {
|
||||||
|
socket.destroy();
|
||||||
|
}
|
||||||
|
tcpSockets.clear();
|
||||||
|
tcpServer.close(() => {
|
||||||
|
log.info({}, "TCP Server stopped");
|
||||||
|
});
|
||||||
|
isServerRunning = false;
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "tcp",
|
||||||
|
subModule: "create_server",
|
||||||
|
message: "TCP server stopped.",
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
room: "",
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
export const restartTCPServer = async () => {
|
||||||
|
if (!isServerRunning) {
|
||||||
|
startTCPServer();
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "warn",
|
||||||
|
module: "tcp",
|
||||||
|
subModule: "create_server",
|
||||||
|
message: "Server is not running will try to start it",
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
room: "",
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
for (const socket of tcpSockets) {
|
||||||
|
socket.destroy();
|
||||||
|
}
|
||||||
|
tcpSockets.clear();
|
||||||
|
tcpServer.close(() => {
|
||||||
|
log.info({}, "TCP Server stopped");
|
||||||
|
});
|
||||||
|
isServerRunning = false;
|
||||||
|
|
||||||
|
await delay(1500);
|
||||||
|
|
||||||
|
startTCPServer();
|
||||||
|
}
|
||||||
|
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "tcp",
|
||||||
|
subModule: "create_server",
|
||||||
|
message: "TCP server has been restarted.",
|
||||||
|
data: [],
|
||||||
|
notify: false,
|
||||||
|
room: "",
|
||||||
|
});
|
||||||
|
};
|
||||||
19
backend/tcpServer/tcpRestart.route.ts
Normal file
19
backend/tcpServer/tcpRestart.route.ts
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { restartTCPServer } from "./tcp.server.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/restart", async (_, res) => {
|
||||||
|
const connect = await restartTCPServer();
|
||||||
|
apiReturn(res, {
|
||||||
|
success: connect.success,
|
||||||
|
level: connect.success ? "info" : "error",
|
||||||
|
module: "tcp",
|
||||||
|
subModule: "post",
|
||||||
|
message: "TCP Server has been restarted",
|
||||||
|
data: connect.data,
|
||||||
|
status: connect.success ? 200 : 400,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
export default r;
|
||||||
20
backend/tcpServer/tcpStart.route.ts
Normal file
20
backend/tcpServer/tcpStart.route.ts
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { startTCPServer } from "./tcp.server.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/start", async (_, res) => {
|
||||||
|
const connect = await startTCPServer();
|
||||||
|
apiReturn(res, {
|
||||||
|
success: connect.success,
|
||||||
|
level: connect.success ? "info" : "error",
|
||||||
|
module: "routes",
|
||||||
|
subModule: "prodSql",
|
||||||
|
message: connect.message,
|
||||||
|
data: connect.data,
|
||||||
|
status: connect.success ? 200 : 400,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
export default r;
|
||||||
20
backend/tcpServer/tcpStop.route.ts
Normal file
20
backend/tcpServer/tcpStop.route.ts
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
import { Router } from "express";
|
||||||
|
import { apiReturn } from "../utils/returnHelper.utils.js";
|
||||||
|
import { stopTCPServer } from "./tcp.server.js";
|
||||||
|
|
||||||
|
const r = Router();
|
||||||
|
|
||||||
|
r.post("/stop", async (_, res) => {
|
||||||
|
const connect = await stopTCPServer();
|
||||||
|
apiReturn(res, {
|
||||||
|
success: connect.success,
|
||||||
|
level: connect.success ? "info" : "error",
|
||||||
|
module: "routes",
|
||||||
|
subModule: "prodSql",
|
||||||
|
message: connect.message,
|
||||||
|
data: [],
|
||||||
|
status: connect.success ? 200 : 400,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
export default r;
|
||||||
9
backend/types/purhcaseTypes.ts
Normal file
9
backend/types/purhcaseTypes.ts
Normal file
@@ -0,0 +1,9 @@
|
|||||||
|
export type GpStatus = {
|
||||||
|
id: string;
|
||||||
|
req: string;
|
||||||
|
};
|
||||||
|
|
||||||
|
export type StatusUpdate = {
|
||||||
|
id: string;
|
||||||
|
approvedStatus: string;
|
||||||
|
};
|
||||||
@@ -9,6 +9,7 @@ export const allowedOrigins = [
|
|||||||
"http://localhost:4000",
|
"http://localhost:4000",
|
||||||
"http://localhost:4001",
|
"http://localhost:4001",
|
||||||
"http://localhost:5500",
|
"http://localhost:5500",
|
||||||
|
"http://localhost:8081",
|
||||||
"https://admin.socket.io",
|
"https://admin.socket.io",
|
||||||
"https://electron-socket-io-playground.vercel.app",
|
"https://electron-socket-io-playground.vercel.app",
|
||||||
`${process.env.URL}`,
|
`${process.env.URL}`,
|
||||||
|
|||||||
@@ -3,6 +3,7 @@ import { eq } from "drizzle-orm";
|
|||||||
import { db } from "../db/db.controller.js";
|
import { db } from "../db/db.controller.js";
|
||||||
import { jobAuditLog } from "../db/schema/auditLog.schema.js";
|
import { jobAuditLog } from "../db/schema/auditLog.schema.js";
|
||||||
import { createLogger } from "../logger/logger.controller.js";
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
import type { ReturnHelper } from "./returnHelper.utils.js";
|
||||||
|
|
||||||
// example createJob
|
// example createJob
|
||||||
// createCronJob("test Cron", "*/5 * * * * *", async () => {
|
// createCronJob("test Cron", "*/5 * * * * *", async () => {
|
||||||
@@ -45,7 +46,7 @@ const cronStats: Record<string, { created: number; replaced: number }> = {};
|
|||||||
export const createCronJob = async (
|
export const createCronJob = async (
|
||||||
name: string,
|
name: string,
|
||||||
schedule: string, // cron string with 8 8 IE: */5 * * * * * every 5th second
|
schedule: string, // cron string with 8 8 IE: */5 * * * * * every 5th second
|
||||||
task: () => Promise<void>, // what function are we passing over
|
task: () => Promise<void | ReturnHelper>, // what function are we passing over
|
||||||
source = "unknown",
|
source = "unknown",
|
||||||
) => {
|
) => {
|
||||||
// get the timezone based on the os timezone set
|
// get the timezone based on the os timezone set
|
||||||
|
|||||||
73
backend/utils/mailViews/qualityBlocking.hbs
Normal file
73
backend/utils/mailViews/qualityBlocking.hbs
Normal file
@@ -0,0 +1,73 @@
|
|||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8" />
|
||||||
|
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||||
|
|
||||||
|
{{!-- <link rel="stylesheet" href="styles/styles.css" /> --}}
|
||||||
|
<style>
|
||||||
|
.email-wrapper {
|
||||||
|
max-width: 80%; /* Limit width to 80% of the window */
|
||||||
|
margin: 0 auto; /* Center the content horizontally */
|
||||||
|
}
|
||||||
|
.email-table {
|
||||||
|
width: 100%;
|
||||||
|
border-collapse: collapse;
|
||||||
|
}
|
||||||
|
.email-table td {
|
||||||
|
vertical-align: top;
|
||||||
|
padding: 10px;
|
||||||
|
border: 1px solid #000;
|
||||||
|
border-radius: 25px; /* Rounded corners */
|
||||||
|
background-color: #f0f0f0; /* Optional: Add a background color */
|
||||||
|
}
|
||||||
|
.email-table h2 {
|
||||||
|
margin: 0;
|
||||||
|
}
|
||||||
|
.remarks {
|
||||||
|
border: 1px solid black;
|
||||||
|
padding: 10px;
|
||||||
|
background-color: #f0f0f0;
|
||||||
|
border-radius: 25px;
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<div class="email-wrapper">
|
||||||
|
<p>All,</p>
|
||||||
|
<p>Please see the new blocking order that was created.</p>
|
||||||
|
<div>
|
||||||
|
<div class="email-table">
|
||||||
|
<table>
|
||||||
|
<tr>
|
||||||
|
<td>
|
||||||
|
<p><strong>Blocking number: </strong>{{items.blockingNumber}}</p>
|
||||||
|
<p><strong>Blocking Date: </strong>{{items.blockingDate}}</p>
|
||||||
|
<p><strong>Article: </strong>{{items.av}}</p>
|
||||||
|
<p><strong>Production Lot: </strong>{{items.lotNumber}}</p>
|
||||||
|
<p><strong>Line: </strong>{{items.line}}</p>
|
||||||
|
</td>
|
||||||
|
<td>
|
||||||
|
<p><strong>Customer: </strong>{{items.customer}}</p>
|
||||||
|
<p><strong>Blocked pieces /LUs: </strong>{{items.peicesAndLoadingUnits}}</p>
|
||||||
|
<p><strong>Main defect group: </strong>{{items.mainDefectGroup}}</p>
|
||||||
|
<p><strong>Main defect: </strong>{{items.mainDefect}}</p>
|
||||||
|
</td>
|
||||||
|
</tr>
|
||||||
|
</table>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
<div class="remarks">
|
||||||
|
<h4>Remarks:</h4>
|
||||||
|
<p>{{items.remark}}</p>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
<br>
|
||||||
|
<p>For further questions please reach out to quality.</p>
|
||||||
|
<p>Thank you,</p>
|
||||||
|
<p>Quality Department</p>
|
||||||
|
</p>
|
||||||
|
</div>
|
||||||
|
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
35
backend/utils/mailViews/serverCrash.hbs
Normal file
35
backend/utils/mailViews/serverCrash.hbs
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8" />
|
||||||
|
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||||
|
{{!--<title>Order Summary</title> --}}
|
||||||
|
{{> styles}}
|
||||||
|
<style>
|
||||||
|
pre {
|
||||||
|
background-color: #f8f9fa;
|
||||||
|
color: #d63384;
|
||||||
|
padding: 10px;
|
||||||
|
border-radius: 5px;
|
||||||
|
white-space: pre-wrap;
|
||||||
|
font-family: monospace;
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
{{!-- <link rel="stylesheet" href="styles/styles.css" /> --}}
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<h3>{{plant}},<br/> Has encountered an unexpected error.</h1>
|
||||||
|
<p>
|
||||||
|
Please see below the stack error from the crash.
|
||||||
|
</p>
|
||||||
|
<hr/>
|
||||||
|
<div>
|
||||||
|
<h3>Error Message: </h3>
|
||||||
|
<p>{{error.message}}</p>
|
||||||
|
</div>
|
||||||
|
<hr/>
|
||||||
|
<div>
|
||||||
|
<h3>Stack trace</h3>
|
||||||
|
<pre>{{{error.stack}}}</pre>
|
||||||
|
</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
36
backend/utils/mailViews/serverCritialIssue.hbs
Normal file
36
backend/utils/mailViews/serverCritialIssue.hbs
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
<html lang="en">
|
||||||
|
<head>
|
||||||
|
<meta charset="UTF-8" />
|
||||||
|
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
|
||||||
|
{{!--<title>Order Summary</title> --}}
|
||||||
|
{{> styles}}
|
||||||
|
<style>
|
||||||
|
pre {
|
||||||
|
background-color: #f8f9fa;
|
||||||
|
color: #d63384;
|
||||||
|
padding: 10px;
|
||||||
|
border-radius: 5px;
|
||||||
|
white-space: pre-wrap;
|
||||||
|
font-family: monospace;
|
||||||
|
}
|
||||||
|
</style>
|
||||||
|
{{!-- <link rel="stylesheet" href="styles/styles.css" /> --}}
|
||||||
|
</head>
|
||||||
|
<body>
|
||||||
|
<h3>{{plant}},<br/> Has encountered an error.</h1>
|
||||||
|
<p>
|
||||||
|
The below error came from Module: {{module}}, Submodule: {{submodule}}.
|
||||||
|
</p>
|
||||||
|
<p>The error below is considered to be critical and should be addressed</p>
|
||||||
|
<hr/>
|
||||||
|
<div>
|
||||||
|
<h3>Error Message: </h3>
|
||||||
|
<p>{{message}}</p>
|
||||||
|
</div>
|
||||||
|
<hr/>
|
||||||
|
<div>
|
||||||
|
<h3>Stack trace</h3>
|
||||||
|
<pre>{{{error}}}</pre>
|
||||||
|
</div>
|
||||||
|
</body>
|
||||||
|
</html>
|
||||||
41
backend/utils/pgConnectToLst.utils.ts
Normal file
41
backend/utils/pgConnectToLst.utils.ts
Normal file
@@ -0,0 +1,41 @@
|
|||||||
|
import pkg from "pg";
|
||||||
|
const { Pool } = pkg;
|
||||||
|
|
||||||
|
const baseConfig = {
|
||||||
|
host: process.env.DATABASE_HOST ?? "localhost",
|
||||||
|
port: parseInt(process.env.DATABASE_PORT ?? "5433", 10),
|
||||||
|
user: process.env.DATABASE_USER,
|
||||||
|
password: process.env.DATABASE_PASSWORD,
|
||||||
|
};
|
||||||
|
|
||||||
|
// Pools (one per DB)
|
||||||
|
const v1Pool = new Pool({
|
||||||
|
...baseConfig,
|
||||||
|
database: "lst",
|
||||||
|
});
|
||||||
|
|
||||||
|
const v2Pool = new Pool({
|
||||||
|
...baseConfig,
|
||||||
|
database: "lst_db",
|
||||||
|
});
|
||||||
|
|
||||||
|
// Query helpers
|
||||||
|
export const v1QueryRun = async (query: string, params?: any[]) => {
|
||||||
|
try {
|
||||||
|
const res = await v1Pool.query(query, params);
|
||||||
|
return res;
|
||||||
|
} catch (err) {
|
||||||
|
console.error("V1 query error:", err);
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
export const v2QueryRun = async (query: string, params?: any[]) => {
|
||||||
|
try {
|
||||||
|
const res = await v2Pool.query(query, params);
|
||||||
|
return res;
|
||||||
|
} catch (err) {
|
||||||
|
console.error("V2 query error:", err);
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
};
|
||||||
124
backend/utils/prodEndpoint.utils.ts
Normal file
124
backend/utils/prodEndpoint.utils.ts
Normal file
@@ -0,0 +1,124 @@
|
|||||||
|
import https from "node:https";
|
||||||
|
import axios from "axios";
|
||||||
|
import { returnFunc } from "./returnHelper.utils.js";
|
||||||
|
import { tryCatch } from "./trycatch.utils.js";
|
||||||
|
|
||||||
|
type bodyData = any;
|
||||||
|
|
||||||
|
type Data = {
|
||||||
|
endpoint: string;
|
||||||
|
data?: bodyData[];
|
||||||
|
method: "post" | "get" | "delete" | "patch";
|
||||||
|
};
|
||||||
|
|
||||||
|
// type ApiResponse<T = unknown> = {
|
||||||
|
// status: number;
|
||||||
|
// statusText: string;
|
||||||
|
// data: T;
|
||||||
|
// };
|
||||||
|
|
||||||
|
// create the test server stuff
|
||||||
|
const testServers = [
|
||||||
|
{ token: "test1", port: 8940 },
|
||||||
|
{ token: "test2", port: 8941 },
|
||||||
|
{ token: "test3", port: 8942 },
|
||||||
|
];
|
||||||
|
|
||||||
|
const agent = new https.Agent({
|
||||||
|
rejectUnauthorized: false,
|
||||||
|
});
|
||||||
|
|
||||||
|
export const prodEndpointCreation = async (endpoint: string) => {
|
||||||
|
let url = "";
|
||||||
|
//get the plant token
|
||||||
|
const plantToken = process.env.PROD_PLANT_TOKEN ?? "test1";
|
||||||
|
|
||||||
|
// check if we are a test server
|
||||||
|
const testServer = testServers.some((server) => server.token === plantToken);
|
||||||
|
|
||||||
|
// await db
|
||||||
|
// .select()
|
||||||
|
// .from(settings)
|
||||||
|
// .where(eq(settings.name, "dbServer"));
|
||||||
|
|
||||||
|
if (testServer) {
|
||||||
|
//filter out what testserver we are
|
||||||
|
const test = testServers.filter((t) => t.token === plantToken);
|
||||||
|
// "https://usmcd1vms036.alpla.net:8942/application/public/v1.0/DemandManagement/ORDERS"
|
||||||
|
url = `https://${process.env.PROD_SERVER}.alpla.net:${test[0]?.port}/application${endpoint}`;
|
||||||
|
return url;
|
||||||
|
} else {
|
||||||
|
url = `https://${plantToken}prod.alpla.net/application${endpoint}`;
|
||||||
|
return url;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
*
|
||||||
|
* @param data
|
||||||
|
* @param timeoutDelay
|
||||||
|
* @returns
|
||||||
|
*/
|
||||||
|
export const runProdApi = async (data: Data) => {
|
||||||
|
const url = await prodEndpointCreation(data.endpoint);
|
||||||
|
|
||||||
|
const { data: d, error } = await tryCatch(
|
||||||
|
axios({
|
||||||
|
method: data.method as string,
|
||||||
|
url,
|
||||||
|
data: data.data ? data.data[0] : undefined,
|
||||||
|
headers: {
|
||||||
|
"X-API-Key": process.env.TEC_API_KEY || "",
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
},
|
||||||
|
validateStatus: () => true,
|
||||||
|
httpsAgent: agent,
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
|
||||||
|
switch (d?.status) {
|
||||||
|
case 200:
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "info",
|
||||||
|
module: "utils",
|
||||||
|
subModule: "prodEndpoint",
|
||||||
|
message: "Data from prod endpoint",
|
||||||
|
data: d.data,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
|
||||||
|
case 401:
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "utils",
|
||||||
|
subModule: "prodEndpoint",
|
||||||
|
message: "Data from prod endpoint",
|
||||||
|
data: d.data,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
case 400:
|
||||||
|
return returnFunc({
|
||||||
|
success: false,
|
||||||
|
level: "error",
|
||||||
|
module: "utils",
|
||||||
|
subModule: "prodEndpoint",
|
||||||
|
message: "Data from prod endpoint",
|
||||||
|
data: d.data,
|
||||||
|
notify: false,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
if (error) {
|
||||||
|
return returnFunc({
|
||||||
|
success: true,
|
||||||
|
level: "error",
|
||||||
|
module: "utils",
|
||||||
|
subModule: "prodEndpoint",
|
||||||
|
message: "Failed to get data from the prod endpoint",
|
||||||
|
data: error as any,
|
||||||
|
notify: true,
|
||||||
|
});
|
||||||
|
}
|
||||||
|
};
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
import type { Response } from "express";
|
import type { Response } from "express";
|
||||||
import { createLogger } from "../logger/logger.controller.js";
|
import { createLogger } from "../logger/logger.controller.js";
|
||||||
|
|
||||||
interface Data<T = unknown[]> {
|
export interface ReturnHelper<T = unknown[]> {
|
||||||
success: boolean;
|
success: boolean;
|
||||||
module:
|
module:
|
||||||
| "system"
|
| "system"
|
||||||
@@ -12,27 +12,12 @@ interface Data<T = unknown[]> {
|
|||||||
| "opendock"
|
| "opendock"
|
||||||
| "notification"
|
| "notification"
|
||||||
| "email"
|
| "email"
|
||||||
| "purchase";
|
| "purchase"
|
||||||
subModule:
|
| "tcp"
|
||||||
| "db"
|
| "logistics";
|
||||||
| "labeling"
|
subModule: string;
|
||||||
| "printer"
|
|
||||||
| "prodSql"
|
level: "info" | "error" | "debug" | "fatal" | "warn";
|
||||||
| "query"
|
|
||||||
| "sendmail"
|
|
||||||
| "auth"
|
|
||||||
| "datamart"
|
|
||||||
| "jobs"
|
|
||||||
| "apt"
|
|
||||||
| "settings"
|
|
||||||
| "get"
|
|
||||||
| "update"
|
|
||||||
| "delete"
|
|
||||||
| "post"
|
|
||||||
| "notification"
|
|
||||||
| "delete"
|
|
||||||
| "printing";
|
|
||||||
level: "info" | "error" | "debug" | "fatal";
|
|
||||||
message: string;
|
message: string;
|
||||||
room?: string;
|
room?: string;
|
||||||
data?: T;
|
data?: T;
|
||||||
@@ -53,7 +38,7 @@ interface Data<T = unknown[]> {
|
|||||||
* data: [] the data that will be passed back
|
* data: [] the data that will be passed back
|
||||||
* notify: false by default this is to send a notification to a users email to alert them of an issue.
|
* notify: false by default this is to send a notification to a users email to alert them of an issue.
|
||||||
*/
|
*/
|
||||||
export const returnFunc = (data: Data) => {
|
export const returnFunc = (data: ReturnHelper) => {
|
||||||
const notify = data.notify ? data.notify : false;
|
const notify = data.notify ? data.notify : false;
|
||||||
const room = data.room ?? data.room;
|
const room = data.room ?? data.room;
|
||||||
const log = createLogger({ module: data.module, subModule: data.subModule });
|
const log = createLogger({ module: data.module, subModule: data.subModule });
|
||||||
@@ -63,13 +48,14 @@ export const returnFunc = (data: Data) => {
|
|||||||
log.info({ notify: notify, room }, data.message);
|
log.info({ notify: notify, room }, data.message);
|
||||||
break;
|
break;
|
||||||
case "error":
|
case "error":
|
||||||
log.error({ notify: notify, error: data.data, room }, data.message);
|
log.error({ notify: notify, stack: data.data ?? [], room }, data.message);
|
||||||
|
|
||||||
break;
|
break;
|
||||||
case "debug":
|
case "debug":
|
||||||
log.debug({ notify: notify, room }, data.message);
|
log.debug({ notify: notify, stack: data.data ?? [], room }, data.message);
|
||||||
break;
|
break;
|
||||||
case "fatal":
|
case "fatal":
|
||||||
log.fatal({ notify: notify, room }, data.message);
|
log.fatal({ notify: notify, stack: data.data ?? [], room }, data.message);
|
||||||
}
|
}
|
||||||
|
|
||||||
// api section to return
|
// api section to return
|
||||||
@@ -85,7 +71,7 @@ export const returnFunc = (data: Data) => {
|
|||||||
|
|
||||||
export function apiReturn(
|
export function apiReturn(
|
||||||
res: Response,
|
res: Response,
|
||||||
opts: Data & { status?: number },
|
opts: ReturnHelper & { status?: number },
|
||||||
optional?: unknown, // leave this as unknown so we can pass an object or an array over.
|
optional?: unknown, // leave this as unknown so we can pass an object or an array over.
|
||||||
): Response {
|
): Response {
|
||||||
const result = returnFunc(opts);
|
const result = returnFunc(opts);
|
||||||
|
|||||||
@@ -88,7 +88,7 @@ export const sendEmail = async (data: EmailData) => {
|
|||||||
level: "error",
|
level: "error",
|
||||||
module: "utils",
|
module: "utils",
|
||||||
subModule: "sendmail",
|
subModule: "sendmail",
|
||||||
message: `Error sending Email to : ${data.email}`,
|
message: `Error sending Email to : ${data.email}, Error: ${error.message}`,
|
||||||
data: [{ error: error }],
|
data: [{ error: error }],
|
||||||
notify: false,
|
notify: false,
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -5,13 +5,17 @@ meta {
|
|||||||
}
|
}
|
||||||
|
|
||||||
get {
|
get {
|
||||||
url: {{url}}/api/datamart/:name
|
url: {{url}}/api/datamart/:name?historical=x
|
||||||
body: none
|
body: none
|
||||||
auth: inherit
|
auth: inherit
|
||||||
}
|
}
|
||||||
|
|
||||||
|
params:query {
|
||||||
|
historical: x
|
||||||
|
}
|
||||||
|
|
||||||
params:path {
|
params:path {
|
||||||
name: activeArticles
|
name: inventory
|
||||||
}
|
}
|
||||||
|
|
||||||
settings {
|
settings {
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
vars {
|
vars {
|
||||||
url: http://localhost:3600/lst
|
url: http://localhost:3000/lst
|
||||||
readerIp: 10.44.14.215
|
readerIp: 10.44.14.215
|
||||||
}
|
}
|
||||||
vars:secret [
|
vars:secret [
|
||||||
|
|||||||
@@ -14,7 +14,7 @@ body:json {
|
|||||||
{
|
{
|
||||||
"userId":"m6AbQXFwOXoX3YKLfwWgq2LIdDqS5jqv",
|
"userId":"m6AbQXFwOXoX3YKLfwWgq2LIdDqS5jqv",
|
||||||
"notificationId": "0399eb2a-39df-48b7-9f1c-d233cec94d2e",
|
"notificationId": "0399eb2a-39df-48b7-9f1c-d233cec94d2e",
|
||||||
"emails": ["blake.mattes@alpla.com","cowchmonkey@gmail.com"]
|
"emails": ["blake.matthes@alpla.com","blake.matthes@alpla.com"]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
1486
frontend/package-lock.json
generated
1486
frontend/package-lock.json
generated
File diff suppressed because it is too large
Load Diff
@@ -26,6 +26,8 @@
|
|||||||
"radix-ui": "^1.4.3",
|
"radix-ui": "^1.4.3",
|
||||||
"react": "^19.1.1",
|
"react": "^19.1.1",
|
||||||
"react-dom": "^19.1.1",
|
"react-dom": "^19.1.1",
|
||||||
|
"react-markdown": "^10.1.0",
|
||||||
|
"remark-gfm": "^4.0.1",
|
||||||
"shadcn": "^4.0.8",
|
"shadcn": "^4.0.8",
|
||||||
"socket.io-client": "^4.8.3",
|
"socket.io-client": "^4.8.3",
|
||||||
"sonner": "^2.0.7",
|
"sonner": "^2.0.7",
|
||||||
@@ -36,6 +38,7 @@
|
|||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@eslint/js": "^9.36.0",
|
"@eslint/js": "^9.36.0",
|
||||||
|
"@tailwindcss/typography": "^0.5.19",
|
||||||
"@tanstack/router-plugin": "^1.166.7",
|
"@tanstack/router-plugin": "^1.166.7",
|
||||||
"@types/react": "^19.1.13",
|
"@types/react": "^19.1.13",
|
||||||
"@types/react-dom": "^19.1.9",
|
"@types/react-dom": "^19.1.9",
|
||||||
|
|||||||
BIN
frontend/public/imgs/docs/notifications/dk_profile.png
Normal file
BIN
frontend/public/imgs/docs/notifications/dk_profile.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 5.8 KiB |
Binary file not shown.
|
After Width: | Height: | Size: 27 KiB |
Binary file not shown.
|
After Width: | Height: | Size: 20 KiB |
BIN
frontend/public/imgs/docs/notifications/lt_profile.png
Normal file
BIN
frontend/public/imgs/docs/notifications/lt_profile.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 5.9 KiB |
BIN
frontend/public/imgs/docs/notifications/lt_qualityBlocking.png
Normal file
BIN
frontend/public/imgs/docs/notifications/lt_qualityBlocking.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 31 KiB |
BIN
frontend/public/imgs/docs/notifications/lt_reprints.png
Normal file
BIN
frontend/public/imgs/docs/notifications/lt_reprints.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 21 KiB |
105
frontend/src/components/Sidebar/DocBar.tsx
Normal file
105
frontend/src/components/Sidebar/DocBar.tsx
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
import { Link, useRouterState } from "@tanstack/react-router";
|
||||||
|
import { ChevronRight } from "lucide-react";
|
||||||
|
import {
|
||||||
|
Collapsible,
|
||||||
|
CollapsibleContent,
|
||||||
|
CollapsibleTrigger,
|
||||||
|
} from "../ui/collapsible";
|
||||||
|
|
||||||
|
import {
|
||||||
|
SidebarGroup,
|
||||||
|
SidebarGroupContent,
|
||||||
|
SidebarGroupLabel,
|
||||||
|
SidebarMenu,
|
||||||
|
SidebarMenuButton,
|
||||||
|
SidebarMenuItem,
|
||||||
|
SidebarMenuSub,
|
||||||
|
SidebarMenuSubButton,
|
||||||
|
SidebarMenuSubItem,
|
||||||
|
useSidebar,
|
||||||
|
} from "../ui/sidebar";
|
||||||
|
|
||||||
|
const docs = [
|
||||||
|
{
|
||||||
|
title: "Notifications",
|
||||||
|
url: "/intro",
|
||||||
|
//icon,
|
||||||
|
isActive: window.location.pathname.includes("notifications") ?? false,
|
||||||
|
items: [
|
||||||
|
{
|
||||||
|
title: "Reprints",
|
||||||
|
url: "/reprints",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
title: "New Blocking order",
|
||||||
|
url: "/qualityBlocking",
|
||||||
|
},
|
||||||
|
],
|
||||||
|
},
|
||||||
|
];
|
||||||
|
export default function DocBar() {
|
||||||
|
const { setOpen } = useSidebar();
|
||||||
|
const pathname = useRouterState({
|
||||||
|
select: (s) => s.location.pathname,
|
||||||
|
});
|
||||||
|
|
||||||
|
const isNotifications = pathname.includes("notifications");
|
||||||
|
|
||||||
|
return (
|
||||||
|
<SidebarGroup>
|
||||||
|
<SidebarGroupLabel>Docs</SidebarGroupLabel>
|
||||||
|
<SidebarGroupContent>
|
||||||
|
<SidebarMenu>
|
||||||
|
<SidebarMenuItem key={"docs"}>
|
||||||
|
<SidebarMenuButton asChild>
|
||||||
|
<Link to={"/docs"} onClick={() => setOpen(false)}>
|
||||||
|
{/* <item.icon /> */}
|
||||||
|
<span>{"Intro"}</span>
|
||||||
|
</Link>
|
||||||
|
</SidebarMenuButton>
|
||||||
|
</SidebarMenuItem>
|
||||||
|
</SidebarMenu>
|
||||||
|
<SidebarMenu>
|
||||||
|
{docs.map((item) => (
|
||||||
|
<Collapsible
|
||||||
|
key={item.title}
|
||||||
|
asChild
|
||||||
|
defaultOpen={isNotifications}
|
||||||
|
className="group/collapsible"
|
||||||
|
>
|
||||||
|
<SidebarMenuItem>
|
||||||
|
<CollapsibleTrigger asChild>
|
||||||
|
<SidebarMenuButton tooltip={item.title}>
|
||||||
|
<Link
|
||||||
|
to={"/docs/$"}
|
||||||
|
params={{ _splat: `notifications${item.url}` }}
|
||||||
|
>
|
||||||
|
{item.title}
|
||||||
|
</Link>
|
||||||
|
<ChevronRight className="ml-auto transition-transform duration-200 group-data-[state=open]/collapsible:rotate-90" />
|
||||||
|
</SidebarMenuButton>
|
||||||
|
</CollapsibleTrigger>
|
||||||
|
<CollapsibleContent>
|
||||||
|
<SidebarMenuSub>
|
||||||
|
{item.items?.map((subItem) => (
|
||||||
|
<SidebarMenuSubItem key={subItem.title}>
|
||||||
|
<SidebarMenuSubButton asChild>
|
||||||
|
<Link
|
||||||
|
to={"/docs/$"}
|
||||||
|
params={{ _splat: `notifications${subItem.url}` }}
|
||||||
|
>
|
||||||
|
{subItem.title}
|
||||||
|
</Link>
|
||||||
|
</SidebarMenuSubButton>
|
||||||
|
</SidebarMenuSubItem>
|
||||||
|
))}
|
||||||
|
</SidebarMenuSub>
|
||||||
|
</CollapsibleContent>
|
||||||
|
</SidebarMenuItem>
|
||||||
|
</Collapsible>
|
||||||
|
))}
|
||||||
|
</SidebarMenu>
|
||||||
|
</SidebarGroupContent>
|
||||||
|
</SidebarGroup>
|
||||||
|
);
|
||||||
|
}
|
||||||
@@ -7,6 +7,7 @@ import {
|
|||||||
} from "@/components/ui/sidebar";
|
} from "@/components/ui/sidebar";
|
||||||
import { useSession } from "@/lib/auth-client";
|
import { useSession } from "@/lib/auth-client";
|
||||||
import AdminSidebar from "./AdminBar";
|
import AdminSidebar from "./AdminBar";
|
||||||
|
import DocBar from "./DocBar";
|
||||||
|
|
||||||
export function AppSidebar() {
|
export function AppSidebar() {
|
||||||
const { data: session } = useSession();
|
const { data: session } = useSession();
|
||||||
@@ -21,6 +22,7 @@ export function AppSidebar() {
|
|||||||
<SidebarMenu>
|
<SidebarMenu>
|
||||||
<SidebarMenuItem>
|
<SidebarMenuItem>
|
||||||
<SidebarContent>
|
<SidebarContent>
|
||||||
|
<DocBar/>
|
||||||
{session &&
|
{session &&
|
||||||
(session.user.role === "admin" ||
|
(session.user.role === "admin" ||
|
||||||
session.user.role === "systemAdmin") && (
|
session.user.role === "systemAdmin") && (
|
||||||
|
|||||||
76
frontend/src/components/ui/alert.tsx
Normal file
76
frontend/src/components/ui/alert.tsx
Normal file
@@ -0,0 +1,76 @@
|
|||||||
|
import * as React from "react"
|
||||||
|
import { cva, type VariantProps } from "class-variance-authority"
|
||||||
|
|
||||||
|
import { cn } from "@/lib/utils"
|
||||||
|
|
||||||
|
const alertVariants = cva(
|
||||||
|
"group/alert relative grid w-full gap-0.5 rounded-lg border px-2.5 py-2 text-left text-sm has-data-[slot=alert-action]:relative has-data-[slot=alert-action]:pr-18 has-[>svg]:grid-cols-[auto_1fr] has-[>svg]:gap-x-2 *:[svg]:row-span-2 *:[svg]:translate-y-0.5 *:[svg]:text-current *:[svg:not([class*='size-'])]:size-4",
|
||||||
|
{
|
||||||
|
variants: {
|
||||||
|
variant: {
|
||||||
|
default: "bg-card text-card-foreground",
|
||||||
|
destructive:
|
||||||
|
"bg-card text-destructive *:data-[slot=alert-description]:text-destructive/90 *:[svg]:text-current",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
defaultVariants: {
|
||||||
|
variant: "default",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
function Alert({
|
||||||
|
className,
|
||||||
|
variant,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<"div"> & VariantProps<typeof alertVariants>) {
|
||||||
|
return (
|
||||||
|
<div
|
||||||
|
data-slot="alert"
|
||||||
|
role="alert"
|
||||||
|
className={cn(alertVariants({ variant }), className)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function AlertTitle({ className, ...props }: React.ComponentProps<"div">) {
|
||||||
|
return (
|
||||||
|
<div
|
||||||
|
data-slot="alert-title"
|
||||||
|
className={cn(
|
||||||
|
"font-medium group-has-[>svg]/alert:col-start-2 [&_a]:underline [&_a]:underline-offset-3 [&_a]:hover:text-foreground",
|
||||||
|
className
|
||||||
|
)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function AlertDescription({
|
||||||
|
className,
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<"div">) {
|
||||||
|
return (
|
||||||
|
<div
|
||||||
|
data-slot="alert-description"
|
||||||
|
className={cn(
|
||||||
|
"text-sm text-balance text-muted-foreground md:text-pretty [&_a]:underline [&_a]:underline-offset-3 [&_a]:hover:text-foreground [&_p:not(:last-child)]:mb-4",
|
||||||
|
className
|
||||||
|
)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function AlertAction({ className, ...props }: React.ComponentProps<"div">) {
|
||||||
|
return (
|
||||||
|
<div
|
||||||
|
data-slot="alert-action"
|
||||||
|
className={cn("absolute top-2 right-2", className)}
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
export { Alert, AlertTitle, AlertDescription, AlertAction }
|
||||||
31
frontend/src/components/ui/collapsible.tsx
Normal file
31
frontend/src/components/ui/collapsible.tsx
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
import { Collapsible as CollapsiblePrimitive } from "radix-ui"
|
||||||
|
|
||||||
|
function Collapsible({
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof CollapsiblePrimitive.Root>) {
|
||||||
|
return <CollapsiblePrimitive.Root data-slot="collapsible" {...props} />
|
||||||
|
}
|
||||||
|
|
||||||
|
function CollapsibleTrigger({
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof CollapsiblePrimitive.CollapsibleTrigger>) {
|
||||||
|
return (
|
||||||
|
<CollapsiblePrimitive.CollapsibleTrigger
|
||||||
|
data-slot="collapsible-trigger"
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
function CollapsibleContent({
|
||||||
|
...props
|
||||||
|
}: React.ComponentProps<typeof CollapsiblePrimitive.CollapsibleContent>) {
|
||||||
|
return (
|
||||||
|
<CollapsiblePrimitive.CollapsibleContent
|
||||||
|
data-slot="collapsible-content"
|
||||||
|
{...props}
|
||||||
|
/>
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
export { Collapsible, CollapsibleTrigger, CollapsibleContent }
|
||||||
62
frontend/src/docs/notifications/intro.tsx
Normal file
62
frontend/src/docs/notifications/intro.tsx
Normal file
@@ -0,0 +1,62 @@
|
|||||||
|
export default function into() {
|
||||||
|
return (
|
||||||
|
<div className="mx-auto w-full max-w-4xl px-6 py-8">
|
||||||
|
<h1 className="text-3xl underline p-2">Notifications</h1>
|
||||||
|
|
||||||
|
<p className="p-2">
|
||||||
|
All notifications are a subscription based, please open the menu and
|
||||||
|
select the notification you would like to know more info about
|
||||||
|
</p>
|
||||||
|
|
||||||
|
<hr />
|
||||||
|
<p>To subscribe to a notification</p>
|
||||||
|
<ol className="list-decimal list-inside">
|
||||||
|
<li>Click on your profile</li>
|
||||||
|
|
||||||
|
<img
|
||||||
|
src="/lst/app/imgs/docs/notifications/lt_profile.png"
|
||||||
|
alt="Reprint notification example"
|
||||||
|
className="m-2 rounded-lg border-2"
|
||||||
|
/>
|
||||||
|
<li>Click account</li>
|
||||||
|
<li>Select the notification you would like to subscribe to.</li>
|
||||||
|
<img
|
||||||
|
src="/lst/app/imgs/docs/notifications/lt_notification_select.png"
|
||||||
|
alt="Reprint notification example"
|
||||||
|
className="m-2 rounded-lg border-2"
|
||||||
|
/>
|
||||||
|
<li>
|
||||||
|
If you want to have more people on the notification you can add more
|
||||||
|
emails by clicking the add email button.{" "}
|
||||||
|
<p className="text-sm underline">
|
||||||
|
Please note that each user can subscribe on there own so you do not
|
||||||
|
need to add others unless you want to add them.
|
||||||
|
</p>
|
||||||
|
</li>
|
||||||
|
<li>When you are ready click subscribe</li>
|
||||||
|
</ol>
|
||||||
|
<br />
|
||||||
|
<p className="">
|
||||||
|
NOTE: you can select the same notification and add more people or just
|
||||||
|
your self only, when you do this it will override you current
|
||||||
|
subscription and add / remove the emails
|
||||||
|
</p>
|
||||||
|
<hr className="m-2" />
|
||||||
|
<div>
|
||||||
|
<p>
|
||||||
|
The table at the bottom of your profile is where all of your current
|
||||||
|
subscriptions will be at.
|
||||||
|
</p>
|
||||||
|
<p>
|
||||||
|
Clicking the trash can will remove the notifications from sending you
|
||||||
|
emails
|
||||||
|
</p>
|
||||||
|
<img
|
||||||
|
src="/lst/app/imgs/docs/notifications/lt_notification_table.png"
|
||||||
|
alt="Reprint notification example"
|
||||||
|
className="m-2 rounded-lg border-2"
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
19
frontend/src/docs/notifications/qualityBlocking.tsx
Normal file
19
frontend/src/docs/notifications/qualityBlocking.tsx
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
export default function reprints() {
|
||||||
|
return (
|
||||||
|
<div className="mx-auto w-full max-w-4xl px-6 py-8">
|
||||||
|
<h1 className="text-3xl underline p-2">Quality Blocking</h1>
|
||||||
|
|
||||||
|
<p className="p-2">
|
||||||
|
When a new blocking order is created a new alert will be sent out to all
|
||||||
|
users subscribed. if there are multiple blocking orders created between
|
||||||
|
checks you can expect to get multiple emails. below you will see an
|
||||||
|
example of a blocking email that is sent out
|
||||||
|
</p>
|
||||||
|
<img
|
||||||
|
src="/lst/app/imgs/docs/notifications/lt_qualityBlocking.png"
|
||||||
|
alt="Reprint notification example"
|
||||||
|
className="m-2 rounded-lg border-2"
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
18
frontend/src/docs/notifications/reprints.tsx
Normal file
18
frontend/src/docs/notifications/reprints.tsx
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
export default function reprints() {
|
||||||
|
return (
|
||||||
|
<div className="mx-auto w-full max-w-4xl px-6 py-8">
|
||||||
|
<h1 className="text-3xl underline p-2">Reprints</h1>
|
||||||
|
|
||||||
|
<p className="p-2">
|
||||||
|
The reprint alert will monitor for labels that have been printed within
|
||||||
|
a defined time. when a label is printed in the defined time an email
|
||||||
|
will sent out that looks similar to the below
|
||||||
|
</p>
|
||||||
|
<img
|
||||||
|
src="/lst/app/imgs/docs/notifications/lt_reprints.png"
|
||||||
|
alt="Reprint notification example"
|
||||||
|
className="m-2 rounded-lg border-2"
|
||||||
|
/>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
26
frontend/src/lib/docs.ts
Normal file
26
frontend/src/lib/docs.ts
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
import type { ComponentType } from "react";
|
||||||
|
|
||||||
|
const modules = import.meta.glob("../docs/**/*.tsx", {
|
||||||
|
eager: true,
|
||||||
|
});
|
||||||
|
|
||||||
|
type DocModule = {
|
||||||
|
default: ComponentType;
|
||||||
|
};
|
||||||
|
|
||||||
|
const docsMap: Record<string, ComponentType> = {};
|
||||||
|
|
||||||
|
for (const path in modules) {
|
||||||
|
const mod = modules[path] as DocModule;
|
||||||
|
|
||||||
|
const slug = path
|
||||||
|
.replace("../docs/", "")
|
||||||
|
.replace(".tsx", "");
|
||||||
|
|
||||||
|
// "notifications/intro"
|
||||||
|
docsMap[slug] = mod.default;
|
||||||
|
}
|
||||||
|
|
||||||
|
export function getDoc(slug: string) {
|
||||||
|
return docsMap[slug];
|
||||||
|
}
|
||||||
@@ -11,6 +11,8 @@
|
|||||||
import { Route as rootRouteImport } from './routes/__root'
|
import { Route as rootRouteImport } from './routes/__root'
|
||||||
import { Route as AboutRouteImport } from './routes/about'
|
import { Route as AboutRouteImport } from './routes/about'
|
||||||
import { Route as IndexRouteImport } from './routes/index'
|
import { Route as IndexRouteImport } from './routes/index'
|
||||||
|
import { Route as DocsIndexRouteImport } from './routes/docs/index'
|
||||||
|
import { Route as DocsSplatRouteImport } from './routes/docs/$'
|
||||||
import { Route as AdminSettingsRouteImport } from './routes/admin/settings'
|
import { Route as AdminSettingsRouteImport } from './routes/admin/settings'
|
||||||
import { Route as AdminNotificationsRouteImport } from './routes/admin/notifications'
|
import { Route as AdminNotificationsRouteImport } from './routes/admin/notifications'
|
||||||
import { Route as AdminLogsRouteImport } from './routes/admin/logs'
|
import { Route as AdminLogsRouteImport } from './routes/admin/logs'
|
||||||
@@ -29,6 +31,16 @@ const IndexRoute = IndexRouteImport.update({
|
|||||||
path: '/',
|
path: '/',
|
||||||
getParentRoute: () => rootRouteImport,
|
getParentRoute: () => rootRouteImport,
|
||||||
} as any)
|
} as any)
|
||||||
|
const DocsIndexRoute = DocsIndexRouteImport.update({
|
||||||
|
id: '/docs/',
|
||||||
|
path: '/docs/',
|
||||||
|
getParentRoute: () => rootRouteImport,
|
||||||
|
} as any)
|
||||||
|
const DocsSplatRoute = DocsSplatRouteImport.update({
|
||||||
|
id: '/docs/$',
|
||||||
|
path: '/docs/$',
|
||||||
|
getParentRoute: () => rootRouteImport,
|
||||||
|
} as any)
|
||||||
const AdminSettingsRoute = AdminSettingsRouteImport.update({
|
const AdminSettingsRoute = AdminSettingsRouteImport.update({
|
||||||
id: '/admin/settings',
|
id: '/admin/settings',
|
||||||
path: '/admin/settings',
|
path: '/admin/settings',
|
||||||
@@ -72,6 +84,8 @@ export interface FileRoutesByFullPath {
|
|||||||
'/admin/logs': typeof AdminLogsRoute
|
'/admin/logs': typeof AdminLogsRoute
|
||||||
'/admin/notifications': typeof AdminNotificationsRoute
|
'/admin/notifications': typeof AdminNotificationsRoute
|
||||||
'/admin/settings': typeof AdminSettingsRoute
|
'/admin/settings': typeof AdminSettingsRoute
|
||||||
|
'/docs/$': typeof DocsSplatRoute
|
||||||
|
'/docs/': typeof DocsIndexRoute
|
||||||
'/user/profile': typeof authUserProfileRoute
|
'/user/profile': typeof authUserProfileRoute
|
||||||
'/user/resetpassword': typeof authUserResetpasswordRoute
|
'/user/resetpassword': typeof authUserResetpasswordRoute
|
||||||
'/user/signup': typeof authUserSignupRoute
|
'/user/signup': typeof authUserSignupRoute
|
||||||
@@ -83,6 +97,8 @@ export interface FileRoutesByTo {
|
|||||||
'/admin/logs': typeof AdminLogsRoute
|
'/admin/logs': typeof AdminLogsRoute
|
||||||
'/admin/notifications': typeof AdminNotificationsRoute
|
'/admin/notifications': typeof AdminNotificationsRoute
|
||||||
'/admin/settings': typeof AdminSettingsRoute
|
'/admin/settings': typeof AdminSettingsRoute
|
||||||
|
'/docs/$': typeof DocsSplatRoute
|
||||||
|
'/docs': typeof DocsIndexRoute
|
||||||
'/user/profile': typeof authUserProfileRoute
|
'/user/profile': typeof authUserProfileRoute
|
||||||
'/user/resetpassword': typeof authUserResetpasswordRoute
|
'/user/resetpassword': typeof authUserResetpasswordRoute
|
||||||
'/user/signup': typeof authUserSignupRoute
|
'/user/signup': typeof authUserSignupRoute
|
||||||
@@ -95,6 +111,8 @@ export interface FileRoutesById {
|
|||||||
'/admin/logs': typeof AdminLogsRoute
|
'/admin/logs': typeof AdminLogsRoute
|
||||||
'/admin/notifications': typeof AdminNotificationsRoute
|
'/admin/notifications': typeof AdminNotificationsRoute
|
||||||
'/admin/settings': typeof AdminSettingsRoute
|
'/admin/settings': typeof AdminSettingsRoute
|
||||||
|
'/docs/$': typeof DocsSplatRoute
|
||||||
|
'/docs/': typeof DocsIndexRoute
|
||||||
'/(auth)/user/profile': typeof authUserProfileRoute
|
'/(auth)/user/profile': typeof authUserProfileRoute
|
||||||
'/(auth)/user/resetpassword': typeof authUserResetpasswordRoute
|
'/(auth)/user/resetpassword': typeof authUserResetpasswordRoute
|
||||||
'/(auth)/user/signup': typeof authUserSignupRoute
|
'/(auth)/user/signup': typeof authUserSignupRoute
|
||||||
@@ -108,6 +126,8 @@ export interface FileRouteTypes {
|
|||||||
| '/admin/logs'
|
| '/admin/logs'
|
||||||
| '/admin/notifications'
|
| '/admin/notifications'
|
||||||
| '/admin/settings'
|
| '/admin/settings'
|
||||||
|
| '/docs/$'
|
||||||
|
| '/docs/'
|
||||||
| '/user/profile'
|
| '/user/profile'
|
||||||
| '/user/resetpassword'
|
| '/user/resetpassword'
|
||||||
| '/user/signup'
|
| '/user/signup'
|
||||||
@@ -119,6 +139,8 @@ export interface FileRouteTypes {
|
|||||||
| '/admin/logs'
|
| '/admin/logs'
|
||||||
| '/admin/notifications'
|
| '/admin/notifications'
|
||||||
| '/admin/settings'
|
| '/admin/settings'
|
||||||
|
| '/docs/$'
|
||||||
|
| '/docs'
|
||||||
| '/user/profile'
|
| '/user/profile'
|
||||||
| '/user/resetpassword'
|
| '/user/resetpassword'
|
||||||
| '/user/signup'
|
| '/user/signup'
|
||||||
@@ -130,6 +152,8 @@ export interface FileRouteTypes {
|
|||||||
| '/admin/logs'
|
| '/admin/logs'
|
||||||
| '/admin/notifications'
|
| '/admin/notifications'
|
||||||
| '/admin/settings'
|
| '/admin/settings'
|
||||||
|
| '/docs/$'
|
||||||
|
| '/docs/'
|
||||||
| '/(auth)/user/profile'
|
| '/(auth)/user/profile'
|
||||||
| '/(auth)/user/resetpassword'
|
| '/(auth)/user/resetpassword'
|
||||||
| '/(auth)/user/signup'
|
| '/(auth)/user/signup'
|
||||||
@@ -142,6 +166,8 @@ export interface RootRouteChildren {
|
|||||||
AdminLogsRoute: typeof AdminLogsRoute
|
AdminLogsRoute: typeof AdminLogsRoute
|
||||||
AdminNotificationsRoute: typeof AdminNotificationsRoute
|
AdminNotificationsRoute: typeof AdminNotificationsRoute
|
||||||
AdminSettingsRoute: typeof AdminSettingsRoute
|
AdminSettingsRoute: typeof AdminSettingsRoute
|
||||||
|
DocsSplatRoute: typeof DocsSplatRoute
|
||||||
|
DocsIndexRoute: typeof DocsIndexRoute
|
||||||
authUserProfileRoute: typeof authUserProfileRoute
|
authUserProfileRoute: typeof authUserProfileRoute
|
||||||
authUserResetpasswordRoute: typeof authUserResetpasswordRoute
|
authUserResetpasswordRoute: typeof authUserResetpasswordRoute
|
||||||
authUserSignupRoute: typeof authUserSignupRoute
|
authUserSignupRoute: typeof authUserSignupRoute
|
||||||
@@ -163,6 +189,20 @@ declare module '@tanstack/react-router' {
|
|||||||
preLoaderRoute: typeof IndexRouteImport
|
preLoaderRoute: typeof IndexRouteImport
|
||||||
parentRoute: typeof rootRouteImport
|
parentRoute: typeof rootRouteImport
|
||||||
}
|
}
|
||||||
|
'/docs/': {
|
||||||
|
id: '/docs/'
|
||||||
|
path: '/docs'
|
||||||
|
fullPath: '/docs/'
|
||||||
|
preLoaderRoute: typeof DocsIndexRouteImport
|
||||||
|
parentRoute: typeof rootRouteImport
|
||||||
|
}
|
||||||
|
'/docs/$': {
|
||||||
|
id: '/docs/$'
|
||||||
|
path: '/docs/$'
|
||||||
|
fullPath: '/docs/$'
|
||||||
|
preLoaderRoute: typeof DocsSplatRouteImport
|
||||||
|
parentRoute: typeof rootRouteImport
|
||||||
|
}
|
||||||
'/admin/settings': {
|
'/admin/settings': {
|
||||||
id: '/admin/settings'
|
id: '/admin/settings'
|
||||||
path: '/admin/settings'
|
path: '/admin/settings'
|
||||||
@@ -222,6 +262,8 @@ const rootRouteChildren: RootRouteChildren = {
|
|||||||
AdminLogsRoute: AdminLogsRoute,
|
AdminLogsRoute: AdminLogsRoute,
|
||||||
AdminNotificationsRoute: AdminNotificationsRoute,
|
AdminNotificationsRoute: AdminNotificationsRoute,
|
||||||
AdminSettingsRoute: AdminSettingsRoute,
|
AdminSettingsRoute: AdminSettingsRoute,
|
||||||
|
DocsSplatRoute: DocsSplatRoute,
|
||||||
|
DocsIndexRoute: DocsIndexRoute,
|
||||||
authUserProfileRoute: authUserProfileRoute,
|
authUserProfileRoute: authUserProfileRoute,
|
||||||
authUserResetpasswordRoute: authUserResetpasswordRoute,
|
authUserResetpasswordRoute: authUserResetpasswordRoute,
|
||||||
authUserSignupRoute: authUserSignupRoute,
|
authUserSignupRoute: authUserSignupRoute,
|
||||||
|
|||||||
31
frontend/src/routes/docs/$.tsx
Normal file
31
frontend/src/routes/docs/$.tsx
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
import { createFileRoute, Link } from "@tanstack/react-router";
|
||||||
|
import { getDoc } from "../../lib/docs";
|
||||||
|
|
||||||
|
export const Route = createFileRoute("/docs/$")({
|
||||||
|
component: RouteComponent,
|
||||||
|
});
|
||||||
|
|
||||||
|
function RouteComponent() {
|
||||||
|
const { _splat } = Route.useParams();
|
||||||
|
const slug = _splat || "";
|
||||||
|
|
||||||
|
const Doc = getDoc(slug);
|
||||||
|
|
||||||
|
if (!Doc) {
|
||||||
|
return (
|
||||||
|
<div>
|
||||||
|
<p>
|
||||||
|
You Have reached a doc page that dose not seem to exist please
|
||||||
|
validate and come back
|
||||||
|
</p>
|
||||||
|
<Link to="/docs">Docs Home</Link>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
return (
|
||||||
|
<div className="mx-auto w-full max-w-4xl px-6 py-8">
|
||||||
|
<Doc />
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
100
frontend/src/routes/docs/index.tsx
Normal file
100
frontend/src/routes/docs/index.tsx
Normal file
@@ -0,0 +1,100 @@
|
|||||||
|
import { createFileRoute, Link } from "@tanstack/react-router";
|
||||||
|
|
||||||
|
export const Route = createFileRoute("/docs/")({
|
||||||
|
component: RouteComponent,
|
||||||
|
});
|
||||||
|
|
||||||
|
function RouteComponent() {
|
||||||
|
return (
|
||||||
|
<div className="mx-auto w-full max-w-4xl px-6 py-8">
|
||||||
|
<h1 className="text-3xl underline p-2">Logistics Support Tool Intro</h1>
|
||||||
|
<h2 className="text-2xl shadow-2xl p-2">What is lst</h2>
|
||||||
|
<p className="p-2">
|
||||||
|
Lst is a logistics support tool, and aid to ALPLAprod All data in here
|
||||||
|
is just to be treated as an aid and can still be completed manually in
|
||||||
|
alplaprod. These docs are here to help show what LST has to offer as
|
||||||
|
well as the manual process via alpla prod.
|
||||||
|
</p>
|
||||||
|
<hr />
|
||||||
|
<h2 className="text-2xl shadow-2xl p-2">What dose LST offer</h2>
|
||||||
|
<ul className="list-disc list-inside">
|
||||||
|
<li>One click print</li>
|
||||||
|
<ul className="list-disc list-inside indent-8">
|
||||||
|
<li>Controls printing of labels</li>
|
||||||
|
<li>devices that can be used</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
<li>Printer control</li>
|
||||||
|
<li>plc control</li>
|
||||||
|
<li>ame palletizer control</li>
|
||||||
|
</ul>
|
||||||
|
<li>considers more business logic than alplaprod</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
<li>
|
||||||
|
enough material is needed in the system to create the next pallet
|
||||||
|
</li>
|
||||||
|
<li>this will be the same for packaging as well.</li>
|
||||||
|
</ul>
|
||||||
|
<li>special processes</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
<li>in-house delivery triggered once booked in</li>
|
||||||
|
<li>stop gap on printing labels at specific times</li>
|
||||||
|
<li>per line delay in printing</li>
|
||||||
|
</ul>
|
||||||
|
</ul>
|
||||||
|
<li>Silos Management</li>
|
||||||
|
<ul className="list-disc list-inside indent-8">
|
||||||
|
<li>Silo adjustments per location</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
<li>Charts for the last 10 adjustments</li>
|
||||||
|
<li>Historical data</li>
|
||||||
|
<li>Comments on per adjustment</li>
|
||||||
|
<li>Automatic email for more than 5% deviation</li>
|
||||||
|
</ul>
|
||||||
|
<li>Attach silo</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
<li>Only shows machines not attached to this silo</li>
|
||||||
|
</ul>
|
||||||
|
<li>Detach silo</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
Only shows machines that are attached to the silo.
|
||||||
|
</ul>
|
||||||
|
</ul>
|
||||||
|
<li>TMS integration</li>
|
||||||
|
<ul className="list-disc list-inside indent-8">
|
||||||
|
<li>integration with TI to auto add in orders</li>
|
||||||
|
<ul className="list-disc list-inside indent-16">
|
||||||
|
<li>orders are based on a time defined per plant.</li>
|
||||||
|
<li>carriers can be auto set.</li>
|
||||||
|
</ul>
|
||||||
|
</ul>
|
||||||
|
<li>
|
||||||
|
<Link
|
||||||
|
to={"/docs/$"}
|
||||||
|
params={{ _splat: "notifications/intro" }}
|
||||||
|
className="underline"
|
||||||
|
>
|
||||||
|
Notifications
|
||||||
|
</Link>
|
||||||
|
</li>
|
||||||
|
<ul className="list-disc list-inside indent-8">
|
||||||
|
<li>Automated alerts</li>
|
||||||
|
<li>Subscription based</li>
|
||||||
|
<li>Processes notifications</li>
|
||||||
|
</ul>
|
||||||
|
<li>Datamart</li>
|
||||||
|
<ul className="list-disc list-inside indent-8">
|
||||||
|
<li>queries that can be pulled via excel</li>
|
||||||
|
<li>queries are created to allow better views for the plants</li>
|
||||||
|
<li>Faster customer reports</li>
|
||||||
|
</ul>
|
||||||
|
<li>Fake EDI (Demand Management)</li>
|
||||||
|
<ul className="list-disc list-inside indent-8">
|
||||||
|
<li>Orders in (standard template)</li>
|
||||||
|
<li>Customer specific orders templates per plant</li>
|
||||||
|
<li>Forecast (standard Template)</li>
|
||||||
|
<li>Customer specific forecast per plant</li>
|
||||||
|
</ul>
|
||||||
|
</ul>
|
||||||
|
</div>
|
||||||
|
);
|
||||||
|
}
|
||||||
@@ -4,6 +4,7 @@ import { tanstackRouter } from "@tanstack/router-plugin/vite";
|
|||||||
import react from "@vitejs/plugin-react-swc";
|
import react from "@vitejs/plugin-react-swc";
|
||||||
import { defineConfig } from "vite";
|
import { defineConfig } from "vite";
|
||||||
|
|
||||||
|
|
||||||
// https://vite.dev/config/
|
// https://vite.dev/config/
|
||||||
export default defineConfig({
|
export default defineConfig({
|
||||||
plugins: [
|
plugins: [
|
||||||
|
|||||||
43
lstMobile/.gitignore
vendored
Normal file
43
lstMobile/.gitignore
vendored
Normal file
@@ -0,0 +1,43 @@
|
|||||||
|
# Learn more https://docs.github.com/en/get-started/getting-started-with-git/ignoring-files
|
||||||
|
|
||||||
|
# dependencies
|
||||||
|
node_modules/
|
||||||
|
|
||||||
|
# Expo
|
||||||
|
.expo/
|
||||||
|
dist/
|
||||||
|
web-build/
|
||||||
|
expo-env.d.ts
|
||||||
|
|
||||||
|
# Native
|
||||||
|
.kotlin/
|
||||||
|
*.orig.*
|
||||||
|
*.jks
|
||||||
|
*.p8
|
||||||
|
*.p12
|
||||||
|
*.key
|
||||||
|
*.mobileprovision
|
||||||
|
|
||||||
|
# Metro
|
||||||
|
.metro-health-check*
|
||||||
|
|
||||||
|
# debug
|
||||||
|
npm-debug.*
|
||||||
|
yarn-debug.*
|
||||||
|
yarn-error.*
|
||||||
|
|
||||||
|
# macOS
|
||||||
|
.DS_Store
|
||||||
|
*.pem
|
||||||
|
|
||||||
|
# local env files
|
||||||
|
.env*.local
|
||||||
|
|
||||||
|
# typescript
|
||||||
|
*.tsbuildinfo
|
||||||
|
|
||||||
|
app-example
|
||||||
|
|
||||||
|
# generated native folders
|
||||||
|
/ios
|
||||||
|
/android
|
||||||
1
lstMobile/.vscode/extensions.json
vendored
Normal file
1
lstMobile/.vscode/extensions.json
vendored
Normal file
@@ -0,0 +1 @@
|
|||||||
|
{ "recommendations": ["expo.vscode-expo-tools"] }
|
||||||
7
lstMobile/.vscode/settings.json
vendored
Normal file
7
lstMobile/.vscode/settings.json
vendored
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
{
|
||||||
|
"editor.codeActionsOnSave": {
|
||||||
|
"source.fixAll": "explicit",
|
||||||
|
"source.organizeImports": "explicit",
|
||||||
|
"source.sortMembers": "explicit"
|
||||||
|
}
|
||||||
|
}
|
||||||
56
lstMobile/README.md
Normal file
56
lstMobile/README.md
Normal file
@@ -0,0 +1,56 @@
|
|||||||
|
# Welcome to your Expo app 👋
|
||||||
|
|
||||||
|
This is an [Expo](https://expo.dev) project created with [`create-expo-app`](https://www.npmjs.com/package/create-expo-app).
|
||||||
|
|
||||||
|
## Get started
|
||||||
|
|
||||||
|
1. Install dependencies
|
||||||
|
|
||||||
|
```bash
|
||||||
|
npm install
|
||||||
|
```
|
||||||
|
|
||||||
|
2. Start the app
|
||||||
|
|
||||||
|
```bash
|
||||||
|
npx expo start
|
||||||
|
```
|
||||||
|
|
||||||
|
In the output, you'll find options to open the app in a
|
||||||
|
|
||||||
|
- [development build](https://docs.expo.dev/develop/development-builds/introduction/)
|
||||||
|
- [Android emulator](https://docs.expo.dev/workflow/android-studio-emulator/)
|
||||||
|
- [iOS simulator](https://docs.expo.dev/workflow/ios-simulator/)
|
||||||
|
- [Expo Go](https://expo.dev/go), a limited sandbox for trying out app development with Expo
|
||||||
|
|
||||||
|
You can start developing by editing the files inside the **app** directory. This project uses [file-based routing](https://docs.expo.dev/router/introduction).
|
||||||
|
|
||||||
|
## Get a fresh project
|
||||||
|
|
||||||
|
When you're ready, run:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
npm run reset-project
|
||||||
|
```
|
||||||
|
|
||||||
|
This command will move the starter code to the **app-example** directory and create a blank **app** directory where you can start developing.
|
||||||
|
|
||||||
|
### Other setup steps
|
||||||
|
|
||||||
|
- To set up ESLint for linting, run `npx expo lint`, or follow our guide on ["Using ESLint and Prettier"](https://docs.expo.dev/guides/using-eslint/)
|
||||||
|
- If you'd like to set up unit testing, follow our guide on ["Unit Testing with Jest"](https://docs.expo.dev/develop/unit-testing/)
|
||||||
|
- Learn more about the TypeScript setup in this template in our guide on ["Using TypeScript"](https://docs.expo.dev/guides/typescript/)
|
||||||
|
|
||||||
|
## Learn more
|
||||||
|
|
||||||
|
To learn more about developing your project with Expo, look at the following resources:
|
||||||
|
|
||||||
|
- [Expo documentation](https://docs.expo.dev/): Learn fundamentals, or go into advanced topics with our [guides](https://docs.expo.dev/guides).
|
||||||
|
- [Learn Expo tutorial](https://docs.expo.dev/tutorial/introduction/): Follow a step-by-step tutorial where you'll create a project that runs on Android, iOS, and the web.
|
||||||
|
|
||||||
|
## Join the community
|
||||||
|
|
||||||
|
Join our community of developers creating universal apps.
|
||||||
|
|
||||||
|
- [Expo on GitHub](https://github.com/expo/expo): View our open source platform and contribute.
|
||||||
|
- [Discord community](https://chat.expo.dev): Chat with Expo users and ask questions.
|
||||||
47
lstMobile/app.json
Normal file
47
lstMobile/app.json
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
{
|
||||||
|
"expo": {
|
||||||
|
"name": "LST mobile",
|
||||||
|
"slug": "lst-mobile",
|
||||||
|
"version": "0.0.1-alpha",
|
||||||
|
"orientation": "portrait",
|
||||||
|
"icon": "./assets/images/icon.png",
|
||||||
|
"scheme": "lstmobile",
|
||||||
|
"userInterfaceStyle": "automatic",
|
||||||
|
"ios": {
|
||||||
|
"icon": "./assets/expo.icon"
|
||||||
|
},
|
||||||
|
"android": {
|
||||||
|
"adaptiveIcon": {
|
||||||
|
"backgroundColor": "#E6F4FE",
|
||||||
|
"foregroundImage": "./assets/images/android-icon-foreground.png",
|
||||||
|
"backgroundImage": "./assets/images/android-icon-background.png",
|
||||||
|
"monochromeImage": "./assets/images/android-icon-monochrome.png",
|
||||||
|
"package": "net.alpla.lst.mobile",
|
||||||
|
"versionCode": 1
|
||||||
|
},
|
||||||
|
"predictiveBackGestureEnabled": false,
|
||||||
|
"package": "com.anonymous.lstMobile"
|
||||||
|
},
|
||||||
|
"web": {
|
||||||
|
"output": "static",
|
||||||
|
"favicon": "./assets/images/favicon.png"
|
||||||
|
},
|
||||||
|
"plugins": [
|
||||||
|
"expo-router",
|
||||||
|
[
|
||||||
|
"expo-splash-screen",
|
||||||
|
{
|
||||||
|
"backgroundColor": "#208AEF",
|
||||||
|
"android": {
|
||||||
|
"image": "./assets/images/splash-icon.png",
|
||||||
|
"imageWidth": 76
|
||||||
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
|
],
|
||||||
|
"experiments": {
|
||||||
|
"typedRoutes": true,
|
||||||
|
"reactCompiler": true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user