diff --git a/.env.example b/.env.example
index ff2b0a0..b36254e 100644
--- a/.env.example
+++ b/.env.example
@@ -15,3 +15,5 @@ REDIS_CONNECTION=local
REDIS_HOST=127.0.0.1
REDIS_PORT=6379
REDIS_PASSWORD=
+BASE_URL_LOG="http://172.16.5.7:8080"
+FOLDER_LOGS="/logs"
diff --git a/.gitignore b/.gitignore
index d84c174..0335270 100644
--- a/.gitignore
+++ b/.gitignore
@@ -8,3 +8,4 @@ app/utils/indexSN.txt
app/store
tmp
.npm
+logs
diff --git a/README.md b/README.md
index e69de29..46d6c9b 100644
--- a/README.md
+++ b/README.md
@@ -0,0 +1,9 @@
+# LOG SERVICE
+
+### 1. Copy .env.example -> .env
+
+### 2. npm install
+
+### 3. node ace migration:run
+
+### 4. npm run dev
diff --git a/app/Controllers/Http/ErpsController.ts b/app/Controllers/Http/ErpsController.ts
index aac4b28..8012a5b 100644
--- a/app/Controllers/Http/ErpsController.ts
+++ b/app/Controllers/Http/ErpsController.ts
@@ -1,6 +1,11 @@
import fs from "fs";
import type { HttpContextContract } from "@ioc:Adonis/Core/HttpContext";
import axios from "axios";
+import Env from "@ioc:Adonis/Core/Env";
+const path = require("path");
+
+const BASE_URL = Env.get("BASE_URL_LOG");
+const BASE_URL_AUTO = Env.get(`${BASE_URL}/AUTOlog/`);
export default class ErpsController {
/**
@@ -17,10 +22,8 @@ export default class ErpsController {
try {
// console.log("check!")
const listLog: string[] = [];
- const response = await axios.get("http://172.16.5.7:8080/");
- const responseAUTO = await axios.get(
- "http://172.16.5.7:8080/AUTOlog/"
- );
+ const response = await axios.get(BASE_URL);
+ const responseAUTO = await axios.get(BASE_URL_AUTO);
let data1 = response.data
.split("\n")
@@ -49,8 +52,8 @@ export default class ErpsController {
.slice(u.search(""))
.split(">")[1]
.includes("AUTO")
- ? "http://172.16.5.7:8080/AUTOlog/"
- : "http://172.16.5.7:8080/") +
+ ? BASE_URL_AUTO
+ : BASE_URL + "/") +
u.slice(u.search("")).split(">")[1] +
" "
);
@@ -63,235 +66,156 @@ export default class ErpsController {
}
};
+ const fetchWithRetry = async (url) => {
+ let retries = 0;
+ const MAX_RETRIES = 10;
+ while (retries < MAX_RETRIES) {
+ try {
+ const response = await axios.get(url);
+ return response.data;
+ } catch (error) {
+ if (error.code) {
+ retries++;
+ continue;
+ } else {
+ console.error("Error fetching file:", error);
+ return null;
+ }
+ }
+ }
+ return null;
+ };
+
+ const extractInfoFromLine = (line, index, fName) => {
+ const cleanedLine = line.replace(/\r/g, "");
+ const results: any = [];
+
+ const getWarehouse = (fileName) =>
+ /(US(?!A)|-US|\.US|US-)/.test(fileName) ? "US" : "AU";
+
+ const extractSN = (str) =>
+ str?.replace(/[!@#$%^&*()_+{}\[\]:;<>,.?~\\/]/g, "").trim();
+
+ // PID & SN
+ if (
+ cleanedLine.includes("PID:") &&
+ cleanedLine.includes("SN:") &&
+ !cleanedLine.includes("%")
+ ) {
+ const parts = cleanedLine.split(",");
+ const SN = extractSN(
+ parts.find((p) => p.includes("SN:"))?.split(":")[1] ?? ""
+ );
+ const PID = extractSN(
+ parts.find((p) => p.includes("PID:"))?.split(":")[1] ?? ""
+ );
+ const VID = extractSN(
+ parts.find((p) => p.includes("VID:"))?.split(":")[1] ?? ""
+ );
+
+ if (SN && SN !== "N/A" && SN.length > 4) {
+ results.push({
+ PID,
+ VID,
+ SN,
+ line: [index + 1],
+ fileName: fName,
+ warehouse: getWarehouse(fName),
+ });
+ }
+ }
+
+ // Serial Number
+ if (cleanedLine.includes("Serial Number")) {
+ const PCB_SN = extractSN(cleanedLine.split(":")[1]);
+ if (PCB_SN) {
+ results.push({
+ PID: "",
+ VID: "",
+ SN: PCB_SN,
+ line: [index + 1],
+ fileName: fName,
+ warehouse: getWarehouse(fName),
+ });
+ }
+ }
+
+ // Processor board ID
+ if (cleanedLine.includes("Processor board ID")) {
+ const PBID = extractSN(cleanedLine.split(" ").pop());
+ if (PBID?.length >= 8) {
+ results.push({
+ PID: "",
+ VID: "",
+ SN: PBID,
+ line: [index + 1],
+ fileName: fName,
+ warehouse: getWarehouse(fName),
+ });
+ }
+ }
+
+ return results;
+ };
+
+ const mergeLines = (output, lineResult) => {
+ lineResult.forEach((entry) => {
+ const existing = output.find((o) => o.SN === entry.SN);
+ if (existing) {
+ existing.line = [...new Set([...existing.line, ...entry.line])];
+ if (entry.PID) existing.PID = entry.PID;
+ if (entry.VID) existing.VID = entry.VID;
+ } else {
+ output.push(entry);
+ }
+ });
+ };
+
const fetchFiles = async (from, to) => {
try {
const urls = await getListLog(from, to);
- let report = [];
+ if (!urls || urls.length === 0) {
+ console.log("No logs found");
+ return [];
+ }
const fileContents = await Promise.all(
- urls.map(async (url) => {
- const maxRetries = 10;
- let retries = 0;
- while (retries < maxRetries) {
- try {
- const response = await axios.get(url?.split(" ")[0]);
- return response.data;
- } catch (error) {
- if (error.code !== "") {
- //=== "ETIMEDOUT" || error.code === "ECONNRESET"
- // console.log("Connection timed out. Retrying...");
- retries++;
- } else {
- console.error("Error fetching file:", error);
- return;
- }
- }
- }
- })
+ urls.map((u) => fetchWithRetry(u?.split(" ")[0]))
);
- // Handle the file contents
+ let report: any = [];
+
fileContents.forEach((content, index) => {
- console.log(`Content of file ${index + 1}:`);
- const arrayLine = content?.split("\n");
- let output = [];
- if (arrayLine !== undefined) {
- for (let i = 0; i < arrayLine.length; i++) {
- let SN = arrayLine[i]
- ?.split("SN:")[1]
- ?.trim()
- .replace(/[!@#$%^&*()_+{}\[\]:;<>,.?~\\/]/g, "");
- if (
- arrayLine[i].search("PID:") !== -1 &&
- arrayLine[i].search("SN:") !== -1 &&
- arrayLine[i].search("%") === -1 &&
- arrayLine[i]
- ?.split(",")[2]
- ?.split(":")[1]
- ?.replace("\r", "")
- .trim() !== "" &&
- SN !== "N/A" && SN.length>4
- ) {
- if (output.some((u) => u.SN === SN)) {
- output.map((u, index) => {
- if (u.SN === SN) {
- output[index].PID =
- arrayLine[i]?.split("VID:")[0] !== undefined
- ? arrayLine[i]
- ?.split("VID:")[0]
- ?.slice(
- arrayLine[i]?.split("VID:")[0]?.search("PID")
- )
- ?.split(":")[1]
- ?.split(",")[0]
- ?.trim()
- : "";
- (output[index].VID =
- arrayLine[i]?.split("SN:")[0] !== undefined
- ? arrayLine[i]
- ?.split("SN:")[0]
- ?.split("VID:")[1]
- ?.split(",")[0]
- ?.trim()
- : ""),
- (output[index].line = output[index].line.concat([
- i + 1,
- ]));
- }
- });
- } else {
- let fName = urls[index]
- ?.split("/")
- [urls[index]?.split("/")?.length - 1]?.trim();
- output.push({
- PID:
- arrayLine[i]?.split("VID:")[0] !== undefined
- ? arrayLine[i]
- ?.split("VID:")[0]
- ?.slice(
- arrayLine[i]?.split("VID:")[0]?.search("PID")
- )
- ?.split(":")[1]
- ?.split(",")[0]
- ?.trim()
- : "",
- VID:
- arrayLine[i]?.split("SN:")[0] !== undefined
- ? arrayLine[i]
- ?.split("SN:")[0]
- ?.split("VID:")[1]
- ?.split(",")[0]
- ?.trim()
- : "",
- SN:
- arrayLine[i].split("SN:")[1] !== undefined
- ? SN.search(" ") !== -1
- ? SN?.split(" ")[0]
- : SN
- : "",
- line: [i + 1],
- fileName: fName,
- warehouse:
- (fName.search("-US") !== -1 ||
- fName.search(".US") !== -1 ||
- fName.search("US-") !== -1) &&
- fName.search("AUS") === -1
- ? "US"
- : "AU",
- });
- }
- }
+ const lines = content?.split("\n");
+ if (!lines) return;
- if (arrayLine[i].search("Serial Number") !== -1) {
- let PCB_SN = arrayLine[i]
- ?.split("Serial Number")[1]
- .split(":")[1]
- ?.replace("\r", "")
- .trim()
- .replace(/[!@#$%^&*()_+{}\[\]:;<>,.?~\\/]/g, "");
- if (
- //Neu SN da nam trong output
- output.some((u) => u.SN === PCB_SN)
- ) {
- output.map((u, index) => {
- if (u.SN === PCB_SN) {
- output[index].line = output[index].line.concat([i + 1]);
- }
- });
- } else {
- // if (
- // /^[A-Z0-9-]{5,}$/.test(
- // PCB_SN
- // )
- // ) {
- let fName = urls[index]
- ?.split("/")
- [urls[index]?.split("/")?.length - 1]?.trim();
- output.push({
- PID: "",
- VID: "",
- SN:
- PCB_SN?.search(" ") !== -1
- ? PCB_SN?.split(" ")[0]
- : PCB_SN,
- line: [i + 1],
- fileName: fName,
- warehouse:
- (fName.search("-US") !== -1 ||
- fName.search(".US") !== -1 ||
- fName.search("US-") !== -1) &&
- fName.search("AUS") === -1
- ? "US"
- : "AU",
- });
- // }
- }
- }
+ const fName = path.basename(urls[index] ?? "").trim();
+ const output = [];
- if (arrayLine[i].search("Processor board ID") !== -1) {
- let PBID = arrayLine[i]
- ?.split(" ")
- [arrayLine[i]?.split(" ").length - 1]?.replace("\r", "")
- .trim()
- .replace(/[!@#$%^&*()_+{}\[\]:;<>,.?~\\/]/g, "");
- if (
- //Neu SN da nam trong output
- output.some((u) => u.SN === PBID)
- ) {
- output.map((u, index) => {
- if (u.SN === PBID) {
- output[index].line = output[index].line.concat([i + 1]);
- }
- });
- } else {
- if (PBID?.length >= 8) {
- let fName = urls[index]
- ?.split("/")
- [urls[index]?.split("/")?.length - 1]?.trim();
- output.push({
- PID: "",
- VID: "",
- SN: PBID?.search(" ") !== -1 ? PBID?.split(" ")[0] : PBID,
- line: [i + 1],
- fileName: fName,
- warehouse:
- (fName.search("-US") !== -1 ||
- fName.search(".US") !== -1 ||
- fName.search("US-") !== -1) &&
- fName.search("AUS") === -1
- ? "US"
- : "AU",
- });
- }
- }
- }
- }
- report = report.concat(output);
- }
+ lines.forEach((line, i) => {
+ const lineResult = extractInfoFromLine(line, i, fName);
+ mergeLines(output, lineResult);
+ });
+
+ report = [...report, ...output];
});
- fs.writeFile(
+ const finalReport = report.filter(
+ (i) => i.SN && /^[A-Z0-9-]{5,}$/.test(i.SN) && i.PID
+ );
+
+ fs.writeFileSync(
"./app/utils/indexSN.txt",
- JSON.stringify(report)
- .replace(/,{/g, "\n,{")
- .replace(/\\u0000/g, ""),
- function (err) {
- if (err) {
- return console.error(err);
- }
- console.log("Write loggg !");
- }
+ JSON.stringify(finalReport, null, 2).replace(/\\u0000/g, ""),
+ "utf-8"
);
- // console.log(report);
- return report.filter(
- (i) => i.SN !== "" && /^[A-Z0-9-]{5,}$/.test(i.SN) === true && i.PID !==""
- );
- // }, 15000);
+ console.log("Write loggg !");
+ return finalReport;
} catch (error) {
- response
- .status(500)
- .send({ mess: "GET INFORMATION FAIL", error: error });
+ console.error("GET INFORMATION FAIL", error);
+ throw error;
}
};
+
const result = await fetchFiles(from, to);
response.status(200).json(result);
}
@@ -309,7 +233,7 @@ export default class ErpsController {
try {
let fName =
fileName.search("AUTO") !== -1 ? "AUTOlog/" + fileName : fileName;
- const res = await axios.get("http://172.16.5.7:8080/" + fName);
+ const res = await axios.get(BASE_URL + "/" + fName);
const arrayLine = res?.data?.split("\n");
if (range >= line) {
@@ -323,9 +247,7 @@ export default class ErpsController {
}
} catch (error) {
console.log(error);
- response
- .status(202)
- .send({ mess: "FILE NOT FOUND", error: error });
+ response.status(202).send({ mess: "FILE NOT FOUND", error: error });
}
}
diff --git a/app/Controllers/Http/LogsController.ts b/app/Controllers/Http/LogsController.ts
index 8e0a82d..324050b 100644
--- a/app/Controllers/Http/LogsController.ts
+++ b/app/Controllers/Http/LogsController.ts
@@ -1,162 +1,142 @@
-import type { HttpContextContract } from '@ioc:Adonis/Core/HttpContext'
-import Database from '@ioc:Adonis/Lucid/Database';
-import KeyValue from 'App/Models/KeyValue';
-import LogDetectFile from 'App/Models/LogDetectFile';
-import axios from 'axios';
+import type { HttpContextContract } from "@ioc:Adonis/Core/HttpContext";
+import Database from "@ioc:Adonis/Lucid/Database";
+import KeyValue from "App/Models/KeyValue";
+import LogDetectFile from "App/Models/LogDetectFile";
+import axios from "axios";
+import Env from "@ioc:Adonis/Core/Env";
+
+const BASE_URL = Env.get("BASE_URL_LOG");
+
+// Utility function for version detection
+const checkSpecialVersion = (paragraph: string): string => {
+ try {
+ const patterns = [
+ /\(CAT[1-9]K.*Version 16\.9\.[2-9]/,
+ /\(CAT[1-9]K.*Version 1[7-9]\.[0-9]\.[0-9]/,
+ /\(CAT[1-9]K.*Version [2-9][0-9]\.[0-9]\.[0-9]/,
+ ];
+ for (const regex of patterns) {
+ const match = paragraph.match(regex);
+ if (match) return match[0];
+ }
+ return "";
+ } catch {
+ return "";
+ }
+};
export default class LogsController {
public async showLog({ request, response }: HttpContextContract) {
try {
- // fghf
- const checkSpecialVersion = (paragraph) => {
- try {
- const regex = /\(CAT[1-9]K.*Version 16\.9\.[2-9]/;
- const regex1 =
- /\(CAT[1-9]K.*Version 1[7-9]\.[0-9]\.[0-9]/;
- const regex2 =
- /\(CAT[1-9]K.*Version [2-9][0-9]\.[0-9]\.[0-9]/;
- // Use the regular expression to find the match
- const match = paragraph.match(regex);
- const match1 = paragraph.match(regex1);
- const match2 = paragraph.match(regex2);
- if (match || match1 || match2) {
- if (match) {
- return match[0];
- }
+ const fileName = request.params().name;
+ const fileDetect = await LogDetectFile.findBy("file_name", fileName);
- if (match1) {
- return match1[0];
- }
+ if (!fileDetect) {
+ return response.status(203).send("FILE NOT FOUND");
+ }
- if (match2) {
- return match2[0];
- }
- } else {
- return "";
- }
- } catch (error) {
- console.log(error);
- }
- };
-
- let fileDetect = await LogDetectFile.findBy(
- "file_name",
- request.params().name
- );
- let logsDetect = await Database.rawQuery(
- "select * from log_reports where id_file = " + fileDetect?.id_ldf
+ const logsDetect = await Database.rawQuery(
+ "SELECT * FROM log_reports WHERE id_file = ?",
+ [fileDetect.id_ldf]
);
- let modelSpecialDetected = [];
- let issueSpecialDetected = [];
- let listLine = logsDetect[0]
- .map((obj) => obj.line)
- .filter((value, index, self) => {
- return self.indexOf(value) === index;
- })
- .sort((a, b) => a - b);
-
- const content = await axios.get(
- request.params().name.search("AUTO") !== -1
- ? "http://172.16.5.7:8080/AUTOlog/" + request.params().name
- : "http://172.16.5.7:8080/" + request.params().name
- );
- let allValue = await KeyValue.all();
- let listKeyValues = allValue.filter(
- (i) =>
- i.$original.key === "MODEL_SPECIAL" ||
- i.$original.key === "CATCH_FAULTY"
+ const lines = [...new Set(logsDetect[0].map((obj) => obj.line))].sort(
+ (a, b) => a - b
);
- let MODEL_SPECIAL = allValue
- .filter((i) => i.$original.key === "MODEL_SPECIAL")
- .map((obj) => obj.$original.value);
+ const logUrl = fileName.includes("AUTO")
+ ? `${BASE_URL}/AUTOlog/${fileName}`
+ : `${BASE_URL}/${fileName}`;
- let listExcludeErr = allValue
- .filter((i) => i.$original.key === "EXCLUDE_ERR")
- .map((obj) => obj.$original.value);
+ const content = await axios.get(logUrl);
+ const allKeyValues = await KeyValue.all();
- let data = content.data.split("\n");
- data.map(async (line, index) => {
- data[index] = index + 1 + "|-|" + line;
+ const keyValueMap = allKeyValues.reduce((acc, { $original }) => {
+ acc[$original.key] = acc[$original.key] || [];
+ acc[$original.key].push($original.value);
+ return acc;
+ }, {} as Record);
- if (checkSpecialVersion(line) !== "") {
- const specialVersion = checkSpecialVersion(line)
- data[index] =
- data[index].slice(0, data[index].indexOf(specialVersion))
- +
+ const MODEL_SPECIAL = keyValueMap["MODEL_SPECIAL"] || [];
+ const CATCH_FAULTY = keyValueMap["CATCH_FAULTY"] || [];
+ const EXCLUDE_ERR = keyValueMap["EXCLUDE_ERR"] || [];
+
+ const rawData = content.data.split("\n");
+ const processedData: string[] = [];
+ const modelSpecialDetected: string[] = [];
+ const issueSpecialDetected: string[] = [];
+
+ for (let i = 0; i < rawData.length; i++) {
+ let line = rawData[i];
+ let numberedLine = `${i + 1}|-|${line}`;
+ const specialVersion = checkSpecialVersion(line);
+
+ if (specialVersion) {
+ const index = numberedLine.indexOf(specialVersion);
+ numberedLine =
+ numberedLine.slice(0, index) +
"|-|" +
specialVersion +
"|-|" +
- data[index].slice(data[index].indexOf(specialVersion) + specialVersion.length);
+ numberedLine.slice(index + specialVersion.length);
}
- listKeyValues
- .map((obj) => obj.$original.value)
- .map(async (value) => {
- if (
- line.indexOf(value) !== -1 &&
- listExcludeErr.filter((err) => line.includes(err)).length === 0
- ) {
- data[index] =
- data[index].slice(0, data[index].indexOf(value)) +
- "|-|" +
- value +
- "|-|" +
- data[index].slice(data[index].indexOf(value) + value.length);
- // }
- }
- });
- });
-
- listLine.map(async (u) => {
- if (
- listExcludeErr.filter((err) => data[u - 1].includes(err)).length === 0
- ) {
-
- if (MODEL_SPECIAL.filter((i) => data[u - 1].includes(i)).length > 0) {
- modelSpecialDetected.push(data[u - 1]);
- } else {
- if (checkSpecialVersion(data[u - 1]) !== "") {
- modelSpecialDetected.push(data[u - 1]);
- } else {
- issueSpecialDetected.push(data[u - 1]);
- }
+ for (const value of [...MODEL_SPECIAL, ...CATCH_FAULTY]) {
+ if (
+ line.includes(value) &&
+ !EXCLUDE_ERR.some((err) => line.includes(err))
+ ) {
+ const index = numberedLine.indexOf(value);
+ numberedLine =
+ numberedLine.slice(0, index) +
+ "|-|" +
+ value +
+ "|-|" +
+ numberedLine.slice(index + value.length);
+ break;
}
}
- });
- let modelSpecial =
- modelSpecialDetected.length > 0 ? modelSpecialDetected.join("\n") : "";
- let issueItem =
- issueSpecialDetected.length > 0 ? issueSpecialDetected.join("\n") : "";
+
+ processedData.push(numberedLine);
+ }
+
+ for (const lineIndex of lines) {
+ const line = processedData[lineIndex - 1];
+ if (EXCLUDE_ERR.some((err) => line.includes(err))) continue;
+
+ if (
+ MODEL_SPECIAL.some((model) => line.includes(model)) ||
+ checkSpecialVersion(line)
+ ) {
+ modelSpecialDetected.push(line);
+ } else {
+ issueSpecialDetected.push(line);
+ }
+ }
+
response.status(200).send({
- modelSpecial: modelSpecial,
- issueItem: issueItem,
- contentFile: data.join("\n"),
+ modelSpecial: modelSpecialDetected.join("\n"),
+ issueItem: issueSpecialDetected.join("\n"),
+ contentFile: processedData.join("\n"),
});
} catch (error) {
- // console.error(error);
response.status(203).send("FILE NOT FOUND");
}
}
- public async getAllLogDetect({ request, response }: HttpContextContract) {
+ public async getAllLogDetect({ response }: HttpContextContract) {
try {
- let fileDetect = await LogDetectFile.all();
- let listFiles = fileDetect.map((obj) => obj.file_name);
- response.status(200).send(listFiles);
- } catch (error) {
+ const files = await LogDetectFile.all();
+ const fileNames = files.map((file) => file.file_name);
+ response.status(200).send(fileNames);
+ } catch {
response.status(203).send("NO FILE");
}
}
- public async store({ }: HttpContextContract) { }
-
- public async show({ }: HttpContextContract) { }
-
- public async edit({ }: HttpContextContract) { }
-
- public async update({ }: HttpContextContract) { }
-
- public async destroy({ }: HttpContextContract) { }
+ public async store({}: HttpContextContract) {}
+ public async show({}: HttpContextContract) {}
+ public async edit({}: HttpContextContract) {}
+ public async update({}: HttpContextContract) {}
+ public async destroy({}: HttpContextContract) {}
}
diff --git a/app/Controllers/Http/UsersController.ts b/app/Controllers/Http/UsersController.ts
index 1b5f982..8abe983 100644
--- a/app/Controllers/Http/UsersController.ts
+++ b/app/Controllers/Http/UsersController.ts
@@ -23,7 +23,9 @@ export default class UsersController {
const data = await request.validate({ schema: userSchema });
const user = await Users.create(data);
- return response.status(201).send({ mess: "REGISTER_SUCCESS" });
+ return response
+ .status(201)
+ .send({ mess: "REGISTER_SUCCESS", data: user });
} else {
return response.status(203).send({ mess: "USERNAME EXIST" });
}
diff --git a/app/Controllers/Http/ValuesController.ts b/app/Controllers/Http/ValuesController.ts
index 5291e89..bdb6fa5 100644
--- a/app/Controllers/Http/ValuesController.ts
+++ b/app/Controllers/Http/ValuesController.ts
@@ -14,7 +14,7 @@ export default class ValuesController {
value: request.all().value,
model: "All",
});
- response.status(200).send("ADD VALUE SUCCESS!");
+ response.status(200).send({ mess: "ADD VALUE SUCCESS!", data: value });
} catch (error) {
response.status(500).send("ADD VALUE FAIL!");
}
@@ -29,7 +29,7 @@ export default class ValuesController {
const value = await KeyValue.findOrFail(request.all().id);
value.value = request.all().value;
await value.save();
- response.status(200).send("EDIT VALUE SUCCESS!");
+ response.status(200).send({ mess: "EDIT VALUE SUCCESS!", data: value });
} catch (error) {
response.status(500).send("EDIT VALUE FAIL!");
}
diff --git a/app/utils/runtimeCheckLogs.ts b/app/utils/runtimeCheckLogs.ts
index 2aba843..1eea4a4 100644
--- a/app/utils/runtimeCheckLogs.ts
+++ b/app/utils/runtimeCheckLogs.ts
@@ -11,7 +11,7 @@ import { getListLineByItem } from "./getListLineByItem";
import { sendMessToZulip } from "./sendMessToZulip";
export async function runtimeCheckLogs(folderPath) {
try {
- let cacheFile = []
+ let cacheFile: any = [];
let fileList = Array();
let fileList_old = Array();
// Function to update the list of files
@@ -34,34 +34,20 @@ export async function runtimeCheckLogs(folderPath) {
});
}
- const checkSpecialVersion = (paragraph) => {
+ const checkSpecialVersion = (paragraph: string): string => {
try {
- const regex = /\(CAT[1-9]K.*Version 16\.9\.[2-9]/;
- const regex1 =
- /\(CAT[1-9]K.*Version 1[7-9]\.[0-9]\.[0-9]/;
- const regex2 =
- /\(CAT[1-9]K.*Version [2-9][0-9]\.[0-9]\.[0-9]/;
- // Use the regular expression to find the match
- const match = paragraph.match(regex);
- const match1 = paragraph.match(regex1);
- const match2 = paragraph.match(regex2);
- if (match || match1 || match2) {
- if (match) {
- return match[0];
- }
-
- if (match1) {
- return match1[0];
- }
-
- if (match2) {
- return match2[0];
- }
- } else {
- return "";
+ const patterns = [
+ /\(CAT[1-9]K.*Version 16\.9\.[2-9]/,
+ /\(CAT[1-9]K.*Version 1[7-9]\.[0-9]\.[0-9]/,
+ /\(CAT[1-9]K.*Version [2-9][0-9]\.[0-9]\.[0-9]/,
+ ];
+ for (const regex of patterns) {
+ const match = paragraph.match(regex);
+ if (match) return match[0];
}
- } catch (error) {
- console.log(error);
+ return "";
+ } catch {
+ return "";
}
};
// Watch the folder for new files
@@ -73,9 +59,9 @@ export async function runtimeCheckLogs(folderPath) {
if (
filePath?.split(".")[filePath.split(".").length - 1] === "log" &&
filePath.split("/")[filePath.split("/").length - 1]?.split("-")[0] ===
- //localhost
- // filePath.split("\\")[filePath.split("\\").length - 1]?.split("-")[0] ===
- moment(Date.now()).format("YYYYMMDD").toString()
+ //localhost
+ // filePath.split("\\")[filePath.split("\\").length - 1]?.split("-")[0] ===
+ moment(Date.now()).format("YYYYMMDD").toString()
) {
//add information file to database
console.log("New file added: ", filePath);
@@ -102,7 +88,9 @@ export async function runtimeCheckLogs(folderPath) {
try {
//only check new file ---> fileList - fileList_old = new file
let listFileWatch = fileList
- ?.filter((i) => fileList_old.includes(i) === false && i.includes(".log"))
+ ?.filter(
+ (i) => fileList_old.includes(i) === false && i.includes(".log")
+ )
.map((file) => folderPath + "/" + file);
const watcher = chokidar.watch(listFileWatch, {
persistent: true,
@@ -119,7 +107,7 @@ export async function runtimeCheckLogs(folderPath) {
const fileName = path.split("/")[path.split("/").length - 1];
// const fileName = path.split("\\")[path.split("\\").length - 1];
const filePath = path;
- let lines = [];
+ let lines: any = [];
const today = DateTime.now().toFormat("yyyy-MM-dd");
let allFile = await LogDetectFile.query().whereRaw(
`DATE(created_at) = ?`,
@@ -132,7 +120,11 @@ export async function runtimeCheckLogs(folderPath) {
[today]
);
- cacheFile = cacheFile.concat(allFile.filter(obj2 => !cacheFile.some(obj1 => obj1.id_ldf === obj2.id_ldf)));
+ cacheFile = cacheFile.concat(
+ allFile.filter(
+ (obj2) => !cacheFile.some((obj1) => obj1.id_ldf === obj2.id_ldf)
+ )
+ );
//get information file
let fileDetect = allFile?.filter(
@@ -198,13 +190,10 @@ export async function runtimeCheckLogs(folderPath) {
report.$original.detected_content === value
);
- if (
- log?.id_ldf === "" ||
- log?.id_ldf === null ||
- log?.id_ldf === undefined
- ) {
-
- let logBackup = cacheFile.filter((i) => i.file_name === fileName)[0]
+ if (!log?.id_ldf) {
+ let logBackup = cacheFile.filter(
+ (i) => i.file_name === fileName
+ )[0];
if (logBackup !== undefined) {
let checkLog2 = allReport?.filter(
(report) =>
@@ -221,7 +210,7 @@ export async function runtimeCheckLogs(folderPath) {
lines.push(index + 1);
}
} else {
- console.log("ERROR CHECK ", fileName)
+ console.log("ERROR CHECK ", fileName);
}
}
if (checkLog?.length === 0) {
@@ -233,24 +222,19 @@ export async function runtimeCheckLogs(folderPath) {
lines.push(index + 1);
}
}
-
});
if (
checkSpecialVersion(line) !== "" &&
listExcludeErr.filter((err) => line.includes(err)).length ===
- 0
+ 0
) {
let checkVersion = checkSpecialVersion(line);
let log = allFile?.filter(
(i) => i.$original.file_name === fileName
)[0];
- if (
- log?.id_ldf === "" ||
- log?.id_ldf === null ||
- log?.id_ldf === undefined
- ) {
- console.log("ERROR CHECK ", fileName)
+ if (!log?.id_ldf) {
+ console.log("ERROR CHECK ", fileName);
}
let checkLog = allReport?.filter(
(report) =>
@@ -276,7 +260,6 @@ export async function runtimeCheckLogs(folderPath) {
listKeyValues
.map((obj) => obj.$original.value)
.map(async (value) => {
-
if (
line.includes(value) &&
listExcludeErr.filter((err) => line.includes(err))
@@ -285,12 +268,8 @@ export async function runtimeCheckLogs(folderPath) {
let log = allFile?.filter(
(i) => i.$original.file_name === fileName
)[0];
- if (
- log?.id_ldf === "" ||
- log?.id_ldf === null ||
- log?.id_ldf === undefined
- ) {
- console.log("ERROR CHECK ", fileName)
+ if (!log?.id_ldf) {
+ console.log("ERROR CHECK ", fileName);
}
let checkLog = allReport?.filter(
(report) =>
@@ -320,12 +299,8 @@ export async function runtimeCheckLogs(folderPath) {
let log = allFile?.filter(
(i) => i.$original.file_name === fileName
)[0];
- if (
- log?.id_ldf === "" ||
- log?.id_ldf === null ||
- log?.id_ldf === undefined
- ) {
- console.log("ERROR CHECK ", fileName)
+ if (!log?.id_ldf) {
+ console.log("ERROR CHECK ", fileName);
}
let checkLog = allReport?.filter(
(report) =>
@@ -355,10 +330,10 @@ export async function runtimeCheckLogs(folderPath) {
} else {
console.log(
fileName +
- "has changed(" +
- contentFile.length +
- ") ---SOS---" +
- lines.length
+ "has changed(" +
+ contentFile.length +
+ ") ---SOS---" +
+ lines.length
);
let allReport_new = await LogReport.query().whereRaw(
`DATE(created_at) = ?`,
@@ -403,12 +378,12 @@ export async function runtimeCheckLogs(folderPath) {
contentFile[line - 1]?.replace(
log.detected_content,
"[" +
- log.detected_content +
- "](https://logs.danielvu.com/logs/" +
- fileName +
- "#" +
- line +
- ")"
+ log.detected_content +
+ "](https://logs.danielvu.com/logs/" +
+ fileName +
+ "#" +
+ line +
+ ")"
);
});
content =
@@ -428,25 +403,30 @@ export async function runtimeCheckLogs(folderPath) {
")\n";
});
- let icon = ""
- if (checkSpecialVersion(content) !== "" || important.filter((i) => content.includes(i)).length > 0) {
- icon += "------------\n\n:no_entry: :no_entry:**" +
+ let icon = "";
+ if (
+ checkSpecialVersion(content) !== "" ||
+ important.filter((i) => content.includes(i)).length > 0
+ ) {
+ icon +=
+ "------------\n\n:no_entry: :no_entry:**" +
fileName +
- "**:no_entry: :no_entry:"
+ "**:no_entry: :no_entry:";
} else {
- icon += "------------\n\n:warning: :warning: **" + fileName + "**";
+ icon +=
+ "------------\n\n:warning: :warning: **" + fileName + "**";
}
sendMessToZulip(
"stream",
Env.get("ZULIP_STREAM_ALERT"),
Env.get("ZULIP_TOPIC_ALERT"),
icon +
- "\n\n" +
- content +
- "\n\n" +
- spoiler +
- "\n\n***Issue found:***\n" +
- issueFound
+ "\n\n" +
+ content +
+ "\n\n" +
+ spoiler +
+ "\n\n***Issue found:***\n" +
+ issueFound
);
}
}, 3000);
diff --git a/app/utils/sendDeviceInfor.ts b/app/utils/sendDeviceInfor.ts
index 3b33a16..1f07b0d 100644
--- a/app/utils/sendDeviceInfor.ts
+++ b/app/utils/sendDeviceInfor.ts
@@ -20,8 +20,8 @@ export const sendDeviceInfora = async () => {
.filter((i) => i.$attributes.key === "MEMORY_DEFAULT")
.map((obj) => obj.$attributes.value);
- console.log(memDefault)
- let listInformation = [];
+ console.log(memDefault);
+ let listInformation: any = [];
let dataFile = await LogDetectFile.all();
let html = "";
@@ -195,36 +195,38 @@ export const sendDeviceInfora = async () => {
(line) =>
line.includes("bytes of memory") ||
line.includes("bytes of physical memory")
- )
- const mathches = stringMem[0]?.match(regexSPE)
+ );
+ const mathches = stringMem[0]?.match(regexSPE);
- if(mathches && stringMem[0]?.includes("CHASSIS")){
- console.log(mathches)
- memDefaultForPID =
+ if (mathches && stringMem[0]?.includes("CHASSIS")) {
+ console.log(mathches);
+ memDefaultForPID =
memDefault.filter((i) =>
- mathches[0].trim().includes(i.split(":")[0])
+ mathches[0].trim().includes(i.split(":")[0])
)[0] !== undefined
? memDefault.filter((i) =>
- mathches[0].trim().includes(i.split(":")[0])
+ mathches[0].trim().includes(i.split(":")[0])
)[0]
: mathches[0].trim() + ":N/A:N/A";
- PID = mathches[0].trim();
- SN = showInventoryContent.filter((i)=>i.includes(mathches[0].trim()))[0]?.split("SN:")[1]?.replace(/,/g, "").trim();
- VID = showInventoryContent.filter((i)=>i.includes(mathches[0].trim()))[0]?.split("VID:")[1]
+ PID = mathches[0].trim();
+ SN = showInventoryContent
+ .filter((i) => i.includes(mathches[0].trim()))[0]
+ ?.split("SN:")[1]
+ ?.replace(/,/g, "")
+ .trim();
+ VID = showInventoryContent
+ .filter((i) => i.includes(mathches[0].trim()))[0]
+ ?.split("VID:")[1]
?.split("SN:")[0]
?.replace(/,/g, "")
.trim();
- }
+ }
let RAM =
- stringMem
- .join("
")
- .match(regexMemory) !== null
+ stringMem.join("
").match(regexMemory) !== null
? (
parseInt(
- stringMem
- .join("
")
- .match(regexMemory)[0]
+ stringMem.join("
").match(regexMemory)[0]
) /
1024 /
1024
@@ -312,8 +314,10 @@ export const sendDeviceInfora = async () => {
let specialModel = allKeyValue
.filter((i) => i.key === "MODEL_SPECIAL")
.map((obj) => obj.$attributes.value);
- let pattern = /[\x00-\x20\x7F]/g;
- listInformation = listInformation.filter((i)=>i.PID.match(pattern)===null)
+ let pattern = /[\x00-\x20\x7F]/g;
+ listInformation = listInformation.filter(
+ (i) => i.PID.match(pattern) === null
+ );
listInformation = listInformation.filter(
(i) =>
i.RAM !== "" ||
diff --git a/database/migrations/1692586422677_users.ts b/database/migrations/1692586422677_users.ts
index e5e3d9b..9fcccea 100644
--- a/database/migrations/1692586422677_users.ts
+++ b/database/migrations/1692586422677_users.ts
@@ -9,8 +9,7 @@ export default class extends BaseSchema {
table.string("username", 50).notNullable();
table.string("password", 300).notNullable();
- table.timestamp("created_at", { useTz: true });
- table.timestamp("updated_at", { useTz: true });
+ table.timestamps(true, true);
});
}
diff --git a/database/migrations/1692588201278_key_values.ts b/database/migrations/1692588201278_key_values.ts
index 8a2b872..2ee8ddf 100644
--- a/database/migrations/1692588201278_key_values.ts
+++ b/database/migrations/1692588201278_key_values.ts
@@ -1,23 +1,22 @@
-import BaseSchema from '@ioc:Adonis/Lucid/Schema'
+import BaseSchema from "@ioc:Adonis/Lucid/Schema";
export default class extends BaseSchema {
- protected tableName = 'key_values'
+ protected tableName = "key_values";
- public async up () {
+ public async up() {
this.schema.createTable(this.tableName, (table) => {
- table.increments('id_key').primary()
- table.string("key", 100).notNullable()
- table.string("value", 200).notNullable()
- table.string("model", 30).notNullable().defaultTo("All")
+ table.increments("id_key").primary();
+ table.string("key", 100).notNullable();
+ table.string("value", 200).notNullable();
+ table.string("model", 30).notNullable().defaultTo("All");
/**
* Uses timestamptz for PostgreSQL and DATETIME2 for MSSQL
*/
- table.timestamp('created_at', { useTz: true })
- table.timestamp('updated_at', { useTz: true })
- })
+ table.timestamps(true, true);
+ });
}
- public async down () {
- this.schema.dropTable(this.tableName)
+ public async down() {
+ this.schema.dropTable(this.tableName);
}
}
diff --git a/database/migrations/1692588436912_log_detect_files.ts b/database/migrations/1692588436912_log_detect_files.ts
index 99aec7c..272e186 100644
--- a/database/migrations/1692588436912_log_detect_files.ts
+++ b/database/migrations/1692588436912_log_detect_files.ts
@@ -1,21 +1,20 @@
-import BaseSchema from '@ioc:Adonis/Lucid/Schema'
+import BaseSchema from "@ioc:Adonis/Lucid/Schema";
export default class extends BaseSchema {
- protected tableName = 'log_detect_files'
+ protected tableName = "log_detect_files";
- public async up () {
+ public async up() {
this.schema.createTable(this.tableName, (table) => {
- table.increments('id_ldf').primary()
- table.string("file_name", 100).notNullable()
+ table.increments("id_ldf").primary();
+ table.string("file_name", 100).notNullable();
/**
* Uses timestamptz for PostgreSQL and DATETIME2 for MSSQL
*/
- table.timestamp('created_at', { useTz: true })
- table.timestamp('updated_at', { useTz: true })
- })
+ table.timestamps(true, true);
+ });
}
- public async down () {
- this.schema.dropTable(this.tableName)
+ public async down() {
+ this.schema.dropTable(this.tableName);
}
}
diff --git a/database/migrations/1692588646452_log_reports.ts b/database/migrations/1692588646452_log_reports.ts
index d03efbf..82e1125 100644
--- a/database/migrations/1692588646452_log_reports.ts
+++ b/database/migrations/1692588646452_log_reports.ts
@@ -1,24 +1,23 @@
-import BaseSchema from '@ioc:Adonis/Lucid/Schema'
+import BaseSchema from "@ioc:Adonis/Lucid/Schema";
export default class extends BaseSchema {
- protected tableName = 'log_reports'
+ protected tableName = "log_reports";
- public async up () {
+ public async up() {
this.schema.createTable(this.tableName, (table) => {
- table.increments('id_report').primary()
- table.string("detected_content", 200).notNullable()
- table.integer("line", 6).notNullable()
- table.integer("id_file").notNullable()
+ table.increments("id_report").primary();
+ table.string("detected_content", 200).notNullable();
+ table.integer("line", 6).notNullable();
+ table.integer("id_file").notNullable();
/**
* Uses timestamptz for PostgreSQL and DATETIME2 for MSSQL
*/
- table.timestamp('created_at', { useTz: true })
- table.timestamp('updated_at', { useTz: true })
- })
+ table.timestamps(true, true);
+ });
}
- public async down () {
- this.schema.dropTable(this.tableName)
+ public async down() {
+ this.schema.dropTable(this.tableName);
}
}
diff --git a/database/migrations/1696904362937_info_devices.ts b/database/migrations/1696904362937_info_devices.ts
index 8d02fad..b06c3cf 100644
--- a/database/migrations/1696904362937_info_devices.ts
+++ b/database/migrations/1696904362937_info_devices.ts
@@ -1,21 +1,20 @@
-import BaseSchema from '@ioc:Adonis/Lucid/Schema'
+import BaseSchema from "@ioc:Adonis/Lucid/Schema";
export default class extends BaseSchema {
- protected tableName = 'info_devices'
+ protected tableName = "info_devices";
- public async up () {
+ public async up() {
this.schema.createTable(this.tableName, (table) => {
- table.increments('id')
+ table.increments("id");
/**
* Uses timestamptz for PostgreSQL and DATETIME2 for MSSQL
*/
- table.timestamp('created_at', { useTz: true })
- table.timestamp('updated_at', { useTz: true })
- })
+ table.timestamps(true, true);
+ });
}
- public async down () {
- this.schema.dropTable(this.tableName)
+ public async down() {
+ this.schema.dropTable(this.tableName);
}
}