mirror of
https://github.com/godotengine/godot-team-reports.git
synced 2025-12-31 13:48:17 +03:00
Refactor database composer code to be in line with other pages
This commit is contained in:
505
compose-db.js
505
compose-db.js
@@ -1,23 +1,13 @@
|
||||
const fs = require('fs').promises;
|
||||
const fsConstants = require('fs').constants;
|
||||
const path = require('path');
|
||||
const fetch = require('node-fetch');
|
||||
|
||||
const teams = {};
|
||||
const reviewers = {};
|
||||
const authors = {};
|
||||
const pulls = [];
|
||||
|
||||
const PULLS_PER_PAGE = 100;
|
||||
let page_count = 1;
|
||||
let last_cursor = "";
|
||||
|
||||
const ExitCodes = {
|
||||
"RequestFailure": 1,
|
||||
"ParseFailure": 2,
|
||||
};
|
||||
|
||||
const API_REPOSITORY_ID = `owner:"godotengine" name:"godot"`;
|
||||
const PULLS_PER_PAGE = 100;
|
||||
const API_RATE_LIMIT = `
|
||||
rateLimit {
|
||||
limit
|
||||
@@ -36,7 +26,48 @@ const GH_MAGIC_KEYWORDS = [
|
||||
const GH_MAGIC_RE = RegExp("(" + GH_MAGIC_KEYWORDS.join("|") + ") ([a-z0-9-_]+/[a-z0-9-_]+)?#([0-9]+)", "gi");
|
||||
const GH_MAGIC_FULL_RE = RegExp("(" + GH_MAGIC_KEYWORDS.join("|") + ") https://github.com/([a-z0-9-_]+/[a-z0-9-_]+)/issues/([0-9]+)", "gi");
|
||||
|
||||
async function fetchGithub(query) {
|
||||
class DataFetcher {
|
||||
constructor(data_owner, data_repo) {
|
||||
this.api_repository_id = `owner:"${data_owner}" name:"${data_repo}"`;
|
||||
|
||||
this.page_count = 1;
|
||||
this.last_cursor = "";
|
||||
}
|
||||
|
||||
async _logResponse(data, name) {
|
||||
try {
|
||||
try {
|
||||
await fs.access("logs", fsConstants.R_OK | fsConstants.W_OK);
|
||||
} catch (err) {
|
||||
await fs.mkdir("logs");
|
||||
}
|
||||
|
||||
await fs.writeFile(`logs/${name}.json`, JSON.stringify(data, null, 4), {encoding: "utf-8"});
|
||||
} catch (err) {
|
||||
console.error("Error saving log file: " + err);
|
||||
}
|
||||
}
|
||||
|
||||
_handleResponseErrors(res) {
|
||||
console.warn(` Failed to get pull requests for '${this.api_repository_id}'; server responded with ${res.status} ${res.statusText}`);
|
||||
const retry_header = res.headers.get("Retry-After");
|
||||
if (retry_header) {
|
||||
console.log(` Retry after: ${retry_header}`);
|
||||
}
|
||||
}
|
||||
|
||||
_handleDataErrors(data) {
|
||||
if (typeof data["errors"] === "undefined") {
|
||||
return;
|
||||
}
|
||||
|
||||
console.warn(` Server handled the request, but there were errors:`);
|
||||
data.errors.forEach((item) => {
|
||||
console.log(` [${item.type}] ${item.message}`);
|
||||
});
|
||||
}
|
||||
|
||||
async fetchGithub(query) {
|
||||
const init = {};
|
||||
init.method = "POST";
|
||||
init.headers = {};
|
||||
@@ -55,55 +86,7 @@ async function fetchGithub(query) {
|
||||
return await fetch("https://api.github.com/graphql", init);
|
||||
}
|
||||
|
||||
async function logResponse(data, name) {
|
||||
try {
|
||||
try {
|
||||
await fs.access("logs", fsConstants.R_OK | fsConstants.W_OK);
|
||||
} catch (err) {
|
||||
await fs.mkdir("logs");
|
||||
}
|
||||
|
||||
await fs.writeFile(`logs/${name}.json`, JSON.stringify(data, null, 4), {encoding: "utf-8"});
|
||||
} catch (err) {
|
||||
console.error("Error saving log file: " + err);
|
||||
}
|
||||
}
|
||||
|
||||
function handleResponseErrors(res) {
|
||||
console.warn(` Failed to get pull requests for '${API_REPOSITORY_ID}'; server responded with ${res.status} ${res.statusText}`);
|
||||
const retry_header = res.headers.get("Retry-After");
|
||||
if (retry_header) {
|
||||
console.log(` Retry after: ${retry_header}`);
|
||||
}
|
||||
}
|
||||
|
||||
function handleDataErrors(data) {
|
||||
if (typeof data["errors"] === "undefined") {
|
||||
return;
|
||||
}
|
||||
|
||||
console.warn(` Server handled the request, but there were errors:`);
|
||||
data.errors.forEach((item) => {
|
||||
console.log(` [${item.type}] ${item.message}`);
|
||||
});
|
||||
}
|
||||
|
||||
function mapNodes(object) {
|
||||
return object.edges.map((item) => item["node"])
|
||||
}
|
||||
|
||||
function sluggifyTeam(name) {
|
||||
let slug = name
|
||||
.toLowerCase()
|
||||
// Replace runs of non-alphanumerical characters with '-'; '_' is also allowed.
|
||||
.replace(/[^0-9a-z_]+/g, "-")
|
||||
// Trim trailing '-' characters.
|
||||
.replace(/[-]+$/, "");
|
||||
|
||||
return slug;
|
||||
}
|
||||
|
||||
async function checkRates() {
|
||||
async checkRates() {
|
||||
try {
|
||||
const query = `
|
||||
query {
|
||||
@@ -111,16 +94,16 @@ async function checkRates() {
|
||||
}
|
||||
`;
|
||||
|
||||
const res = await fetchGithub(query);
|
||||
const res = await this.fetchGithub(query);
|
||||
if (res.status !== 200) {
|
||||
handleResponseErrors(res);
|
||||
this._handleResponseErrors(res);
|
||||
process.exitCode = ExitCodes.RequestFailure;
|
||||
return;
|
||||
}
|
||||
|
||||
const data = await res.json();
|
||||
await logResponse(data, "_rate_limit");
|
||||
handleDataErrors(data);
|
||||
await this._logResponse(data, "_rate_limit");
|
||||
this._handleDataErrors(data);
|
||||
|
||||
const rate_limit = data.data["rateLimit"];
|
||||
console.log(` [$${rate_limit.cost}] Available API calls: ${rate_limit.remaining}/${rate_limit.limit}; resets at ${rate_limit.resetAt}`);
|
||||
@@ -131,12 +114,12 @@ async function checkRates() {
|
||||
}
|
||||
}
|
||||
|
||||
async function fetchPulls(page) {
|
||||
async fetchPulls(page) {
|
||||
try {
|
||||
let after_cursor = "";
|
||||
let after_text = "initial";
|
||||
if (last_cursor !== "") {
|
||||
after_cursor = `after: "${last_cursor}"`;
|
||||
if (this.last_cursor !== "") {
|
||||
after_cursor = `after: "${this.last_cursor}"`;
|
||||
after_text = after_cursor;
|
||||
}
|
||||
|
||||
@@ -145,7 +128,7 @@ async function fetchPulls(page) {
|
||||
const query = `
|
||||
query {
|
||||
${API_RATE_LIMIT}
|
||||
repository(${API_REPOSITORY_ID}) {
|
||||
repository(${this.api_repository_id}) {
|
||||
pullRequests(first:${PULLS_PER_PAGE} ${after_cursor} states: OPEN) {
|
||||
totalCount
|
||||
pageInfo {
|
||||
@@ -227,21 +210,21 @@ async function fetchPulls(page) {
|
||||
`;
|
||||
|
||||
let page_text = page;
|
||||
if (page_count > 1) {
|
||||
page_text = `${page}/${page_count}`;
|
||||
if (this.page_count > 1) {
|
||||
page_text = `${page}/${this.page_count}`;
|
||||
}
|
||||
console.log(` Requesting page ${page_text} of pull request data (${after_text}).`);
|
||||
|
||||
const res = await fetchGithub(query);
|
||||
const res = await this.fetchGithub(query);
|
||||
if (res.status !== 200) {
|
||||
handleResponseErrors(res);
|
||||
this._handleResponseErrors(res);
|
||||
process.exitCode = ExitCodes.RequestFailure;
|
||||
return [];
|
||||
}
|
||||
|
||||
const data = await res.json();
|
||||
await logResponse(data, `data_page_${page}`);
|
||||
handleDataErrors(data);
|
||||
await this._logResponse(data, `data_page_${page}`);
|
||||
this._handleDataErrors(data);
|
||||
|
||||
const rate_limit = data.data["rateLimit"];
|
||||
const repository = data.data["repository"];
|
||||
@@ -249,8 +232,8 @@ async function fetchPulls(page) {
|
||||
|
||||
console.log(` [$${rate_limit.cost}] Retrieved ${pulls_data.length} pull requests; processing...`);
|
||||
|
||||
last_cursor = repository.pullRequests.pageInfo.endCursor;
|
||||
page_count = Math.ceil(repository.pullRequests.totalCount / PULLS_PER_PAGE);
|
||||
this.last_cursor = repository.pullRequests.pageInfo.endCursor;
|
||||
this.page_count = Math.ceil(repository.pullRequests.totalCount / PULLS_PER_PAGE);
|
||||
|
||||
return pulls_data;
|
||||
} catch (err) {
|
||||
@@ -259,162 +242,28 @@ async function fetchPulls(page) {
|
||||
return [];
|
||||
}
|
||||
}
|
||||
|
||||
function processPulls(pullsRaw) {
|
||||
try {
|
||||
pullsRaw.forEach((item) => {
|
||||
// Compile basic information about a PR.
|
||||
let pr = {
|
||||
"id": item.id,
|
||||
"public_id": item.number,
|
||||
"url": item.url,
|
||||
"diff_url": `${item.url}.diff`,
|
||||
"patch_url": `${item.url}.patch`,
|
||||
|
||||
"title": item.title,
|
||||
"state": item.state,
|
||||
"is_draft": item.isDraft,
|
||||
"authored_by": null,
|
||||
"created_at": item.createdAt,
|
||||
"updated_at": item.updatedAt,
|
||||
|
||||
"target_branch": item.baseRef.name,
|
||||
|
||||
"mergeable_state": item.mergeable,
|
||||
"mergeable_reason": 'UNKNOWN', //item.mergeStateStatus,
|
||||
"labels": [],
|
||||
"milestone": null,
|
||||
"links": [],
|
||||
|
||||
"teams": [],
|
||||
"reviewers": [],
|
||||
};
|
||||
|
||||
// Compose and link author information.
|
||||
const author = {
|
||||
"id": "",
|
||||
"user": "ghost",
|
||||
"avatar": "https://avatars.githubusercontent.com/u/10137?v=4",
|
||||
"url": "https://github.com/ghost",
|
||||
"pull_count": 0,
|
||||
};
|
||||
if (item.author != null) {
|
||||
author["id"] = item.author.id;
|
||||
author["user"] = item.author.login;
|
||||
author["avatar"] = item.author.avatarUrl;
|
||||
author["url"] = item.author.url;
|
||||
}
|
||||
pr.authored_by = author.id;
|
||||
|
||||
// Store the author if they haven't been stored.
|
||||
if (typeof authors[author.id] === "undefined") {
|
||||
authors[author.id] = author;
|
||||
}
|
||||
authors[author.id].pull_count++;
|
||||
|
||||
// Add the milestone, if available.
|
||||
if (item.milestone) {
|
||||
pr.milestone = {
|
||||
"id": item.milestone.id,
|
||||
"title": item.milestone.title,
|
||||
"url": item.milestone.url,
|
||||
};
|
||||
}
|
||||
|
||||
// Add labels, if available.
|
||||
let labels = mapNodes(item.labels);
|
||||
labels.forEach((labelItem) => {
|
||||
pr.labels.push({
|
||||
"id": labelItem.id,
|
||||
"name": labelItem.name,
|
||||
"color": "#" + labelItem.color
|
||||
});
|
||||
});
|
||||
pr.labels.sort((a, b) => {
|
||||
if (a.name > b.name) return 1;
|
||||
if (a.name < b.name) return -1;
|
||||
return 0;
|
||||
});
|
||||
|
||||
// Look for linked issues in the body.
|
||||
pr.links = extractLinkedIssues(item.body);
|
||||
|
||||
// Extract requested reviewers.
|
||||
let review_requests = mapNodes(item.reviewRequests).map(it => it.requestedReviewer);
|
||||
|
||||
// Add teams, if available.
|
||||
let requested_teams = review_requests.filter(it => it && it["__typename"] === "Team");
|
||||
if (requested_teams.length > 0) {
|
||||
requested_teams.forEach((teamItem) => {
|
||||
const team = {
|
||||
"id": teamItem.id,
|
||||
"name": teamItem.name,
|
||||
"avatar": teamItem.avatarUrl,
|
||||
"slug": sluggifyTeam(teamItem.name),
|
||||
"pull_count": 0,
|
||||
};
|
||||
|
||||
// Store the team if it hasn't been stored before.
|
||||
if (typeof teams[team.id] == "undefined") {
|
||||
teams[team.id] = team;
|
||||
}
|
||||
teams[team.id].pull_count++;
|
||||
|
||||
// Reference the team.
|
||||
pr.teams.push(team.id);
|
||||
});
|
||||
} else {
|
||||
// If there are no teams, use a fake "empty" team to track those PRs as well.
|
||||
const team = {
|
||||
"id": "",
|
||||
"name": "No team assigned",
|
||||
"avatar": "",
|
||||
"slug": "_",
|
||||
"pull_count": 0,
|
||||
};
|
||||
|
||||
// Store the team if it hasn't been stored before.
|
||||
if (typeof teams[team.id] === "undefined") {
|
||||
teams[team.id] = team;
|
||||
}
|
||||
teams[team.id].pull_count++;
|
||||
|
||||
// Reference the team.
|
||||
pr.teams.push(team.id);
|
||||
class DataProcessor {
|
||||
constructor() {
|
||||
this.teams = {};
|
||||
this.reviewers = {};
|
||||
this.authors = {};
|
||||
this.pulls = [];
|
||||
}
|
||||
|
||||
// Add individual reviewers, if available
|
||||
let requested_reviewers = review_requests.filter(it => it && it["__typename"] === "User");
|
||||
if (requested_reviewers.length > 0) {
|
||||
requested_reviewers.forEach((reviewerItem) => {
|
||||
const reviewer = {
|
||||
"id": reviewerItem.id,
|
||||
"name": reviewerItem.login,
|
||||
"avatar": reviewerItem.avatarUrl,
|
||||
"slug": reviewerItem.login,
|
||||
"pull_count": 0,
|
||||
};
|
||||
_sluggifyTeam(name) {
|
||||
let slug = name
|
||||
.toLowerCase()
|
||||
// Replace runs of non-alphanumerical characters with '-'; '_' is also allowed.
|
||||
.replace(/[^0-9a-z_]+/g, "-")
|
||||
// Trim trailing '-' characters.
|
||||
.replace(/[-]+$/, "");
|
||||
|
||||
// Store the reviewer if it hasn't been stored before.
|
||||
if (typeof reviewers[reviewer.id] == "undefined") {
|
||||
reviewers[reviewer.id] = reviewer;
|
||||
}
|
||||
reviewers[reviewer.id].pull_count++;
|
||||
|
||||
// Reference the reviewer.
|
||||
pr.reviewers.push(reviewer.id);
|
||||
});
|
||||
return slug;
|
||||
}
|
||||
|
||||
pulls.push(pr);
|
||||
});
|
||||
} catch (err) {
|
||||
console.error(" Error parsing pull request data: " + err);
|
||||
process.exitCode = ExitCodes.ParseFailure;
|
||||
}
|
||||
}
|
||||
|
||||
function extractLinkedIssues(pullBody) {
|
||||
_extractLinkedIssues(pullBody) {
|
||||
const links = [];
|
||||
if (!pullBody) {
|
||||
return links;
|
||||
@@ -462,29 +311,203 @@ function extractLinkedIssues(pullBody) {
|
||||
return links;
|
||||
}
|
||||
|
||||
function checkForExit() {
|
||||
processPulls(pullsRaw) {
|
||||
try {
|
||||
pullsRaw.forEach((item) => {
|
||||
// Compile basic information about a PR.
|
||||
let pr = {
|
||||
"id": item.id,
|
||||
"public_id": item.number,
|
||||
"url": item.url,
|
||||
"diff_url": `${item.url}.diff`,
|
||||
"patch_url": `${item.url}.patch`,
|
||||
|
||||
"title": item.title,
|
||||
"state": item.state,
|
||||
"is_draft": item.isDraft,
|
||||
"authored_by": null,
|
||||
"created_at": item.createdAt,
|
||||
"updated_at": item.updatedAt,
|
||||
|
||||
"target_branch": item.baseRef.name,
|
||||
|
||||
"mergeable_state": item.mergeable,
|
||||
"mergeable_reason": 'UNKNOWN', //item.mergeStateStatus,
|
||||
"labels": [],
|
||||
"milestone": null,
|
||||
"links": [],
|
||||
|
||||
"teams": [],
|
||||
"reviewers": [],
|
||||
};
|
||||
|
||||
// Compose and link author information.
|
||||
const author = {
|
||||
"id": "",
|
||||
"user": "ghost",
|
||||
"avatar": "https://avatars.githubusercontent.com/u/10137?v=4",
|
||||
"url": "https://github.com/ghost",
|
||||
"pull_count": 0,
|
||||
};
|
||||
if (item.author != null) {
|
||||
author["id"] = item.author.id;
|
||||
author["user"] = item.author.login;
|
||||
author["avatar"] = item.author.avatarUrl;
|
||||
author["url"] = item.author.url;
|
||||
}
|
||||
pr.authored_by = author.id;
|
||||
|
||||
// Store the author if they haven't been stored.
|
||||
if (typeof this.authors[author.id] === "undefined") {
|
||||
this.authors[author.id] = author;
|
||||
}
|
||||
this.authors[author.id].pull_count++;
|
||||
|
||||
// Add the milestone, if available.
|
||||
if (item.milestone) {
|
||||
pr.milestone = {
|
||||
"id": item.milestone.id,
|
||||
"title": item.milestone.title,
|
||||
"url": item.milestone.url,
|
||||
};
|
||||
}
|
||||
|
||||
// Add labels, if available.
|
||||
let labels = mapNodes(item.labels);
|
||||
labels.forEach((labelItem) => {
|
||||
pr.labels.push({
|
||||
"id": labelItem.id,
|
||||
"name": labelItem.name,
|
||||
"color": "#" + labelItem.color
|
||||
});
|
||||
});
|
||||
pr.labels.sort((a, b) => {
|
||||
if (a.name > b.name) return 1;
|
||||
if (a.name < b.name) return -1;
|
||||
return 0;
|
||||
});
|
||||
|
||||
// Look for linked issues in the body.
|
||||
pr.links = this._extractLinkedIssues(item.body);
|
||||
|
||||
// Extract requested reviewers.
|
||||
let review_requests = mapNodes(item.reviewRequests).map(it => it.requestedReviewer);
|
||||
|
||||
// Add teams, if available.
|
||||
let requested_teams = review_requests.filter(it => it && it["__typename"] === "Team");
|
||||
if (requested_teams.length > 0) {
|
||||
requested_teams.forEach((teamItem) => {
|
||||
const team = {
|
||||
"id": teamItem.id,
|
||||
"name": teamItem.name,
|
||||
"avatar": teamItem.avatarUrl,
|
||||
"slug": this._sluggifyTeam(teamItem.name),
|
||||
"pull_count": 0,
|
||||
};
|
||||
|
||||
// Store the team if it hasn't been stored before.
|
||||
if (typeof this.teams[team.id] == "undefined") {
|
||||
this.teams[team.id] = team;
|
||||
}
|
||||
this.teams[team.id].pull_count++;
|
||||
|
||||
// Reference the team.
|
||||
pr.teams.push(team.id);
|
||||
});
|
||||
} else {
|
||||
// If there are no teams, use a fake "empty" team to track those PRs as well.
|
||||
const team = {
|
||||
"id": "",
|
||||
"name": "No team assigned",
|
||||
"avatar": "",
|
||||
"slug": "_",
|
||||
"pull_count": 0,
|
||||
};
|
||||
|
||||
// Store the team if it hasn't been stored before.
|
||||
if (typeof this.teams[team.id] === "undefined") {
|
||||
this.teams[team.id] = team;
|
||||
}
|
||||
this.teams[team.id].pull_count++;
|
||||
|
||||
// Reference the team.
|
||||
pr.teams.push(team.id);
|
||||
}
|
||||
|
||||
// Add individual reviewers, if available
|
||||
let requested_reviewers = review_requests.filter(it => it && it["__typename"] === "User");
|
||||
if (requested_reviewers.length > 0) {
|
||||
requested_reviewers.forEach((reviewerItem) => {
|
||||
const reviewer = {
|
||||
"id": reviewerItem.id,
|
||||
"name": reviewerItem.login,
|
||||
"avatar": reviewerItem.avatarUrl,
|
||||
"slug": reviewerItem.login,
|
||||
"pull_count": 0,
|
||||
};
|
||||
|
||||
// Store the reviewer if it hasn't been stored before.
|
||||
if (typeof this.reviewers[reviewer.id] == "undefined") {
|
||||
this.reviewers[reviewer.id] = reviewer;
|
||||
}
|
||||
this.reviewers[reviewer.id].pull_count++;
|
||||
|
||||
// Reference the reviewer.
|
||||
pr.reviewers.push(reviewer.id);
|
||||
});
|
||||
}
|
||||
|
||||
this.pulls.push(pr);
|
||||
});
|
||||
} catch (err) {
|
||||
console.error(" Error parsing pull request data: " + err);
|
||||
process.exitCode = ExitCodes.ParseFailure;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function mapNodes(object) {
|
||||
return object.edges.map((item) => item["node"])
|
||||
}
|
||||
|
||||
async function main() {
|
||||
// Internal utility methods.
|
||||
const checkForExit = () => {
|
||||
if (process.exitCode > 0) {
|
||||
process.exit();
|
||||
}
|
||||
}
|
||||
|
||||
async function delay(msec) {
|
||||
const delay = async (msec) => {
|
||||
return new Promise(resolve => setTimeout(resolve, msec));
|
||||
}
|
||||
|
||||
async function main() {
|
||||
console.log("[*] Building local pull request database.");
|
||||
|
||||
let data_owner = "godotengine";
|
||||
let data_repo = "godot";
|
||||
process.argv.forEach((arg) => {
|
||||
if (arg.indexOf("owner:") === 0) {
|
||||
data_owner = arg.substring(6);
|
||||
}
|
||||
if (arg.indexOf("repo:") === 0) {
|
||||
data_repo = arg.substring(5);
|
||||
}
|
||||
});
|
||||
|
||||
console.log(`[*] Configured for the "${data_owner}/${data_repo}" repository.`);
|
||||
const dataFetcher = new DataFetcher(data_owner, data_repo);
|
||||
const dataProcessor = new DataProcessor();
|
||||
|
||||
console.log("[*] Checking the rate limits before.")
|
||||
await checkRates();
|
||||
await dataFetcher.checkRates();
|
||||
checkForExit();
|
||||
|
||||
console.log("[*] Fetching pull request data from GitHub.");
|
||||
// Pages are starting with 1 for better presentation.
|
||||
let page = 1;
|
||||
while (page <= page_count) {
|
||||
const pullsRaw = await fetchPulls(page);
|
||||
processPulls(pullsRaw);
|
||||
while (page <= dataFetcher.page_count) {
|
||||
const pullsRaw = await dataFetcher.fetchPulls(page);
|
||||
dataProcessor.processPulls(pullsRaw);
|
||||
checkForExit();
|
||||
page++;
|
||||
|
||||
@@ -494,20 +517,22 @@ async function main() {
|
||||
}
|
||||
|
||||
console.log("[*] Checking the rate limits after.")
|
||||
await checkRates();
|
||||
await dataFetcher.checkRates();
|
||||
checkForExit();
|
||||
|
||||
console.log("[*] Finalizing database.")
|
||||
const output = {
|
||||
"generated_at": Date.now(),
|
||||
"teams": teams,
|
||||
"reviewers": reviewers,
|
||||
"authors": authors,
|
||||
"pulls": pulls,
|
||||
"teams": dataProcessor.teams,
|
||||
"reviewers": dataProcessor.reviewers,
|
||||
"authors": dataProcessor.authors,
|
||||
"pulls": dataProcessor.pulls,
|
||||
};
|
||||
try {
|
||||
console.log("[*] Storing database to file.")
|
||||
console.log("[*] Storing database to file.");
|
||||
// NOTE: The repository owner and name are not respected here, the file will be overwritten.
|
||||
await fs.writeFile("out/data.json", JSON.stringify(output), {encoding: "utf-8"});
|
||||
console.log("[*] Database built.");
|
||||
} catch (err) {
|
||||
console.error("Error saving database file: " + err);
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user