10 Commits

9 changed files with 935 additions and 924 deletions

View File

@@ -20,7 +20,7 @@ WORKDIR /app
COPY package*.json ./ COPY package*.json ./
RUN npm ci --omit=dev RUN npm ci --omit=dev
COPY --from=builder /app/dist ./dist COPY --from=builder /app/dist /app/dist
USER node USER node

1430
package-lock.json generated

File diff suppressed because it is too large Load Diff

View File

@@ -25,14 +25,11 @@
"@aws-sdk/client-s3": "^3.964.0", "@aws-sdk/client-s3": "^3.964.0",
"@aws-sdk/lib-storage": "^3.964.0", "@aws-sdk/lib-storage": "^3.964.0",
"@owlboard/backend-data-contracts": "^0.1.9", "@owlboard/backend-data-contracts": "^0.1.9",
"mongodb": "^7.0.0",
"nats": "^2.29.3", "nats": "^2.29.3",
"readline": "^1.3.0", "readline": "^1.3.0"
"xxhashjs": "^0.2.2"
}, },
"devDependencies": { "devDependencies": {
"@types/node": "^25.0.3", "@types/node": "^25.0.3",
"@types/xxhashjs": "^0.2.4",
"tsx": "^4.21.0", "tsx": "^4.21.0",
"typescript": "^5.9.3" "typescript": "^5.9.3"
} }

79
src/config.ts Normal file
View File

@@ -0,0 +1,79 @@
import { readFileSync, existsSync } from "node:fs"
import { join } from "node:path"
import { log } from "./logger.js";
interface Config {
Mq: Mq,
S3: S3,
Mongo: Mongo,
General: GeneralConfig,
}
export interface Mq {
MQ_USER: string;
MQ_PASS: string;
MQ_URL: string;
MQ_TOPIC: string;
}
export interface S3 {
S3_ENDPOINT: string;
S3_BUCKET: string;
S3_ACCESS_KEY: string;
S3_SECRET_KEY: string;
S3_REGION: string;
}
export interface Mongo {
MONGO_URI: string;
MONGO_DB: string;
MONGO_USER: string;
MONGO_PASS: string;
}
export interface GeneralConfig {
TOC: string;
}
export function ConfigLoader(): Config {
const cfg: Config = {
Mq: {
MQ_URL: getConfig("MQ_URL"),
MQ_TOPIC: getConfig("MQ_TOPIC"),
MQ_USER: getConfig("MQ_USER"),
MQ_PASS: getConfig("MQ_PASS"),
},
S3: {
S3_ENDPOINT: getConfig("S3_ENDPOINT"),
S3_BUCKET: getConfig("S3_BUCKET"),
S3_ACCESS_KEY: getConfig("S3_ACCESS_KEY"),
S3_SECRET_KEY: getConfig("S3_SECRET_KEY"),
S3_REGION: getConfig("S3_REGION"),
},
Mongo: {
MONGO_URI: getConfig("MONGO_URI"),
MONGO_DB: getConfig("MONGO_DB"),
MONGO_USER: getConfig("MONGO_USER"),
MONGO_PASS: getConfig("MONGO_PASS"),
},
General: {
TOC: getConfig("TOC"),
}
};
return cfg
}
function getConfig(key: string): string {
const filePath = join("/etc/secrets", key);
if (existsSync(filePath)) {
try {
return readFileSync(filePath, "utf8").trim();
} catch (err) {
log("ERROR", `Error reading secret file at ${filePath}: ${err}`);
}
}
return process.env[key] || ""
}

View File

@@ -1,62 +0,0 @@
// Functions to check the current applied version
// updating will be handled by the processing service
import { MongoClient } from "mongodb";
import { log } from "./logger";
const uri = process.env.MONGO_URI || "";
const db = process.env.MONGO_DB || "";
const user = process.env.MONGO_USER || "";
const pass = process.env.MONGO_PASS || "";
const collection = "data_ingress_meta";
if(!uri || !db || !user || !pass) {
log('ERROR', "Missing MONGO Configuration - EXIT CODE: 35");
process.exit(35);
} else {
log("DEBUG", `MongoDB Connection`, {
uri: uri,
db: db,
collection: collection,
user: user,
pass: "****",
});
};
const CONNECTION_URI = `mongodb://${encodeURIComponent(user)}:${encodeURIComponent(pass)}@${uri}`;
let mongoClient: MongoClient | null = null;
async function getMongoClient() {
if (mongoClient) return mongoClient;
mongoClient = new MongoClient(CONNECTION_URI);
await mongoClient.connect();
return mongoClient;
}
export async function isPackageProcessed(serviceName: string, packageName: string): Promise<boolean> {
try {
const client = await getMongoClient();
const database = client.db(db);
const coll = database.collection(collection);
const result = await coll.findOne({ service_name: serviceName });
if (!result) {
log('INFO', `No metadata found for ${serviceName}. Fetching PIS Data`);
return false;
}
if (result.latest_package === packageName) {
log('INFO', 'No update needed');
return true;
}
log('INFO', `Version mismatch. DB: ${result.latest_package}, Current: ${packageName}. Update required`)
return false;
} catch (err) {
log('ERROR', 'Failed to check Mongo for version state:', err);
process.exit(1);
}
}

View File

@@ -4,13 +4,39 @@ import { processAndStore } from './sss.js'
import { getLatestPackageName, getRequestStream } from './sources/gitea.js' import { getLatestPackageName, getRequestStream } from './sources/gitea.js'
import { processPisStream } from './process.js' import { processPisStream } from './process.js'
import { isPackageProcessed } from './database.js' import { ConfigLoader } from './config.js'
import { natsManager } from './nats.js'
async function exit(exitCode: string | number=0): Promise<void> {
log("INFO", `Exiting with code: ${exitCode}`);
try {
await natsManager.close();
} catch (err) {
log("ERROR", `Error during cleanup: ${err}`);
process.exit(1)
}
process.exit(exitCode);
}
async function main() { async function main() {
const SERVICE_NAME = process.env.SERVICE_NAME; const SERVICE_NAME = process.env.SERVICE_NAME;
if (!SERVICE_NAME) { if (!SERVICE_NAME) {
log('ERROR', "SERVICE_NAME env variable must be set"); log('ERROR', "SERVICE_NAME env variable must be set");
process.exit(1); process.exitCode = 1;
return;
}
const CURRENT_VERSION_KEY: string = `${SERVICE_NAME}-current-version`;
const config = ConfigLoader();
try {
log('INFO', `Initialising NATS`);
await natsManager.connect(config.Mq);
} catch (err) {
log('ERROR', `Unable to connect to NATS: ${err}`);
} }
try { try {
@@ -19,26 +45,34 @@ async function main() {
if (!packageInfo.assets[0]?.browser_download_url) { if (!packageInfo.assets[0]?.browser_download_url) {
log('ERROR', `No attachments found for release ${packageInfo.name}`); log('ERROR', `No attachments found for release ${packageInfo.name}`);
process.exit(9); process.exitCode = 1;
return;
} }
if (await isPackageProcessed(SERVICE_NAME, packageInfo.name)) { const lastAppliedVersion: string | null = await natsManager.getState(CURRENT_VERSION_KEY);
log('INFO', `Database matches latest release. Exiting`); if (lastAppliedVersion === packageInfo.name) {
process.exit(0); log('INFO', "No new data, exiting");
return;
} }
const inputStream: Readable = await getRequestStream(packageInfo.assets[0].browser_download_url); const inputStream: Readable = await getRequestStream(packageInfo.assets[0].browser_download_url);
const objectGenerator = processPisStream(inputStream); const objectGenerator = processPisStream(config.General, inputStream);
const filename = `${packageInfo.name.replace(/\s+/g, '_')}_pis_data_ndjson`; const filename = `${packageInfo.name.replace(/\s+/g, '_')}_pis_data_ndjson`;
log('DEBUG', `Processing stream to: ${filename}`); log('DEBUG', `Processing stream to: ${filename}`);
await processAndStore(objectGenerator, filename); await processAndStore(config.S3, objectGenerator, filename);
log('DEBUG', 'Done'); log('DEBUG', 'Done');
log('DEBUG', "Sending message to NATS");
await natsManager.sendFileUpdateMessage(filename, packageInfo.name, SERVICE_NAME);
await natsManager.setState(CURRENT_VERSION_KEY, packageInfo.name);
} catch (err) { } catch (err) {
log('ERROR', 'Fatal error in pipeline: ', err); log('ERROR', 'Fatal error in pipeline: ', err);
process.exit(7); process.exitCode = 1;
} finally {
await exit(process.exitCode || 0);
} }
} }

View File

@@ -1,62 +1,136 @@
import { connect, JSONCodec } from "nats"; import { connect, JSONCodec, StringCodec } from "nats";
import type { ConnectionOptions, NatsConnection, Payload } from "nats"; import type { ConnectionOptions, NatsConnection, JetStreamClient, KV } from "nats";
import { log } from "./logger"; import { log } from "./logger.js";
import { hostname } from "node:os"; import { hostname } from "node:os";
import type { MQFileUpdate } from "@owlboard/backend-data-contracts/dist/data-ingress_mq-file-update"; import type { MQFileUpdate } from "@owlboard/backend-data-contracts/dist/data-ingress_mq-file-update";
import type { Mq } from "./config.js";
const jc = JSONCodec(); const jc = JSONCodec();
const sc = StringCodec();
async function getNatsConnection(): Promise<NatsConnection> { class NatsManager {
const serverUrl = process.env.MQ_URL || "nats://localhost:4222"; private nc: NatsConnection | null = null;
private js: JetStreamClient | null = null;
private kv: KV | null = null;
private bucketName = "INGRESS_STATES";
const options: ConnectionOptions = { /**
servers: serverUrl, * Opens connection to NATS
name: hostname(), */
reconnect: true, async connect(cfg: Mq): Promise<void> {
maxReconnectAttempts: -1, if (this.nc) return;
};
if (process.env.MQ_USER && process.env.MQ_PASS) { const options: ConnectionOptions = {
options.user = process.env.MQ_USER; servers: cfg.MQ_URL,
options.pass = process.env.MQ_PASS; name: hostname(),
log("INFO", "NATS: Using username/password authentication"); reconnect: true,
} else { maxReconnectAttempts: -1,
log("INFO", "NATS: Connecting without authentication"); waitOnFirstConnect: true,
};
if (cfg.MQ_USER && cfg.MQ_PASS) {
options.user = cfg.MQ_USER;
options.pass = cfg.MQ_PASS;
log("INFO", "NATS: Using auth credentials");
}
try {
this.nc = await connect(options);
this.js = this.nc.jetstream();
log("INFO", `NATS: Connected to ${cfg.MQ_URL}`);
// Handle connection close events
this.nc.closed().then((err) => {
if (err) {
log("ERROR", `NATS: Connection closed: ${err}`);
} else {
log("INFO", "NATS: Connection ended")
}
this.nc = null;
this.js = null;
this.kv = null;
});
} catch (err) {
log("ERROR", `NATS: Initial connection failed: ${err}`);
throw err;
}
} }
return await connect(options) /**
} * Accessor for the KV store
*/
// Send Message Function here to send the message to NATS private async getKV(): Promise<KV> {
export async function sendFileUpdateMessage(path: string, version: string): Promise<boolean> { if (!this.js) throw new Error("NATS: JetStream not initialized. Call connect() first.");
const serviceName: string = "pis-data-ingress"; if (!this.kv) {
const serviceId: string = hostname(); this.kv = await this.js.views.kv(this.bucketName);
const message: MQFileUpdate = {
service_name: "pis-data-ingress",
service_id: serviceId,
sent_timestamp: Math.floor(Date.now() / 1000),
data_type: "file",
data_kind: "pis",
payload: {
version: version,
filepath: path,
} }
}; return this.kv;
}
let nats: NatsConnection | undefined;
try { /**
const nats: NatsConnection = await getNatsConnection(); * Get the last recorded state/hash for a service
*/
async getState(key: string): Promise<string | null> {
try {
const store = await this.getKV();
const entry = await store.get(key);
return entry ? sc.decode(entry.value) : null;
} catch (err) {
log("ERROR", `NATS: KV Get failed for ${key}: ${err}`);
return null;
}
}
/**
* Set the current state/hash for a service
*/
async setState(key: string, value: string): Promise<void> {
try {
const store = await this.getKV();
await store.put(key, sc.encode(value));
} catch (err) {
log("ERROR", `NATS: KV Set failed for ${key}: ${err}`);
throw err;
}
}
/**
* Publishes message to the JetStream
*/
async sendFileUpdateMessage(path: string, version: string, serviceName: string): Promise<boolean> {
if (!this.js) throw new Error("NATS: JetStream not initialized");
const message: MQFileUpdate = {
service_name: serviceName,
service_id: hostname(),
sent_timestamp: Math.floor(Date.now() / 1000),
data_type: "file",
data_kind: "pis",
payload: {
version: version,
filepath: path,
}
};
const subject = `ingress.file.${message.data_kind}`; const subject = `ingress.file.${message.data_kind}`;
nats.publish(subject, jc.encode(message)); try {
await this.js.publish(subject, jc.encode(message));
await nats.drain(); log("INFO", `NATS: Message published to ${subject}`);
return true return true;
} catch (err) { } catch (err) {
log("ERROR", `NATS: Failed to send message: ${err}`); log("ERROR", `NATS: Failed to publish to JetStream: ${err}`);
if (nats) {nats.close()} return false;
return false; }
} }
}
async close() {
if (this.nc) {
await this.nc.drain();
this.nc = null;
}
}
}
// Export instance
export const natsManager = new NatsManager();

View File

@@ -1,8 +1,8 @@
import { Readable } from 'node:stream'; import { Readable } from 'node:stream';
import { createInterface } from 'node:readline'; import { createInterface } from 'node:readline';
import XXH from 'xxhashjs';
import { log } from './logger.js'; import { log } from './logger.js';
import { DataIngressPisData } from '@owlboard/backend-data-contracts'; import { DataIngressPisData } from '@owlboard/backend-data-contracts';
import type { GeneralConfig } from './config.js';
const BASE_URL = process.env.BASEURL || 'https://owlboard.info' const BASE_URL = process.env.BASEURL || 'https://owlboard.info'
@@ -18,8 +18,8 @@ interface InputRecord {
stops: string[]; stops: string[];
} }
export async function* processPisStream(inputStream: Readable) { export async function* processPisStream(cfg: GeneralConfig, inputStream: Readable) {
const TOC = process.env.TOC; const TOC = cfg.TOC;
if (!TOC) { if (!TOC) {
log('ERROR', "TOC not set: Exit code 19"); log('ERROR', "TOC not set: Exit code 19");
process.exit(19); process.exit(19);
@@ -35,19 +35,19 @@ export async function* processPisStream(inputStream: Readable) {
if (!line.trim()) continue; if (!line.trim()) continue;
const record = JSON.parse(line) as InputRecord; const record = JSON.parse(line) as InputRecord;
log('DEBUG', `JSON Line Parsed: ${JSON.stringify(record)}`) // log('DEBUG', `JSON Line Parsed: ${JSON.stringify(record)}`)
const crsHash = XXH.h64(record.stops.join('|'), SEED); const normalisedStops = record.stops.map(stop => stop.toUpperCase());
const tiplocStops = await mapStopsToTiploc(record.stops);
const tiplocHash = XXH.h64(tiplocStops.join('|'), SEED); const tiplocStops = await mapStopsToTiploc(normalisedStops);
const data: DataIngressPisData.PisObjects = { const data: DataIngressPisData.PisObjects = {
code: record.code, code: record.code,
toc: TOC.toLowerCase(), toc: TOC.toLowerCase(),
crsStops: record.stops, crsStops: normalisedStops,
crsHash: crsHash.toString(), crsHash: "",
tiplocStops: tiplocStops, tiplocStops: tiplocStops,
tiplocHash: tiplocHash.toString(), tiplocHash: "",
} }
yield data; yield data;
@@ -63,8 +63,9 @@ async function mapStopsToTiploc(crsStops: string[]): Promise<string[]> {
// Cache Miss // Cache Miss
try { try {
const tiploc = await fetchTiplocFromApi(crs); const tiploc = await fetchTiplocFromApi(crs);
tiplocCache.set(crs, tiploc); const normalisedTiploc = tiploc.toUpperCase();
return tiploc; tiplocCache.set(crs, normalisedTiploc);
return normalisedTiploc;
} catch (err) { } catch (err) {
log('ERROR', `Failed lookup for: ${crs}`, err); log('ERROR', `Failed lookup for: ${crs}`, err);
process.exit(99); process.exit(99);

View File

@@ -1,11 +1,13 @@
import { S3Client } from "@aws-sdk/client-s3"; import { S3Client, CreateBucketCommand, HeadBucketCommand, S3ServiceException } from "@aws-sdk/client-s3";
import { Upload } from "@aws-sdk/lib-storage"; import { Upload } from "@aws-sdk/lib-storage";
import { createWriteStream } from "node:fs"; import { createWriteStream } from "node:fs";
import { Readable } from "node:stream"; import { Readable } from "node:stream";
import { DataIngressPisData } from "@owlboard/backend-data-contracts"; import { DataIngressPisData } from "@owlboard/backend-data-contracts";
import { log } from "./logger.js"; import { log } from "./logger.js";
import type { S3 } from "./config.js";
export async function processAndStore( export async function processAndStore(
cfg: S3,
generator: AsyncGenerator<DataIngressPisData.PisObjects>, generator: AsyncGenerator<DataIngressPisData.PisObjects>,
filename: string filename: string
) { ) {
@@ -15,29 +17,45 @@ export async function processAndStore(
} }
})()); })());
const useS3 = process.env.S3_ENDPOINT && process.env.S3_BUCKET; const useS3 = cfg.S3_ENDPOINT && cfg.S3_BUCKET;
if (useS3) { if (useS3) {
if (!process.env.S3_ENDPOINT || process.env.S3_BUCKET || !process.env.S3_REGION || !process.env.S3_ACCESS_KEY || !process.env.S3_SECRET_KEY) { if (!cfg.S3_ENDPOINT || !cfg.S3_BUCKET || !cfg.S3_REGION || !cfg.S3_ACCESS_KEY || !cfg.S3_SECRET_KEY) {
log("DEBUG", "Missing required variables for S3 Support - EXIT CODE 24"); log("DEBUG", "Missing required variables for S3 Support - EXIT CODE 24");
process.exit(24); process.exit(24);
} }
log('INFO', `Streaming to S3: ${process.env.S3_BUCKET}/${filename}`);
log("DEBUG", "Opening connection to S3 Server");
const client = new S3Client({ const client = new S3Client({
endpoint: process.env.S3_ENDPOINT, endpoint: cfg.S3_ENDPOINT,
region: process.env.S3_REGION, region: cfg.S3_REGION,
credentials: { credentials: {
accessKeyId: process.env.S3_ACCESS_KEY!, accessKeyId: cfg.S3_ACCESS_KEY!,
secretAccessKey: process.env.S3_SECRET_KEY, secretAccessKey: cfg.S3_SECRET_KEY,
}, },
forcePathStyle: true, forcePathStyle: true,
}); });
// Check bucket exists, and create if needed
try {
await client.send(new HeadBucketCommand({ Bucket: cfg.S3_BUCKET }));
} catch (err) {
if (err instanceof S3ServiceException && err.$metadata.httpStatusCode === 404) {
log('INFO', `Bucket ${cfg.S3_BUCKET} does not exist, creating...`);
await client.send(new CreateBucketCommand({ Bucket: cfg.S3_BUCKET }));
} else {
log(`ERROR`, `Failed to verify bucket: ${cfg.S3_BUCKET}`);
process.exit(21);
}
}
log('INFO', `Streaming to S3: ${cfg.S3_BUCKET}/${filename}`);
const upload = new Upload({ const upload = new Upload({
client, client,
params: { params: {
Bucket: process.env.S3_BUCKET, Bucket: cfg.S3_BUCKET,
Key: filename, Key: filename,
Body: ndjsonStream, Body: ndjsonStream,
ContentType: "application/x-ndjson", ContentType: "application/x-ndjson",