🔧 kafka-powered autodownload loop

This commit is contained in:
2024-06-03 17:18:22 -04:00
parent 12e46334da
commit 60e5b6f61b
2 changed files with 125 additions and 48 deletions

View File

@@ -2,6 +2,25 @@
import { Context, Service, ServiceBroker, ServiceSchema, Errors } from "moleculer"; import { Context, Service, ServiceBroker, ServiceSchema, Errors } from "moleculer";
import { Kafka } from "kafkajs"; import { Kafka } from "kafkajs";
interface Comic {
wanted: {
markEntireVolumeWanted?: boolean;
issues?: Array<any>;
volume: {
id: string;
name: string;
};
};
}
interface PaginatedResult {
wantedComics: Comic[];
total: number;
page: number;
limit: number;
pages: number;
}
export default class AutoDownloadService extends Service { export default class AutoDownloadService extends Service {
private kafkaProducer: any; private kafkaProducer: any;
@@ -18,35 +37,60 @@ export default class AutoDownloadService extends Service {
rest: "POST /searchWantedComics", rest: "POST /searchWantedComics",
handler: async (ctx: Context<{}>) => { handler: async (ctx: Context<{}>) => {
try { try {
const wantedComics: any = await this.broker.call( let page = 1;
"library.getComicsMarkedAsWanted", const limit = this.BATCH_SIZE;
{},
);
this.logger.info("Fetched wanted comics:", wantedComics.length);
for (const comic of wantedComics) { while (true) {
if (comic.wanted.markEntireVolumeWanted) { const result: PaginatedResult = await this.broker.call(
const issues: any = await this.broker.call( "library.getComicsMarkedAsWanted",
"comicvine.getIssuesForVolume", { page, limit },
{ );
volumeId: comic.wanted.volume.id,
}, if (!result || !result.wantedComics) {
this.logger.error("Invalid response structure", result);
throw new Errors.MoleculerError(
"Invalid response structure from getComicsMarkedAsWanted",
500,
"INVALID_RESPONSE_STRUCTURE",
); );
for (const issue of issues) { }
await this.produceJobToKafka(
comic.wanted.volume.name, this.logger.info(
issue, `Fetched ${result.wantedComics.length} comics from page ${page} of ${result.pages}`,
); );
}
} else if (comic.wanted.issues && comic.wanted.issues.length > 0) { for (const comic of result.wantedComics) {
for (const issue of comic.wanted.issues) { if (comic.wanted.markEntireVolumeWanted) {
await this.produceJobToKafka( const issues: any = await this.broker.call(
comic.wanted.volume?.name, "comicvine.getIssuesForVolume",
issue, {
volumeId: comic.wanted.volume.id,
},
); );
for (const issue of issues) {
await this.produceJobToKafka(
comic.wanted.volume.name,
issue,
);
}
} else if (
comic.wanted.issues &&
comic.wanted.issues.length > 0
) {
for (const issue of comic.wanted.issues) {
await this.produceJobToKafka(
comic.wanted.volume?.name,
issue,
);
}
} }
} }
if (page >= result.pages) break;
page += 1;
} }
return { success: true, message: "Processing started." };
} catch (error) { } catch (error) {
this.logger.error("Error in searchWantedComics:", error); this.logger.error("Error in searchWantedComics:", error);
throw new Errors.MoleculerError( throw new Errors.MoleculerError(

View File

@@ -3,7 +3,20 @@ import { Service, ServiceBroker, ServiceSchema } from "moleculer";
import { Kafka, EachMessagePayload, logLevel } from "kafkajs"; import { Kafka, EachMessagePayload, logLevel } from "kafkajs";
import { isUndefined } from "lodash"; import { isUndefined } from "lodash";
import io from "socket.io-client"; import io from "socket.io-client";
interface SearchResult {
result: {
id: string;
// Add other relevant fields
};
search_id: string;
// Add other relevant fields
}
interface SearchResultPayload {
groupedResult: SearchResult;
updatedResult: SearchResult;
instanceId: string;
}
export default class ComicProcessorService extends Service { export default class ComicProcessorService extends Service {
private kafkaConsumer: any; private kafkaConsumer: any;
private socketIOInstance: any; private socketIOInstance: any;
@@ -136,35 +149,55 @@ export default class ComicProcessorService extends Service {
this.logger.info("Socket.IO connected successfully."); this.logger.info("Socket.IO connected successfully.");
}); });
this.socketIOInstance.on("searchResultAdded", (data: any) => { this.socketIOInstance.on(
this.logger.info( "searchResultAdded",
"Received search result added:", ({ groupedResult, instanceId }: SearchResultPayload) => {
JSON.stringify(data, null, 4), this.logger.info(
); "Received search result added:",
this.airDCPPSearchResults.push(data); JSON.stringify(groupedResult, null, 4),
});
this.socketIOInstance.on("searchResultUpdated", async (data: any) => {
this.logger.info(
"Received search result update:",
JSON.stringify(data, null, 4),
);
if (
!isUndefined(data.result) &&
!isUndefined(this.airDCPPSearchResults.result)
) {
const toReplaceIndex = this.airDCPPSearchResults.findIndex(
(element: any) => {
return element?.result.id === data.result.id;
},
); );
this.airDCPPSearchResults[toReplaceIndex] = data.result; this.airDCPPSearchResults.push({
} groupedResult: groupedResult.result,
}); instanceId,
});
},
);
this.socketIOInstance.on(
"searchResultUpdated",
async ({ updatedResult, instanceId }: SearchResultPayload) => {
this.logger.info(
"Received search result update:",
JSON.stringify(updatedResult, null, 4),
);
if (
!isUndefined(updatedResult.result) &&
!isUndefined(this.airDCPPSearchResults.result)
) {
const toReplaceIndex = this.airDCPPSearchResults.findIndex(
(element: any) => {
return element?.result.id === updatedResult.result.id;
},
);
this.airDCPPSearchResults[toReplaceIndex] = {
result: updatedResult.result,
instanceId,
};
}
},
);
this.socketIOInstance.on("searchComplete", async () => { this.socketIOInstance.on("searchComplete", async () => {
// Ensure results are not empty before producing to Kafka // Ensure results are not empty before producing to Kafka
if (this.airDCPPSearchResults.length > 0) { if (this.airDCPPSearchResults.length > 0) {
await this.produceResultsToKafka(this.airDCPPSearchResults, []); const results = this.airDCPPSearchResults.reduce((acc: any, item: any) => {
const key = item.instanceId;
if (!acc[key]) {
acc[key] = [];
}
acc[key].push(item);
return acc;
}, {});
await this.produceResultsToKafka(results, []);
} else { } else {
this.logger.warn( this.logger.warn(
"AirDC++ search results are empty, not producing to Kafka.", "AirDC++ search results are empty, not producing to Kafka.",