143 lines
4.2 KiB
TypeScript
143 lines
4.2 KiB
TypeScript
/*
|
|
* 🦉 Bubo Reader
|
|
* ====
|
|
* Dead simple feed reader (RSS + JSON) that renders an HTML
|
|
* page with links to content from feeds organized by site
|
|
*
|
|
* Code: https://github.com/georgemandis/bubo-rss
|
|
* Copyright (c) 2019 George Mandis (https://george.mand.is)
|
|
* Version: 1.0.1 (11/14/2021)
|
|
* Licensed under the MIT License (http://opensource.org/licenses/MIT)
|
|
*/
|
|
|
|
import fetch from "node-fetch";
|
|
import Parser from "rss-parser";
|
|
import { Feeds, FeedItem } from "./@types/bubo";
|
|
import { Response } from "node-fetch";
|
|
import { render } from "./renderer.js";
|
|
import { getLink, getTitle, getTimestamp, parseFeed, getFeedList } from "./utilities.js";
|
|
import { writeFile } from "fs/promises";
|
|
import chalk from "chalk";
|
|
|
|
const parser = new Parser();
|
|
const feedList = await getFeedList();
|
|
const feedListLength = Object.entries(feedList).flat(2).length - Object.keys(feedList).length;
|
|
|
|
/**
|
|
* contentFromAllFeeds = Contains normalized, aggregated feed data and is passed to template renderer at the end
|
|
* errors = Contains errors from parsing feeds and is also passed to template.
|
|
*/
|
|
const contentFromAllFeeds: Feeds = {};
|
|
const errors: unknown[] = [];
|
|
|
|
// benchmarking data + utility
|
|
const initTime = Date.now();
|
|
const benchmark = (startTime: number) => chalk.cyanBright.bold(`(${(Date.now() - startTime) / 1000} seconds)`);
|
|
|
|
/**
|
|
* These values are used to control throttling/batching the fetches:
|
|
* - MAX_CONNECTION = max number of fetches to contain in a batch
|
|
* - DELAY_MS = the delay in milliseconds between batches
|
|
*/
|
|
const MAX_CONNECTIONS = Infinity;
|
|
const DELAY_MS = 850;
|
|
|
|
const error = chalk.bold.red;
|
|
const success = chalk.bold.green;
|
|
|
|
// keeping tally of total feeds fetched and parsed so we can compare
|
|
// to feedListLength and know when we're finished.
|
|
let completed = 0;
|
|
|
|
|
|
/**
|
|
* finishBuild
|
|
* --
|
|
* function that gets called when all the feeds are through fetching
|
|
* and we want to build the static output.
|
|
*/
|
|
const finishBuild: () => void = async () => {
|
|
console.log("\nDone fetching everything!");
|
|
|
|
// generate the static HTML output from our template renderer
|
|
const output = render({
|
|
data: contentFromAllFeeds,
|
|
errors: errors
|
|
});
|
|
|
|
// write the output to public/index.html
|
|
await writeFile("./public/index.html", output);
|
|
console.log(`Finished writing to output. ${benchmark(initTime)}`);
|
|
};
|
|
|
|
/**
|
|
* processFeed
|
|
* --
|
|
* Process an individual feed and normalize its items
|
|
* @param { group, feed, startTime}
|
|
* @returns Promise<void>
|
|
*/
|
|
const processFeed = (
|
|
{
|
|
group, feed, startTime
|
|
}: { group: string; feed: string, startTime: number }
|
|
) => async (response: Response): Promise<void> => {
|
|
const body = await parseFeed(response);
|
|
completed++;
|
|
// skip to the next one if this didn't work out
|
|
if (!body) return;
|
|
|
|
try {
|
|
const contents: FeedItem =
|
|
(typeof body === "string" ? (await parser.parseString(body)) : body) as FeedItem;
|
|
|
|
contents.feed = feed;
|
|
contents.title = getTitle(contents);
|
|
contents.link = getLink(contents);
|
|
|
|
// try to normalize date attribute naming
|
|
contents?.items?.forEach((item) => {
|
|
item.timestamp = getTimestamp(item);
|
|
item.title = getTitle(item);
|
|
item.link = getLink(item);
|
|
});
|
|
|
|
contentFromAllFeeds[group].push(contents as object);
|
|
console.log(`${success("Successfully fetched:")} ${feed} ${benchmark(startTime)}`);
|
|
|
|
} catch (err) {
|
|
console.log(`${error("Error processing:")} ${feed} ${benchmark(startTime)}`);
|
|
errors.push(`Error processing: ${feed} | ${err}`);
|
|
}
|
|
|
|
// if this is the last feed, go ahead and build the output
|
|
(completed === feedListLength) && finishBuild();
|
|
};
|
|
|
|
|
|
// go through each group of feeds and process
|
|
const processFeeds = () => {
|
|
let idx = 0;
|
|
|
|
for (const [group, feeds] of Object.entries(feedList)) {
|
|
contentFromAllFeeds[group] = [];
|
|
|
|
for (const feed of feeds) {
|
|
const startTime = Date.now();
|
|
setTimeout(() => {
|
|
console.log(`Fetching: ${feed}...`);
|
|
|
|
fetch(feed).then(processFeed({ group, feed, startTime })).catch(err => {
|
|
console.log(error(`Error fetching ${feed} ${benchmark(startTime)}`));
|
|
errors.push(`Error fetching ${feed} ${err.toString()}`);
|
|
});
|
|
|
|
}, (idx % (feedListLength / MAX_CONNECTIONS)) * DELAY_MS);
|
|
idx++;
|
|
}
|
|
|
|
}
|
|
};
|
|
|
|
|
|
processFeeds(); |