Convert the streaming server to ESM (#29389)

Co-authored-by: Claire <claire.github-309c@sitedethib.com>
pull/2650/head
Renaud Chaput 2024-02-27 15:59:20 +01:00 committed by GitHub
parent bc4c5ed918
commit 036f5a05e3
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
8 changed files with 65 additions and 78 deletions

View File

@ -1,4 +1,8 @@
/* eslint-disable import/no-commonjs */
// @ts-check
// @ts-ignore - This needs to be a CJS file (eslint does not yet support ESM configs), and TS is complaining we use require
const { defineConfig } = require('eslint-define-config');
module.exports = defineConfig({
@ -22,22 +26,18 @@ module.exports = defineConfig({
// to maintain.
'no-delete-var': 'off',
// The streaming server is written in commonjs, not ESM for now:
'import/no-commonjs': 'off',
// This overrides the base configuration for this rule to pick up
// dependencies for the streaming server from the correct package.json file.
'import/no-extraneous-dependencies': [
'error',
{
devDependencies: [
'streaming/.eslintrc.js',
],
devDependencies: ['streaming/.eslintrc.cjs'],
optionalDependencies: false,
peerDependencies: false,
includeTypes: true,
packageDir: __dirname,
},
],
'import/extensions': ['error', 'always'],
},
});

View File

@ -5,15 +5,14 @@
* override it in let statements.
* @type {string}
*/
const UNEXPECTED_ERROR_MESSAGE = 'An unexpected error occurred';
exports.UNKNOWN_ERROR_MESSAGE = UNEXPECTED_ERROR_MESSAGE;
export const UNEXPECTED_ERROR_MESSAGE = 'An unexpected error occurred';
/**
* Extracts the status and message properties from the error object, if
* available for public use. The `unknown` is for catch statements
* @param {Error | AuthenticationError | RequestError | unknown} err
*/
exports.extractStatusAndMessage = function(err) {
export function extractStatusAndMessage(err) {
let statusCode = 500;
let errorMessage = UNEXPECTED_ERROR_MESSAGE;
if (err instanceof AuthenticationError || err instanceof RequestError) {
@ -22,9 +21,9 @@ exports.extractStatusAndMessage = function(err) {
}
return { statusCode, errorMessage };
};
}
class RequestError extends Error {
export class RequestError extends Error {
/**
* @param {string} message
*/
@ -35,9 +34,7 @@ class RequestError extends Error {
}
}
exports.RequestError = RequestError;
class AuthenticationError extends Error {
export class AuthenticationError extends Error {
/**
* @param {string} message
*/
@ -47,5 +44,3 @@ class AuthenticationError extends Error {
this.status = 401;
}
}
exports.AuthenticationError = AuthenticationError;

View File

@ -1,32 +1,36 @@
// @ts-check
const fs = require('fs');
const http = require('http');
const path = require('path');
const url = require('url');
import fs from 'node:fs';
import http from 'node:http';
import path from 'node:path';
import url from 'node:url';
const cors = require('cors');
const dotenv = require('dotenv');
const express = require('express');
const { Redis } = require('ioredis');
const { JSDOM } = require('jsdom');
const pg = require('pg');
const dbUrlToConfig = require('pg-connection-string').parse;
const WebSocket = require('ws');
import cors from 'cors';
import dotenv from 'dotenv';
import express from 'express';
import { Redis } from 'ioredis';
import { JSDOM } from 'jsdom';
import pg from 'pg';
import pgConnectionString from 'pg-connection-string';
import WebSocket from 'ws';
const errors = require('./errors');
const { AuthenticationError, RequestError } = require('./errors');
const { logger, httpLogger, initializeLogLevel, attachWebsocketHttpLogger, createWebsocketLogger } = require('./logging');
const { setupMetrics } = require('./metrics');
const { isTruthy, normalizeHashtag, firstParam } = require("./utils");
import { AuthenticationError, RequestError, extractStatusAndMessage as extractErrorStatusAndMessage } from './errors.js';
import { logger, httpLogger, initializeLogLevel, attachWebsocketHttpLogger, createWebsocketLogger } from './logging.js';
import { setupMetrics } from './metrics.js';
import { isTruthy, normalizeHashtag, firstParam } from './utils.js';
const environment = process.env.NODE_ENV || 'development';
// Correctly detect and load .env or .env.production file based on environment:
const dotenvFile = environment === 'production' ? '.env.production' : '.env';
const dotenvFilePath = path.resolve(
url.fileURLToPath(
new URL(path.join('..', dotenvFile), import.meta.url)
)
);
dotenv.config({
path: path.resolve(__dirname, path.join('..', dotenvFile))
path: dotenvFilePath
});
initializeLogLevel(process.env, environment);
@ -143,7 +147,7 @@ const pgConfigFromEnv = (env) => {
let baseConfig = {};
if (env.DATABASE_URL) {
const parsedUrl = dbUrlToConfig(env.DATABASE_URL);
const parsedUrl = pgConnectionString.parse(env.DATABASE_URL);
// The result of dbUrlToConfig from pg-connection-string is not type
// compatible with pg.PoolConfig, since parts of the connection URL may be
@ -326,7 +330,7 @@ const startServer = async () => {
// Unfortunately for using the on('upgrade') setup, we need to manually
// write a HTTP Response to the Socket to close the connection upgrade
// attempt, so the following code is to handle all of that.
const {statusCode, errorMessage } = errors.extractStatusAndMessage(err);
const {statusCode, errorMessage } = extractErrorStatusAndMessage(err);
/** @type {Record<string, string | number | import('pino-http').ReqId>} */
const headers = {
@ -748,7 +752,7 @@ const startServer = async () => {
return;
}
const {statusCode, errorMessage } = errors.extractStatusAndMessage(err);
const {statusCode, errorMessage } = extractErrorStatusAndMessage(err);
res.writeHead(statusCode, { 'Content-Type': 'application/json' });
res.end(JSON.stringify({ error: errorMessage }));
@ -1155,7 +1159,7 @@ const startServer = async () => {
// @ts-ignore
streamFrom(channelIds, req, req.log, onSend, onEnd, 'eventsource', options.needsFiltering);
}).catch(err => {
const {statusCode, errorMessage } = errors.extractStatusAndMessage(err);
const {statusCode, errorMessage } = extractErrorStatusAndMessage(err);
res.log.info({ err }, 'Eventsource subscription error');
@ -1353,7 +1357,7 @@ const startServer = async () => {
stopHeartbeat,
};
}).catch(err => {
const {statusCode, errorMessage } = errors.extractStatusAndMessage(err);
const {statusCode, errorMessage } = extractErrorStatusAndMessage(err);
logger.error({ err }, 'Websocket subscription error');
@ -1482,13 +1486,15 @@ const startServer = async () => {
// Decrement the metrics for connected clients:
connectedClients.labels({ type: 'websocket' }).dec();
// We need to delete the session object as to ensure it correctly gets
// We need to unassign the session object as to ensure it correctly gets
// garbage collected, without doing this we could accidentally hold on to
// references to the websocket, the request, and the logger, causing
// memory leaks.
//
// @ts-ignore
delete session;
// This is commented out because `delete` only operated on object properties
// It needs to be replaced by `session = undefined`, but it requires every calls to
// `session` to check for it, thus a significant refactor
// delete session;
});
// Note: immediately after the `error` event is emitted, the `close` event

View File

@ -1,6 +1,6 @@
const { pino } = require('pino');
const { pinoHttp, stdSerializers: pinoHttpSerializers } = require('pino-http');
const uuid = require('uuid');
import { pino } from 'pino';
import { pinoHttp, stdSerializers as pinoHttpSerializers } from 'pino-http';
import * as uuid from 'uuid';
/**
* Generates the Request ID for logging and setting on responses
@ -36,7 +36,7 @@ function sanitizeRequestLog(req) {
return log;
}
const logger = pino({
export const logger = pino({
name: "streaming",
// Reformat the log level to a string:
formatters: {
@ -59,7 +59,7 @@ const logger = pino({
}
});
const httpLogger = pinoHttp({
export const httpLogger = pinoHttp({
logger,
genReqId: generateRequestId,
serializers: {
@ -71,7 +71,7 @@ const httpLogger = pinoHttp({
* Attaches a logger to the request object received by http upgrade handlers
* @param {http.IncomingMessage} request
*/
function attachWebsocketHttpLogger(request) {
export function attachWebsocketHttpLogger(request) {
generateRequestId(request);
request.log = logger.child({
@ -84,7 +84,7 @@ function attachWebsocketHttpLogger(request) {
* @param {http.IncomingMessage} request
* @param {import('./index.js').ResolvedAccount} resolvedAccount
*/
function createWebsocketLogger(request, resolvedAccount) {
export function createWebsocketLogger(request, resolvedAccount) {
// ensure the request.id is always present.
generateRequestId(request);
@ -98,17 +98,12 @@ function createWebsocketLogger(request, resolvedAccount) {
});
}
exports.logger = logger;
exports.httpLogger = httpLogger;
exports.attachWebsocketHttpLogger = attachWebsocketHttpLogger;
exports.createWebsocketLogger = createWebsocketLogger;
/**
* Initializes the log level based on the environment
* @param {Object<string, any>} env
* @param {string} environment
*/
exports.initializeLogLevel = function initializeLogLevel(env, environment) {
export function initializeLogLevel(env, environment) {
if (env.LOG_LEVEL && Object.keys(logger.levels.values).includes(env.LOG_LEVEL)) {
logger.level = env.LOG_LEVEL;
} else if (environment === 'development') {
@ -116,4 +111,4 @@ exports.initializeLogLevel = function initializeLogLevel(env, environment) {
} else {
logger.level = 'info';
}
};
}

View File

@ -1,6 +1,6 @@
// @ts-check
const metrics = require('prom-client');
import metrics from 'prom-client';
/**
* @typedef StreamingMetrics
@ -18,7 +18,7 @@ const metrics = require('prom-client');
* @param {import('pg').Pool} pgPool
* @returns {StreamingMetrics}
*/
function setupMetrics(channels, pgPool) {
export function setupMetrics(channels, pgPool) {
// Collect metrics from Node.js
metrics.collectDefaultMetrics();
@ -101,5 +101,3 @@ function setupMetrics(channels, pgPool) {
messagesSent,
};
}
exports.setupMetrics = setupMetrics;

View File

@ -7,6 +7,7 @@
},
"description": "Mastodon's Streaming Server",
"private": true,
"type": "module",
"repository": {
"type": "git",
"url": "https://github.com/mastodon/mastodon.git"

View File

@ -2,11 +2,11 @@
"extends": "../tsconfig.json",
"compilerOptions": {
"target": "esnext",
"module": "CommonJS",
"moduleResolution": "node",
"module": "NodeNext",
"moduleResolution": "NodeNext",
"noUnusedParameters": false,
"tsBuildInfoFile": "../tmp/cache/streaming/tsconfig.tsbuildinfo",
"paths": {},
},
"include": ["./*.js", "./.eslintrc.js"],
"include": ["./*.js", "./.eslintrc.cjs"],
}

View File

@ -16,11 +16,9 @@ const FALSE_VALUES = [
* @param {any} value
* @returns {boolean}
*/
const isTruthy = value =>
value && !FALSE_VALUES.includes(value);
exports.isTruthy = isTruthy;
export function isTruthy(value) {
return value && !FALSE_VALUES.includes(value);
}
/**
* See app/lib/ascii_folder.rb for the canon definitions
@ -33,7 +31,7 @@ const EQUIVALENT_ASCII_CHARS = 'AAAAAAaaaaaaAaAaAaCcCcCcCcCcDdDdDdEEEEeeeeEeEeEe
* @param {string} str
* @returns {string}
*/
function foldToASCII(str) {
export function foldToASCII(str) {
const regex = new RegExp(NON_ASCII_CHARS.split('').join('|'), 'g');
return str.replace(regex, function(match) {
@ -42,28 +40,22 @@ function foldToASCII(str) {
});
}
exports.foldToASCII = foldToASCII;
/**
* @param {string} str
* @returns {string}
*/
function normalizeHashtag(str) {
export function normalizeHashtag(str) {
return foldToASCII(str.normalize('NFKC').toLowerCase()).replace(/[^\p{L}\p{N}_\u00b7\u200c]/gu, '');
}
exports.normalizeHashtag = normalizeHashtag;
/**
* @param {string|string[]} arrayOrString
* @returns {string}
*/
function firstParam(arrayOrString) {
export function firstParam(arrayOrString) {
if (Array.isArray(arrayOrString)) {
return arrayOrString[0];
} else {
return arrayOrString;
}
}
exports.firstParam = firstParam;