388 lines
13 KiB
TypeScript
388 lines
13 KiB
TypeScript
import { Job } from "bullmq";
|
|
import { db } from "db/init.ts";
|
|
import { getLatestVideoSnapshot, getVideosNearMilestone } from "db/snapshot.ts";
|
|
import {
|
|
bulkGetVideosWithoutProcessingSchedules,
|
|
bulkScheduleSnapshot,
|
|
bulkSetSnapshotStatus,
|
|
findClosestSnapshot,
|
|
findSnapshotBefore, getAllVideosWithoutActiveSnapshotSchedule,
|
|
getBulkSnapshotsInNextSecond,
|
|
getLatestSnapshot,
|
|
getSnapshotsInNextSecond,
|
|
getVideosWithoutActiveSnapshotSchedule,
|
|
scheduleSnapshot,
|
|
setSnapshotStatus,
|
|
snapshotScheduleExists,
|
|
videoHasProcessingSchedule,
|
|
} from "db/snapshotSchedule.ts";
|
|
import { Client } from "https://deno.land/x/postgres@v0.19.3/mod.ts";
|
|
import { HOUR, MINUTE, SECOND, WEEK } from "$std/datetime/constants.ts";
|
|
import logger from "log/logger.ts";
|
|
import { SnapshotQueue } from "mq/index.ts";
|
|
import { insertVideoSnapshot } from "mq/task/getVideoStats.ts";
|
|
import { NetSchedulerError } from "@core/net/delegate.ts";
|
|
import { getBiliVideoStatus, setBiliVideoStatus } from "db/allData.ts";
|
|
import { truncate } from "utils/truncate.ts";
|
|
import { lockManager } from "mq/lockManager.ts";
|
|
import { getSongsPublihsedAt } from "db/songs.ts";
|
|
import { bulkGetVideoStats } from "net/bulkGetVideoStats.ts";
|
|
import { getAdjustedShortTermETA } from "../scheduling.ts";
|
|
import {SnapshotScheduleType} from "@core/db/schema";
|
|
|
|
const priorityMap: { [key: string]: number } = {
|
|
"milestone": 1,
|
|
"normal": 3,
|
|
};
|
|
|
|
const snapshotTypeToTaskMap: { [key: string]: string } = {
|
|
"milestone": "snapshotMilestoneVideo",
|
|
"normal": "snapshotVideo",
|
|
"new": "snapshotMilestoneVideo",
|
|
};
|
|
|
|
export const bulkSnapshotTickWorker = async (_job: Job) => {
|
|
const client = await db.connect();
|
|
try {
|
|
const schedules = await getBulkSnapshotsInNextSecond(client);
|
|
const count = schedules.length;
|
|
const groups = Math.ceil(count / 30);
|
|
for (let i = 0; i < groups; i++) {
|
|
const group = schedules.slice(i * 30, (i + 1) * 30);
|
|
const aids = group.map((schedule) => Number(schedule.aid));
|
|
const filteredAids = await bulkGetVideosWithoutProcessingSchedules(client, aids);
|
|
if (filteredAids.length === 0) continue;
|
|
await bulkSetSnapshotStatus(client, filteredAids, "processing");
|
|
await SnapshotQueue.add("bulkSnapshotVideo", {
|
|
schedules: group,
|
|
}, { priority: 3 });
|
|
}
|
|
return `OK`
|
|
} catch (e) {
|
|
logger.error(e as Error);
|
|
} finally {
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
export const snapshotTickWorker = async (_job: Job) => {
|
|
const client = await db.connect();
|
|
try {
|
|
const schedules = await getSnapshotsInNextSecond(client);
|
|
for (const schedule of schedules) {
|
|
if (await videoHasProcessingSchedule(client, Number(schedule.aid))) {
|
|
continue;
|
|
}
|
|
let priority = 3;
|
|
if (schedule.type && priorityMap[schedule.type]) {
|
|
priority = priorityMap[schedule.type];
|
|
}
|
|
const aid = Number(schedule.aid);
|
|
await setSnapshotStatus(client, schedule.id, "processing");
|
|
await SnapshotQueue.add("snapshotVideo", {
|
|
aid: aid,
|
|
id: Number(schedule.id),
|
|
type: schedule.type ?? "normal",
|
|
}, { priority });
|
|
}
|
|
return `OK`;
|
|
} catch (e) {
|
|
logger.error(e as Error);
|
|
} finally {
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
export const closetMilestone = (views: number) => {
|
|
if (views < 100000) return 100000;
|
|
if (views < 1000000) return 1000000;
|
|
return 10000000;
|
|
};
|
|
|
|
export const collectMilestoneSnapshotsWorker = async (_job: Job) => {
|
|
const client = await db.connect();
|
|
try {
|
|
const videos = await getVideosNearMilestone(client);
|
|
for (const video of videos) {
|
|
const aid = Number(video.aid);
|
|
const eta = await getAdjustedShortTermETA(client, aid);
|
|
if (eta > 144) continue;
|
|
const now = Date.now();
|
|
const scheduledNextSnapshotDelay = eta * HOUR;
|
|
const maxInterval = 4 * HOUR;
|
|
const minInterval = 1 * SECOND;
|
|
const delay = truncate(scheduledNextSnapshotDelay, minInterval, maxInterval);
|
|
const targetTime = now + delay;
|
|
await scheduleSnapshot(client, aid, "milestone", targetTime);
|
|
logger.log(`Scheduled milestone snapshot for aid ${aid} in ${(delay / MINUTE).toFixed(2)} mins.`, "mq");
|
|
}
|
|
} catch (e) {
|
|
logger.error(e as Error, "mq", "fn:collectMilestoneSnapshotsWorker");
|
|
} finally {
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
const getRegularSnapshotInterval = async (client: Client, aid: number) => {
|
|
const now = Date.now();
|
|
const date = new Date(now - 24 * HOUR);
|
|
let oldSnapshot = await findSnapshotBefore(client, aid, date);
|
|
if (!oldSnapshot) oldSnapshot = await findClosestSnapshot(client, aid, date);
|
|
const latestSnapshot = await getLatestSnapshot(client, aid);
|
|
if (!oldSnapshot || !latestSnapshot) return 0;
|
|
if (oldSnapshot.created_at === latestSnapshot.created_at) return 0;
|
|
const hoursDiff = (latestSnapshot.created_at - oldSnapshot.created_at) / HOUR;
|
|
if (hoursDiff < 8) return 24;
|
|
const viewsDiff = latestSnapshot.views - oldSnapshot.views;
|
|
if (viewsDiff === 0) return 72;
|
|
const speedPerDay = viewsDiff / (hoursDiff + 0.001) * 24;
|
|
if (speedPerDay < 6) return 36;
|
|
if (speedPerDay < 120) return 24;
|
|
if (speedPerDay < 320) return 12;
|
|
return 6;
|
|
};
|
|
|
|
export const archiveSnapshotsWorker = async (_job: Job) => {
|
|
const client = await db.connect();
|
|
const startedAt = Date.now();
|
|
if (await lockManager.isLocked("dispatchArchiveSnapshots")) {
|
|
logger.log("dispatchArchiveSnapshots is already running", "mq");
|
|
client.release();
|
|
return;
|
|
}
|
|
await lockManager.acquireLock("dispatchArchiveSnapshots", 30 * 60);
|
|
try {
|
|
const aids = await getAllVideosWithoutActiveSnapshotSchedule(client);
|
|
for (const rawAid of aids) {
|
|
const aid = Number(rawAid);
|
|
const latestSnapshot = await getLatestVideoSnapshot(client, aid);
|
|
const now = Date.now();
|
|
const lastSnapshotedAt = latestSnapshot?.time ?? now;
|
|
const interval = 168;
|
|
logger.log(`Scheduled regular snapshot for aid ${aid} in ${interval} hours.`, "mq");
|
|
const targetTime = lastSnapshotedAt + interval * HOUR;
|
|
await scheduleSnapshot(client, aid, "archive", targetTime);
|
|
if (now - startedAt > 250 * MINUTE) {
|
|
return;
|
|
}
|
|
}
|
|
} catch (e) {
|
|
logger.error(e as Error, "mq", "fn:archiveSnapshotsWorker");
|
|
} finally {
|
|
await lockManager.releaseLock("dispatchArchiveSnapshots");
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
export const regularSnapshotsWorker = async (_job: Job) => {
|
|
const client = await db.connect();
|
|
const startedAt = Date.now();
|
|
if (await lockManager.isLocked("dispatchRegularSnapshots")) {
|
|
logger.log("dispatchRegularSnapshots is already running", "mq");
|
|
client.release();
|
|
return;
|
|
}
|
|
await lockManager.acquireLock("dispatchRegularSnapshots", 30 * 60);
|
|
try {
|
|
const aids = await getVideosWithoutActiveSnapshotSchedule(client);
|
|
for (const rawAid of aids) {
|
|
const aid = Number(rawAid);
|
|
const latestSnapshot = await getLatestVideoSnapshot(client, aid);
|
|
const now = Date.now();
|
|
const lastSnapshotedAt = latestSnapshot?.time ?? now;
|
|
const interval = await getRegularSnapshotInterval(client, aid);
|
|
logger.log(`Scheduled regular snapshot for aid ${aid} in ${interval} hours.`, "mq");
|
|
const targetTime = truncate(lastSnapshotedAt + interval * HOUR, now + 1, now + 100000 * WEEK);
|
|
await scheduleSnapshot(client, aid, "normal", targetTime);
|
|
if (now - startedAt > 25 * MINUTE) {
|
|
return;
|
|
}
|
|
}
|
|
} catch (e) {
|
|
logger.error(e as Error, "mq", "fn:regularSnapshotsWorker");
|
|
} finally {
|
|
await lockManager.releaseLock("dispatchRegularSnapshots");
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
export const takeBulkSnapshotForVideosWorker = async (job: Job) => {
|
|
const schedules: SnapshotScheduleType[] = job.data.schedules;
|
|
const ids = schedules.map((schedule) => Number(schedule.id));
|
|
const aidsToFetch: number[] = [];
|
|
const client = await db.connect();
|
|
try {
|
|
for (const schedule of schedules) {
|
|
const aid = Number(schedule.aid);
|
|
const id = Number(schedule.id);
|
|
const exists = await snapshotScheduleExists(client, id);
|
|
if (!exists) {
|
|
continue;
|
|
}
|
|
aidsToFetch.push(aid);
|
|
}
|
|
const data = await bulkGetVideoStats(aidsToFetch);
|
|
if (typeof data === "number") {
|
|
await bulkSetSnapshotStatus(client, ids, "failed");
|
|
await bulkScheduleSnapshot(client, aidsToFetch, "normal", Date.now() + 15 * SECOND);
|
|
return `GET_BILI_STATUS_${data}`;
|
|
}
|
|
for (const video of data) {
|
|
const aid = video.id;
|
|
const stat = video.cnt_info;
|
|
const views = stat.play;
|
|
const danmakus = stat.danmaku;
|
|
const replies = stat.reply;
|
|
const likes = stat.thumb_up;
|
|
const coins = stat.coin;
|
|
const shares = stat.share;
|
|
const favorites = stat.collect;
|
|
const query: string = `
|
|
INSERT INTO video_snapshot (aid, views, danmakus, replies, likes, coins, shares, favorites)
|
|
VALUES ($1, $2, $3, $4, $5, $6, $7, $8)
|
|
`;
|
|
await client.queryObject(
|
|
query,
|
|
[aid, views, danmakus, replies, likes, coins, shares, favorites],
|
|
);
|
|
|
|
logger.log(`Taken snapshot for video ${aid} in bulk.`, "net", "fn:takeBulkSnapshotForVideosWorker");
|
|
}
|
|
await bulkSetSnapshotStatus(client, ids, "completed");
|
|
|
|
for (const schedule of schedules) {
|
|
const aid = Number(schedule.aid);
|
|
const type = schedule.type;
|
|
if (type == 'archive') continue;
|
|
const interval = await getRegularSnapshotInterval(client, aid);
|
|
logger.log(`Scheduled regular snapshot for aid ${aid} in ${interval} hours.`, "mq");
|
|
await scheduleSnapshot(client, aid, "normal", Date.now() + interval * HOUR);
|
|
}
|
|
return `DONE`;
|
|
} catch (e) {
|
|
if (e instanceof NetSchedulerError && e.code === "NO_PROXY_AVAILABLE") {
|
|
logger.warn(
|
|
`No available proxy for bulk request now.`,
|
|
"mq",
|
|
"fn:takeBulkSnapshotForVideosWorker",
|
|
);
|
|
await bulkSetSnapshotStatus(client, ids, "completed");
|
|
await bulkScheduleSnapshot(client, aidsToFetch, "normal", Date.now() + 20 * MINUTE * Math.random());
|
|
return;
|
|
}
|
|
logger.error(e as Error, "mq", "fn:takeBulkSnapshotForVideosWorker");
|
|
await bulkSetSnapshotStatus(client, ids, "failed");
|
|
} finally {
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
export const takeSnapshotForVideoWorker = async (job: Job) => {
|
|
const id = job.data.id;
|
|
const aid = Number(job.data.aid);
|
|
const type = job.data.type;
|
|
const task = snapshotTypeToTaskMap[type] ?? "snapshotVideo";
|
|
const client = await db.connect();
|
|
const retryInterval = type === "milestone" ? 5 * SECOND : 2 * MINUTE;
|
|
const exists = await snapshotScheduleExists(client, id);
|
|
if (!exists) {
|
|
client.release();
|
|
return;
|
|
}
|
|
const status = await getBiliVideoStatus(client, aid);
|
|
if (status !== 0) {
|
|
client.release();
|
|
return `REFUSE_WORKING_BILI_STATUS_${status}`;
|
|
}
|
|
try {
|
|
await setSnapshotStatus(client, id, "processing");
|
|
const stat = await insertVideoSnapshot(client, aid, task);
|
|
if (typeof stat === "number") {
|
|
await setBiliVideoStatus(client, aid, stat);
|
|
await setSnapshotStatus(client, id, "completed");
|
|
return `GET_BILI_STATUS_${stat}`;
|
|
}
|
|
await setSnapshotStatus(client, id, "completed");
|
|
if (type === "normal") {
|
|
const interval = await getRegularSnapshotInterval(client, aid);
|
|
logger.log(`Scheduled regular snapshot for aid ${aid} in ${interval} hours.`, "mq");
|
|
await scheduleSnapshot(client, aid, type, Date.now() + interval * HOUR);
|
|
return `DONE`;
|
|
} else if (type === "new") {
|
|
const publihsedAt = await getSongsPublihsedAt(client, aid);
|
|
const timeSincePublished = stat.time - publihsedAt!;
|
|
const viewsPerHour = stat.views / timeSincePublished * HOUR;
|
|
if (timeSincePublished > 48 * HOUR) {
|
|
return `DONE`;
|
|
}
|
|
if (timeSincePublished > 2 * HOUR && viewsPerHour < 10) {
|
|
return `DONE`;
|
|
}
|
|
let intervalMins = 240;
|
|
if (viewsPerHour > 50) {
|
|
intervalMins = 120;
|
|
}
|
|
if (viewsPerHour > 100) {
|
|
intervalMins = 60;
|
|
}
|
|
if (viewsPerHour > 1000) {
|
|
intervalMins = 15;
|
|
}
|
|
await scheduleSnapshot(client, aid, type, Date.now() + intervalMins * MINUTE, true);
|
|
}
|
|
if (type !== "milestone") return `DONE`;
|
|
const eta = await getAdjustedShortTermETA(client, aid);
|
|
if (eta > 144) return "ETA_TOO_LONG";
|
|
const now = Date.now();
|
|
const targetTime = now + eta * HOUR;
|
|
await scheduleSnapshot(client, aid, type, targetTime);
|
|
await setSnapshotStatus(client, id, "completed");
|
|
return `DONE`;
|
|
} catch (e) {
|
|
if (e instanceof NetSchedulerError && e.code === "NO_PROXY_AVAILABLE") {
|
|
logger.warn(
|
|
`No available proxy for aid ${job.data.aid}.`,
|
|
"mq",
|
|
"fn:takeSnapshotForVideoWorker",
|
|
);
|
|
await setSnapshotStatus(client, id, "completed");
|
|
await scheduleSnapshot(client, aid, type, Date.now() + retryInterval);
|
|
return;
|
|
}
|
|
logger.error(e as Error, "mq", "fn:takeSnapshotForVideoWorker");
|
|
await setSnapshotStatus(client, id, "failed");
|
|
} finally {
|
|
client.release();
|
|
}
|
|
};
|
|
|
|
export const scheduleCleanupWorker = async (_job: Job) => {
|
|
const client = await db.connect();
|
|
try {
|
|
const query = `
|
|
SELECT id, aid, type
|
|
FROM snapshot_schedule
|
|
WHERE status IN ('pending', 'processing')
|
|
AND started_at < NOW() - INTERVAL '30 minutes'
|
|
`;
|
|
const { rows } = await client.queryObject<{ id: bigint; aid: bigint; type: string }>(query);
|
|
if (rows.length === 0) return;
|
|
for (const row of rows) {
|
|
const id = Number(row.id);
|
|
const aid = Number(row.aid);
|
|
const type = row.type;
|
|
await setSnapshotStatus(client, id, "timeout");
|
|
await scheduleSnapshot(client, aid, type, Date.now() + 10 * SECOND);
|
|
logger.log(
|
|
`Schedule ${id} has no response received for 5 minutes, rescheduled.`,
|
|
"mq",
|
|
"fn:scheduleCleanupWorker",
|
|
);
|
|
}
|
|
} catch (e) {
|
|
logger.error(e as Error, "mq", "fn:scheduleCleanupWorker");
|
|
} finally {
|
|
client.release();
|
|
}
|
|
};
|