mirror of
https://git.mirrors.martin98.com/https://github.com/mendableai/firecrawl
synced 2025-08-12 08:28:58 +08:00
feat(queue-worker/afterJobDone): improved ccq insert logic
This commit is contained in:
parent
f41af8241e
commit
97c635676d
@ -54,6 +54,7 @@ import { scrapeOptions } from "../controllers/v1/types";
|
|||||||
import {
|
import {
|
||||||
cleanOldConcurrencyLimitEntries,
|
cleanOldConcurrencyLimitEntries,
|
||||||
cleanOldCrawlConcurrencyLimitEntries,
|
cleanOldCrawlConcurrencyLimitEntries,
|
||||||
|
getConcurrencyLimitActiveJobs,
|
||||||
pushConcurrencyLimitActiveJob,
|
pushConcurrencyLimitActiveJob,
|
||||||
pushCrawlConcurrencyLimitActiveJob,
|
pushCrawlConcurrencyLimitActiveJob,
|
||||||
removeConcurrencyLimitActiveJob,
|
removeConcurrencyLimitActiveJob,
|
||||||
@ -86,6 +87,7 @@ import http from "http";
|
|||||||
import https from "https";
|
import https from "https";
|
||||||
import { cacheableLookup } from "../scraper/scrapeURL/lib/cacheableLookup";
|
import { cacheableLookup } from "../scraper/scrapeURL/lib/cacheableLookup";
|
||||||
import { robustFetch } from "../scraper/scrapeURL/lib/fetch";
|
import { robustFetch } from "../scraper/scrapeURL/lib/fetch";
|
||||||
|
import { RateLimiterMode } from "../types";
|
||||||
|
|
||||||
configDotenv();
|
configDotenv();
|
||||||
|
|
||||||
@ -805,11 +807,16 @@ const workerFun = async (
|
|||||||
}
|
}
|
||||||
|
|
||||||
if (job.id && job.data && job.data.team_id) {
|
if (job.id && job.data && job.data.team_id) {
|
||||||
await removeConcurrencyLimitActiveJob(job.data.team_id, job.id);
|
const maxConcurrency = (await getACUCTeam(job.data.team_id, false, true, job.data.is_extract ? RateLimiterMode.Extract : RateLimiterMode.Crawl))?.concurrency ?? 2;
|
||||||
cleanOldConcurrencyLimitEntries(job.data.team_id);
|
|
||||||
|
|
||||||
// No need to check if we're under the limit here -- if the current job is finished,
|
await removeConcurrencyLimitActiveJob(job.data.team_id, job.id);
|
||||||
// we are 1 under the limit, assuming the job insertion logic never over-inserts. - MG
|
await cleanOldConcurrencyLimitEntries(job.data.team_id);
|
||||||
|
|
||||||
|
// Check if we're under the concurrency limit before adding a new job
|
||||||
|
const currentActiveConcurrency = (await getConcurrencyLimitActiveJobs(job.data.team_id)).length;
|
||||||
|
const concurrencyLimited = currentActiveConcurrency >= maxConcurrency;
|
||||||
|
|
||||||
|
if (!concurrencyLimited) {
|
||||||
const nextJob = await takeConcurrencyLimitedJob(job.data.team_id);
|
const nextJob = await takeConcurrencyLimitedJob(job.data.team_id);
|
||||||
if (nextJob !== null) {
|
if (nextJob !== null) {
|
||||||
await pushConcurrencyLimitActiveJob(
|
await pushConcurrencyLimitActiveJob(
|
||||||
@ -833,6 +840,7 @@ const workerFun = async (
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if (job.data && job.data.sentry && Sentry.isInitialized()) {
|
if (job.data && job.data.sentry && Sentry.isInitialized()) {
|
||||||
Sentry.continueTrace(
|
Sentry.continueTrace(
|
||||||
|
Loading…
x
Reference in New Issue
Block a user