mirror of
https://git.mirrors.martin98.com/https://github.com/mendableai/firecrawl
synced 2025-08-16 04:15:57 +08:00
Fix Supabase client configuration errors when USE_DB_AUTHENTICATION is false (#1534)
* Fix Supabase client configuration errors when USE_DB_AUTHENTICATION is false Co-Authored-By: hello@sideguide.dev <hello+firecrawl@sideguide.dev> * Add USE_DB_AUTHENTICATION checks to map and search controllers Add test for USE_DB_AUTHENTICATION=false Add USE_DB_AUTHENTICATION checks to billing services Add USE_DB_AUTHENTICATION checks to batch_billing.ts Add USE_DB_AUTHENTICATION checks to cached-docs.ts Add USE_DB_AUTHENTICATION checks to supabase-jobs.ts Add USE_DB_AUTHENTICATION checks to team-id-sync.ts Add USE_DB_AUTHENTICATION checks to test-suite log.ts Add USE_DB_AUTHENTICATION checks to idempotency services Co-Authored-By: hello@sideguide.dev <hello+firecrawl@sideguide.dev> * Revert "Add USE_DB_AUTHENTICATION checks to map and search controllers" This reverts commit 834a5d51a68c74ada67800fa3a0aa45bde22d745. --------- Co-authored-by: Devin AI <158243242+devin-ai-integration[bot]@users.noreply.github.com> Co-authored-by: hello@sideguide.dev <hello+firecrawl@sideguide.dev> Co-authored-by: Nicolas <nicolascamara29@gmail.com> Co-authored-by: Gergő Móricz <mo.geryy@gmail.com>
This commit is contained in:
parent
526165e1b9
commit
ab30c8e4ac
@ -147,19 +147,23 @@ async function finishCrawlIfNeeded(job: Job & { id: string }, sc: StoredCrawl) {
|
|||||||
"crawl:" + job.data.crawl_id + ":visited_unique",
|
"crawl:" + job.data.crawl_id + ":visited_unique",
|
||||||
),
|
),
|
||||||
);
|
);
|
||||||
|
|
||||||
logger.info("Visited URLs", {
|
logger.info("Visited URLs", {
|
||||||
visitedUrls: visitedUrls.size,
|
visitedUrls: visitedUrls.size,
|
||||||
});
|
});
|
||||||
|
|
||||||
const lastUrls: string[] = (
|
let lastUrls: string[] = [];
|
||||||
(
|
const useDbAuthentication = process.env.USE_DB_AUTHENTICATION === "true";
|
||||||
await supabase_service.rpc("diff_get_last_crawl_urls", {
|
if (useDbAuthentication) {
|
||||||
i_team_id: job.data.team_id,
|
lastUrls = (
|
||||||
i_url: sc.originUrl!,
|
(
|
||||||
})
|
await supabase_service.rpc("diff_get_last_crawl_urls", {
|
||||||
).data ?? []
|
i_team_id: job.data.team_id,
|
||||||
).map((x) => x.url);
|
i_url: sc.originUrl!,
|
||||||
|
})
|
||||||
|
).data ?? []
|
||||||
|
).map((x) => x.url);
|
||||||
|
}
|
||||||
|
|
||||||
const lastUrlsSet = new Set(lastUrls);
|
const lastUrlsSet = new Set(lastUrls);
|
||||||
|
|
||||||
@ -257,7 +261,8 @@ async function finishCrawlIfNeeded(job: Job & { id: string }, sc: StoredCrawl) {
|
|||||||
if (
|
if (
|
||||||
visitedUrls.length > 0 &&
|
visitedUrls.length > 0 &&
|
||||||
job.data.crawlerOptions !== null &&
|
job.data.crawlerOptions !== null &&
|
||||||
originUrl
|
originUrl &&
|
||||||
|
process.env.USE_DB_AUTHENTICATION === "true"
|
||||||
) {
|
) {
|
||||||
// Queue the indexing job instead of doing it directly
|
// Queue the indexing job instead of doing it directly
|
||||||
await getIndexQueue().add(
|
await getIndexQueue().add(
|
||||||
|
Loading…
x
Reference in New Issue
Block a user