From 9e8c629ff4d9424adc0a0458b5266beffdc55f21 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?M=C3=B3ricz=20Gerg=C5=91?= Date: Thu, 9 Jan 2025 09:50:54 +0100 Subject: [PATCH] fix(log_job): don't redact with auth header --- apps/api/src/services/logging/log_job.ts | 27 ++++++++++++------------ 1 file changed, 14 insertions(+), 13 deletions(-) diff --git a/apps/api/src/services/logging/log_job.ts b/apps/api/src/services/logging/log_job.ts index ca0c6b00..b367e0b6 100644 --- a/apps/api/src/services/logging/log_job.ts +++ b/apps/api/src/services/logging/log_job.ts @@ -27,19 +27,20 @@ export async function logJob(job: FirecrawlJob, force: boolean = false) { } // Redact any pages that have an authorization header - if ( - job.scrapeOptions && - job.scrapeOptions.headers && - job.scrapeOptions.headers["Authorization"] - ) { - job.scrapeOptions.headers["Authorization"] = "REDACTED"; - job.docs = [ - { - content: "REDACTED DUE TO AUTHORIZATION HEADER", - html: "REDACTED DUE TO AUTHORIZATION HEADER", - }, - ]; - } + // actually, Don't. we use the db to retrieve results now. this breaks authed crawls - mogery + // if ( + // job.scrapeOptions && + // job.scrapeOptions.headers && + // job.scrapeOptions.headers["Authorization"] + // ) { + // job.scrapeOptions.headers["Authorization"] = "REDACTED"; + // job.docs = [ + // { + // content: "REDACTED DUE TO AUTHORIZATION HEADER", + // html: "REDACTED DUE TO AUTHORIZATION HEADER", + // }, + // ]; + // } const jobColumn = { job_id: job.job_id ? job.job_id : null, success: job.success,