mirror of
https://git.mirrors.martin98.com/https://github.com/mendableai/firecrawl
synced 2025-08-12 16:49:00 +08:00
Nick:
This commit is contained in:
parent
a3f3168212
commit
9ba1ae9ae1
@ -278,14 +278,14 @@ v1Router.get(
|
|||||||
|
|
||||||
v1Router.post(
|
v1Router.post(
|
||||||
"/deep-research",
|
"/deep-research",
|
||||||
authMiddleware(RateLimiterMode.Extract),
|
authMiddleware(RateLimiterMode.Crawl),
|
||||||
checkCreditsMiddleware(1),
|
checkCreditsMiddleware(1),
|
||||||
wrap(deepResearchController),
|
wrap(deepResearchController),
|
||||||
);
|
);
|
||||||
|
|
||||||
v1Router.get(
|
v1Router.get(
|
||||||
"/deep-research/:jobId",
|
"/deep-research/:jobId",
|
||||||
authMiddleware(RateLimiterMode.ExtractStatus),
|
authMiddleware(RateLimiterMode.CrawlStatus),
|
||||||
wrap(deepResearchStatusController),
|
wrap(deepResearchStatusController),
|
||||||
);
|
);
|
||||||
|
|
||||||
|
@ -1843,7 +1843,12 @@ class FirecrawlApp:
|
|||||||
__experimental_stream=experimental_stream
|
__experimental_stream=experimental_stream
|
||||||
)
|
)
|
||||||
|
|
||||||
response = self.async_generate_llms_text(url, params)
|
response = self.async_generate_llms_text(
|
||||||
|
url,
|
||||||
|
max_urls=max_urls,
|
||||||
|
show_full_text=show_full_text,
|
||||||
|
experimental_stream=experimental_stream
|
||||||
|
)
|
||||||
if not response.get('success') or 'id' not in response:
|
if not response.get('success') or 'id' not in response:
|
||||||
return response
|
return response
|
||||||
|
|
||||||
@ -2219,7 +2224,14 @@ class FirecrawlApp:
|
|||||||
research_params['__experimental_streamSteps'] = __experimental_stream_steps
|
research_params['__experimental_streamSteps'] = __experimental_stream_steps
|
||||||
research_params = DeepResearchParams(**research_params)
|
research_params = DeepResearchParams(**research_params)
|
||||||
|
|
||||||
response = self.async_deep_research(query, research_params)
|
response = self.async_deep_research(
|
||||||
|
query,
|
||||||
|
max_depth=max_depth,
|
||||||
|
time_limit=time_limit,
|
||||||
|
max_urls=max_urls,
|
||||||
|
analysis_prompt=analysis_prompt,
|
||||||
|
system_prompt=system_prompt
|
||||||
|
)
|
||||||
if not response.get('success') or 'id' not in response:
|
if not response.get('success') or 'id' not in response:
|
||||||
return response
|
return response
|
||||||
|
|
||||||
@ -3529,7 +3541,12 @@ class AsyncFirecrawlApp(FirecrawlApp):
|
|||||||
if experimental_stream is not None:
|
if experimental_stream is not None:
|
||||||
params['__experimental_stream'] = experimental_stream
|
params['__experimental_stream'] = experimental_stream
|
||||||
|
|
||||||
response = await self.async_generate_llms_text(url, params)
|
response = await self.async_generate_llms_text(
|
||||||
|
url,
|
||||||
|
max_urls=max_urls,
|
||||||
|
show_full_text=show_full_text,
|
||||||
|
experimental_stream=experimental_stream
|
||||||
|
)
|
||||||
if not response.get('success') or 'id' not in response:
|
if not response.get('success') or 'id' not in response:
|
||||||
return response
|
return response
|
||||||
|
|
||||||
@ -3678,7 +3695,14 @@ class AsyncFirecrawlApp(FirecrawlApp):
|
|||||||
research_params['__experimental_streamSteps'] = __experimental_stream_steps
|
research_params['__experimental_streamSteps'] = __experimental_stream_steps
|
||||||
research_params = DeepResearchParams(**research_params)
|
research_params = DeepResearchParams(**research_params)
|
||||||
|
|
||||||
response = await self.async_deep_research(query, research_params)
|
response = await self.async_deep_research(
|
||||||
|
query,
|
||||||
|
max_depth=max_depth,
|
||||||
|
time_limit=time_limit,
|
||||||
|
max_urls=max_urls,
|
||||||
|
analysis_prompt=analysis_prompt,
|
||||||
|
system_prompt=system_prompt
|
||||||
|
)
|
||||||
if not response.get('success') or 'id' not in response:
|
if not response.get('success') or 'id' not in response:
|
||||||
return response
|
return response
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user