feat(rust): update rust sdk to support new features (#1446)

* chore(rust-sdk): cargo fmt

* feat(rust-sdk): implement search api + example + test

* feat(rust-sdk): implement crawl cancel api + example + test

* feat(rust-sdk): implement crawl check errors api + example + test

* feat(rust-sdk): implement batch crawl + test + example

+ Fix MapOptions

* feat(rust-sdk): implement extract api + test + example

* feat(rust-sdk): implement llmtxt api + test + example

* chore(rust-sdk): correct mock tests

* chore(rust-sdk): prep for cargo distribution
This commit is contained in:
kkharji 2025-04-18 07:59:59 +03:00 committed by GitHub
parent 33aece8e96
commit f2c01340d1
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
20 changed files with 4350 additions and 125 deletions

1377
apps/rust-sdk/Cargo.lock generated

File diff suppressed because it is too large Load Diff

View File

@ -1,13 +1,13 @@
[package]
name = "firecrawl"
author= "Mendable.ai"
version = "1.0.0"
version = "1.1.0"
edition = "2021"
license = "MIT"
homepage = "https://www.firecrawl.dev/"
repository ="https://github.com/mendableai/firecrawl"
description = "Rust SDK for Firecrawl API."
authors = ["Gergő Móricz <mogery@firecrawl.dev>", "sanix-darker <sanixdk@gmail.com>"]
authors = ["Gergő Móricz <mogery@firecrawl.dev>", "sanix-darker <sanixdk@gmail.com>", "kkharji <kkharji@protonmail.com>"]
[lib]
path = "src/lib.rs"
@ -23,12 +23,18 @@ log = "^0.4"
thiserror = "^1.0"
uuid = { version = "^1.10", features = ["v4"] }
tokio = { version = "^1", features = ["full"] }
futures = "0.3.31"
schemars = "0.8.22"
[dev-dependencies]
clippy = "^0.0.302"
assert_matches = "^1.5"
dotenvy = "^0.15"
tokio = { version = "1", features = ["full"] }
mockito = "1.7.0"
clap = { version ="4.5.35", features = ["derive"] }
axum = { version = "0.8.3", features = ["tokio", "macros"] }
bat = "0.25.0"
[build-dependencies]
tokio = { version = "1", features = ["full"] }

View File

@ -0,0 +1,175 @@
use clap::{Parser, Subcommand};
use firecrawl::{
batch_scrape::{BatchScrapeParams, WebhookOptions},
map::MapOptions,
scrape::{ScrapeFormats, ScrapeOptions},
FirecrawlApp,
};
use serde_json::Value;
use std::error::Error;
use std::net::SocketAddr;
use std::sync::Arc;
use tokio::sync::Mutex;
// Store webhook responses
struct WebhookState {
responses: Vec<Value>,
}
#[derive(Parser)]
#[command(version, about, long_about = None)]
struct Cli {
#[command(subcommand)]
command: Commands,
}
#[derive(Subcommand)]
enum Commands {
/// Multiple URL scraping with webhook monitoring
Basic,
}
async fn create_firecrawl_app() -> Result<FirecrawlApp, Box<dyn Error>> {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
FirecrawlApp::new_selfhosted(api_url, None::<&str>).map_err(|e| e.into())
}
// Start webhook server and return its address
async fn start_webhook_server(
port: u16,
state: Arc<Mutex<WebhookState>>,
) -> Result<String, Box<dyn Error>> {
let state = state.clone();
use axum::routing::post;
use axum::Json;
let app = axum::Router::new().route(
"/",
post(move |body: Json<Value>| {
let state = state.clone();
async move {
state.lock().await.responses.push(body.0.clone());
match serde_json::to_string_pretty(&body.0) {
Ok(data) => println!(
"Received webhook: {}",
serde_json::to_string_pretty(&data).unwrap()
),
Err(_) => println!("Received webhook: {}", body.0),
}
"OK"
}
}),
);
let addr = SocketAddr::from(([0, 0, 0, 0], port));
let webhook_url = format!("http://host.docker.internal:{}", port);
tokio::spawn(async move {
let listener = tokio::net::TcpListener::bind(addr)
.await
.inspect_err(|err| println!("{err:?}"))
.unwrap();
if let Err(e) = axum::serve(listener, app).await {
eprintln!("Webhook server error: {}", e);
}
});
println!("Webhook server running at {}", webhook_url);
Ok(webhook_url)
}
#[tokio::main]
async fn main() -> Result<(), Box<dyn Error>> {
let cli = Cli::parse();
let firecrawl = create_firecrawl_app().await?;
let state = Arc::new(Mutex::new(WebhookState { responses: vec![] }));
let webhook_url = start_webhook_server(39120, state.clone()).await?;
match cli.command {
Commands::Basic => {
let mut urls = Vec::new();
let url_one = "https://invalid-url.url/";
println!("Mapping: {}", url_one);
match firecrawl.map_url(url_one, None).await {
Ok(mapped_urls) => urls.extend(mapped_urls),
Err(e) => println!("Error mapping {}: {}", url_one, e),
}
let url_two = "https://www.devjobsscanner.com";
println!("Mapping: {}", url_two);
match firecrawl
.map_url(
url_two,
Some(MapOptions {
search: Some("rust".into()),
limit: Some(20),
..Default::default()
}),
)
.await
{
Ok(mapped_urls) => urls.extend(mapped_urls),
Err(e) => println!("Error mapping {}: {}", url_two, e),
}
test_multiple_urls(&firecrawl, urls, &webhook_url).await?;
// Give time for webhooks to arrive
tokio::time::sleep(tokio::time::Duration::from_secs(5)).await;
println!(
"Received {} webhook responses",
state.lock().await.responses.len()
);
}
}
Ok(())
}
async fn test_multiple_urls(
app: &FirecrawlApp,
urls: Vec<String>,
webhook_url: &str,
) -> Result<(), Box<dyn Error>> {
println!("Testing batch scraping of {} URLs", urls.len());
let webhook = WebhookOptions {
url: webhook_url.to_string(),
headers: None,
auth_token: None,
};
let params = BatchScrapeParams {
urls,
webhook: Some(webhook),
ignore_invalid_urls: true,
options: Some(ScrapeOptions {
formats: Some(vec![ScrapeFormats::Markdown, ScrapeFormats::Links]),
..Default::default()
}),
..Default::default()
};
let batch = app.async_batch_scrape_urls(params).await?;
println!("Batch job started: {}", batch.id);
// Poll status periodically
loop {
let status = app.check_batch_scrape_status(&batch.id).await?;
println!("Progress: {}/{} pages", status.completed, status.total);
if status.completed >= status.total {
println!("Batch job completed!");
break;
}
tokio::time::sleep(tokio::time::Duration::from_secs(2)).await;
}
Ok(())
}

View File

@ -0,0 +1,33 @@
use firecrawl::FirecrawlApp;
use std::error::Error;
use std::time::Duration;
#[tokio::main]
async fn main() -> Result<(), Box<dyn Error>> {
// Get API URL from environment
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
// Create the FirecrawlApp instance
let firecrawl = FirecrawlApp::new_selfhosted(api_url, None::<&str>)?;
// Start a crawl job
println!("Starting a crawl job...");
let crawl_response = firecrawl
.crawl_url_async("https://example.com", None)
.await?;
println!("Crawl job started with ID: {}", crawl_response.id);
// Wait for a moment to let the crawl job start
println!("Waiting for a moment...");
tokio::time::sleep(Duration::from_secs(2)).await;
// Cancel the crawl job
println!("Cancelling the crawl job...");
let cancel_response = firecrawl.cancel_crawl(&crawl_response.id).await?;
println!("Cancellation result:");
println!(" Status: {:?}", cancel_response.status);
Ok(())
}

View File

@ -0,0 +1,59 @@
use firecrawl::FirecrawlApp;
use std::error::Error;
#[tokio::main]
async fn main() -> Result<(), Box<dyn Error>> {
// Get API URL from environment
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
// Create the FirecrawlApp instance
let firecrawl = FirecrawlApp::new_selfhosted(api_url, None::<&str>)?;
// Start a crawl job that will likely have some errors (invalid URL format)
println!("Starting a crawl job...");
let crawl_response = firecrawl
.crawl_url_async("https://no-wer-agg.invalid", None)
.await?;
println!("Crawl job started with ID: {}", crawl_response.id);
println!("Let it do it's thing...");
tokio::time::sleep(tokio::time::Duration::from_secs(3)).await;
// Check the crawl errors
println!("Checking for crawl errors...");
match firecrawl.check_crawl_errors(&crawl_response.id).await {
Ok(error_response) => {
println!("Crawl errors response:");
println!(" Number of errors: {}", error_response.errors.len());
if !error_response.errors.is_empty() {
println!("\nDetailed errors:");
for (i, error) in error_response.errors.iter().enumerate() {
println!("Error #{}", i + 1);
println!(" ID: {}", error.id);
if let Some(timestamp) = &error.timestamp {
println!(" Timestamp: {}", timestamp);
}
println!(" URL: {}", error.url);
println!(" Error: {}", error.error);
}
}
println!(
"\nRobots.txt blocked URLs: {}",
error_response.robots_blocked.len()
);
for (i, url) in error_response.robots_blocked.iter().enumerate() {
println!(" {}. {}", i + 1, url);
}
}
Err(e) => {
println!("Failed to check crawl errors: {}", e);
}
}
let cancel = firecrawl.cancel_crawl(&crawl_response.id).await?;
println!("Cancel: {}", cancel.status);
Ok(())
}

View File

@ -1,4 +1,8 @@
use firecrawl::{crawl::CrawlOptions, scrape::{ExtractOptions, ScrapeFormats, ScrapeOptions}, FirecrawlApp};
use firecrawl::{
crawl::CrawlOptions,
scrape::{ExtractOptions, ScrapeFormats, ScrapeOptions},
FirecrawlApp,
};
use serde_json::json;
#[tokio::main]
@ -23,12 +27,13 @@ async fn main() {
..Default::default()
};
let crawl_result = app
.crawl_url("https://mendable.ai", crawl_options)
.await;
let crawl_result = app.crawl_url("https://mendable.ai", crawl_options).await;
match crawl_result {
Ok(data) => println!("Crawl Result (used {} credits):\n{:#?}", data.credits_used, data.data),
Ok(data) => println!(
"Crawl Result (used {} credits):\n{:#?}",
data.credits_used, data.data
),
Err(e) => eprintln!("Crawl failed: {}", e),
}
@ -61,7 +66,8 @@ async fn main() {
extract: ExtractOptions {
schema: json_schema.into(),
..Default::default()
}.into(),
}
.into(),
..Default::default()
};
@ -75,9 +81,7 @@ async fn main() {
}
// Map a website (Alpha)
let map_result = app
.map_url("https://firecrawl.dev", None)
.await;
let map_result = app.map_url("https://firecrawl.dev", None).await;
match map_result {
Ok(data) => println!("Mapped URLs: {:#?}", data),

View File

@ -0,0 +1,237 @@
use firecrawl::{extract::ExtractParams, FirecrawlApp};
use serde_json::json;
use std::error::Error;
use clap::{Parser, ValueEnum};
#[derive(Parser)]
#[command(author, version, about, long_about = None)]
struct Args {
#[arg(value_enum)]
command: Examples,
}
#[derive(Copy, Clone, PartialEq, Eq, ValueEnum)]
enum Examples {
Basic,
Schema,
JsonSchema,
}
#[tokio::main]
async fn main() -> Result<(), Box<dyn Error>> {
let args = Args::parse();
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let firecrawl = FirecrawlApp::new_selfhosted(api_url, None::<&str>)?;
let urls = vec![
"https://www.firecrawl.dev/".to_string(),
"https://betteruptime.com".to_string(),
];
match args.command {
Examples::Basic => {
println!("Example 1: Extracting with URLs and prompt");
let extract_params = ExtractParams {
prompt: Some(
"Extract Product promise, consice descirption and category".to_string(),
),
url_trace: Some(true),
..Default::default()
};
println!("Starting asynchronous extraction job...");
let response = firecrawl
.async_extract(ExtractParams {
urls: Some(urls.iter().map(|u| u.to_string()).collect()),
prompt: extract_params.prompt.clone(),
url_trace: extract_params.url_trace,
..Default::default()
})
.await?;
println!("Extract job initiated:");
println!(" Job ID: {}", response.id);
println!("\nChecking extract status...");
for _ in 0..5 {
let response = firecrawl.get_extract_status(&response.id).await?;
println!("Extract status: {}", response.status);
if let Some(url_trace) = &response.url_trace {
println!("URL traces:");
for trace in url_trace {
println!(" URL: {}", trace.url);
println!(" Status: {}", trace.status);
}
}
println!("Extract data: {:#?}", response.data);
if response.status == "completed" {
break;
}
tokio::time::sleep(tokio::time::Duration::from_secs(2)).await;
}
}
Examples::Schema => {
println!("Example 2: Extracting with schema");
let schema = json!({
"type": "object",
"properties": {
"category": { "type": "string" },
"promise": { "type": "string" },
"descirption": { "type": "string" }
},
"required": ["category", "promise", "description"]
});
println!("Starting synchronous extraction job...");
match firecrawl
.extract(ExtractParams {
urls: urls.into(),
schema: Some(schema),
..Default::default()
})
.await
{
Ok(result) => {
println!("Extraction completed successfully!");
println!("Status: {}", result.status);
if let Some(data) = result.data {
println!("\nExtracted data:");
println!(" Title: {}", data["title"]);
if let Some(desc) = data.get("description") {
println!(" Description: {}", desc);
}
println!(
" Content (preview): {:.100}...",
data["content"].as_str().unwrap_or("N/A")
);
}
if let Some(sources) = result.sources {
println!("\nSources:");
for (field, urls) in sources {
println!(" {}: {}", field, urls.join(", "));
}
}
}
Err(e) => {
println!("Extraction failed: {}", e);
}
}
}
Examples::JsonSchema => {
println!("Example 3: Using JsonSchema derive");
/// A comprehensive analysis of given product
#[derive(serde::Serialize, serde::Deserialize, schemars::JsonSchema)]
struct ProductAnalysis {
/// The full name of the product
product_name: String,
/// The company/brand behind the product
brand: String,
/// The general price range (e.g. "Premium", "$10-50", "Enterprise")
price_range: String,
/// The main customer segments this product targets
target_audience: Vec<String>,
/// Primary benefits and value propositions of the product
key_benefits: Vec<String>,
/// Distinctive features that set this product apart from competitors
unique_selling_points: Vec<String>,
/// Direct comparisons with competing products/services
competitor_comparison: Vec<String>,
/// Technologies, frameworks, or platforms used (if applicable)
tech_stack: Option<Vec<String>>,
/// Aggregated review data and sentiment analysis
reviews_summary: ReviewsSummary,
// /// Score from 0-10 indicating product-market fit based on analysis
// market_fit_score: f32, // NOTE: Breaks
/// Assessment of future growth prospects (e.g. "High", "Moderate", "Limited")
growth_potential: String,
/// Relevant compliance standards and certifications
regulatory_compliance: Option<Vec<String>>,
}
/// Aggregated analysis of product reviews from multiple sources
#[derive(serde::Serialize, serde::Deserialize, schemars::JsonSchema)]
struct ReviewsSummary {
/// Overall sentiment from review analysis (e.g. "Highly Positive", "Mixed", "Negative")
sentiment_analysis: String,
/// Most frequently mentioned positive aspects
common_praises: Vec<String>,
/// Most frequently mentioned criticisms or issues
common_complaints: Vec<String>,
/// Platforms or websites where reviews were sourced from
review_sources: Vec<String>,
}
println!("Starting extraction with derived schema...");
match firecrawl
.extract_with_schemars::<ProductAnalysis>(ExtractParams {
urls: urls.into(),
..Default::default()
})
.await
{
Ok(result) => {
println!("Extraction completed!");
println!("Status: {}", result.status);
if let Some(data) = result.data {
if let Ok(analysis) = serde_json::from_value::<ProductAnalysis>(data) {
println!("\nExtracted Product Analysis:");
println!(" Product: {}", analysis.product_name);
println!(" Brand: {}", analysis.brand);
println!(" Price Range: {}", analysis.price_range);
println!(" Target Audience:");
for audience in analysis.target_audience {
println!(" - {}", audience);
}
println!(" Key Benefits:");
for benefit in analysis.key_benefits {
println!(" - {}", benefit);
}
println!(" USPs:");
for usp in analysis.unique_selling_points {
println!(" - {}", usp);
}
println!("\n Reviews Summary:");
println!(
" Sentiment: {}",
analysis.reviews_summary.sentiment_analysis
);
println!(" Common Praises:");
for praise in analysis.reviews_summary.common_praises {
println!(" - {}", praise);
}
println!(" Common Complaints:");
for complaint in analysis.reviews_summary.common_complaints {
println!(" - {}", complaint);
}
} else {
println!("Failed to parse extracted data");
}
}
if let Some(sources) = result.sources {
println!("\nSources:");
for (field, urls) in sources {
println!(" {}: {}", field, urls.join(", "));
}
}
}
Err(e) => {
println!("Extraction failed: {}", e);
}
}
}
}
Ok(())
}

View File

@ -0,0 +1,173 @@
#![allow(clippy::option_map_unit_fn)]
use bat::{Input, PrettyPrinter};
use firecrawl::{llmstxt::GenerateLLMsTextParams, FirecrawlApp};
use std::error::Error;
use clap::{Parser, ValueEnum};
#[derive(Copy, Clone, PartialEq, Eq, ValueEnum)]
enum Mode {
Basic,
Pool,
Fulltext,
}
#[derive(Parser)]
#[command(author, version, about, long_about = None)]
struct Args {
/// URL for which to generate LLMs.txt
#[arg(default_value = "https://www.firecrawl.dev/")]
url: String,
#[arg(long, short = 'm', value_enum, default_value = "Mode::Basic")]
mode: Mode,
/// Maximum number of URLs to process
#[arg(long, short = 'd', default_value = "1")]
max_urls: u32,
/// Whether to show the full LLMs-full.txt in the response
#[arg(long, short = 'f', default_value = "false")]
full_text: bool,
/// Experimental streaming option
#[arg(long, short = 's', default_value = "false")]
stream: bool,
}
#[tokio::main]
async fn main() -> Result<(), Box<dyn Error>> {
let args = Args::parse();
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let firecrawl = FirecrawlApp::new_selfhosted(api_url, None::<&str>)?;
let params = GenerateLLMsTextParams {
url: args.url.clone(),
max_urls: args.max_urls,
show_full_text: args.full_text,
experimental_stream: args.stream,
};
match args.mode {
Mode::Basic => {
println!("Example 1: Basic LLMs.txt generation (synchronous)");
println!("Generating LLMs.txt for {}...", args.url);
firecrawl
.generate_llms_text(params)
.await
.inspect(|result| {
println!("Expires at: {}", result.expires_at);
let text = (if args.full_text {
result.data.full.as_ref()
} else {
result.data.compact.as_ref()
})
.expect("LLM Text");
pretty_print_content("Firecrawl Result", text).expect("Print");
})?;
}
Mode::Pool => {
println!("Example 2: Asynchronous LLMs.txt generation with manual polling");
println!("Starting asynchronous LLMs.txt generation job...");
let response = firecrawl.async_generate_llms_text(params).await?;
println!("LLMs.txt generation job initiated:");
println!(" Job ID: {}", response.id);
println!("\nManually polling for status...");
for _ in 0..10 {
let status = firecrawl
.check_generate_llms_text_status(&response.id)
.await?;
match status.status.as_str() {
"completed" => {
println!("LLMs.txt generation completed!");
let text = (if args.full_text {
status.data.full.as_ref()
} else {
status.data.compact.as_ref()
})
.expect("LLM Text");
pretty_print_content("Pool Result", text).expect("Print");
break;
}
"failed" => {
println!(
"LLMs.txt generation failed: {}",
status.error.unwrap_or_default()
);
break;
}
status => println!("Generation status: {}", status),
}
println!("Waiting 2 seconds before checking again...");
tokio::time::sleep(tokio::time::Duration::from_secs(2)).await;
}
}
Mode::Fulltext => {
println!("Example 3: LLMs.txt generation with full text");
println!("Generating LLMs.txt with full text...");
match firecrawl.generate_llms_text(params).await {
Ok(result) => {
println!("LLMs.txt generation completed successfully!");
let llmstxt = result.data.compact.expect("LLMs Text Expected");
let fulltxt = result.data.full.expect("Full LLMs Text Expected");
pretty_print_contents(&[
("LLMs.txt (compact)", llmstxt),
("LLMs.txt (full text)", fulltxt),
])
.expect("Print")
}
Err(e) => {
println!("LLMs.txt generation failed: {}", e);
}
}
}
}
Ok(())
}
/// Pretty prints the provided content with syntax highlighting
fn pretty_print_content(title: &str, content: &str) -> Result<(), Box<dyn Error>> {
PrettyPrinter::new()
.header(true)
.grid(true)
.input(
Input::from_bytes(content.as_bytes())
.title(title)
.name("file.md"),
)
.print()?;
Ok(())
}
/// Pretty prints multiple contents with syntax highlighting
fn pretty_print_contents(title_contents: &[(&'static str, String)]) -> Result<(), Box<dyn Error>> {
let mut inputs = Vec::new();
for (title, content) in title_contents {
inputs.push(
Input::from_bytes(content.as_bytes())
.title(*title)
.name("file.md"),
);
}
PrettyPrinter::new()
.header(true)
.grid(true)
.inputs(inputs)
.print()?;
Ok(())
}

View File

@ -0,0 +1,186 @@
use clap::{Parser, ValueEnum};
use firecrawl::{
search::{SearchParams, SearchResponse},
FirecrawlApp,
};
use std::error::Error;
#[derive(Debug, Parser)]
#[command(author, version, about, long_about = None)]
struct Args {
/// Which example to run
#[arg(value_enum, default_value_t = Examples::All)]
example: Examples,
}
#[derive(Debug, Clone, ValueEnum)]
enum Examples {
All,
Basic,
Advanced,
Geo,
Temporal,
Social,
News,
Academic,
Commercial,
}
#[tokio::main]
async fn main() -> Result<(), Box<dyn Error>> {
let args = Args::parse();
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let firecrawl = FirecrawlApp::new_selfhosted(api_url, None::<&str>)?;
match args.example {
Examples::All => {
run_basic_example(&firecrawl).await?;
run_advanced_example(&firecrawl).await?;
run_geographic_example(&firecrawl).await?;
run_temporal_example(&firecrawl).await?;
run_social_example(&firecrawl).await?;
run_news_example(&firecrawl).await?;
run_academic_example(&firecrawl).await?;
run_commercial_example(&firecrawl).await?;
}
Examples::Basic => run_basic_example(&firecrawl).await?,
Examples::Advanced => run_advanced_example(&firecrawl).await?,
Examples::Geo => run_geographic_example(&firecrawl).await?,
Examples::Temporal => run_temporal_example(&firecrawl).await?,
Examples::Social => run_social_example(&firecrawl).await?,
Examples::News => run_news_example(&firecrawl).await?,
Examples::Academic => run_academic_example(&firecrawl).await?,
Examples::Commercial => run_commercial_example(&firecrawl).await?,
}
Ok(())
}
async fn run_basic_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "rust programming language";
let results = firecrawl.search(query, None).await?;
print_results("Basic Search", query, &results);
Ok(())
}
async fn run_advanced_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "rust web framework site:github.com OR site:gitlab.com";
let params = SearchParams {
query: query.to_string(),
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Advanced Repository Search", query, &results);
Ok(())
}
async fn run_geographic_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "coworking space startup hub";
let params = SearchParams {
query: query.to_string(),
// WARN: Doesn't work with searxng
location: Some("Silicon Valley, CA".to_string()),
// WARN: Doesn't work with searxng
country: Some("us".to_string()),
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Geographic-Specific Search", query, &results);
Ok(())
}
async fn run_temporal_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "artificial intelligence breakthroughs";
let params = SearchParams {
query: query.to_string(),
// WARN: Doesn't work with searxng
tbs: Some("qdr:m1".to_string()),
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Recent AI News", query, &results);
Ok(())
}
async fn run_social_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "viral tech trends site:twitter.com";
let params = SearchParams {
query: query.to_string(),
// WARN: Doesn't work. Maybe searxng related
filter: Some("site:twitter.com OR site:linkedin.com".to_string()),
// WARN: Doesn't work with searxng
tbs: Some("qdr:w".to_string()), // Last week
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Social Media Tech Trends", query, &results);
Ok(())
}
async fn run_news_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query =
"cryptocurrency market analysis site:reuters.com OR site:bloomberg.com OR site:ft.com";
let params = SearchParams {
query: query.to_string(),
// WARN: Doesn't work with searxng
tbs: Some("qdr:d".to_string()), // Last 24 hours
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Financial News Search", query, &results);
Ok(())
}
async fn run_academic_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "quantum computing research papers site:arxiv.org OR site:scholar.google.com";
let params = SearchParams {
query: query.to_string(),
// WARN: Doesn't work. Maybe searxng related
// filter: Some("site:arxiv.org OR site:scholar.google.com".to_string()),
// WARN: Doesn't work with searxng
tbs: Some("qdr:y".to_string()), // Last year
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Academic Research Search", query, &results);
Ok(())
}
async fn run_commercial_example(firecrawl: &FirecrawlApp) -> Result<(), Box<dyn Error>> {
let query = "enterprise cloud solutions reviews site:g2.com";
let params = SearchParams {
query: query.to_string(),
limit: Some(5),
..Default::default()
};
let results = firecrawl.search_with_params(params).await?;
print_results("Commercial Product Search", query, &results);
Ok(())
}
fn print_results(name: &str, query: &str, results: &SearchResponse) {
let sec = "=".repeat(70);
println!("\n{sec}");
println!("🔍 {name}");
println!("🔎 Query: \"{query}\"");
println!("{sec}");
for (i, doc) in results.data.iter().enumerate() {
println!("{}. 📌 Title: {}", i + 1, doc.title);
println!(" - 🔗 URL: {}", doc.url);
println!(" - 📝 Description: \"{:.40}\"...", doc.description);
}
if let Some(warning) = &results.warning {
println!("\n⚠️ Warning: {warning}");
}
println!("{sec}\n");
}

View File

@ -0,0 +1,494 @@
use std::collections::HashMap;
use serde::{Deserialize, Serialize};
use crate::{
crawl::{CrawlErrorsResponse, CrawlStatus, CrawlStatusTypes},
scrape::ScrapeOptions,
FirecrawlApp, FirecrawlError, API_VERSION,
};
#[serde_with::skip_serializing_none]
#[derive(Deserialize, Serialize, Debug, Default, Clone)]
#[serde(rename_all = "camelCase")]
pub struct BatchScrapeParams {
/// List of URLs to scrape
pub urls: Vec<String>,
/// Scrape options to apply to all URLs
#[serde(flatten)]
pub options: Option<ScrapeOptions>,
/// Whether to ignore invalid URLs
#[serde(rename = "ignoreInvalidURLs")]
pub ignore_invalid_urls: bool,
/// ID of an existing job to append these URLs to
pub append_to_id: Option<String>,
/// Webhook configuration
pub webhook: Option<WebhookOptions>,
/// Idempotency key to send to the crawl endpoint.
#[serde(skip)]
pub idempotency_key: Option<String>,
}
/// Options for webhook notifications
#[serde_with::skip_serializing_none]
#[derive(Deserialize, Serialize, Debug, Default, Clone)]
#[serde(rename_all = "camelCase")]
pub struct WebhookOptions {
/// URL to send webhook notifications to
pub url: String,
/// Custom headers to include in webhook requests
pub headers: Option<HashMap<String, String>>,
/// Authentication token for the webhook
pub auth_token: Option<String>,
}
impl From<&str> for WebhookOptions {
fn from(url: &str) -> Self {
Self {
url: url.to_string(),
headers: None,
auth_token: None,
}
}
}
/// Response from initiating a batch scrape job
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct BatchScrapeResponse {
/// Whether the request was successful
pub success: bool,
/// The ID of the batch scrape job
pub id: String,
/// URL to get the status of the batch scrape job
pub url: String,
/// List of URLs that were invalid and could not be processed
pub invalid_urls: Option<Vec<String>>,
}
impl From<String> for WebhookOptions {
fn from(url: String) -> Self {
Self {
url,
headers: None,
auth_token: None,
}
}
}
impl FirecrawlApp {
/// Initiates an asynchronous batch scrape job
pub async fn async_batch_scrape_urls(
&self,
params: BatchScrapeParams,
) -> Result<BatchScrapeResponse, FirecrawlError> {
let headers = self.prepare_headers(params.idempotency_key.as_ref());
let response = self
.client
.post(format!("{}{}/batch/scrape", self.api_url, API_VERSION))
.headers(headers)
.json(&params)
.send()
.await
.map_err(|e| FirecrawlError::HttpError("Initiating batch scrape job".to_string(), e))?;
self.handle_response(response, "initiate batch scrape job")
.await
}
/// Initiates a batch scrape job and waits for completion
pub async fn batch_scrape_urls(
&self,
params: BatchScrapeParams,
poll_interval: Option<u64>,
) -> Result<CrawlStatus, FirecrawlError> {
let poll_interval_ms = poll_interval.unwrap_or(2000);
let response = self.async_batch_scrape_urls(params).await?;
self.monitor_batch_job_status(&response.id, poll_interval_ms)
.await
}
/// Checks the status of a batch scrape job
pub async fn check_batch_scrape_status(
&self,
id: impl AsRef<str>,
) -> Result<CrawlStatus, FirecrawlError> {
let response = self
.client
.get(format!(
"{}{}/batch/scrape/{}",
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(
format!("Checking status of batch scrape {}", id.as_ref()),
e,
)
})?;
let mut status: CrawlStatus = self
.handle_response(
response,
format!("Checking status of batch scrape {}", id.as_ref()),
)
.await?;
if status.status == CrawlStatusTypes::Completed {
while let Some(next) = status.next.clone() {
let new_status = self.check_batch_scrape_status_next(next).await?;
status.data.extend_from_slice(&new_status.data);
status.next = new_status.next;
}
}
Ok(status)
}
/// Helper function to paginate through batch scrape status results
async fn check_batch_scrape_status_next(
&self,
next: impl AsRef<str>,
) -> Result<CrawlStatus, FirecrawlError> {
let response = self
.client
.get(next.as_ref())
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(
format!("Paginating batch scrape using URL {:?}", next.as_ref()),
e,
)
})?;
self.handle_response(
response,
format!("Paginating batch scrape using URL {:?}", next.as_ref()),
)
.await
}
/// Check for errors in a batch scrape job
pub async fn check_batch_scrape_errors(
&self,
id: impl AsRef<str>,
) -> Result<CrawlErrorsResponse, FirecrawlError> {
let response = self
.client
.get(format!(
"{}{}/batch/scrape/{}/errors",
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(
format!("Checking errors for batch scrape {}", id.as_ref()),
e,
)
})?;
self.handle_response(
response,
format!("Checking errors for batch scrape {}", id.as_ref()),
)
.await
}
/// Helper function to poll for batch job status until completion
async fn monitor_batch_job_status(
&self,
id: &str,
poll_interval: u64,
) -> Result<CrawlStatus, FirecrawlError> {
loop {
let status_data = self.check_batch_scrape_status(id).await?;
match status_data.status {
CrawlStatusTypes::Completed => {
break Ok(status_data);
}
CrawlStatusTypes::Scraping => {
tokio::time::sleep(tokio::time::Duration::from_millis(poll_interval)).await;
}
CrawlStatusTypes::Failed => {
break Err(FirecrawlError::CrawlJobFailed(
"Batch scrape job failed".into(),
status_data,
));
}
CrawlStatusTypes::Cancelled => {
break Err(FirecrawlError::CrawlJobFailed(
"Batch scrape job was cancelled".into(),
status_data,
));
}
}
}
}
}
#[cfg(test)]
mod tests {
use super::*;
use serde_json::json;
#[tokio::test]
#[ignore = "Makes real network request"]
async fn test_real_batch_scrape() {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let app = FirecrawlApp::new_selfhosted(api_url, None::<&str>).unwrap();
// Start a batch scrape job
let params = BatchScrapeParams {
urls: vec![
"https://example.com".to_string(),
"https://example.org".to_string(),
],
ignore_invalid_urls: true,
..Default::default()
};
let response = app.async_batch_scrape_urls(params).await.unwrap();
assert!(response.success);
assert!(!response.id.is_empty());
assert!(!response.url.is_empty());
}
#[tokio::test]
async fn test_async_batch_scrape_with_mock() {
let mut server = mockito::Server::new_async().await;
// Set up the mock
let mock = server
.mock("POST", "/v1/batch/scrape")
// Remove the match_body expectation which might be causing issues
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "batch-123",
"url": "https://api.example.com/v1/batch/batch-123",
"invalidUrls": []
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = BatchScrapeParams {
urls: vec![
"https://example.com".to_string(),
"https://example.org".to_string(),
],
ignore_invalid_urls: true,
..Default::default()
};
let response = app.async_batch_scrape_urls(params).await.unwrap();
assert!(response.success);
assert_eq!(response.id, "batch-123");
assert_eq!(response.url, "https://api.example.com/v1/batch/batch-123");
assert!(response.invalid_urls.unwrap_or_default().is_empty());
mock.assert();
}
#[tokio::test]
async fn test_batch_scrape_with_webhook() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/batch/scrape")
// Remove the match_body expectation to simplify
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "batch-123",
"url": "https://api.example.com/v1/batch/batch-123"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = BatchScrapeParams {
urls: vec!["https://example.com".to_string()],
webhook: Some("https://webhook.example.com/notify".into()),
..Default::default()
};
let response = app.async_batch_scrape_urls(params).await.unwrap();
assert!(response.success);
assert_eq!(response.id, "batch-123");
mock.assert();
}
#[tokio::test]
async fn test_check_batch_scrape_status_with_mock() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("GET", "/v1/batch/scrape/batch-123")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"status": "completed",
"total": 2,
"completed": 2,
"creditsUsed": 2,
"expiresAt": "2023-12-31T23:59:59Z",
"data": [
{
"metadata": {
"sourceURL": "https://example.com",
"statusCode": 200
},
"markdown": "Example Domain content"
},
{
"metadata": {
"sourceURL": "https://example.org",
"statusCode": 200
},
"markdown": "Another example content"
}
]
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let status = app.check_batch_scrape_status("batch-123").await.unwrap();
assert_eq!(status.total, 2);
assert_eq!(status.completed, 2);
assert_eq!(status.data.len(), 2);
assert_eq!(status.data[0].metadata.source_url, "https://example.com");
assert_eq!(status.data[1].metadata.source_url, "https://example.org");
mock.assert();
}
#[tokio::test]
async fn test_check_batch_scrape_errors_with_mock() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("GET", "/v1/batch/scrape/batch-123/errors")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"errors": [
{
"id": "error1",
"timestamp": "2023-01-01T00:00:00Z",
"url": "https://invalid.example.com",
"error": "Failed to load page"
}
],
"robotsBlocked": [
"https://example.com/admin"
]
})
.to_string(),
)
.create_async()
.await;
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let errors = app.check_batch_scrape_errors("batch-123").await.unwrap();
assert_eq!(errors.errors.len(), 1);
assert_eq!(errors.errors[0].url, "https://invalid.example.com");
assert_eq!(errors.robots_blocked.len(), 1);
assert_eq!(errors.robots_blocked[0], "https://example.com/admin");
mock.assert();
}
#[tokio::test]
async fn test_batch_scrape_with_invalid_urls() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/batch/scrape")
// Remove the match_body expectation
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "batch-123",
"url": "https://api.example.com/v1/batch/batch-123",
"invalidUrls": ["invalid-url"]
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = BatchScrapeParams {
urls: vec!["https://example.com".to_string(), "invalid-url".to_string()],
ignore_invalid_urls: true,
..Default::default()
};
let response = app.async_batch_scrape_urls(params).await.unwrap();
assert!(response.success);
assert_eq!(response.id, "batch-123");
assert_eq!(response.invalid_urls, Some(vec!["invalid-url".to_string()]));
mock.assert();
}
#[tokio::test]
async fn test_batch_scrape_error_response() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/batch/scrape")
.with_status(400)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": false,
"error": "No valid URLs provided"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = BatchScrapeParams::default();
let result = app.async_batch_scrape_urls(params).await;
assert!(result.is_err());
mock.assert();
}
}

View File

@ -2,7 +2,11 @@ use std::collections::HashMap;
use serde::{Deserialize, Serialize};
use crate::{document::Document, scrape::{ScrapeFormats, ScrapeOptions}, FirecrawlApp, FirecrawlError, API_VERSION};
use crate::{
document::Document,
scrape::{ScrapeFormats, ScrapeOptions},
FirecrawlApp, FirecrawlError, API_VERSION,
};
#[derive(Deserialize, Serialize, Clone, Copy, Debug)]
pub enum CrawlScrapeFormats {
@ -81,7 +85,9 @@ pub struct CrawlScrapeOptions {
impl From<CrawlScrapeOptions> for ScrapeOptions {
fn from(value: CrawlScrapeOptions) -> Self {
ScrapeOptions {
formats: value.formats.map(|formats| formats.into_iter().map(|x| x.into()).collect()),
formats: value
.formats
.map(|formats| formats.into_iter().map(|x| x.into()).collect()),
only_main_content: value.only_main_content,
include_tags: value.include_tags,
exclude_tags: value.exclude_tags,
@ -200,6 +206,29 @@ pub struct CrawlStatus {
pub data: Vec<Document>,
}
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct CrawlError {
pub id: String,
pub timestamp: Option<String>,
pub url: String,
pub error: String,
}
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct CrawlErrorsResponse {
pub errors: Vec<CrawlError>,
#[serde(rename = "robotsBlocked")]
pub robots_blocked: Vec<String>,
}
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct CancelCrawlResponse {
pub status: String,
}
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct CrawlAsyncResponse {
@ -228,14 +257,15 @@ impl FirecrawlApp {
let response = self
.client
.post(&format!("{}{}/crawl", self.api_url, API_VERSION))
.post(format!("{}{}/crawl", self.api_url, API_VERSION))
.headers(headers.clone())
.json(&body)
.send()
.await
.map_err(|e| FirecrawlError::HttpError(format!("Crawling {:?}", url.as_ref()), e))?;
self.handle_response::<CrawlAsyncResponse>(response, "start crawl job").await
self.handle_response::<CrawlAsyncResponse>(response, "start crawl job")
.await
}
/// Performs a crawl job for a URL using the Firecrawl API, waiting for the end result. This may take a long time depending on the size of the target page and your options (namely `CrawlOptions.limit`).
@ -245,38 +275,65 @@ impl FirecrawlApp {
options: impl Into<Option<CrawlOptions>>,
) -> Result<CrawlStatus, FirecrawlError> {
let options = options.into();
let poll_interval = options.as_ref().and_then(|x| x.poll_interval).unwrap_or(2000);
let poll_interval = options
.as_ref()
.and_then(|x| x.poll_interval)
.unwrap_or(2000);
let res = self.crawl_url_async(url, options).await?;
self.monitor_job_status(&res.id, poll_interval).await
}
async fn check_crawl_status_next(&self, next: impl AsRef<str>) -> Result<CrawlStatus, FirecrawlError> {
async fn check_crawl_status_next(
&self,
next: impl AsRef<str>,
) -> Result<CrawlStatus, FirecrawlError> {
let response = self
.client
.get(next.as_ref())
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| FirecrawlError::HttpError(format!("Paginating crawl using URL {:?}", next.as_ref()), e))?;
.map_err(|e| {
FirecrawlError::HttpError(
format!("Paginating crawl using URL {:?}", next.as_ref()),
e,
)
})?;
self.handle_response(response, format!("Paginating crawl using URL {:?}", next.as_ref())).await
self.handle_response(
response,
format!("Paginating crawl using URL {:?}", next.as_ref()),
)
.await
}
/// Checks for the status of a crawl, based on the crawl's ID. To be used in conjunction with `FirecrawlApp::crawl_url_async`.
pub async fn check_crawl_status(&self, id: impl AsRef<str>) -> Result<CrawlStatus, FirecrawlError> {
pub async fn check_crawl_status(
&self,
id: impl AsRef<str>,
) -> Result<CrawlStatus, FirecrawlError> {
let response = self
.client
.get(&format!(
.get(format!(
"{}{}/crawl/{}",
self.api_url, API_VERSION, id.as_ref()
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| FirecrawlError::HttpError(format!("Checking status of crawl {}", id.as_ref()), e))?;
.map_err(|e| {
FirecrawlError::HttpError(format!("Checking status of crawl {}", id.as_ref()), e)
})?;
let mut status: CrawlStatus = self.handle_response(response, format!("Checking status of crawl {}", id.as_ref())).await?;
let mut status: CrawlStatus = self
.handle_response(
response,
format!("Checking status of crawl {}", id.as_ref()),
)
.await?;
if status.status == CrawlStatusTypes::Completed {
while let Some(next) = status.next {
@ -304,16 +361,240 @@ impl FirecrawlApp {
tokio::time::sleep(tokio::time::Duration::from_millis(poll_interval)).await;
}
CrawlStatusTypes::Failed => {
break Err(FirecrawlError::CrawlJobFailed(format!(
"Crawl job failed."
), status_data));
break Err(FirecrawlError::CrawlJobFailed(
"Crawl job failed".into(),
status_data,
));
}
CrawlStatusTypes::Cancelled => {
break Err(FirecrawlError::CrawlJobFailed(format!(
"Crawl job was cancelled."
), status_data));
break Err(FirecrawlError::CrawlJobFailed(
"Crawl job was cancelled.".into(),
status_data,
));
}
}
}
}
/// Cancel an asynchronous crawl job using the Firecrawl API.
///
/// # Returns
///
/// A response indicating whether the cancellation was successful, or a FirecrawlError if the request fails.
pub async fn cancel_crawl(
&self,
id: impl AsRef<str>,
) -> Result<CancelCrawlResponse, FirecrawlError> {
let response = self
.client
.delete(format!(
"{}{}/crawl/{}",
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(format!("Cancelling crawl {}", id.as_ref()), e)
})?;
self.handle_response(response, "crawl_cancel").await
}
/// Returns information about crawl errors.
///
/// # Returns
///
/// A response containing information about crawl errors, or a FirecrawlError if the request fails.
pub async fn check_crawl_errors(
&self,
id: impl AsRef<str>,
) -> Result<CrawlErrorsResponse, FirecrawlError> {
let response = self
.client
.get(format!(
"{}{}/crawl/{}/errors",
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(format!("Checking errors for crawl {}", id.as_ref()), e)
})?;
self.handle_response(response, "crawl_check").await
}
}
#[cfg(test)]
mod tests {
use super::*;
use serde_json::json;
#[tokio::test]
#[ignore = "Makes real network request"]
async fn test_real_cancel_crawl() {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let app = FirecrawlApp::new_selfhosted(api_url, None::<&str>).unwrap();
// First start a crawl job
let crawl_response = app
.crawl_url_async("https://example.com", None)
.await
.unwrap();
// Then cancel it
let cancel_response = app.cancel_crawl(crawl_response.id).await.unwrap();
assert_eq!(cancel_response.status, "cancelled");
}
#[tokio::test]
async fn test_cancel_crawl_with_mock() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for the cancel request
let mock = server
.mock("DELETE", "/v1/crawl/test-crawl-id")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": null,
"status": "cancelled"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let response = app.cancel_crawl("test-crawl-id").await.unwrap();
assert_eq!(response.status, "cancelled");
mock.assert();
}
#[tokio::test]
async fn test_cancel_crawl_error_response() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for an error response
let mock = server
.mock("DELETE", "/v1/crawl/invalid-id")
.with_status(404)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": false,
"error": "Crawl job not found"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let result = app.cancel_crawl("invalid-id").await;
assert!(result.is_err());
mock.assert();
}
#[tokio::test]
#[ignore = "Makes real network request"]
async fn test_real_check_crawl_errors() {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let app = FirecrawlApp::new_selfhosted(api_url, None::<&str>).unwrap();
// First start a crawl job
let crawl_response = app
.crawl_url_async("https://no-wer-agg.invalid", None)
.await
.unwrap();
// Check for errors
let errors_response = app.check_crawl_errors(crawl_response.id).await.unwrap();
println!("{errors_response:?}");
tokio::time::sleep(tokio::time::Duration::from_secs(3)).await;
assert!(
!errors_response.errors.is_empty(),
"WARN: Error returned related to Supabase not in my environment. It may fail"
);
}
#[tokio::test]
async fn test_check_crawl_errors_with_mock() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for the check errors request
let mock = server
.mock("GET", "/v1/crawl/test-crawl-id/errors")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"errors": [
{
"id": "error1",
"timestamp": "2023-01-01T00:00:00Z",
"url": "https://example.com/error-page",
"error": "Failed to load page"
}
],
"robotsBlocked": [
"https://example.com/blocked-by-robots"
]
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let response = app.check_crawl_errors("test-crawl-id").await.unwrap();
assert_eq!(response.errors.len(), 1);
assert_eq!(response.errors[0].id, "error1");
assert_eq!(response.errors[0].url, "https://example.com/error-page");
assert_eq!(response.errors[0].error, "Failed to load page");
assert_eq!(response.robots_blocked.len(), 1);
assert_eq!(
response.robots_blocked[0],
"https://example.com/blocked-by-robots"
);
mock.assert();
}
#[tokio::test]
async fn test_check_crawl_errors_error_response() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for an error response
let mock = server
.mock("GET", "/v1/crawl/invalid-id/errors")
.with_status(404)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": false,
"error": "Crawl job not found"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let result = app.check_crawl_errors("invalid-id").await;
assert!(result.is_err());
mock.assert();
}
}

View File

@ -83,4 +83,3 @@ pub struct Document {
/// The warning message will contain any errors encountered during the extraction.
pub warning: Option<String>,
}

View File

@ -42,4 +42,6 @@ pub enum FirecrawlError {
APIError(String, FirecrawlAPIError),
#[error("Crawl job failed: {0}")]
CrawlJobFailed(String, CrawlStatus),
#[error("Missuse: {0}")]
Missuse(String),
}

View File

@ -0,0 +1,596 @@
use std::collections::HashMap;
use schemars::schema_for;
use serde::{Deserialize, Serialize};
use serde_json::Value;
use crate::{FirecrawlApp, FirecrawlError, API_VERSION};
/// Parameters for extract requests
#[serde_with::skip_serializing_none]
#[derive(Deserialize, Serialize, Debug, Default, Clone)]
#[serde(rename_all = "camelCase")]
pub struct ExtractParams {
/// URLs to extract information from
pub urls: Option<Vec<String>>,
/// Extraction prompt
pub prompt: Option<String>,
/// Schema for structured output
pub schema: Option<Value>,
/// System prompt for the LLM
pub system_prompt: Option<String>,
/// Allow following external links
pub allow_external_links: Option<bool>,
/// Enable web search for additional information
pub enable_web_search: Option<bool>,
/// Show sources in the response
pub show_sources: Option<bool>,
/// Origin information, defaults to "api-sdk"
pub origin: Option<String>,
/// Timeout in milliseconds, defaults to 60000
pub timeout: Option<u32>,
/// Whether to include URL trace information, defaults to false
pub url_trace: Option<bool>,
/// Whether to ignore sitemap, defaults to false
pub ignore_sitemap: Option<bool>,
/// Whether to include subdomains, defaults to true
pub include_subdomains: Option<bool>,
/// Maximum number of URLs to process
pub limit: Option<u32>,
/// Experimental: Stream steps information
#[serde(rename = "__experimental_streamSteps")]
pub experimental_stream_steps: Option<bool>,
/// Experimental: Include LLM usage information
#[serde(rename = "__experimental_llmUsage")]
pub experimental_llm_usage: Option<bool>,
/// Experimental: Show sources information
#[serde(rename = "__experimental_showSources")]
pub experimental_show_sources: Option<bool>,
/// Experimental: Cache key
#[serde(rename = "__experimental_cacheKey")]
pub experimental_cache_key: Option<String>,
/// Experimental: Cache mode, defaults to "direct"
#[serde(rename = "__experimental_cacheMode")]
pub experimental_cache_mode: Option<String>,
}
/// Response from initiating an extract operation
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct ExtractResponse {
/// Whether the request was successful
pub success: bool,
/// The ID of the extract job
pub id: String,
/// URL trace information if requested
pub url_trace: Option<Vec<URLTrace>>,
}
/// Information about URL processing during extraction
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct URLTrace {
/// The URL being processed
pub url: String,
/// Status of processing this URL
pub status: String,
/// Timing information for URL processing
pub timing: URLTraceTiming,
/// Error message if processing failed
pub error: Option<String>,
/// Warning message if there were issues
pub warning: Option<String>,
/// Content statistics
pub content_stats: Option<ContentStats>,
/// Relevance score for this URL (0-1)
pub relevance_score: Option<f64>,
/// Whether this URL was used in the final completion
pub used_in_completion: Option<bool>,
/// Fields extracted from this URL
pub extracted_fields: Option<Vec<String>>,
}
/// Timing information for URL processing
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct URLTraceTiming {
/// When the URL was discovered
pub discovered_at: String,
/// When scraping began for this URL
pub scraped_at: Option<String>,
/// When processing was completed for this URL
pub completed_at: Option<String>,
}
/// Statistics about processed content
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct ContentStats {
/// Length of the raw content in characters
pub raw_content_length: u32,
/// Length of the processed content in characters
pub processed_content_length: u32,
/// Number of tokens used for this content
pub tokens_used: u32,
}
/// Response for extract status check
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct ExtractStatusResponse {
/// Whether the request was successful
pub success: bool,
/// Status of the extract job: "pending", "processing", "completed", "failed"
pub status: String,
/// Extracted data, present when status is "completed"
pub data: Option<Value>,
/// Error message if the job failed
pub error: Option<String>,
/// URL trace information if requested
pub url_trace: Option<Vec<URLTrace>>,
/// Sources information if requested
pub sources: Option<HashMap<String, Vec<String>>>,
}
impl FirecrawlApp {
/// Extracts information from URLs using the Firecrawl API.
///
/// This is the synchronous version that polls until completion.
///
/// Either `params.prompt` or `params.schema` must be provided.
pub async fn extract(
&self,
params: impl Into<ExtractParams>,
) -> Result<ExtractStatusResponse, FirecrawlError> {
let mut params = params.into();
// Validation: Either prompt or schema must be provided
if params.prompt.is_none() && params.schema.is_none() {
return Err(FirecrawlError::APIError(
"Extract validation".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: "Either prompt or schema must be provided".to_string(),
details: None,
},
));
}
// Set default origin if not provided
if params.origin.is_none() {
params.origin = Some("api-sdk".to_string());
}
// Initiate the extract job asynchronously
let response = self.async_extract(params).await?;
// Poll for the result
let poll_interval = 2000; // Default to 2 seconds
self.monitor_extract_job_status(&response.id, poll_interval)
.await
}
pub async fn extract_with_schemars<T>(
&self,
params: impl Into<ExtractParams>,
) -> Result<ExtractStatusResponse, FirecrawlError>
where
T: schemars::JsonSchema,
{
let mut params = params.into();
let schema = schema_for!(T);
let schema_json = serde_json::to_value(schema).map_err(|e| {
FirecrawlError::APIError(
"Schema serialization".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: e.to_string(),
details: None,
},
)
})?;
params.schema = Some(schema_json);
self.extract(params).await
}
/// Initiates an asynchronous extract operation.
///
/// # Arguments
///
/// * `params` - Parameters for the extract request
///
/// # Returns
///
/// A response containing the extract job ID, or a FirecrawlError if the request fails.
///
/// # Notes
///
/// Either `params.urls` or `params.prompt` must be provided.
/// Either `params.prompt` or `params.schema` must be provided.
pub async fn async_extract(
&self,
params: impl Into<ExtractParams>,
) -> Result<ExtractResponse, FirecrawlError> {
let params = params.into();
// Validation: Either URLs or prompt must be provided
if params.urls.is_none() && params.prompt.is_none() {
return Err(FirecrawlError::APIError(
"Extract validation".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: "Either URLs or prompt must be provided".to_string(),
details: None,
},
));
}
// Validation: Either prompt or schema must be provided
if params.prompt.is_none() && params.schema.is_none() {
return Err(FirecrawlError::APIError(
"Extract validation".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: "Either prompt or schema must be provided".to_string(),
details: None,
},
));
}
let headers = self.prepare_headers(None);
let response = self
.client
.post(format!("{}{}/extract", self.api_url, API_VERSION))
.headers(headers)
.json(&params)
.send()
.await
.map_err(|e| FirecrawlError::HttpError("Initiating extract job".to_string(), e))?;
self.handle_response(response, "initiate extract job").await
}
/// Checks the status of an extract job.
///
/// # Arguments
///
/// * `id` - The ID of the extract job
///
/// # Returns
///
/// A response containing the status of the extract job, or a FirecrawlError if the request fails.
pub async fn get_extract_status(
&self,
id: impl AsRef<str>,
) -> Result<ExtractStatusResponse, FirecrawlError> {
let response = self
.client
.get(format!(
"{}{}/extract/{}",
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(format!("Checking status of extract {}", id.as_ref()), e)
})?;
self.handle_response(
response,
format!("Checking status of extract {}", id.as_ref()),
)
.await
}
/// Helper function to poll for extract job status until completion
async fn monitor_extract_job_status(
&self,
id: &str,
poll_interval: u64,
) -> Result<ExtractStatusResponse, FirecrawlError> {
loop {
let status_data = self.get_extract_status(id).await?;
match status_data.status.as_str() {
"completed" => {
break Ok(status_data);
}
"pending" | "processing" => {
tokio::time::sleep(tokio::time::Duration::from_millis(poll_interval)).await;
}
"failed" => {
let error_msg = status_data
.error
.clone()
.unwrap_or_else(|| "Extract job failed".to_string());
break Err(FirecrawlError::APIError(
"Extract job failed".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: error_msg,
details: None,
},
));
}
_ => {
break Err(FirecrawlError::APIError(
"Extract job status".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: format!("Unexpected status: {}", status_data.status),
details: None,
},
));
}
}
}
}
}
#[cfg(test)]
mod tests {
use super::*;
use serde_json::json;
#[tokio::test]
#[ignore = "Makes real network request"]
async fn test_real_extract() {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let app = FirecrawlApp::new_selfhosted(api_url, None::<&str>).unwrap();
// Create extract params
let params = ExtractParams {
urls: Some(vec!["https://example.com".to_string()]),
prompt: Some("Extract the title and main content from this page".to_string()),
schema: None,
origin: Some("test".to_string()),
..Default::default()
};
// Start an extract job
let response = app.async_extract(params).await.unwrap();
assert!(response.success);
assert!(!response.id.is_empty());
}
#[tokio::test]
async fn test_async_extract_with_mock() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for the extract request
let mock = server
.mock("POST", "/v1/extract")
.match_body(mockito::Matcher::PartialJson(json!({
"urls": ["https://example.com"],
"prompt": "Extract the title and main content"
})))
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "extract-123",
"urlTrace": []
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = ExtractParams {
urls: Some(vec!["https://example.com".to_string()]),
prompt: Some("Extract the title and main content".to_string()),
schema: None,
..Default::default()
};
let response = app.async_extract(params).await.unwrap();
assert!(response.success);
assert_eq!(response.id, "extract-123");
assert!(response.url_trace.unwrap_or_default().is_empty());
mock.assert();
}
#[tokio::test]
async fn test_extract_with_schema() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for the extract request with schema
let mock = server
.mock("POST", "/v1/extract")
.match_body(mockito::Matcher::PartialJson(json!({
"urls": ["https://example.com"],
"schema": {
"type": "object",
"properties": {
"title": { "type": "string" },
"content": { "type": "string" }
}
}
})))
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "extract-123"
})
.to_string(),
)
.create();
// Set up the mock for the status request
let status_mock = server
.mock("GET", "/v1/extract/extract-123")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"status": "completed",
"data": {
"title": "Example Domain",
"content": "This domain is for use in illustrative examples in documents."
}
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let urls = Some(vec!["https://example.com".to_string()]);
let params = ExtractParams {
urls,
schema: Some(json!({
"type": "object",
"properties": {
"title": { "type": "string" },
"content": { "type": "string" }
}
})),
..Default::default()
};
let response = app.extract(params).await.unwrap();
assert!(response.success);
assert_eq!(response.status, "completed");
let data = response.data.unwrap();
assert_eq!(data["title"], "Example Domain");
assert_eq!(
data["content"],
"This domain is for use in illustrative examples in documents."
);
mock.assert();
status_mock.assert();
}
#[tokio::test]
async fn test_extract_status_with_mock() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for the status check
let mock = server
.mock("GET", "/v1/extract/extract-123")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"status": "processing",
"urlTrace": [
{
"url": "https://example.com",
"status": "scraping",
"timing": {
"discoveredAt": "2023-01-01T00:00:00Z"
}
}
]
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let status = app.get_extract_status("extract-123").await.unwrap();
assert!(status.success);
assert_eq!(status.status, "processing");
assert_eq!(status.url_trace.unwrap()[0].url, "https://example.com");
mock.assert();
}
#[tokio::test]
async fn test_extract_validation_errors() {
let app = FirecrawlApp::new_selfhosted("https://example.com", Some("test_key")).unwrap();
// Test missing both URLs and prompt
let result = app.async_extract(ExtractParams::default()).await;
assert!(result.is_err());
// Test having URLs but missing both prompt and schema
let params = ExtractParams {
urls: Some(vec!["https://example.com".to_string()]),
..Default::default()
};
let result = app.async_extract(params).await;
assert!(result.is_err());
}
#[tokio::test]
async fn test_extract_api_error() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for an error response
let mock = server
.mock("POST", "/v1/extract")
.with_status(400)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": false,
"error": "Invalid schema format"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = ExtractParams {
urls: Some(vec!["https://example.com".to_string()]),
schema: Some(json!("invalid")), // Invalid schema format
..Default::default()
};
let result = app.async_extract(params).await;
assert!(result.is_err());
mock.assert();
}
}

View File

@ -2,14 +2,18 @@ use reqwest::{Client, Response};
use serde::de::DeserializeOwned;
use serde_json::Value;
pub mod batch_scrape;
pub mod crawl;
pub mod document;
mod error;
pub mod extract;
pub mod llmstxt;
pub mod map;
pub mod scrape;
pub mod search;
pub use error::FirecrawlError;
use error::FirecrawlAPIError;
pub use error::FirecrawlError;
#[derive(Clone, Debug)]
pub struct FirecrawlApp {
@ -26,7 +30,10 @@ impl FirecrawlApp {
FirecrawlApp::new_selfhosted(CLOUD_API_URL, Some(api_key))
}
pub fn new_selfhosted(api_url: impl AsRef<str>, api_key: Option<impl AsRef<str>>) -> Result<Self, FirecrawlError> {
pub fn new_selfhosted(
api_url: impl AsRef<str>,
api_key: Option<impl AsRef<str>>,
) -> Result<Self, FirecrawlError> {
let url = api_url.as_ref().to_string();
if url == CLOUD_API_URL && api_key.is_none() {
@ -36,7 +43,7 @@ impl FirecrawlApp {
success: false,
error: "API key is required for cloud service".to_string(),
details: None,
}
},
));
}
@ -73,27 +80,43 @@ impl FirecrawlApp {
.text()
.await
.map_err(|e| FirecrawlError::ResponseParseErrorText(e))
.and_then(|response_json| serde_json::from_str::<Value>(&response_json).map_err(|e| FirecrawlError::ResponseParseError(e)))
.and_then(|response_json| {
serde_json::from_str::<Value>(&response_json)
.map_err(|e| FirecrawlError::ResponseParseError(e))
.inspect(|data| {
#[cfg(debug_assertions)]
println!("Response JSON: {:#?}", data);
})
})
.and_then(|response_value| {
if response_value["success"].as_bool().unwrap_or(false) {
Ok(serde_json::from_value::<T>(response_value).map_err(|e| FirecrawlError::ResponseParseError(e))?)
if action.as_ref().starts_with("crawl_") // no success in check/cancel crawl responses
|| response_value["success"].as_bool().unwrap_or(false)
{
Ok(serde_json::from_value::<T>(response_value)
.map_err(|e| FirecrawlError::ResponseParseError(e))?)
} else {
Err(FirecrawlError::APIError(
action.as_ref().to_string(),
serde_json::from_value(response_value).map_err(|e| FirecrawlError::ResponseParseError(e))?
serde_json::from_value(response_value)
.map_err(|e| FirecrawlError::ResponseParseError(e))?,
))
}
});
match &response {
Ok(_) => response,
Err(FirecrawlError::ResponseParseError(_)) | Err(FirecrawlError::ResponseParseErrorText(_)) => {
Err(FirecrawlError::ResponseParseError(_))
| Err(FirecrawlError::ResponseParseErrorText(_)) => {
if is_success {
response
} else {
Err(FirecrawlError::HttpRequestFailed(action.as_ref().to_string(), status.as_u16(), status.as_str().to_string()))
Err(FirecrawlError::HttpRequestFailed(
action.as_ref().to_string(),
status.as_u16(),
status.as_str().to_string(),
))
}
}
},
Err(_) => response,
}
}

View File

@ -0,0 +1,426 @@
use serde::{Deserialize, Serialize};
use crate::{FirecrawlApp, FirecrawlError, API_VERSION};
/// Parameters for generating LLMs.txt
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct GenerateLLMsTextParams {
/// URL for which to generate LLMs.txt
pub url: String,
/// Maximum number of URLs to process. Default: 10
pub max_urls: u32,
/// Whether to show the full LLMs-full.txt in the response. Default: false
pub show_full_text: bool,
/// Experimental streaming option
#[serde(rename = "__experimental_stream")]
pub experimental_stream: bool,
}
impl Default for GenerateLLMsTextParams {
fn default() -> Self {
Self {
url: String::new(),
max_urls: 1,
show_full_text: false,
experimental_stream: false,
}
}
}
/// Response from initiating a LLMs.txt generation job
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct GenerateLLMsTextResponse {
/// Whether the request was successful
pub success: bool,
/// Job ID for the LLMs.txt generation
pub id: String,
}
#[derive(Deserialize, Serialize, Debug, Clone, Default)]
pub struct LLMTextData {
#[serde(rename = "llmstxt")]
pub compact: Option<String>,
#[serde(rename = "llmsfulltxt")]
pub full: Option<String>,
}
/// Response from checking the status of a LLMs.txt generation job
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct GenerateLLMsTextStatusResponse {
/// Whether the request was successful
pub success: bool,
/// Status of the job: "pending", "processing", "completed", "failed"
pub status: String,
/// Generated LLMs.txt data, present when status is "completed"
#[serde(default)]
pub data: LLMTextData,
/// Error message if the job failed
pub error: Option<String>,
/// Expiration timestamp for the data
pub expires_at: String,
}
impl FirecrawlApp {
/// Generates LLMs.txt for a given URL and polls until completion.
///
/// # Arguments
///
/// * `params` - Parameters for the LLMs.txt generation
///
/// # Returns
///
/// A response containing the generation results, or a FirecrawlError if the request fails.
pub async fn generate_llms_text(
&self,
params: impl Into<GenerateLLMsTextParams>,
) -> Result<GenerateLLMsTextStatusResponse, FirecrawlError> {
// Initiate the LLMs.txt generation job asynchronously
let response = self.async_generate_llms_text(params).await?;
// Poll for the result
let poll_interval = 2000; // Default to 2 seconds
self.monitor_llms_text_job_status(&response.id, poll_interval)
.await
}
/// Initiates an asynchronous LLMs.txt generation operation.
///
/// # Arguments
///
/// * `params` - Parameters for the LLMs.txt generation
///
/// # Returns
///
/// A response containing the generation job ID, or a FirecrawlError if the request fails.
pub async fn async_generate_llms_text(
&self,
params: impl Into<GenerateLLMsTextParams>,
) -> Result<GenerateLLMsTextResponse, FirecrawlError> {
let params = params.into();
// Validation: URL must be provided
if params.url.is_empty() {
return Err(FirecrawlError::APIError(
"Generate LLMs.txt validation".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: "URL must be provided".to_string(),
details: None,
},
));
}
let headers = self.prepare_headers(None);
let response = self
.client
.post(format!("{}{}/llmstxt", self.api_url, API_VERSION))
.headers(headers)
.json(&params)
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError("Initiating LLMs.txt generation".to_string(), e)
})?;
self.handle_response(response, "initiate LLMs.txt generation")
.await
}
/// Checks the status of a LLMs.txt generation operation.
///
/// # Arguments
///
/// * `id` - The ID of the LLMs.txt generation operation
///
/// # Returns
///
/// A response containing the current status and results of the generation operation,
/// or a FirecrawlError if the request fails.
pub async fn check_generate_llms_text_status(
&self,
id: impl AsRef<str>,
) -> Result<GenerateLLMsTextStatusResponse, FirecrawlError> {
let response = self
.client
.get(format!(
"{}{}/llmstxt/{}",
self.api_url,
API_VERSION,
id.as_ref()
))
.headers(self.prepare_headers(None))
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(
format!("Checking status of LLMs.txt generation {}", id.as_ref()),
e,
)
})?;
self.handle_response(
response,
format!("Checking status of LLMs.txt generation {}", id.as_ref()),
)
.await
}
/// Helper function to poll for LLMs.txt generation job status until completion
async fn monitor_llms_text_job_status(
&self,
id: &str,
poll_interval: u64,
) -> Result<GenerateLLMsTextStatusResponse, FirecrawlError> {
loop {
let status_data = self.check_generate_llms_text_status(id).await?;
match status_data.status.as_str() {
"completed" => {
break Ok(status_data);
}
"pending" | "processing" => {
tokio::time::sleep(tokio::time::Duration::from_millis(poll_interval)).await;
}
"failed" => {
let error_msg = status_data
.error
.clone()
.unwrap_or_else(|| "LLMs.txt generation failed".to_string());
break Err(FirecrawlError::APIError(
"LLMs.txt generation failed".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: error_msg,
details: None,
},
));
}
_ => {
break Err(FirecrawlError::APIError(
"LLMs.txt generation status".to_string(),
crate::error::FirecrawlAPIError {
success: false,
error: format!("Unexpected status: {}", status_data.status),
details: None,
},
));
}
}
}
}
}
#[cfg(test)]
mod tests {
use super::*;
use serde_json::json;
#[tokio::test]
#[ignore = "Makes real network request"]
async fn test_real_generate_llms_text() {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let app = FirecrawlApp::new_selfhosted(api_url, None::<&str>).unwrap();
let params = GenerateLLMsTextParams {
url: "https://example.com".to_string(),
max_urls: 5,
show_full_text: true,
..Default::default()
};
let response = app.async_generate_llms_text(params).await.unwrap();
assert!(response.success);
assert!(!response.id.is_empty());
}
#[tokio::test]
async fn test_async_generate_llms_text_with_mock() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/llmstxt")
.match_body(mockito::Matcher::PartialJson(json!({
"url": "https://example.com",
"maxUrls": 5
})))
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "llmstxt-123"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = GenerateLLMsTextParams {
url: "https://example.com".to_string(),
max_urls: 5,
..Default::default()
};
let response = app.async_generate_llms_text(params).await.unwrap();
assert!(response.success);
assert_eq!(response.id, "llmstxt-123");
mock.assert();
}
#[tokio::test]
async fn test_check_generate_llms_text_status_with_mock() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("GET", "/v1/llmstxt/llmstxt-123")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"status": "processing",
"expiresAt": "2023-01-01T00:00:00Z"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let status = app
.check_generate_llms_text_status("llmstxt-123")
.await
.unwrap();
assert!(status.success);
assert_eq!(status.status, "processing");
assert_eq!(status.expires_at, "2023-01-01T00:00:00Z");
mock.assert();
}
#[tokio::test]
async fn test_generate_llms_text_with_mock() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for the generate request
let mock = server
.mock("POST", "/v1/llmstxt")
.match_body(mockito::Matcher::PartialJson(json!({
"url": "https://example.com",
"showFullText": true
})))
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"id": "llmstxt-123"
})
.to_string(),
)
.create();
// Set up the mock for the status request
let status_mock = server
.mock("GET", "/v1/llmstxt/llmstxt-123")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"status": "completed",
"data": {
"llmstxt": "Allow: /about\nDisallow: /admin\n",
"llmsfulltxt": "# LLMs.txt\n\nAllow: /about\nDisallow: /admin\n"
},
"expiresAt": "2023-01-01T00:00:00Z"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = GenerateLLMsTextParams {
url: "https://example.com".to_string(),
show_full_text: true,
..Default::default()
};
let response = app.generate_llms_text(params).await.unwrap();
assert!(response.success);
assert_eq!(response.status, "completed");
let data = response.data;
assert_eq!(
data.compact,
Some("Allow: /about\nDisallow: /admin\n".into())
);
assert_eq!(
data.full,
Some("# LLMs.txt\n\nAllow: /about\nDisallow: /admin\n".into())
);
mock.assert();
status_mock.assert();
}
#[tokio::test]
async fn test_generate_llms_text_validation_errors() {
let app = FirecrawlApp::new_selfhosted("https://example.com", Some("test_key")).unwrap();
// Test missing URL
let params = GenerateLLMsTextParams {
url: "".to_string(),
..Default::default()
};
let result = app.async_generate_llms_text(params).await;
assert!(result.is_err());
}
#[tokio::test]
async fn test_generate_llms_text_api_error() {
let mut server = mockito::Server::new_async().await;
// Set up the mock for an error response
let mock = server
.mock("POST", "/v1/llmstxt")
.with_status(400)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": false,
"error": "Invalid URL format"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = GenerateLLMsTextParams {
url: "not-a-valid-url".to_string(),
..Default::default()
};
let result = app.async_generate_llms_text(params).await;
assert!(result.is_err());
mock.assert();
}
}

View File

@ -16,7 +16,7 @@ pub struct MapOptions {
pub include_subdomains: Option<bool>,
/// Maximum number of links to return (default: `5000`)
pub exclude_tags: Option<u32>,
pub limit: Option<u32>,
}
#[derive(Deserialize, Serialize, Debug, Default)]
@ -59,7 +59,9 @@ impl FirecrawlApp {
.await
.map_err(|e| FirecrawlError::HttpError(format!("Mapping {:?}", url.as_ref()), e))?;
let response = self.handle_response::<MapResponse>(response, "scrape URL").await?;
let response = self
.handle_response::<MapResponse>(response, "scrape URL")
.await?;
Ok(response.links)
}

View File

@ -43,7 +43,7 @@ pub enum ScrapeFormats {
}
#[serde_with::skip_serializing_none]
#[derive(Deserialize, Serialize, Debug, Default)]
#[derive(Deserialize, Serialize, Debug, Default, Clone)]
#[serde(rename_all = "camelCase")]
pub struct ExtractOptions {
/// Schema the output should adhere to, provided in JSON Schema format.
@ -56,7 +56,7 @@ pub struct ExtractOptions {
}
#[serde_with::skip_serializing_none]
#[derive(Deserialize, Serialize, Debug, Default)]
#[derive(Deserialize, Serialize, Debug, Default, Clone)]
#[serde(rename_all = "camelCase")]
pub struct ScrapeOptions {
/// Formats to extract from the page. (default: `[ Markdown ]`)
@ -131,7 +131,9 @@ impl FirecrawlApp {
.await
.map_err(|e| FirecrawlError::HttpError(format!("Scraping {:?}", url.as_ref()), e))?;
let response = self.handle_response::<ScrapeResponse>(response, "scrape URL").await?;
let response = self
.handle_response::<ScrapeResponse>(response, "scrape URL")
.await?;
Ok(response.data)
}

245
apps/rust-sdk/src/search.rs Normal file
View File

@ -0,0 +1,245 @@
use crate::{scrape::ScrapeOptions, FirecrawlApp, FirecrawlError, API_VERSION};
use serde::{Deserialize, Serialize};
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct SearchParams {
/// The search query string
pub query: String,
/// Maximum number of results to return. Default: 5, Max: 20
pub limit: Option<u32>,
/// Time-based search filter.
#[serde(skip_serializing_if = "Option::is_none")]
pub tbs: Option<String>,
/// Query string to filter search results. Example: "site:example.com"
#[serde(skip_serializing_if = "Option::is_none")]
pub filter: Option<String>,
/// Language code. Default: "en"
pub lang: Option<String>,
/// Country code. Default: "us"
pub country: Option<String>,
/// Geographic location string for local search results
#[serde(skip_serializing_if = "Option::is_none")]
pub location: Option<String>,
/// Origin identifier. Default: "api"
pub origin: Option<String>,
/// Timeout in milliseconds. Default: 60000
pub timeout: Option<u32>,
/// Additional options for webpage scraping behavior
#[serde(skip_serializing_if = "Option::is_none")]
pub scrape_options: Option<ScrapeOptions>,
}
impl Default for SearchParams {
fn default() -> Self {
Self {
query: String::new(),
limit: Some(5),
tbs: None,
filter: None,
lang: Some("en".to_string()),
country: Some("us".to_string()),
location: None,
origin: Some("api".to_string()),
timeout: Some(60000),
scrape_options: None,
}
}
}
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct SearchResponse {
pub success: bool,
pub data: Vec<SearchDocument>,
pub warning: Option<String>,
}
// TODO: Consider merging fields into document::Document (url, title, description) while preserving optionality
/// A document returned from a search or scrape request
#[serde_with::skip_serializing_none]
#[derive(Deserialize, Serialize, Debug, Clone)]
#[serde(rename_all = "camelCase")]
pub struct SearchDocument {
/// Document URL
pub url: String,
/// Document title
pub title: String,
/// Document description
pub description: String,
}
impl FirecrawlApp {
/// Search for content using the Firecrawl API.
///
/// # Arguments
///
/// * `query` - The search query string
/// * `params` - Optional parameters for the search request
///
/// # Returns
///
/// A SearchResponse containing the search results, or a FirecrawlError if the request fails.
pub async fn search(
&self,
query: impl AsRef<str>,
params: impl Into<Option<SearchParams>>,
) -> Result<SearchResponse, FirecrawlError> {
let mut search_params = params.into().unwrap_or_default();
search_params.query = query.as_ref().to_string();
self.search_with_params(search_params).await
}
/// Alternative method that takes SearchParams directly
///
/// # Arguments
///
/// * `params` - Search parameters including the query
///
/// # Returns
///
/// A SearchResponse containing the search results, or a FirecrawlError if the request fails.
pub async fn search_with_params(
&self,
params: SearchParams,
) -> Result<SearchResponse, FirecrawlError> {
let headers = self.prepare_headers(None);
let response = self
.client
.post(format!("{}{}/search", self.api_url, API_VERSION))
.headers(headers)
.json(&params)
.send()
.await
.map_err(|e| {
FirecrawlError::HttpError(format!("Searching with query: {:?}", params.query), e)
})?;
self.handle_response::<SearchResponse>(response, "search")
.await
}
}
#[cfg(test)]
pub mod tests {
use super::*;
use serde_json::json;
#[tokio::test]
#[ignore = "Makes real network request"]
async fn test_real_search() {
let api_url = std::env::var("FIRECRAWL_API_URL")
.expect("Please set the FIRECRAWL_API_URL environment variable");
let app = FirecrawlApp::new_selfhosted(api_url, None::<&str>).unwrap();
let response = app.search("test query", None).await.unwrap();
assert!(response.success);
}
#[tokio::test]
async fn test_search_with_mock() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/search")
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"data": [{
"url": "https://example.com",
"title": "Example Domain",
"description": "...."
}],
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let response = app.search("test", None).await.unwrap();
assert!(response.success);
assert_eq!(response.data.len(), 1);
assert_eq!(response.data[0].url, "https://example.com");
assert_eq!(response.data[0].title, "Example Domain".to_string());
assert_eq!(response.data[0].description, "....".to_string());
mock.assert();
}
#[tokio::test]
async fn test_search_with_params() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/search")
.with_header("content-type", "application/json")
.match_body(mockito::Matcher::Json(json!({
"query": "test",
"limit": 10,
"lang": "fr",
"country": "fr",
"origin": "api",
"timeout": 30000
})))
.with_status(200)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": true,
"data": [],
"warning": "No results found"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let params = SearchParams {
query: "test".to_string(),
limit: Some(10),
lang: Some("fr".to_string()),
country: Some("fr".to_string()),
timeout: Some(30000),
..Default::default()
};
let response = app.search_with_params(params).await.unwrap();
assert!(response.success);
assert_eq!(response.data.len(), 0);
assert_eq!(response.warning, Some("No results found".to_string()));
mock.assert();
}
#[tokio::test]
async fn test_search_error_response() {
let mut server = mockito::Server::new_async().await;
let mock = server
.mock("POST", "/v1/search")
.with_status(400)
.with_header("content-type", "application/json")
.with_body(
json!({
"success": false,
"error": "Invalid query"
})
.to_string(),
)
.create();
let app = FirecrawlApp::new_selfhosted(server.url(), Some("test_key")).unwrap();
let result = app.search("", None).await;
assert!(result.is_err());
mock.assert();
}
#[tokio::test]
async fn test_search_network_error() {
let app = FirecrawlApp::new_selfhosted("http://invalid-url", Some("test_key")).unwrap();
let result = app.search("test", None).await;
assert!(result.is_err());
}
}

View File

@ -1,4 +1,3 @@
use assert_matches::assert_matches;
use dotenvy::dotenv;
use firecrawl::scrape::{ExtractOptions, ScrapeFormats, ScrapeOptions};
use firecrawl::{FirecrawlApp, FirecrawlError};
@ -24,11 +23,8 @@ use std::env;
async fn test_successful_response_with_valid_preview_token() {
dotenv().ok();
let api_url = env::var("API_URL").unwrap();
let app = FirecrawlApp::new_selfhosted(
api_url,
Some(env::var("PREVIEW_TOKEN").unwrap()),
)
.unwrap();
let app =
FirecrawlApp::new_selfhosted(api_url, Some(env::var("PREVIEW_TOKEN").unwrap())).unwrap();
let result = app
.scrape_url("https://roastmywebsite.ai", None)
.await
@ -82,7 +78,8 @@ async fn test_successful_response_for_valid_scrape_with_pdf_file() {
.await
.unwrap();
assert!(result.markdown.is_some());
assert!(result.markdown
assert!(result
.markdown
.unwrap()
.contains("We present spectrophotometric observations of the Broad Line Radio Galaxy"));
}
@ -98,12 +95,12 @@ async fn test_successful_response_for_valid_scrape_with_pdf_file_without_explici
.await
.unwrap();
assert!(result.markdown.is_some());
assert!(result.markdown
assert!(result
.markdown
.unwrap()
.contains("We present spectrophotometric observations of the Broad Line Radio Galaxy"));
}
// #[tokio::test]
// async fn test_should_return_error_for_blocklisted_url() {
// dotenv().ok();
@ -166,7 +163,11 @@ fn test_api_key_requirements() {
match (api_url.contains("api.firecrawl.dev"), api_key) {
(false, _) => {
let result = FirecrawlApp::new_selfhosted(&api_url, None::<String>);
assert!(result.is_ok(), "Local setup failed: {:?}", result.err().unwrap());
assert!(
result.is_ok(),
"Local setup failed: {:?}",
result.err().unwrap()
);
}
(true, None) => {
let result = FirecrawlApp::new_selfhosted(&api_url, None::<String>);