mirror of
https://git.mirrors.martin98.com/https://github.com/mendableai/firecrawl
synced 2025-06-22 18:01:21 +08:00
157 lines
5.4 KiB
Rust
157 lines
5.4 KiB
Rust
use assert_matches::assert_matches;
|
|
use dotenvy::dotenv;
|
|
use firecrawl::scrape::{ExtractOptions, ScrapeFormats, ScrapeOptions};
|
|
use firecrawl::FirecrawlApp;
|
|
use serde_json::json;
|
|
use std::env;
|
|
|
|
#[tokio::test]
|
|
async fn test_blocklisted_url() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let blocklisted_url = "https://facebook.com/fake-test";
|
|
let result = app.scrape_url(blocklisted_url, None).await;
|
|
|
|
assert_matches!(
|
|
result,
|
|
Err(e) if e.to_string().contains("Firecrawl currently does not support social media scraping due to policy restrictions")
|
|
);
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_successful_response_with_valid_preview_token() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let app = FirecrawlApp::new_selfhosted(
|
|
api_url,
|
|
Some("this_is_just_a_preview_token"),
|
|
)
|
|
.unwrap();
|
|
let result = app
|
|
.scrape_url("https://roastmywebsite.ai", None)
|
|
.await
|
|
.unwrap();
|
|
assert!(result.markdown.is_some());
|
|
assert!(result.markdown.unwrap().contains("_Roast_"));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_scrape_url_e2e() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let result = app
|
|
.scrape_url("https://roastmywebsite.ai", None)
|
|
.await
|
|
.unwrap();
|
|
assert!(result.markdown.is_some());
|
|
assert!(result.markdown.unwrap().contains("_Roast_"));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_successful_response_with_valid_api_key_and_include_html() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let params = ScrapeOptions {
|
|
formats: vec! [ ScrapeFormats::Markdown, ScrapeFormats::HTML ].into(),
|
|
..Default::default()
|
|
};
|
|
let result = app
|
|
.scrape_url("https://roastmywebsite.ai", params)
|
|
.await
|
|
.unwrap();
|
|
assert!(result.markdown.is_some());
|
|
assert!(result.html.is_some());
|
|
assert!(result.markdown.unwrap().contains("_Roast_"));
|
|
assert!(result.html.unwrap().contains("<h1"));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_successful_response_for_valid_scrape_with_pdf_file() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let result = app
|
|
.scrape_url("https://arxiv.org/pdf/astro-ph/9301001.pdf", None)
|
|
.await
|
|
.unwrap();
|
|
assert!(result.markdown.is_some());
|
|
assert!(result.markdown
|
|
.unwrap()
|
|
.contains("We present spectrophotometric observations of the Broad Line Radio Galaxy"));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_successful_response_for_valid_scrape_with_pdf_file_without_explicit_extension() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let result = app
|
|
.scrape_url("https://arxiv.org/pdf/astro-ph/9301001", None)
|
|
.await
|
|
.unwrap();
|
|
assert!(result.markdown.is_some());
|
|
assert!(result.markdown
|
|
.unwrap()
|
|
.contains("We present spectrophotometric observations of the Broad Line Radio Galaxy"));
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_should_return_error_for_blocklisted_url() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let blocklisted_url = "https://twitter.com/fake-test";
|
|
let result = app.crawl_url(blocklisted_url, None).await;
|
|
|
|
assert_matches!(
|
|
result,
|
|
Err(e) if e.to_string().contains("Firecrawl currently does not support social media scraping due to policy restrictions.")
|
|
);
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn test_llm_extraction() {
|
|
dotenv().ok();
|
|
let api_url = env::var("API_URL").unwrap();
|
|
let api_key = env::var("TEST_API_KEY").ok();
|
|
let app = FirecrawlApp::new_selfhosted(api_url, api_key).unwrap();
|
|
let options = ScrapeOptions {
|
|
formats: vec! [ ScrapeFormats::Extract ].into(),
|
|
extract: ExtractOptions {
|
|
prompt: "Based on the information on the page, find what the company's mission is and whether it supports SSO, and whether it is open source".to_string().into(),
|
|
schema: json!({
|
|
"type": "object",
|
|
"properties": {
|
|
"company_mission": {"type": "string"},
|
|
"supports_sso": {"type": "boolean"},
|
|
"is_open_source": {"type": "boolean"}
|
|
},
|
|
"required": ["company_mission", "supports_sso", "is_open_source"]
|
|
}).into(),
|
|
..Default::default()
|
|
}.into(),
|
|
..Default::default()
|
|
};
|
|
let result = app
|
|
.scrape_url("https://mendable.ai", options)
|
|
.await
|
|
.unwrap();
|
|
assert!(result.extract.is_some());
|
|
let llm_extraction = &result.extract.unwrap();
|
|
assert!(llm_extraction
|
|
.as_object()
|
|
.unwrap()
|
|
.contains_key("company_mission"));
|
|
assert!(llm_extraction["supports_sso"].is_boolean());
|
|
assert!(llm_extraction["is_open_source"].is_boolean());
|
|
}
|