mirror of
https://github.com/mendableai/firecrawl.git
synced 2024-11-16 11:42:24 +08:00
Nick: sitemap only
This commit is contained in:
parent
431e64e752
commit
f155449458
|
@ -1,10 +1,6 @@
|
||||||
import { Response } from "express";
|
import { Response } from "express";
|
||||||
import { v4 as uuidv4 } from "uuid";
|
import { v4 as uuidv4 } from "uuid";
|
||||||
import {
|
import { mapRequestSchema, RequestWithAuth, scrapeOptions } from "./types";
|
||||||
mapRequestSchema,
|
|
||||||
RequestWithAuth,
|
|
||||||
scrapeOptions,
|
|
||||||
} from "./types";
|
|
||||||
import { crawlToCrawler, StoredCrawl } from "../../lib/crawl-redis";
|
import { crawlToCrawler, StoredCrawl } from "../../lib/crawl-redis";
|
||||||
import { MapResponse, MapRequest } from "./types";
|
import { MapResponse, MapRequest } from "./types";
|
||||||
import { configDotenv } from "dotenv";
|
import { configDotenv } from "dotenv";
|
||||||
|
@ -46,6 +42,7 @@ export async function mapController(
|
||||||
originUrl: req.body.url,
|
originUrl: req.body.url,
|
||||||
crawlerOptions: {
|
crawlerOptions: {
|
||||||
...req.body,
|
...req.body,
|
||||||
|
limit: req.body.sitemapOnly ? 10000000 : limit,
|
||||||
scrapeOptions: undefined,
|
scrapeOptions: undefined,
|
||||||
},
|
},
|
||||||
scrapeOptions: scrapeOptions.parse({}),
|
scrapeOptions: scrapeOptions.parse({}),
|
||||||
|
@ -57,77 +54,92 @@ export async function mapController(
|
||||||
|
|
||||||
const crawler = crawlToCrawler(id, sc);
|
const crawler = crawlToCrawler(id, sc);
|
||||||
|
|
||||||
let urlWithoutWww = req.body.url.replace("www.", "");
|
// If sitemapOnly is true, only get links from sitemap
|
||||||
|
if (req.body.sitemapOnly) {
|
||||||
let mapUrl = req.body.search
|
const sitemap = await crawler.tryGetSitemap();
|
||||||
? `"${req.body.search}" site:${urlWithoutWww}`
|
if (sitemap !== null) {
|
||||||
: `site:${req.body.url}`;
|
sitemap.forEach((x) => {
|
||||||
|
|
||||||
const resultsPerPage = 100;
|
|
||||||
const maxPages = Math.ceil(Math.min(MAX_FIRE_ENGINE_RESULTS, limit) / resultsPerPage);
|
|
||||||
|
|
||||||
const cacheKey = `fireEngineMap:${mapUrl}`;
|
|
||||||
const cachedResult = null;
|
|
||||||
|
|
||||||
let allResults: any[] = [];
|
|
||||||
let pagePromises: Promise<any>[] = [];
|
|
||||||
|
|
||||||
if (cachedResult) {
|
|
||||||
allResults = JSON.parse(cachedResult);
|
|
||||||
} else {
|
|
||||||
const fetchPage = async (page: number) => {
|
|
||||||
return fireEngineMap(mapUrl, {
|
|
||||||
numResults: resultsPerPage,
|
|
||||||
page: page,
|
|
||||||
});
|
|
||||||
};
|
|
||||||
|
|
||||||
pagePromises = Array.from({ length: maxPages }, (_, i) => fetchPage(i + 1));
|
|
||||||
allResults = await Promise.all(pagePromises);
|
|
||||||
|
|
||||||
await redis.set(cacheKey, JSON.stringify(allResults), "EX", 24 * 60 * 60); // Cache for 24 hours
|
|
||||||
}
|
|
||||||
|
|
||||||
// Parallelize sitemap fetch with serper search
|
|
||||||
const [sitemap, ...searchResults] = await Promise.all([
|
|
||||||
req.body.ignoreSitemap ? null : crawler.tryGetSitemap(),
|
|
||||||
...(cachedResult ? [] : pagePromises),
|
|
||||||
]);
|
|
||||||
|
|
||||||
if (!cachedResult) {
|
|
||||||
allResults = searchResults;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (sitemap !== null) {
|
|
||||||
sitemap.forEach((x) => {
|
|
||||||
links.push(x.url);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
let mapResults = allResults
|
|
||||||
.flat()
|
|
||||||
.filter((result) => result !== null && result !== undefined);
|
|
||||||
|
|
||||||
const minumumCutoff = Math.min(MAX_MAP_LIMIT, limit);
|
|
||||||
if (mapResults.length > minumumCutoff) {
|
|
||||||
mapResults = mapResults.slice(0, minumumCutoff);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (mapResults.length > 0) {
|
|
||||||
if (req.body.search) {
|
|
||||||
// Ensure all map results are first, maintaining their order
|
|
||||||
links = [
|
|
||||||
mapResults[0].url,
|
|
||||||
...mapResults.slice(1).map((x) => x.url),
|
|
||||||
...links,
|
|
||||||
];
|
|
||||||
} else {
|
|
||||||
mapResults.map((x) => {
|
|
||||||
links.push(x.url);
|
links.push(x.url);
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
}
|
} else {
|
||||||
|
let urlWithoutWww = req.body.url.replace("www.", "");
|
||||||
|
|
||||||
|
let mapUrl = req.body.search
|
||||||
|
? `"${req.body.search}" site:${urlWithoutWww}`
|
||||||
|
: `site:${req.body.url}`;
|
||||||
|
|
||||||
|
const resultsPerPage = 100;
|
||||||
|
const maxPages = Math.ceil(
|
||||||
|
Math.min(MAX_FIRE_ENGINE_RESULTS, limit) / resultsPerPage
|
||||||
|
);
|
||||||
|
|
||||||
|
const cacheKey = `fireEngineMap:${mapUrl}`;
|
||||||
|
const cachedResult = null;
|
||||||
|
|
||||||
|
let allResults: any[] = [];
|
||||||
|
let pagePromises: Promise<any>[] = [];
|
||||||
|
|
||||||
|
if (cachedResult) {
|
||||||
|
allResults = JSON.parse(cachedResult);
|
||||||
|
} else {
|
||||||
|
const fetchPage = async (page: number) => {
|
||||||
|
return fireEngineMap(mapUrl, {
|
||||||
|
numResults: resultsPerPage,
|
||||||
|
page: page,
|
||||||
|
});
|
||||||
|
};
|
||||||
|
|
||||||
|
pagePromises = Array.from({ length: maxPages }, (_, i) =>
|
||||||
|
fetchPage(i + 1)
|
||||||
|
);
|
||||||
|
allResults = await Promise.all(pagePromises);
|
||||||
|
|
||||||
|
await redis.set(cacheKey, JSON.stringify(allResults), "EX", 24 * 60 * 60); // Cache for 24 hours
|
||||||
|
}
|
||||||
|
|
||||||
|
// Parallelize sitemap fetch with serper search
|
||||||
|
const [sitemap, ...searchResults] = await Promise.all([
|
||||||
|
req.body.ignoreSitemap ? null : crawler.tryGetSitemap(),
|
||||||
|
...(cachedResult ? [] : pagePromises),
|
||||||
|
]);
|
||||||
|
|
||||||
|
if (!cachedResult) {
|
||||||
|
allResults = searchResults;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (sitemap !== null) {
|
||||||
|
sitemap.forEach((x) => {
|
||||||
|
links.push(x.url);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
let mapResults = allResults
|
||||||
|
.flat()
|
||||||
|
.filter((result) => result !== null && result !== undefined);
|
||||||
|
|
||||||
|
const minumumCutoff = Math.min(MAX_MAP_LIMIT, limit);
|
||||||
|
if (mapResults.length > minumumCutoff) {
|
||||||
|
mapResults = mapResults.slice(0, minumumCutoff);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (mapResults.length > 0) {
|
||||||
|
if (req.body.search) {
|
||||||
|
// Ensure all map results are first, maintaining their order
|
||||||
|
links = [
|
||||||
|
mapResults[0].url,
|
||||||
|
...mapResults.slice(1).map((x) => x.url),
|
||||||
|
...links,
|
||||||
|
];
|
||||||
|
} else {
|
||||||
|
mapResults.map((x) => {
|
||||||
|
links.push(x.url);
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
}
|
||||||
// Perform cosine similarity between the search query and the list of links
|
// Perform cosine similarity between the search query and the list of links
|
||||||
if (req.body.search) {
|
if (req.body.search) {
|
||||||
const searchQuery = req.body.search.toLowerCase();
|
const searchQuery = req.body.search.toLowerCase();
|
||||||
|
|
|
@ -261,6 +261,7 @@ export const mapRequestSchema = crawlerOptions.extend({
|
||||||
includeSubdomains: z.boolean().default(true),
|
includeSubdomains: z.boolean().default(true),
|
||||||
search: z.string().optional(),
|
search: z.string().optional(),
|
||||||
ignoreSitemap: z.boolean().default(false),
|
ignoreSitemap: z.boolean().default(false),
|
||||||
|
sitemapOnly: z.boolean().default(false),
|
||||||
limit: z.number().min(1).max(5000).default(5000),
|
limit: z.number().min(1).max(5000).default(5000),
|
||||||
}).strict(strictMessage);
|
}).strict(strictMessage);
|
||||||
|
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"name": "@mendable/firecrawl-js",
|
"name": "@mendable/firecrawl-js",
|
||||||
"version": "1.8.3",
|
"version": "1.8.4",
|
||||||
"description": "JavaScript SDK for Firecrawl API",
|
"description": "JavaScript SDK for Firecrawl API",
|
||||||
"main": "dist/index.js",
|
"main": "dist/index.js",
|
||||||
"types": "dist/index.d.ts",
|
"types": "dist/index.d.ts",
|
||||||
|
|
|
@ -221,6 +221,7 @@ export interface MapParams {
|
||||||
search?: string;
|
search?: string;
|
||||||
ignoreSitemap?: boolean;
|
ignoreSitemap?: boolean;
|
||||||
includeSubdomains?: boolean;
|
includeSubdomains?: boolean;
|
||||||
|
sitemapOnly?: boolean;
|
||||||
limit?: number;
|
limit?: number;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue
Block a user