mirror of
https://github.com/mendableai/firecrawl.git
synced 2024-11-16 11:42:24 +08:00
Update unit tests
This commit is contained in:
parent
f0d4146b42
commit
519ab1aecb
|
@ -655,7 +655,7 @@ describe("E2E Tests for API Routes", () => {
|
|||
const testurls = completedResponse.body.data.map(
|
||||
(item: any) => item.metadata?.sourceURL
|
||||
);
|
||||
console.log(testurls)
|
||||
//console.log(testurls)
|
||||
|
||||
expect(completedResponse.statusCode).toBe(200);
|
||||
expect(completedResponse.body).toHaveProperty("status");
|
||||
|
|
|
@ -73,7 +73,7 @@ describe('WebCrawler maxDepth and filterLinks', () => {
|
|||
const initialUrl = 'http://example.com'; // Set initial URL for this test
|
||||
const enteredMaxCrawledDepth = 0;
|
||||
maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth);
|
||||
console.log(maxCrawledDepth);
|
||||
|
||||
|
||||
crawler = new WebCrawler({
|
||||
initialUrl: initialUrl,
|
||||
|
@ -101,7 +101,7 @@ describe('WebCrawler maxDepth and filterLinks', () => {
|
|||
const initialUrl = 'http://example.com/page1'; // Set initial URL for this test
|
||||
const enteredMaxCrawledDepth = 1;
|
||||
maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth);
|
||||
console.log(maxCrawledDepth);
|
||||
|
||||
|
||||
crawler = new WebCrawler({
|
||||
initialUrl: initialUrl,
|
||||
|
@ -130,7 +130,7 @@ describe('WebCrawler maxDepth and filterLinks', () => {
|
|||
const initialUrl = 'http://example.com/page1'; // Set initial URL for this test
|
||||
const enteredMaxCrawledDepth = 2;
|
||||
maxCrawledDepth = getAdjustedMaxDepth(initialUrl, enteredMaxCrawledDepth);
|
||||
console.log(maxCrawledDepth);
|
||||
|
||||
|
||||
crawler = new WebCrawler({
|
||||
initialUrl: initialUrl,
|
||||
|
|
|
@ -22,15 +22,25 @@ describe('Testing getURLDepth and getAdjustedMaxDepth', () => {
|
|||
|
||||
});
|
||||
|
||||
it('Adjusted maxDepth should return 1 for scrapethissite.com', () => {
|
||||
it('Adjusted maxDepth should return 1 for scrapethissite.com and max depth param of 1', () => {
|
||||
const enteredURL = "https://scrapethissite.com"
|
||||
expect(getAdjustedMaxDepth(enteredURL, 1)).toBe(1);
|
||||
|
||||
});
|
||||
it('Adjusted maxDepth should return 0 for scrapethissite.com and max depth param of 0', () => {
|
||||
const enteredURL = "https://scrapethissite.com"
|
||||
expect(getAdjustedMaxDepth(enteredURL, 0)).toBe(0);
|
||||
|
||||
it('Adjusted maxDepth should return 5 for scrapethissite.com/pages/articles', () => {
|
||||
});
|
||||
|
||||
it('Adjusted maxDepth should return 0 for mendable.ai and max depth param of 0', () => {
|
||||
const enteredURL = "https://mendable.ai"
|
||||
expect(getAdjustedMaxDepth(enteredURL, 0)).toBe(0);
|
||||
});
|
||||
|
||||
it('Adjusted maxDepth should return 4 for scrapethissite.com/pages/articles and max depth param of 2', () => {
|
||||
const enteredURL = "https://scrapethissite.com/pages/articles"
|
||||
expect(getAdjustedMaxDepth(enteredURL, 2)).toBe(5);
|
||||
expect(getAdjustedMaxDepth(enteredURL, 2)).toBe(4);
|
||||
});
|
||||
|
||||
|
||||
|
|
Loading…
Reference in New Issue
Block a user