Skip to content

Commit

Permalink
Reformat everything using biome formatter
Browse files Browse the repository at this point in the history
  • Loading branch information
janbuchar committed Jan 24, 2024
1 parent 533f29b commit cf7b8dc
Show file tree
Hide file tree
Showing 234 changed files with 3,653 additions and 3,944 deletions.
140 changes: 76 additions & 64 deletions .eslintrc.json
Original file line number Diff line number Diff line change
@@ -1,66 +1,78 @@
{
"root": true,
"env": {
"browser": true,
"es2020": true,
"node": true
},
"extends": ["@apify/eslint-config-ts", "prettier"],
"parserOptions": {
"project": "./tsconfig.eslint.json",
"ecmaVersion": 2022
},
"ignorePatterns": [
"node_modules",
"dist",
"coverage",
"**/*.d.ts"
],
"overrides": [
{
"plugins": [
"@typescript-eslint"
],
"files": [
"*.ts"
],
"rules": {
"@typescript-eslint/array-type": "error",
"@typescript-eslint/ban-ts-comment": 0,
"@typescript-eslint/consistent-type-imports": ["error", {
"disallowTypeAnnotations": false
}],
"@typescript-eslint/consistent-type-definitions": ["error", "interface"],
"@typescript-eslint/member-delimiter-style": ["error", {
"multiline": { "delimiter": "semi", "requireLast": true },
"singleline": { "delimiter": "semi", "requireLast": false }
}],
"@typescript-eslint/no-empty-interface": "off",
"no-empty-function": "off",
"@typescript-eslint/no-empty-function": "off",
"@typescript-eslint/no-explicit-any": "off",
"@typescript-eslint/no-floating-promises": "error",
"@typescript-eslint/no-unused-vars": "off",
"@typescript-eslint/comma-dangle": ["error", "always-multiline"]
}
}
],
"rules": {
"quote-props": ["error", "consistent"],
"import/no-extraneous-dependencies": "off",
"max-classes-per-file": 0,
"no-console": "error",
"no-underscore-dangle": 0,
"no-void": 0,
"max-len": ["error", {
"code": 160,
"ignoreUrls": true,
"ignoreComments": true
}],
"import/order": ["error", {
"groups": ["builtin", "external", ["parent", "sibling"], "index", "object"],
"alphabetize": { "order": "asc", "caseInsensitive": true },
"newlines-between": "always"
}]
}
"root": true,
"env": {
"browser": true,
"es2020": true,
"node": true
},
"extends": ["@apify/eslint-config-ts", "prettier"],
"parserOptions": {
"project": "./tsconfig.eslint.json",
"ecmaVersion": 2022
},
"ignorePatterns": ["node_modules", "dist", "coverage", "**/*.d.ts"],
"overrides": [
{
"plugins": ["@typescript-eslint"],
"files": ["*.ts"],
"rules": {
"@typescript-eslint/array-type": "error",
"@typescript-eslint/ban-ts-comment": 0,
"@typescript-eslint/consistent-type-imports": [
"error",
{
"disallowTypeAnnotations": false
}
],
"@typescript-eslint/consistent-type-definitions": [
"error",
"interface"
],
"@typescript-eslint/member-delimiter-style": [
"error",
{
"multiline": { "delimiter": "semi", "requireLast": true },
"singleline": { "delimiter": "semi", "requireLast": false }
}
],
"@typescript-eslint/no-empty-interface": "off",
"no-empty-function": "off",
"@typescript-eslint/no-empty-function": "off",
"@typescript-eslint/no-explicit-any": "off",
"@typescript-eslint/no-floating-promises": "error",
"@typescript-eslint/no-unused-vars": "off",
"@typescript-eslint/comma-dangle": ["error", "always-multiline"]
}
}
],
"rules": {
"quote-props": ["error", "consistent"],
"import/no-extraneous-dependencies": "off",
"max-classes-per-file": 0,
"no-console": "error",
"no-underscore-dangle": 0,
"no-void": 0,
"max-len": [
"error",
{
"code": 160,
"ignoreUrls": true,
"ignoreComments": true
}
],
"import/order": [
"error",
{
"groups": [
"builtin",
"external",
["parent", "sibling"],
"index",
"object"
],
"alphabetize": { "order": "asc", "caseInsensitive": true },
"newlines-between": "always"
}
]
}
}
22 changes: 11 additions & 11 deletions docs/examples/.eslintrc.json
Original file line number Diff line number Diff line change
@@ -1,13 +1,13 @@
{
"root": true,
"extends": "@apify/ts",
"parserOptions": {
"project": "./tsconfig.eslint.json",
"ecmaVersion": 2022
},
"rules": {
"import/extensions": 0,
"import/no-extraneous-dependencies": 0,
"no-console": "off"
}
"root": true,
"extends": "@apify/ts",
"parserOptions": {
"project": "./tsconfig.eslint.json",
"ecmaVersion": 2022
},
"rules": {
"import/extensions": 0,
"import/no-extraneous-dependencies": 0,
"no-console": "off"
}
}
6 changes: 1 addition & 5 deletions docs/examples/add_data_to_dataset.ts
Original file line number Diff line number Diff line change
Expand Up @@ -11,11 +11,7 @@ const crawler = new CheerioCrawler({
},
});

await crawler.addRequests([
'http://www.example.com/page-1',
'http://www.example.com/page-2',
'http://www.example.com/page-3',
]);
await crawler.addRequests(['http://www.example.com/page-1', 'http://www.example.com/page-2', 'http://www.example.com/page-3']);

// Run the crawler
await crawler.run();
7 changes: 1 addition & 6 deletions docs/examples/basic_crawler.ts
Original file line number Diff line number Diff line change
Expand Up @@ -22,12 +22,7 @@ const crawler = new BasicCrawler({
});

// The initial list of URLs to crawl. Here we use just a few hard-coded URLs.
await crawler.addRequests([
'https://www.google.com',
'https://www.example.com',
'https://www.bing.com',
'https://www.wikipedia.com',
]);
await crawler.addRequests(['https://www.google.com', 'https://www.example.com', 'https://www.bing.com', 'https://www.wikipedia.com']);

// Run the crawler and wait for it to finish.
await crawler.run();
Expand Down
4 changes: 1 addition & 3 deletions docs/examples/cheerio_crawler.ts
Original file line number Diff line number Diff line change
Expand Up @@ -57,8 +57,6 @@ const crawler = new CheerioCrawler({
});

// Run the crawler and wait for it to finish.
await crawler.run([
'https://crawlee.dev',
]);
await crawler.run(['https://crawlee.dev']);

log.debug('Crawler finished.');
6 changes: 1 addition & 5 deletions docs/examples/crawl_multiple_urls_cheerio.ts
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,4 @@ const crawler = new CheerioCrawler({
});

// Run the crawler with initial request
await crawler.run([
'http://www.example.com/page-1',
'http://www.example.com/page-2',
'http://www.example.com/page-3',
]);
await crawler.run(['http://www.example.com/page-1', 'http://www.example.com/page-2', 'http://www.example.com/page-3']);
6 changes: 1 addition & 5 deletions docs/examples/crawl_multiple_urls_playwright.ts
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,4 @@ const crawler = new PlaywrightCrawler({
});

// Run the crawler with initial request
await crawler.run([
'http://www.example.com/page-1',
'http://www.example.com/page-2',
'http://www.example.com/page-3',
]);
await crawler.run(['http://www.example.com/page-1', 'http://www.example.com/page-2', 'http://www.example.com/page-3']);
6 changes: 1 addition & 5 deletions docs/examples/crawl_multiple_urls_puppeteer.ts
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,4 @@ const crawler = new PuppeteerCrawler({
});

// Run the crawler with initial request
await crawler.run([
'http://www.example.com/page-1',
'http://www.example.com/page-2',
'http://www.example.com/page-3',
]);
await crawler.run(['http://www.example.com/page-1', 'http://www.example.com/page-2', 'http://www.example.com/page-3']);
9 changes: 3 additions & 6 deletions docs/examples/forms.ts
Original file line number Diff line number Diff line change
Expand Up @@ -17,14 +17,11 @@ await page.select('select#search_language', 'JavaScript');

// Submit the form and wait for full load of next page
console.log('Submit search form');
await Promise.all([
page.waitForNavigation({ waitUntil: 'networkidle2' }),
page.click('#adv_code_search button[type="submit"]'),
]);
await Promise.all([page.waitForNavigation({ waitUntil: 'networkidle2' }), page.click('#adv_code_search button[type="submit"]')]);

// Obtain and print list of search results
const results = await page.$$eval('[data-testid="results-list"] div.search-title > a',
(nodes) => nodes.map((node) => ({
const results = await page.$$eval('[data-testid="results-list"] div.search-title > a', (nodes) =>
nodes.map((node) => ({
url: node.href,
name: node.innerText,
})),
Expand Down
4 changes: 1 addition & 3 deletions docs/examples/http_crawler.ts
Original file line number Diff line number Diff line change
Expand Up @@ -47,8 +47,6 @@ const crawler = new HttpCrawler({
});

// Run the crawler and wait for it to finish.
await crawler.run([
'https://crawlee.dev',
]);
await crawler.run(['https://crawlee.dev']);

log.debug('Crawler finished.');
4 changes: 1 addition & 3 deletions docs/examples/jsdom_crawler.ts
Original file line number Diff line number Diff line change
Expand Up @@ -57,8 +57,6 @@ const crawler = new JSDOMCrawler({
});

// Run the crawler and wait for it to finish.
await crawler.run([
'https://crawlee.dev',
]);
await crawler.run(['https://crawlee.dev']);

log.debug('Crawler finished.');
12 changes: 5 additions & 7 deletions docs/examples/jsdom_crawler_react.ts
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
import { JSDOMCrawler, log } from 'crawlee';

// Create an instance of the JSDOMCrawler class - crawler that automatically
// Create an instance of the JSDOMCrawler class - crawler that automatically
// loads the URLs and parses their HTML using the jsdom library.
const crawler = new JSDOMCrawler({
// Setting the `runScripts` option to `true` allows the crawler to execute client-side
// Setting the `runScripts` option to `true` allows the crawler to execute client-side
// JavaScript code on the page. This is required for some websites (such as the React application in this example), but may pose a security risk.
runScripts: true,
// This function will be called for each crawled URL.
Expand All @@ -18,15 +18,13 @@ const crawler = new JSDOMCrawler({
document.querySelectorAll('button')[18].click(); // =

const result = document.querySelectorAll('.component-display')[0].childNodes[0] as Element;
// The result is passed to the console. Unlike with Playwright or Puppeteer crawlers,
// The result is passed to the console. Unlike with Playwright or Puppeteer crawlers,
// this console call goes to the Node.js console, not the browser console. All the code here runs right in Node.js!
log.info(result.innerHTML); // 2
},
});

// Run the crawler and wait for it to finish.
await crawler.run([
'https://ahfarmer.github.io/calculator/',
]);
await crawler.run(['https://ahfarmer.github.io/calculator/']);

log.debug('Crawler finished.');
log.debug('Crawler finished.');
4 changes: 2 additions & 2 deletions docs/examples/map.ts
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
import { Dataset, KeyValueStore } from 'crawlee';

const dataset = await Dataset.open<{
url: string,
headingCount: number,
url: string;
headingCount: number;
}>();

// Seeding the dataset with some data
Expand Down
6 changes: 1 addition & 5 deletions docs/examples/puppeteer_crawler_crawler_utils_snapshot.ts
Original file line number Diff line number Diff line change
Expand Up @@ -10,11 +10,7 @@ const crawler = new PuppeteerCrawler({
},
});

await crawler.addRequests([
{ url: 'http://www.example.com/page-1' },
{ url: 'http://www.example.com/page-2' },
{ url: 'http://www.example.com/page-3' },
]);
await crawler.addRequests([{ url: 'http://www.example.com/page-1' }, { url: 'http://www.example.com/page-2' }, { url: 'http://www.example.com/page-3' }]);

// Run the crawler
await crawler.run();
6 changes: 1 addition & 5 deletions docs/examples/puppeteer_crawler_page_screenshot.ts
Original file line number Diff line number Diff line change
Expand Up @@ -12,11 +12,7 @@ const crawler = new PuppeteerCrawler({
},
});

await crawler.addRequests([
{ url: 'http://www.example.com/page-1' },
{ url: 'http://www.example.com/page-2' },
{ url: 'http://www.example.com/page-3' },
]);
await crawler.addRequests([{ url: 'http://www.example.com/page-1' }, { url: 'http://www.example.com/page-2' }, { url: 'http://www.example.com/page-3' }]);

// Run the crawler
await crawler.run();
4 changes: 2 additions & 2 deletions docs/examples/reduce.ts
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
import { Dataset, KeyValueStore } from 'crawlee';

const dataset = await Dataset.open<{
url: string,
headingCount: number,
url: string;
headingCount: number;
}>();

// Seeding the dataset with some data
Expand Down
16 changes: 7 additions & 9 deletions docs/guides/avoid_blocking_playwright.ts
Original file line number Diff line number Diff line change
Expand Up @@ -6,16 +6,14 @@ const crawler = new PlaywrightCrawler({
useFingerprints: true, // this is the default
fingerprintOptions: {
fingerprintGeneratorOptions: {
browsers: [{
name: BrowserName.edge,
minVersion: 96,
}],
devices: [
DeviceCategory.desktop,
],
operatingSystems: [
OperatingSystemsName.windows,
browsers: [
{
name: BrowserName.edge,
minVersion: 96,
},
],
devices: [DeviceCategory.desktop],
operatingSystems: [OperatingSystemsName.windows],
},
},
},
Expand Down
13 changes: 3 additions & 10 deletions docs/guides/avoid_blocking_puppeteer.ts
Original file line number Diff line number Diff line change
Expand Up @@ -6,16 +6,9 @@ const crawler = new PuppeteerCrawler({
useFingerprints: true, // this is the default
fingerprintOptions: {
fingerprintGeneratorOptions: {
browsers: [
BrowserName.chrome,
BrowserName.firefox,
],
devices: [
DeviceCategory.mobile,
],
locales: [
'en-US',
],
browsers: [BrowserName.chrome, BrowserName.firefox],
devices: [DeviceCategory.mobile],
locales: ['en-US'],
},
},
},
Expand Down

0 comments on commit cf7b8dc

Please sign in to comment.