mirror of
https://github.com/NikolaiT/se-scraper.git
synced 2024-11-26 01:23:16 +01:00
82 lines
2.4 KiB
JavaScript
82 lines
2.4 KiB
JavaScript
const handler = require('./src/node_scraper.js');
|
|
var fs = require('fs');
|
|
var os = require("os");
|
|
|
|
exports.scrape = async function(config, callback) {
|
|
|
|
// options for scraping
|
|
event = {
|
|
// the user agent to scrape with
|
|
user_agent: 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.110 Safari/537.36',
|
|
// if random_user_agent is set to True, a random user agent is chosen
|
|
random_user_agent: true,
|
|
// whether to select manual settings in visible mode
|
|
set_manual_settings: false,
|
|
// log ip address data
|
|
log_ip_address: false,
|
|
// log http headers
|
|
log_http_headers: false,
|
|
// how long to sleep between requests. a random sleep interval within the range [a,b]
|
|
// is drawn before every request. empty string for no sleeping.
|
|
sleep_range: '[1,1]',
|
|
// which search engine to scrape
|
|
search_engine: 'google',
|
|
compress: false, // compress
|
|
debug: false,
|
|
verbose: false,
|
|
keywords: ['scrapeulous.com'],
|
|
// whether to start the browser in headless mode
|
|
headless: true,
|
|
// the number of pages to scrape for each keyword
|
|
num_pages: 1,
|
|
// path to output file, data will be stored in JSON
|
|
output_file: '',
|
|
// whether to prevent images, css, fonts and media from being loaded
|
|
// will speed up scraping a great deal
|
|
block_assets: true,
|
|
// path to js module that extends functionality
|
|
// this module should export the functions:
|
|
// get_browser, handle_metadata, close_browser
|
|
//custom_func: resolve('examples/pluggable.js'),
|
|
custom_func: '',
|
|
// use a proxy for all connections
|
|
// example: 'socks5://78.94.172.42:1080'
|
|
// example: 'http://118.174.233.10:48400'
|
|
proxy: '',
|
|
// check if headless chrome escapes common detection techniques
|
|
// this is a quick test and should be used for debugging
|
|
test_evasion: false,
|
|
};
|
|
|
|
// overwrite default config
|
|
for (var key in config) {
|
|
event[key] = config[key];
|
|
}
|
|
|
|
if (fs.existsSync(event.keyword_file)) {
|
|
event.keywords = read_keywords_from_file(event.keyword_file);
|
|
}
|
|
|
|
if (!callback) {
|
|
// called when results are ready
|
|
callback = function (err, response) {
|
|
if (err) {
|
|
console.error(err)
|
|
}
|
|
|
|
console.dir(response.results, {depth: null, colors: true});
|
|
}
|
|
}
|
|
|
|
await handler.handler(event, undefined, callback );
|
|
};
|
|
|
|
function read_keywords_from_file(fname) {
|
|
let kws = fs.readFileSync(fname).toString().split(os.EOL);
|
|
// clean keywords
|
|
kws = kws.filter((kw) => {
|
|
return kw.trim().length > 0;
|
|
});
|
|
return kws;
|
|
}
|