{"_id":"crawler","_rev":"172-9205257b5c3b77245e2c159ef662644e","name":"crawler","dist-tags":{"latest":"2.0.2","beta":"2.0.3-beta.3"},"versions":{"0.0.1":{"name":"crawler","version":"0.0.1","keywords":["dom","javascript","crawling","jquery"],"_id":"crawler@0.0.1","maintainers":[{"web":"http://sylvinus.org/","name":"Sylvain Zimmer","email":"sylvain@sylvainzimmer.com"}],"bugs":{"web":"http://github.com/joshfire/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"os":["linux","macos","win"],"cpu":["x86","ppc","x86_64"],"dist":{"shasum":"c22bedc7e72060fe03e71d8515b83dcf9a2fed4e","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.0.1.tgz","integrity":"sha512-tC8cizYwKGhqClkQ7VnV54smEpbIR2MgpgX5gyS7qrK12OodrrUnVwORVRs6c7ChXbaVqv6MOqS+7vQeniB/VA==","signatures":[{"sig":"MEQCICHsYLtPrbPe/nHiJFNCfra9kzNzXWZdbNN2J7Pjr4K+AiBG3psEj8a7yFyFoKZBjSI5jbRQ2ljVxYlaypcVIGSLEw==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node"],"licenses":[{"url":"http://github.com/joshfire/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"_npmVersion":"0.2.8-1","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously.","directories":{"lib":"lib"},"_nodeVersion":"v0.2.4","dependencies":{"jsdom":">= 0.1.20","qunit":">= 0.0.7","request":">= 0.9.5","htmlparser":">= 1.6.2","generic-pool":">= 1.0.2"},"repositories":[{"url":"http://github.com/joshfire/node-crawler.git","type":"git"}],"_nodeSupported":true},"0.0.2":{"name":"crawler","version":"0.0.2","keywords":["dom","javascript","crawling","jquery"],"_id":"crawler@0.0.2","maintainers":[{"web":"http://sylvinus.org/","name":"Sylvain Zimmer","email":"sylvain@sylvainzimmer.com"}],"bugs":{"web":"http://github.com/joshfire/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"os":["linux","macos","win"],"cpu":["x86","ppc","x86_64"],"dist":{"shasum":"74e927fa4eaa99c9c0b14a0b7a09676576b5d42b","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.0.2.tgz","integrity":"sha512-jLDf3PsiUOxOSsge+63CtS5goH80SuQrBekgE678leUOTTYsZQACTpwgVVSGz6Urfi55poqxYWo9+bFrRSJRYQ==","signatures":[{"sig":"MEUCIHPzW1LiXBl6YQrFt2hHY+U3yAwzJLDLtHkksWV/dNnrAiEAvPEsEYu7e2tCZOq0hAfmIW02oQjhmKxfKflOY9HfumA=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node"],"licenses":[{"url":"http://github.com/joshfire/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"http://github.com/joshfire/node-crawler.git","type":"git"},"_npmVersion":"0.2.8-1","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously.","directories":{"lib":"lib"},"_nodeVersion":"v0.2.4","dependencies":{"jsdom":">= 0.1.20","qunit":">= 0.0.7","request":">= 0.9.5","htmlparser":">= 1.6.2","generic-pool":">= 1.0.2"},"_nodeSupported":true},"0.2.0":{"name":"crawler","version":"0.2.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery"],"_id":"crawler@0.2.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"bugs":{"url":"http://github.com/sylvinus/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"dist":{"shasum":"4b1a8085648bf9ed71ee52dae6471ddf0ca70171","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.0.tgz","integrity":"sha512-o6pFM78GWiJwUHvU8YwbKazxCpB5Kzde3Eq7ybB7g2/XFLo58kTqHXKOcVSSGspZv4tdGqUrF3JhXZ2xGnjRjg==","signatures":[{"sig":"MEQCIFlpjmkZ76UuAvCLHLSCFMfIwBe93Nl+hHth20A6nRJlAiBV2DG4ZGJatnFBfHfkjdS67wnSDN45vXhywjDPqiXqoQ==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node >=0.6.x"],"scripts":{"test":"node test/testrunner.js"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously.","directories":{"lib":"lib"},"dependencies":{"iconv":"1.2.3","jsdom":"0.2.15","request":"2.11.1","jschardet":"1.0.2","htmlparser":"1.7.6","underscore":"1.3.3","generic-pool":"2.0.1"},"devDependencies":{"qunit":"0.5.8","express":"2.5.x"}},"0.2.1":{"name":"crawler","version":"0.2.1","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery"],"_id":"crawler@0.2.1","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"bugs":{"url":"http://github.com/sylvinus/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"dist":{"shasum":"531d6c1832920010459f4060c7dfe467ae1a76be","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.1.tgz","integrity":"sha512-VRkV+UZDbJXPPLrtW5z0G/40EgXH8D9lxs+Y1W03BAZYIWZ2Ypm4fuMOAoSRT7RPpvnnjf+ZKS0dDJiJfmafCw==","signatures":[{"sig":"MEYCIQCu2Rbg7m2xB7UWzMqIo1qXjHHHtYR1Y7VPQWRV4wh0QwIhAI5Y3MBEVkCIR3o5sMr2SBkybY0NpjwP5ZFLFAh8yGWJ","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node >=0.6.x"],"scripts":{"test":"node test/testrunner.js"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously.","directories":{"lib":"lib"},"dependencies":{"iconv":"1.2.3","jsdom":"0.2.19","request":"2.11.1","jschardet":"1.0.2","htmlparser":"1.7.6","underscore":"1.3.3","generic-pool":"2.0.1"},"devDependencies":{"qunit":"0.5.8","express":"2.5.x"}},"0.2.2":{"name":"crawler","version":"0.2.2","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery"],"_id":"crawler@0.2.2","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"bugs":{"url":"http://github.com/sylvinus/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"dist":{"shasum":"98baca8ee5cd171b56f4e7ee1e657d86cd4dbba0","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.2.tgz","integrity":"sha512-mob/W0JunftMz59zWYR/la11JgvW+3B1OhDEqbD/QCbDsgvxyV32rlDEYuMeXLvxkNbkfimDORjmyRzg6hiPyw==","signatures":[{"sig":"MEUCIQDUc+XxlpKGJuyq4rzlCfy1GbMIUU7d/WLptDczQ+PzlgIgFYMaxSNdnG+M21z0tF7Elx5ECEI5kYw2hw2yGnyHupk=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node >=0.6.x"],"scripts":{"test":"node test/testrunner.js"},"_npmUser":{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.1.65","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously.","directories":{"lib":"lib"},"dependencies":{"iconv":"1.2.3","jsdom":"0.2.19","request":"2.12.0","jschardet":"1.0.2","htmlparser":"1.7.6","underscore":"1.3.3","generic-pool":"2.0.2"},"devDependencies":{"qunit":"0.5.8","express":"2.5.x","memwatch":"0.1.5"}},"0.2.3":{"name":"crawler","version":"0.2.3","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery"],"_id":"crawler@0.2.3","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"bugs":{"url":"http://github.com/sylvinus/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"dist":{"shasum":"f37a3be9d40085dd0cb672c9f275ba28abc075a6","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.3.tgz","integrity":"sha512-mU49IHbOvGtPfhdjCYLdixAplPY01V1MqXHukhcQYGrVQeG3jP6GT26qJsGrTI/6a8yDw7prkNNI1bzZwPjTqA==","signatures":[{"sig":"MEUCIDEUoLoQQAf7CWWyW2gE4YAlry4EoFNudQAX7XJOq/xNAiEA1WEaT0X8qxa7k7DhF/+Ln1YDCYSi1z3DxxWAD28gimY=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node >=0.6.x"],"scripts":{"test":"node test/testrunner.js"},"_npmUser":{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.1.65","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"1.2.3","jsdom":"0.2.19","request":"2.12.0","jschardet":"1.0.2","htmlparser":"1.7.6","iconv-lite":"0.2.7","underscore":"1.3.3","generic-pool":"2.0.2"},"devDependencies":{"qunit":"0.5.8","express":"2.5.x","memwatch":"0.1.5"},"optionalDependencies":{"iconv":"1.2.3"}},"0.2.4":{"name":"crawler","version":"0.2.4","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery"],"_id":"crawler@0.2.4","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"bugs":{"url":"http://github.com/sylvinus/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"dist":{"shasum":"b138c893c00f76d9aea90d50663b709d25736b75","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.4.tgz","integrity":"sha512-gO+mfOhdBypnB3m89GL9D3ePKALcOmlz8YkOll2h5WSOjtPzhbpzyGQPlXu4ZgbLAmymxMzrBGNTxW1G9w7oSw==","signatures":[{"sig":"MEUCIQDzhVyYZ0vj/28Cf6LIr3JySWyOQSWQWA7WIgv4a+06vAIgBFXRzpwuxduRV/OCRcLs4pdOsPFiXv/FzgfVpXeAEzU=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","engines":["node >=0.6.x"],"scripts":{"test":"node test/testrunner.js"},"_npmUser":{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.1.65","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"1.2.3","jsdom":"0.2.19","request":"2.12.0","jschardet":"1.0.2","htmlparser":"1.7.6","iconv-lite":"0.2.7","underscore":"1.3.3","generic-pool":"2.0.2"},"devDependencies":{"qunit":"0.5.8","express":"2.5.x","memwatch":"0.1.5"},"optionalDependencies":{"iconv":"1.2.3"}},"0.2.5":{"name":"crawler","version":"0.2.5","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery"],"_id":"crawler@0.2.5","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"bugs":{"url":"http://github.com/sylvinus/node-crawler/issues","mail":"sylvain@sylvainzimmer.com"},"dist":{"shasum":"9339d381c74eeb12aa7104af0dcd4543ec3f796b","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.5.tgz","integrity":"sha512-kWVD8CDfW18kD7b1RWozrJZFiQhO27j5njRMrlYRvYbNpps/dj26yNsFvHQumPz8TU2b/9kZ0fDhisiQv4Y1vQ==","signatures":[{"sig":"MEUCIGb/nrmO+iE0OotdvcO+YzHZf8UZGTO+8FBSJHjOU/6OAiEA0yFaxI+9RtuPMlpnkWHYaGBTvOceTW5DN1OMkG8UhNk=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","engines":["node >=0.6.x"],"scripts":{"test":"node test/testrunner.js"},"_npmUser":{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.2.18","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"2.0.6","jsdom":"0.8.2","request":"2.21.0","jschardet":"1.0.2","htmlparser":"1.7.6","iconv-lite":"0.2.8","underscore":"1.4.4","generic-pool":"2.0.3"},"devDependencies":{"qunit":"0.5.16","express":"2.5.x","memwatch":"0.2.2"},"optionalDependencies":{"iconv":"2.0.6"}},"0.2.6":{"name":"crawler","version":"0.2.6","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.2.6","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"a93bf5d8e7967d265a43cff626180f069f342d48","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.6.tgz","integrity":"sha512-+YyfsIhReMIuNH2Ir3wYph4x5EkgU+R0Bo4Llg45NigDzduGNIoxdzPpdNhuG4NV2XtN0AP7ZwNzjnZj7NfqqQ==","signatures":[{"sig":"MEYCIQCzC0FsYiP88wsm6SGj1QnOocjjELrgVMzqOh0jMzp0wAIhAKvCwi8V58m3GGmVhEaUUXtAi0VqpdGdPFiKMG9T32xO","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"a93bf5d8e7967d265a43cff626180f069f342d48","engines":["node >=0.10.x"],"gitHead":"02ae7421264ae719843623efd68e44048a619a00","scripts":{"test":"node test/testrunner.js"},"_npmUser":{"name":"paulvalla","email":"paul.valla+npm@gmail.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.4.14","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"*","jsdom":"^0.10.5","qunit":"^0.6.3","request":"^2.34.0","jschardet":"^1.1.0","htmlparser":"^1.7.7","iconv-lite":"^0.4.0-pre3","underscore":"^1.6.0","compression":"^1.0.2","generic-pool":"^2.0.4"},"devDependencies":{"qunit":"^0.6.4","express":"^4.2.0","memwatch":"^0.2.2"},"optionalDependencies":{"iconv":"*"}},"0.2.7":{"name":"crawler","version":"0.2.7","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.2.7","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"adefa23fe061aecb8de0d1868ae259f9c98606d3","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.2.7.tgz","integrity":"sha512-DcPct5wyrnLJ7A4Pg4Gqiyv1nvVuJ37SsyUF6XoW4tuBfDN4r4RSAEIM8NB/FsBCjXtDH8ULz7O2lFBqbEgC9g==","signatures":[{"sig":"MEUCIQCORHlzVYt9AQZ0cY6uZtAK53q93BKfwGfzR0TkSvvnhgIgcYoKG+ji3xcfrrn5WpGixjNgEW65a1v4vsRArSf8deo=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","engines":["node >=0.10.x"],"scripts":{"test":"node test/testrunner.js"},"_npmUser":{"name":"paulvalla","email":"paul.valla+npm@gmail.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.4.3","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"*","jsdom":"^0.10.5","qunit":"^0.6.3","request":"^2.34.0","jschardet":"^1.1.0","htmlparser":"^1.7.7","iconv-lite":"^0.4.0-pre3","underscore":"^1.6.0","compression":"^1.0.2","generic-pool":"^2.0.4"},"devDependencies":{"qunit":"^0.6.4","express":"^4.2.0","memwatch":"^0.2.2"},"optionalDependencies":{"iconv":"*"}},"0.3.0":{"name":"crawler","version":"0.3.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.3.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"a9d295d56fab4c1675d3602971e71d92b59e12ad","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.3.0.tgz","integrity":"sha512-v98CDNkKJ8MOihpsBBri5D0pS76z6W7q7Trl/k0UxV0NzN5fptql5tP9+vaqINWR63cDCKcrBoJV05q/cZrXDQ==","signatures":[{"sig":"MEUCIQCbXPGmHtQ2vpcQ4SuVd/7fQ5/UO+DIZHKVgciR0TAawgIgNiMrwKvskB9Qy8QMN3ffeneM9Uui/mYQSDRJMaqPQuw=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"a9d295d56fab4c1675d3602971e71d92b59e12ad","engines":["node >=0.8.x"],"gitHead":"e8ad75ebee1258de5f0a1db10e909aa04fa42261","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 test/units/*.js"},"_npmUser":{"name":"paulvalla","email":"paul.valla+npm@gmail.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"2.1.3","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"_nodeVersion":"0.10.26","dependencies":{"iconv":"*","jsdom":"0.11.1","lodash":"2.4.1","request":"2.42.0","jschardet":"1.1.0","iconv-lite":"0.4.4","generic-pool":"2.1.1"},"devDependencies":{"chai":"1.9.1","mocha":"1.21.4"},"optionalDependencies":{"iconv":"*"}},"0.3.1":{"name":"crawler","version":"0.3.1","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.3.1","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"1dae1a7358e7bd9d824e1fc175aece20cab9d3ed","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.3.1.tgz","integrity":"sha512-PSg+7nzOMkfoOZb7ou71fuIhU4Ph3qRpLmw4PBT3rkFBCev87h6fTHxTHWNZ542XzX37rrKVH/R1zByGiZNWUw==","signatures":[{"sig":"MEQCIDpy8BbUlW4BcRi2HrRE2AyXFOYwGJSxPMmrc7VO1dHeAiBI0ZwlodHaACIZKEY0t3eJ5qZ3mLepz+vDzJRrySmKrA==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"1dae1a7358e7bd9d824e1fc175aece20cab9d3ed","engines":["node >=0.8.x"],"gitHead":"de997e489bedabdc362d307959f72f0ac71467c0","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 test/units/*.js"},"_npmUser":{"name":"paulvalla","email":"paul.valla+npm@gmail.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"2.1.3","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"_nodeVersion":"0.10.26","dependencies":{"iconv":"*","jsdom":"0.11.1","lodash":"2.4.1","request":"2.42.0","jschardet":"1.1.0","iconv-lite":"0.4.4","generic-pool":"2.1.1"},"devDependencies":{"chai":"1.9.1","mocha":"1.21.4"},"optionalDependencies":{"iconv":"*"}},"0.4.0":{"name":"crawler","version":"0.4.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.4.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"2525aeef3f841ecf20a3ed3c0d19a8e0bdff91a1","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.4.0.tgz","integrity":"sha512-pJlyXKRsiV5Dc7lRE4BqCj6wpwRlFhx3ic8pk9Y2nr7QoiYkB2bqdt0DJT7WFy86w9i+636Aozfc/JSjed5ImQ==","signatures":[{"sig":"MEUCIQDJ7kz85t8hMjmBVZKZja5SKltJS5/AH9XYp3lEDLSdUAIgZ5qthi8rJhvisgmYCw8wwIy6CrU1GrhFUeBRX6AdI2c=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"2525aeef3f841ecf20a3ed3c0d19a8e0bdff91a1","engines":["node >=0.8.x"],"gitHead":"2daf980381ff46dd18b215e83f4c8aab7e9f4696","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"paulvalla","email":"paul.valla+npm@gmail.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.4.28","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"*","lodash":"2.4.1","cheerio":"0.18.0","request":"2.42.0","jschardet":"1.1.0","iconv-lite":"0.4.4","generic-pool":"2.1.1"},"devDependencies":{"chai":"1.9.2","jsdom":"0.11.1","mocha":"1.21.4","sinon":"1.11.1","mocha-testdata":"1.1.0"},"optionalDependencies":{"iconv":"*"}},"0.4.1":{"name":"crawler","version":"0.4.1","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.4.1","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"51c8c878ca4674aaa1ceb38accaca4241f6d2623","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.4.1.tgz","integrity":"sha512-kFx5hHc6RvIVrdSO8N/qc2Bp6Xrgmw+yfT0G980gWAGbeDJtwKczWu2zF+g0gF9Wx5gjZFhgV8XEgwsx0hVkcA==","signatures":[{"sig":"MEYCIQCQmKuGY/svyp2rYvnLRKCsL7UpKp1JP8raQjvOwkx6LAIhAIlC/vwY8W9EGpmzsNRRtrE6Iwo5t8E5Yl8KEKaT0aDl","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"51c8c878ca4674aaa1ceb38accaca4241f6d2623","engines":["node >=0.8.x"],"gitHead":"c1909ce58472e7f9d85098c21c62fca0515619da","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"paulvalla","email":"paul.valla+npm@gmail.com"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"1.4.28","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"dependencies":{"iconv":"*","lodash":"2.4.1","cheerio":"0.18.0","request":"2.42.0","jschardet":"1.1.0","iconv-lite":"0.4.4","generic-pool":"2.1.1"},"devDependencies":{"chai":"1.9.2","jsdom":"0.11.1","mocha":"1.21.4","sinon":"1.11.1","mocha-testdata":"1.1.0"},"optionalDependencies":{"iconv":"*"}},"0.4.2":{"name":"crawler","version":"0.4.2","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.4.2","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"8247481f6952667f16121e1c8aeefd32730e6e11","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.4.2.tgz","integrity":"sha512-Lv+mgy/CXawjZWlzqX6xuX8YPXJAHTdSnUecLhMcktxdYKQAYVJCbp6nNw7fARfk6O1I9ag6E+X30DS63D1MyQ==","signatures":[{"sig":"MEUCIGUKCvtucZ9p6XwTJPmvBPWILPF6eAG/zguTBfdV+m2oAiEAhTnqhoivW8BncSXCOCNdpzMVj69UgVsCsrHQiUnXS6o=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"8247481f6952667f16121e1c8aeefd32730e6e11","engines":["node >=0.8.x"],"gitHead":"ee41cccdcc573e6739fcbbffe811d13be0353211","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"paulvalla","email":"bonjour@pol.ninja"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"2.5.1","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"_nodeVersion":"0.12.0","dependencies":{"iconv":"*","lodash":"2.4.1","cheerio":"0.18.0","request":"2.42.0","jschardet":"1.1.0","iconv-lite":"0.4.4","generic-pool":"2.1.1"},"devDependencies":{"chai":"1.9.2","jsdom":"3.1.1","mocha":"1.21.4","sinon":"1.11.1","mocha-testdata":"1.1.0"},"optionalDependencies":{"iconv":"*"}},"0.4.3":{"name":"crawler","version":"0.4.3","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler"],"_id":"crawler@0.4.3","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"paul.valla+npm@gmail.com"}],"homepage":"https://github.com/sylvinus/node-crawler","bugs":{"url":"http://github.com/sylvinus/node-crawler/issues"},"dist":{"shasum":"15a7f65b68ab2e55613254490c94d7994a4e2c59","tarball":"https://registry.npmjs.org/crawler/-/crawler-0.4.3.tgz","integrity":"sha512-teONsAfE8Yz/ejnvgrg5UHyIhS5Iy8eA9heNH4ZeZfJvjjWScKuMeHOd05x/j1UkKnUcphrwZuF60lFe6fgRqg==","signatures":[{"sig":"MEQCIHMJJzq45CQ4OZ1Gb/LTrv3SdEujUaHw5Ok+6HjXe/O2AiA3cQxo7B0enC1AaOBpXti5kse4SUd3Hqsiblwt41SPHg==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler","_from":".","_shasum":"15a7f65b68ab2e55613254490c94d7994a4e2c59","engines":["node >=0.8.x"],"gitHead":"587f31a44c5e172a7637923b661dbcc8394b73d7","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"paulvalla","email":"bonjour@pol.ninja"},"licenses":[{"url":"http://github.com/sylvinus/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"https://github.com/sylvinus/node-crawler.git","type":"git"},"_npmVersion":"2.5.1","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously. Scraping should be simple and fun!","directories":{"lib":"lib"},"_nodeVersion":"0.12.0","dependencies":{"iconv":"*","lodash":"2.4.1","cheerio":"0.18.0","request":"2.42.0","jschardet":"1.1.0","iconv-lite":"0.4.4","generic-pool":"2.1.1"},"devDependencies":{"chai":"1.9.2","jsdom":"3.1.1","mocha":"2.2.1","sinon":"1.11.1","mocha-testdata":"1.1.0"},"optionalDependencies":{"iconv":"*"}},"1.0.0":{"name":"crawler","version":"1.0.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.0.0","maintainers":[{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"fb0a0d4643e048d8013a78efafdfb2a8d5264858","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.0.0.tgz","integrity":"sha512-57jsgKJhez0BkgaRXYr7CrhIxAB/l44w5e6lnleFjzhsGMKrzNg0W2+xIa+j44RvNiDt/3AIcWzH9hWpDRuQ9w==","signatures":[{"sig":"MEUCIEQ7DmgbEY3w0Ef3zPcvWny7ElcIxEhfOpaPRi2rv6WhAiEA7dTBNLYoS6Xw3R/D3xELdTocjAjDf0ZDzpYWkLN2nPE=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"fb0a0d4643e048d8013a78efafdfb2a8d5264858","gitHead":"8dd4bc3638fa09ba57915ac75014be18f68b673f","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.8","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.9.1","dependencies":{"lodash":"^3.8.0","cheerio":"^0.22.0","request":"~2.67.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.0","charset-parser":"^0.2.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.0.0.tgz_1480589443446_0.31904673716053367","host":"packages-12-west.internal.npmjs.com"}},"1.0.1":{"name":"crawler","version":"1.0.1","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.0.1","maintainers":[{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"738a756159c8f5a1f1f5fbd8749a1d427427cf2b","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.0.1.tgz","integrity":"sha512-IR1lTdN76GAsHqNptXONoc49R+4ZTlDsjsHLPgrcabJGdZChcvm8V90vFWMhYfFQIravLFO/fgkp42fegOfMbg==","signatures":[{"sig":"MEUCIGn+oM1H2jphwgq6xizBt7II2w+/iEI9m3WaGPny+GMgAiEA5FAmY/1HuEZZd3GFUeEaHYcJKYZk9UB4Wh6f6GhCkPo=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"738a756159c8f5a1f1f5fbd8749a1d427427cf2b","gitHead":"90c8ebe9784d799215cca4da6aad4f3fc854479b","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.8","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.9.1","dependencies":{"lodash":"^3.8.0","cheerio":"^0.22.0","request":"~2.67.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.0","charset-parser":"^0.2.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.0.1.tgz_1481292873466_0.6716562905348837","host":"packages-18-east.internal.npmjs.com"}},"1.0.2":{"name":"crawler","version":"1.0.2","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.0.2","maintainers":[{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"d45c360814e2fe2389e78dc97a0bdac9ded9cbf2","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.0.2.tgz","integrity":"sha512-uzsF4AmiWt6zElsxFDg+PIPxsJr+806ijj9AHLYbBHOApOo1f9w+xxnDRPCYwYvpcB+ErS/AaEeQ5+Nz5yq9+A==","signatures":[{"sig":"MEUCIQDaEQBUqfnBpXWIecZTd+SapjTAmjlX0Tvom1J3zSB9vQIgCriKZ0oRpwR7eHs2ZHyN4Zx7K17RBnmQJRgfR0CzAKs=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"d45c360814e2fe2389e78dc97a0bdac9ded9cbf2","gitHead":"f3b730ca881a433e3ff77828dabff1445889f83f","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.8","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.9.1","dependencies":{"lodash":"^4.17.4","cheerio":"^0.22.0","request":"~2.79.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.0","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.0.2.tgz_1483415455873_0.9277358569670469","host":"packages-18-east.internal.npmjs.com"}},"1.0.3":{"name":"crawler","version":"1.0.3","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.0.3","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"a923a752d68097f2727a28e1352c44c05941cc60","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.0.3.tgz","integrity":"sha512-z14H/DIw6QT9WWQTO4UxGmK3k7Gb7w2VIRagtnD69fmi4IY7a4wARimt1VodLxfGpReYcpKzY1BiJ2E1tBJVHw==","signatures":[{"sig":"MEYCIQDbxRxiKmf+PW8idlgXtCHRDu2ZLUwkZH6MO54mRQDx8AIhAL12onGDV0376+aNKLzhssTcocmvfL4e5x3l2WcI9Gpq","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"a923a752d68097f2727a28e1352c44c05941cc60","gitHead":"e7b8db60256e22bdd26907d7794b5677fa9afa5e","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"darrenqc","email":"darrenqc823@gmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.3","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.3.1","dependencies":{"lodash":"^4.17.4","cheerio":"^0.22.0","request":"~2.79.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.0","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.0.3.tgz_1487655200947_0.7274244653526694","host":"packages-12-west.internal.npmjs.com"}},"1.0.4":{"name":"crawler","version":"1.0.4","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.0.4","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"6beb7106bde85817586575ef181be07c1c21efb3","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.0.4.tgz","integrity":"sha512-WwU4yHuYUp/APl0CH8Z91b1WgmZTgtPTRJQz6sfW7M66yKyRj4Wnhb747Wf4KwMW2fUE0OBNlQADpxPVBmAZLg==","signatures":[{"sig":"MEUCIQCpX1yp5mqhSJ7HmwD14hZscfqt+BM4KE90cLcwsI3eTAIgbH9/ppbtaqRob9bWUDecl39jQ19LNBMs7jZORmKARYU=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"6beb7106bde85817586575ef181be07c1c21efb3","gitHead":"16e372b3aea52ab06ef21a39255067939a03a76b","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.8","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.9.1","dependencies":{"lodash":"^4.17.4","cheerio":"^0.22.0","request":"~2.79.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.0","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.0.4.tgz_1491802533960_0.9176827603951097","host":"packages-18-east.internal.npmjs.com"}},"1.0.5":{"name":"crawler","version":"1.0.5","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.0.5","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"3a95f1bff46763e57ee6161d012b967f8e771684","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.0.5.tgz","integrity":"sha512-DzpBeqscMOHhfK69qhsdtrkXPwAj0nBdBL/JWnD60VRxsLEat4Zri3+VhxgvAVLwDC9D5JwaQ/KGOSZ1dqwJHA==","signatures":[{"sig":"MEYCIQDzIq/ApqyEZGW8RIDoLUhgxn1CHwHZX8rJqSHmM7x2VgIhALt5wOqQgMVOcG34j8O/P+l9gJQm1dsuLF9lBzQn2NGW","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"3a95f1bff46763e57ee6161d012b967f8e771684","gitHead":"0f3a27d4e89d93fcf9c27a417adbba6fab9dde2e","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.8","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.9.1","dependencies":{"lodash":"^4.17.4","cheerio":"^0.22.0","request":"~2.79.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.2","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.0.5.tgz_1494417107443_0.3716903864406049","host":"packages-12-west.internal.npmjs.com"}},"1.1.1":{"name":"crawler","version":"1.1.1","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.1.1","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"7a8de84fece7d976e6b8b65908d9d15a0d2a10e6","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.1.1.tgz","integrity":"sha512-49AHJegQuU3/x8jAbfXJ8BQWKLbWt6EEr+4OpntvuOFHDvhXanGx6nCrhUW95AFCpzuaKRXRG9baACcylHEowA==","signatures":[{"sig":"MEYCIQD9nTddZZjKyyvPH5jAwrZV4MhYghC0WouhMLqzHyY/2AIhAK3Y5y4uSjDlDYHOWphbzLkzw3NUcpFih8EIbqFKn8/c","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"7a8de84fece7d976e6b8b65908d9d15a0d2a10e6","gitHead":"097b008c72ea793fce8e01167663f7ff165f1a9b","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.10","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.11.3","dependencies":{"lodash":"^4.17.4","cheerio":"^0.22.0","request":"~2.83.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.1.1.tgz_1508466933770_0.04619367769919336","host":"s3://npm-registry-packages"}},"1.1.2":{"name":"crawler","version":"1.1.2","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.1.2","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"2bd209f11e7aec396b3765b4f38991e480d4306d","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.1.2.tgz","integrity":"sha512-wxhNQdSOepxz5nBe5k7SelHQtjzMFdyBFuqXt0IKtOzg2+8Ztq0N2vN3jhNkR4yvqMu77u24XuMXxJtKz0GdcA==","signatures":[{"sig":"MEQCIEw4nVY8uXyzDSPJivFLrdyJw4VR2vJ30wAxXZzmk3tZAiBFdFwyFkilaOL3s69POi2IoLy650pwiFT5uF+q5kljJw==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}]},"main":"./lib/crawler.js","_from":".","_shasum":"2bd209f11e7aec396b3765b4f38991e480d4306d","gitHead":"f4344ab6e18a4beb52bf8c9d649e772db5f29259","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.10","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.11.3","dependencies":{"lodash":"^4.17.4","cheerio":"^0.22.0","request":"~2.83.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler-1.1.2.tgz_1512992318722_0.11368034454062581","host":"s3://npm-registry-packages"}},"1.1.3":{"name":"crawler","version":"1.1.3","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.1.3","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"d27ca32c3a14d3441b7088bb719931b496f33c2c","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.1.3.tgz","fileCount":34,"integrity":"sha512-OdCBJAAkuLl41kMeXvXVpzM7fTXoIwS06OH53DrDbWFOFC1iChFBaC11KTPMurdMOnYC8BHyb6boWi2D4UKb1A==","signatures":[{"sig":"MEUCIQDRNvL0cSc2T3/IJcZySkS1eyPkQRKbxZjgv5/4StDrsgIgVkO6I68gnBF3ZhyIDjxP5XDKaBCisIXIQ8dPEtDIjnw=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":339105,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v3.0.4\r\nComment: https://openpgpjs.org\r\n\r\nwsFcBAEBCAAQBQJbBWpSCRA9TVsSAnZWagAATTgP/2W7C3rBJGSrZLFkoAwI\nMRZYfocvjG2OAWvWgJJznBoffRnBauK3wDOQDDb9bvXpc79QofG1+3WeAfow\nVgRe9un4eLsePAgxwJZsbW0VQI5rwUhVG76jzJOnIY5t3V2CvS4ZvYh9wZu6\n3XCKDmr6361A50+Lb+ybMdpbTgGATko7AhoTqhJft35LIBgK/ej4Nk7p8dj9\nk5VOF1105rRAs61BIg/n0Tk6SbgQAXjoV13Mr/1Sw+QP9kdjuVkkotNnqzKd\nyAiGYCxpCYstncPknOCWlqi1dIU3mcSRnzQmDAqQLvlUGm7BvyWFTTrag+7x\n//d/sHCIKalZfLbQi7GbdzM67mRRrYtBV5fk6uKFRud4McGQCfkwDDvyysab\n3QgGFtM9EXKsJoFYEwk26pSPuAKn6FuievIat/F5c05hKoK8JMj0teDlAAJh\nQkNhdTpWwwuRsrlUBIGC8MMh2ORMT8YuLca1hFl7537ogYZD/U+emps4QZYA\niyKNaFUabNkfhPFPwmP7I2r+buA5FY0kQ99deIKz9npw4KJEl5RrpTWiRA31\n0KdquDh74wuoFQ/enAoUd2yDNVl66So8+AxzHvTPxw4zugDrgOc+ChKLxFWw\nYb/Em2MR1m3Y+QjmRTyIGxZvkiu993I/JsAXhDjMuBqFhBd/rv7VAV8OoyAI\nMDrW\r\n=kxub\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","_from":".","_shasum":"d27ca32c3a14d3441b7088bb719931b496f33c2c","gitHead":"65d25f28aaa81d89951693d3a28cdaeb23c8189b","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.10","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.13.1","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.87.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"_hasShrinkwrap":false,"devDependencies":{"chai":"^2.3.0","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.1.3_1527081553530_0.25442065800269975","host":"s3://npm-registry-packages"}},"1.1.4":{"name":"crawler","version":"1.1.4","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.1.4","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"1475ff6a02101403b8f486e45c4113b16e8a623c","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.1.4.tgz","fileCount":36,"integrity":"sha512-7L+jyi//Ni9HgcgQjr2npXSi7r/PXu0c5kzq8g6XMQ185nfEeRWiNQbBrcPYiDor6KjBSOqAZUGuhsPC6EtKWQ==","signatures":[{"sig":"MEQCIHTTGLVOhOWlassmPpgl1ujlAOtX3Aqdi0T2XY13a4LoAiBI/4bKrwkteIwoH7ohIqxW4GeDBpsjWaUIxXTcBs5baA==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":344168,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v3.0.4\r\nComment: https://openpgpjs.org\r\n\r\nwsFcBAEBCAAQBQJbD9JCCRA9TVsSAnZWagAA9JMP/iu09xNL4VZoVx4GxfdY\nvPpp8tJ/C5GZOmMFqa7Va3NxJTmbWdhnLT/lrdI/T2MMGURI3BCVMPIANjLm\nIZy+dyxxkf5Q7xnuPM3irFlmLnZ65u23i434DvYL/Eax0Bx6Fj2J6gy4gzCN\npGhfNPWNqUvoUUZmqfDkXb1FzvseKNqfEOSOE80akvRHD7ju2J+E9t9ySVZ7\n3C7xsv47TwMdxjxPbnHzumLvnPPMfK7DP6LMjhBc3T/6+lTX5cKlGRGkHluL\nTeJrkSTlqXE+mzy1EFWOJyKdn/BFLzGzfkoULPt3+Ij1+kj6KjvuL29f9Sib\nHmXalW71N0xd5+5AHe2ix3FX2WvRx37LMQYUrOTmQn3OAFsDCQpFaCA605fV\nDClf5JmB2CHC8RKa2YyQKeB1TPkVSeQ2zKUsmEjmA0j3RCpSoFvf+i7/2fO3\n3B8g/vn0Yy10xx0VIOlZaMZShu9CBqg2DVAXoaAc1VXPGeyAHl0wSNkYyT3C\nLRRdJqKTNy03r86AOCF+HOxJw+N/FZ7kccxohqzcLOWiYXLzQiu4dMlWCilU\n+7W28SgKt2BdNtSq9+vEYaDB60Oy3VuU4x4m2u+XjeMDfm1UCQuvJFFJnI7D\n40f7/I49Ris1fwt1O/atHen7WmycNCTXsSwtd9sCcPq1PjQdz9CmaLxybuJg\nbqVf\r\n=7quz\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","_from":".","_shasum":"1475ff6a02101403b8f486e45c4113b16e8a623c","gitHead":"a116947cf240444bf10230915dc97d688252b575","scripts":{"test":"./node_modules/mocha/bin/mocha --reporter spec --bail --timeout 10000 tests/*.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.10","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.13.1","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.87.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"_hasShrinkwrap":false,"devDependencies":{"chai":"^2.3.0","nock":"^9.2.6","jsdom":"^9.6.0","mocha":"^2.2.5","sinon":"^1.14.1","whacko":"^0.19.1","mocha-testdata":"^1.1.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.1.4_1527763520825_0.7984914402361953","host":"s3://npm-registry-packages"}},"1.2.0":{"name":"crawler","version":"1.2.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.2.0","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"c0bca14a107be0d1f58d021d12dd797364b0d50f","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.2.0.tgz","fileCount":38,"integrity":"sha512-94q3NfQi9dsPFBcizfwNEhMOAVRgjpTRZcf/BbNh571X6OWXgArCJ2oKPr3MSAi5vNSu06UDrXN/grfVmf5cFw==","signatures":[{"sig":"MEUCIFP9BmmpratPevMgXRBJnSnxYy8pe7RcjoJhLYMa5jfTAiEAjisf4lrvRlVXzVIibXMwPm/H4ltKcEccbwg2HaYaV0E=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":407722,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v3.0.4\r\nComment: https://openpgpjs.org\r\n\r\nwsFcBAEBCAAQBQJbctnkCRA9TVsSAnZWagAAbvIP/0usGY+145RgZLTiWzDn\n/nYY/rrdvklBo1O1MDHo0EQUqkZ3LTNfm7WU9mRv9VKS8wCe/dyXP8lWvo/x\nqzL0tY7VgOGJ4fZnh2w4pGULGtCn/E6irTDbcHIaboUqY5ugirnTpBf4t8pO\nVODlzodwOPF+04RaWUNEOrk7tr+1wCUhH79gtehNarIfu0Ub3u+F5QRM2ekH\nTqJQF5pUhHezDAdHh4zOLFruXdmW92pELmJQRjWyYTQNZl3h3foqH1NUd3QJ\npPh3i4mciIXsR4WtQIPBkDrQmilXeJjkRE3u4ixODph+GmUGVXzPyaH/JGu+\nE4cbwhwX8TQ8zbGlmsTyPQ6yPd1+nhYlG6aIuPcW+8jLxJmW/CWDU1YMoY7y\nX/0pYVVDBBlLM2u+t/yV1BsmjTLjl/6HCTrvZeCosuE3SwRl+b5H2/yt4pI2\na/VjczrS5LzISpDlGrkI/PwgJlOQ+Ez5ewQ0nFNpH308uNesgak0ALWXx12Q\n2dHVqDPU0sVir93vAPs0HyaGMDmfwtp+rvRGfOePh3ixcBYmP7nhp/zINACK\nTmW4n5yKX85Cz/sBbW7rqsGhm0OS8XrIJdzwIL+2vML26Xd4BazD6esv7AiZ\n8R8hcMKe42Kvxvd4b7hzLGhsCcH6UvjmjZr0j9Sxog0Udskw10Wt4nVAe5wc\n61Tv\r\n=i/n/\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","_from":".","_shasum":"c0bca14a107be0d1f58d021d12dd797364b0d50f","gitHead":"f5a972364e36273543328124c5aa814e498a09d8","scripts":{"hint":"eslint ./lib/*.js ./tests/*.js","test":"mocha --timeout=15000 tests/*.test.js","cover":"istanbul cover _mocha --report lcovonly -- --timeout=15000 --reporter spec tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"3.10.10","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"6.14.3","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.87.0","seenreq":"^0.1.7","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3","charset-parser":"^0.2.0"},"engine-strict":{"node":">=4.0.0"},"_hasShrinkwrap":false,"devDependencies":{"chai":"^2.3.0","nock":"^9.2.6","jsdom":"^9.6.0","mocha":"^5.2.0","sinon":"^6.0.0","eslint":"^5.3.0","whacko":"^0.19.1","istanbul":"^0.4.5","coveralls":"^3.0.2","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.2.0_1534253539656_0.7961761682673432","host":"s3://npm-registry-packages"}},"1.2.1":{"name":"crawler","version":"1.2.1","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.2.1","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"396d3fc0e2c227463dcb165bead5c5dd5cf3062a","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.2.1.tgz","fileCount":41,"integrity":"sha512-5hpo+iqAwIl0TXs+c3u5V1gQd4NqsiFDF0QiGZCE/OolLiY1QZA32AaDqAPTTmmxErwYJEoi6lN1j8mO4/WE+A==","signatures":[{"sig":"MEQCIEyBAKqBrq2VmK7phdkJbRlo74NOu17V8woJ5y3YHSHxAiAwOnthv3hqrMOiBmj7ki2DKtlXLtVl+J9kmSqdl5+wfw==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":850604,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v3.0.4\r\nComment: https://openpgpjs.org\r\n\r\nwsFcBAEBCAAQBQJctZwYCRA9TVsSAnZWagAAZ0QP/ivYdccO6ZKTtx0RxpU0\nDsi0s+gaalerLtkuaSwOOWeuvbZA2D7ouN/LCMQDcJt8wuxdl19WDmiopxJ7\nILeBjcJ/kQU6mYapc2AKIQQdOdVAKN+EbLeDn1KzVwBPhsvliHd2S6RwmAVo\nobLMD7UjaGv9hs48EnrdsuJ9ZMl60w6itOjMczTgFamdZ6nKoljGXHAYc4P5\nG9GH/B0D/NSOWJswW+UWoHhGI98zhOtIeHpxb3ACu5yO7gFdZeImiyuDvwsZ\n8MczuEud4HQwXfu9lRIJyU9tnXAooVSOP8wtllKlgcvrs35gvvQaoPt3315W\nSfG/Zs0aqd8yXQpmSpGPKJrrsAWRuMSHr7UotFh5U8EjyR9+J2YXOuYx8ZRf\nYSjzhcyPuVFiehpXs/901AE4HQ9+7IN2Bvo/S8rR1WNmVfj0qR7hd5sHnIlh\nWewMI+xGUeRGNQ6OuN/sCX2MyGQasrSTG+VdPHZgHAe+4dTUBKmH4cdtIz4U\ne3wxDMLXiVPyCqCd2srAHnEcwRKNwUGmb/NN/SQSvfgyo4FfmCEQLAaEL2SY\nqP/jsLGQlVAoe7ILdYQp9KlCy5GgWuVrdyIYlQEW/gwRLAAO/Ohua6uK57Rx\nL/hvZbUH2L0Xi6AYEyNxX0F29bwUcFnMTUxOP/Z9xoDPZ+2/NXoRZ0h4gmyJ\nsn0H\r\n=4kNV\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","gitHead":"34bcfb45d303e332d564a4a56c843868c52a6141","scripts":{"hint":"eslint ./lib/*.js ./tests/*.js","test":"mocha --timeout=15000 tests/*.test.js","cover":"nyc --reporter=lcovonly --reporter=text --reporter=text-summary mocha --timeout=15000 --reporter spec tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"6.9.0","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"10.14.0","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.88.0","seenreq":"^3.0.0","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3"},"engine-strict":{"node":">=4.0.0"},"_hasShrinkwrap":false,"devDependencies":{"nyc":"^13.1.0","chai":"^4.2.0","nock":"^10.0.6","jsdom":"^9.6.0","mocha":"^6.1.0","sinon":"^7.0.0","eslint":"^5.0.0","whacko":"^0.19.1","coveralls":"^3.0.2","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.2.1_1555405847201_0.6610304112426113","host":"s3://npm-registry-packages"}},"1.2.2":{"name":"crawler","version":"1.2.2","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.2.2","maintainers":[{"name":"darrenqc","email":"darrenqc823@gmail.com"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"ee07a8fc29018b2ce9022a32bee5c67ada0ca99a","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.2.2.tgz","fileCount":41,"integrity":"sha512-d2DRjp2/KobEnZtKHEYcpwxmTzQtGTHpI8wvcy61+GXdWXBbzOaHG9fw+gm/C9ADeE7r//O/epe+p6gtK7tUWw==","signatures":[{"sig":"MEYCIQCn7Ip6t6i00au/CfeUtw0zsJNh/aN2j+aZ33zK+4M4DgIhAIznrrErbb1nq0hZSecqVlEzOlWiqCYwxa3lg6TtfSEH","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":851981,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v3.0.4\r\nComment: https://openpgpjs.org\r\n\r\nwsFcBAEBCAAQBQJe35ZtCRA9TVsSAnZWagAAcFQP/jsdeMtJ/iaaoOORbn1u\nYzp2z6I754chA6Nj7BVVtIMfsfNaBwwVXGvTf9tpKtaX2iE2iM3TrTeAy0r2\nLeGW2YZCnv7EKpa7Y1GyGwMifZggVLc1RECvs3Fb2kaem1h60ddDLgdEE8t7\nr62A0ZuUl5fwEsR9TVGn12GUz6s4s90F44DVnEbAkHdfG3+o5WtFSHWMLps3\nmch6IbV6S/vMnEDUCLyogzthDJ7D0frTvMA5Jfz2nZN0mM8/iT+KPKCsuUvw\nm//JUsAANmLYlNSUEYgBBMtClc+KzMygQp+762LjVDMpBz8ZLC5SInCoor6x\ntFDhjrehVFViQKo8vV0AcYZXaOxnr+ri6ytuwCFUAB4EQuC30vELhmK37dS3\nnp5DfFK8y+ZxSuBaE3aMefKQHmWMn/pbQraRxESTxznzmnyLpOyRyYDa6vwy\nts1luIxbpmNuAbtnKDHsUyq3xQk79/lgYU3oLYUfZWHda0ZJSwJeX2gnkKCp\nubGiqrongkmPe7c3vaAMrHezKoNXlxfS0l2B7h0wDjaoZFl1o31eo6bw9ZDR\nAvvgOnuD2PYKoFlF0aV0RqZv+iCeiocL5JJUwjsh3BtpNJlS6VyMLl3mWyCt\n+3zfajuE+zX85Y3GxklXQPOeCTzjmxZhvzrGCB+fyyfxxsLj9Ng13JtVqfxb\nPphe\r\n=BnCQ\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","gitHead":"e6f43788097ab36c022b04266ed305bd699e81d5","scripts":{"hint":"eslint ./lib/*.js ./tests/*.js","test":"mocha --timeout=15000 tests/*.test.js","cover":"nyc --reporter=lcovonly --reporter=text --reporter=text-summary mocha --timeout=15000 --reporter spec tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"6.13.4","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"10.19.0","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.88.0","seenreq":"^3.0.0","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3"},"engine-strict":{"node":">=4.0.0"},"_hasShrinkwrap":false,"devDependencies":{"nyc":"^13.1.0","chai":"^4.2.0","nock":"^10.0.6","jsdom":"^9.6.0","mocha":"^6.1.0","sinon":"^7.0.0","eslint":"^5.0.0","whacko":"^0.19.1","coveralls":"^3.0.2","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.2.2_1591711340901_0.29447721275793004","host":"s3://npm-registry-packages"}},"1.3.0":{"name":"crawler","version":"1.3.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.3.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"d42c747beee8f45ed2e0213c736336bdeef9f449","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.3.0.tgz","fileCount":43,"integrity":"sha512-FQ3io73liPNHhaGkMJi6BPwkZI16ShOTz3jHBUr8oLilVViRFo8mYKRQyNRzCxAU0U5VlokY+Eq9UwZ4IrzXdA==","signatures":[{"sig":"MEQCIHFHc9ubk0eiivgCbb0mse1qhj+BrcXrNRZYREwNtuzxAiAnNTvKFFTUFIwlfzraPn7w+S4e89dpTo1QLUL0DjgICA==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":772339,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v3.0.13\r\nComment: https://openpgpjs.org\r\n\r\nwsFcBAEBCAAQBQJfu7jhCRA9TVsSAnZWagAAys0P/1MovBya1PdkeaTtRKRS\n/dIlPO4hUWI+iVsCyc2u+QapioHF8d2JVnINghtO6wSpfJ8+hmnTfYD3jnCn\nNtP78fSwU0bjTWEfCNQ+e+ZQwEabs2HCMfhsctUC2VLb50o3CRNK2bbhGfAy\nFjty8JsT2EGXzbKyVYmNN2o+tBpBVWuGJbM548nOxDv3jwO/YoaAYihawb1b\nmBkRQw6sUPTTGCmG+CkVzrnTZDl1gbsezIddMLQomaAp+suyjuCDoGyOxlAi\nJaOryULyF80XwaUtT6RYFLkDD2ZxR18QuN8i9W+L7m0623yZFfqNtrc7dRei\nOhibyu3yE45ZEkRxZFfLhINx1go2RKssXNAJ3Id6JpfWTH1kReXzjWGfKgEy\nqMpwR6UkOEE7VlfTGJu7ycEMwc7V1aoICh5T9W3jtHJL/ArfUSL3FFyUEkGS\nkaEFO9Do61BaXh9VfhezPxKgZ8GLMX7iDduOmNoyFm0d9dg7/kAIJPPDCzHQ\nvOtwjUPx1EeLkCR7nPPpNkOU3/ovKCvV41GWLogYDlEa9m4kzi6kjXsAwyRB\n96EmY7LAepcgs16sOBdX14eDb1AHFIm+iQ4duRuU81SegX+xsWECfbE0feJU\nFloLmxzrS9JnndxlJ4qWZTlHk6b1RYmZquo4DtY+lIVOa3VBoPnAPpBvJWFu\nK/mB\r\n=/+K9\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","gitHead":"70ec0efd6a7d4b49894b56c84d7c44259bd023e0","scripts":{"hint":"eslint ./lib/*.js ./tests/*.js","test":"mocha --timeout=15000 tests/*.test.js","cover":"nyc --reporter=lcovonly --reporter=text --reporter=text-summary mocha --timeout=15000 --reporter spec tests/*.test.js","http2test":"mocha --timeout=15000 tests/http2*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"6.14.6","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"10.22.0","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.88.0","seenreq":"^3.0.0","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3"},"engine-strict":{"node":">=10.0.0"},"_hasShrinkwrap":false,"devDependencies":{"nyc":"^13.1.0","chai":"^4.2.0","nock":"^13.0.5","jsdom":"^9.6.0","mocha":"^6.1.0","sinon":"^7.0.0","eslint":"^5.0.0","whacko":"^0.19.1","coveralls":"^3.0.2","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.3.0_1606138080790_0.5941714072469979","host":"s3://npm-registry-packages"}},"1.4.0":{"name":"crawler","version":"1.4.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.4.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"5d3e3cea57d687593c3b56fef3099a72e0fdef69","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.4.0.tgz","fileCount":48,"integrity":"sha512-rFg8/AtpQ5c8SItxwamlP4fGClnDd/n5GOdFakq0KtCdTzXL2HqzROyrAfEWPYuFzX2n+avafe3hQiFGKKOLXw==","signatures":[{"sig":"MEYCIQD8ISe4FECDilGxChDN90wjf4pjDk0+4J0Htd5Ws0gb1gIhAPaRH5Kr32ut8khVckE+X2dZsCfixKt9/QqLt+tWbdSw","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":923509,"npm-signature":"-----BEGIN PGP SIGNATURE-----\r\nVersion: OpenPGP.js v4.10.10\r\nComment: https://openpgpjs.org\r\n\r\nwsFzBAEBCAAGBQJjrk9vACEJED1NWxICdlZqFiEECWMYAoorWMhJKdjhPU1b\r\nEgJ2VmoBZA//RvNl7u+xnfn707IZojB3BSJaOUrqZ3NXTQzG2GUXwOOWYPfO\r\nJ3Wia0rjt0pa+HhDeOmMc7zsjraUquIlF6m9xWkTvtwZP7sNE9iIf2h3DBjk\r\nwpyIQDjjYh23ijdd2lZRtfZQbldL32647FyAlS0VfwMN06aY36cF4FcWZEhy\r\nwhHbC2LjbniaSNtDySy5oOClJ4l/ZnxXjbNG/HNs7LZO/MBOD1wDyulSwroJ\r\n3lj/ZuJj+6ClRfwdkxX5X//OiJD57gYX25OehwQyOiFHwVeRBgxhT2JwP7t5\r\nhAMBHqLdBaQJf6Nm+IR2CEt6udih0zGyosGYCAN0WEr35GFWL8GR3LkOSbcB\r\nXljYxvVNGRePBeJ9+Abb+UPPOOUkqay9sBwEmyFBDt0ERWMqd42mTguHHn19\r\nQX5HJgLKQHetLj8QsQRc0aZkWNYIFbamOT7OEHIrna8/FjPt84KtEtQJP748\r\nMNdapL9nqib3TLfr3DmWe2X19YwkVmHjrxCNGwJwDhh6koEZUCcMmqsxykBb\r\nm43T5cDStagfONLGmS/OB1sWQAvWdVLOxWv2eryHwQVEy7M/BOi5xCWzM7C5\r\nFpe9x//ov9rudSXdpt5NCTVd7Hp8LV++kdUp5xxKSdQuPzw+2t5xxtfod6oE\r\nZSHo5g3AxL9j0Z8MJDx7ZUMiUroGISuWoFA=\r\n=rnfo\r\n-----END PGP SIGNATURE-----\r\n"},"main":"./lib/crawler.js","gitHead":"801b1aceed3c4a0ab064cd8bdf1d5edc253aa1dc","scripts":{"hint":"eslint ./lib/*.js ./tests/*.js","test":"mocha --timeout=15000 tests/*.test.js","cover":"nyc --reporter=lcovonly --reporter=text --reporter=text-summary mocha --timeout=15000 --reporter spec tests/*.test.js","http2test":"mocha --timeout=15000 tests/http2*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"8.19.3","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"16.19.0","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.88.0","seenreq":"^3.0.0","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3"},"engine-strict":{"node":">=10.0.0"},"_hasShrinkwrap":false,"devDependencies":{"nyc":"^13.1.0","chai":"^4.2.0","nock":"^13.0.5","jsdom":"^9.6.0","mocha":"^6.1.0","sinon":"^7.0.0","eslint":"^5.0.0","whacko":"^0.19.1","coveralls":"^3.0.2","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.4.0_1672367983680_0.6672167741681738","host":"s3://npm-registry-packages"}},"1.5.0":{"name":"crawler","version":"1.5.0","keywords":["dom","javascript","crawling","spider","scraper","scraping","jquery","crawler","nodejs"],"_id":"crawler@1.5.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"27fc8c80d2785b798e0f8fa59675401de4f4ff70","tarball":"https://registry.npmjs.org/crawler/-/crawler-1.5.0.tgz","fileCount":40,"integrity":"sha512-0iybajRv0RqIogn60CmglSZ3nmDEUs01cMqBF4ZdSNpYEtJEkEexDuxov/neHgYYrF+PDZnmVygRAkEd/SYIdA==","signatures":[{"sig":"MEQCICrJXdC7sUHlcTgFeq4wiuTXXPpbeGNza/UcDro2JWPXAiBTJsm7kxTP6DzFHMNXe7IzYUXWS8064r2Bj4apn11uBg==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":239665},"main":"./lib/crawler.js","gitHead":"8459ab1b2d37b61230b5e655f67890dc0a337515","scripts":{"hint":"eslint ./lib/*.js ./tests/*.js","test":"mocha --timeout=15000 tests/*.test.js","cover":"nyc --reporter=lcovonly --reporter=text --reporter=text-summary mocha --timeout=15000 --reporter spec tests/*.test.js","http2test":"mocha --timeout=15000 tests/http2*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"licenses":[{"url":"http://github.com/bda-research/node-crawler/blob/master/LICENSE.txt","type":"MIT"}],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"9.6.7","description":"Crawler is a web spider written with Nodejs. It gives you the full power of jQuery on the server to parse a big number of pages as they are downloaded, asynchronously","directories":{"test":"tests"},"_nodeVersion":"18.17.1","dependencies":{"lodash":"^4.17.10","cheerio":"^0.22.0","request":"~2.88.0","seenreq":"^3.0.0","type-is":"^1.6.14","iconv-lite":"^0.4.8","bottleneckp":"~1.1.3"},"engine-strict":{"node":">=10.0.0"},"_hasShrinkwrap":false,"devDependencies":{"nyc":"^13.1.0","chai":"^4.2.0","nock":"^13.0.5","jsdom":"^9.6.0","mocha":"^6.1.0","sinon":"^7.0.0","eslint":"^5.0.0","whacko":"^0.19.1","coveralls":"^3.0.2","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_1.5.0_1704356193189_0.2833559945859103","host":"s3://npm-registry-packages"}},"2.0.0-beta.0":{"name":"crawler","version":"2.0.0-beta.0","keywords":[],"author":"","license":"MIT","_id":"crawler@2.0.0-beta.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"dist":{"shasum":"3a368b68328727df7e331147280cde9b60f659ae","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.0-beta.0.tgz","fileCount":36,"integrity":"sha512-b0AYMbo9sWkxCUyyK5kVNgWlMJetBT563UlNoME+JfjkpWtviBrH6/ksOMbqYiGb1IBpAi6hxUn5p6fsbDvuvg==","signatures":[{"sig":"MEYCIQDRCXO4lfMCgVV2W9XIo5ypnI56ZZUyk3BUP7zkoac5IQIhAJcbTRVzLBKeHPFAPnj6XtLPmNnWp9VHKdF6J8FLak3X","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":107982},"type":"module","engines":{"node":">=16"},"exports":"./index.js","gitHead":"5e2ec9fdbd2721f49140616cea922cd9a6aea59f","scripts":{"test":"npx mocha --timeout=15000 tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"_npmVersion":"10.2.4","description":"New TypeScript Crawler Test version","directories":{},"_nodeVersion":"20.11.1","dependencies":{"got":"^13.0.0","tslog":"^4.9.2","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"tsx":"^4.7.3","chai":"^5.1.0","nock":"^13.5.4","mocha":"^10.4.0","sinon":"^17.0.1","@types/got":"^9.6.12","@types/node":"^20.10.6","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.0-beta.0_1715308528929_0.013021909358180483","host":"s3://npm-registry-packages"}},"2.0.0-beta.1":{"name":"crawler","version":"2.0.0-beta.1","keywords":[],"author":"","license":"MIT","_id":"crawler@2.0.0-beta.1","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"dist":{"shasum":"eb8fa0249d0d66e6b15f23c8280aa91a99d72f17","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.0-beta.1.tgz","fileCount":55,"integrity":"sha512-AOV3G7LQ0sbvUsPBvxlFYjalkULeyx10ccNvfFbxcUPdL4UsVuPmd00jdltB/U1mNAOX+gK3+eTVhFvXlcuUPA==","signatures":[{"sig":"MEUCIQCZWng7r9WLhxmOmHQhRSqB1ZxQAfSw35/C27Z7gz0wDAIgRC4YRktwLJH+ePMbuirou5HtZ2x9LE96wmJYiguNhic=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":83934},"type":"module","engines":{"node":">=16"},"exports":"./dist/index.js","gitHead":"d756dbb66afa46fc39ce5be553179d77c9be43a5","scripts":{"test":"npx mocha --timeout=15000 tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"_npmVersion":"10.2.4","description":"New TypeScript Crawler Test version","directories":{},"_nodeVersion":"20.11.1","dependencies":{"got":"^13.0.0","tslog":"^4.9.2","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"tsx":"^4.7.3","chai":"^5.1.0","nock":"^13.5.4","mocha":"^10.4.0","sinon":"^17.0.1","@types/got":"^9.6.12","@types/node":"^20.10.6","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.0-beta.1_1715419959081_0.2636325629133698","host":"s3://npm-registry-packages"}},"2.0.0-beta.3":{"name":"crawler","version":"2.0.0-beta.3","keywords":[],"author":"","license":"MIT","_id":"crawler@2.0.0-beta.3","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"dist":{"shasum":"f9ced094488e829b86afd57ab7d531dc823e0c7c","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.0-beta.3.tgz","fileCount":55,"integrity":"sha512-1QKsiZcL0mIIb+A8/XP60QOC6b3aLdXnaitd63JXMIcjYbGAjE3xK4B8tBx1nyukI1HHxDEhWuSC2HlChTJDJg==","signatures":[{"sig":"MEUCIQCk76cG3odLUeeaSikIFbiCqZXSpCj1fJJWb5T421YwpQIgC1CaYpSwiqVvhHzarOyCvyUXG76Q2uTZbZ4sj6hTYuI=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":105742},"type":"module","engines":{"node":">=16"},"exports":"./dist/index.js","gitHead":"834058ec3fd93699d72800ef3e9ea646808d81bc","scripts":{"test":"npx mocha --timeout=15000 tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"_npmVersion":"10.2.4","description":"New TypeScript Crawler Test version","directories":{},"_nodeVersion":"20.11.1","dependencies":{"got":"^13.0.0","tslog":"^4.9.2","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"tsx":"^4.7.3","chai":"^5.1.0","nock":"^13.5.4","mocha":"^10.4.0","sinon":"^17.0.1","@types/got":"^9.6.12","@types/node":"^20.10.6","tough-cookie":"^4.1.4","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.0-beta.3_1715657904278_0.5720169151499346","host":"s3://npm-registry-packages"}},"2.0.0-beta.4":{"name":"crawler","version":"2.0.0-beta.4","keywords":[],"license":"MIT","_id":"crawler@2.0.0-beta.4","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"7b7357950533bc295773d1ec7e21fc7a28307997","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.0-beta.4.tgz","fileCount":55,"integrity":"sha512-eoWNFO8FAIeugBI8fKnZHOI7VNmrz/+SLp9+hmHiOo4is5isA4qpzRTLSGgS96409w+L8RMZ3EE9B++c/N+8Sw==","signatures":[{"sig":"MEUCIQDzOcj0xMTyim+tP41iE26KBUJ0zMM6mFU+kEjW4qXUngIgPbok8Lj2ezr+rHsJIQimWPB8/kGcQbV8k+DtYF+sNTc=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":104290},"type":"module","engines":{"node":">=16.20.2"},"exports":"./dist/index.js","gitHead":"61dfed0a45eae08e4b466e817a21e12aed85e649","scripts":{"test":"npx mocha --timeout=15000 tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"8.19.4","description":"New TypeScript Crawler Test version","directories":{},"_nodeVersion":"16.20.2","dependencies":{"got":"^13.0.0","tslog":"^4.9.2","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"tsx":"^4.7.3","chai":"^5.1.0","nock":"^13.5.4","mocha":"^10.4.0","sinon":"^17.0.1","@types/got":"^9.6.12","@types/node":"^20.10.6","tough-cookie":"^4.1.4","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.0-beta.4_1715768103315_0.08091510203508201","host":"s3://npm-registry-packages"}},"2.0.0-beta.5":{"name":"crawler","version":"2.0.0-beta.5","keywords":[],"license":"MIT","_id":"crawler@2.0.0-beta.5","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"578bcc5e02ea8b752aaede948ebad13c001bbdcf","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.0-beta.5.tgz","fileCount":55,"integrity":"sha512-ps7exGkPOUUMOzb4ow662UzpsbdgtsD5SMX70aLcXnstVJuLo83xWWMnnNuG8JyXSaEDQiyAA6PglH964p5KmQ==","signatures":[{"sig":"MEUCIQCPZhY1fK20r4TslA8e+fIE4WhFn5kXDzAY4SVdU6L7fQIgVtIsFG4EsGfgavq8crxx6jFpGvuT/4kWn35kMtAYToA=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":105046},"type":"module","engines":{"node":">=16.20.2"},"exports":"./dist/index.js","gitHead":"11300b880645c4f7c946899a30499181f1dec51f","scripts":{"test":"npx mocha --timeout=15000 tests/*.test.js"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"10.7.0","description":"New TypeScript Crawler Test version","directories":{},"_nodeVersion":"18.20.3","dependencies":{"got":"^13.0.0","tslog":"^4.9.2","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"tsx":"^4.7.3","chai":"^5.1.0","nock":"^13.5.4","mocha":"^10.4.0","sinon":"^17.0.1","@types/got":"^9.6.12","@types/node":"^20.10.6","tough-cookie":"^4.1.4","mocha-testdata":"^1.2.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.0-beta.5_1716785634892_0.9970062214541204","host":"s3://npm-registry-packages"}},"2.0.0":{"name":"crawler","version":"2.0.0","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"license":"MIT","_id":"crawler@2.0.0","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"dist":{"shasum":"8946593ab91d58f633a1329c44c5faad7af2d3aa","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.0.tgz","fileCount":57,"integrity":"sha512-KMZRzM9/S+UaRwiWXtfiHixoCzSrN/HpdN2SJYTmlo7O1+Bj0U1pitlqGKHnXpj/6QvWvnWbmpMYn4PYtl8U/g==","signatures":[{"sig":"MEQCICSktSfZYKouJ0xvuadxXXAczSspeU3vFffgeAYN8a0hAiBOP4DBic7cCcvoNouOGlJbVgDmg6DpdseZlWavZ71MEg==","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":129327},"type":"module","mocha":{"exit":true,"files":["test/**/*.js"],"timeout":15000},"engines":{"node":">=18"},"exports":"./dist/index.js","gitHead":"f066751b2ba13b9f23ce1112f7d196b94c0119fb","scripts":{"test":"mocha","build":"tsc"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"10.7.0","description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and seamless HTTP/2 support.","directories":{},"_nodeVersion":"18.20.3","dependencies":{"got":"^14.4.0","tslog":"^4.9.2","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"devDependencies":{"tsx":"^4.11.2","chai":"^5.1.1","nock":"^13.5.4","mocha":"^10.4.0","sinon":"^17.0.1","eslint":"~9.4.0","globals":"^15.3.0","@eslint/js":"^9.4.0","@types/got":"^9.6.12","typescript":"^5.4.5","@types/node":"^20.14.2","tough-cookie":"^4.1.4","mocha-testdata":"^1.2.0","typescript-eslint":"8.0.0-alpha.27"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.0_1717649707756_0.0703449726367511","host":"s3://npm-registry-packages"}},"2.0.1":{"name":"crawler","version":"2.0.1","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"license":"MIT","_id":"crawler@2.0.1","homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"c8":{"clean":true,"reporter":["lcov","html","text"]},"ava":{"files":["test/*.js","!test/*test.js"],"timeout":"20s","verbose":true,"failFast":true,"extensions":{"js":true}},"dist":{"shasum":"2f181eba459cf100896f630625edb9fa2afdfcaa","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.1.tgz","fileCount":55,"integrity":"sha512-Y/qLN7XTrrHBKDOmj3g4PT8M72t1XmAwCkLCYdB8NNWI26Rszr5yzsN1BjBDHo1JmoDyi/+Ui4f/goB2U74USg==","signatures":[{"sig":"MEUCIQCr0JxjtmrJ3HhAY6ctP7SlD0Q30+dhYUGNPYVS2C5s4wIgO8pNzK2DGDDihtpvDigxzA+eQxMlcGdZp+BGHuxnRAo=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":106253},"type":"module","engines":{"node":">=18"},"exports":"./dist/index.js","gitHead":"366fdc419fec183408a36c5496d8540b7c43c1e0","scripts":{"test":"ava","build":"tsc","cover":"c8 ava"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"10.7.0","description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and HTTP/2 support.","directories":{},"_nodeVersion":"18.20.3","dependencies":{"got":"^14.4.1","sinon":"^18.0.0","tslog":"^4.9.3","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"devDependencies":{"c8":"^10.1.2","ava":"^6.1.3","tsx":"^4.15.7","nock":"^13.5.4","eslint":"~9.4.0","globals":"^15.6.0","@eslint/js":"^9.5.0","@types/got":"^9.6.12","typescript":"^5.5.2","@types/node":"^20.14.7","tough-cookie":"^4.1.4","typescript-eslint":"8.0.0-alpha.27"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.1_1719282801345_0.3667544168063579","host":"s3://npm-registry-packages"}},"2.0.2":{"name":"crawler","version":"2.0.2","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"license":"MIT","_id":"crawler@2.0.2","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"c8":{"clean":true,"reporter":["lcov","html","text"]},"ava":{"files":["test/*.js","!test/*test.js"],"timeout":"20s","verbose":true,"extensions":{"js":true}},"dist":{"shasum":"703b7f76fe5f716408707be8842f30573c54e16f","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.2.tgz","fileCount":55,"integrity":"sha512-im/ZTxSMlnGoQ+52uyTBpa8IdYqLwBCpdjG4BSfXwrgkGHvSYA1uXzO9vAB2N7foUqboS+WF0R0f1mXaV8ZQcw==","signatures":[{"sig":"MEUCICGrKgpYOPxejpcexijSL9LQT/yI9ftA7QrUBB8q7UhSAiEA16GWukrFtUijMhvpnwEI38E2rqzhcVAUS6wLHkz4Ud0=","keyid":"SHA256:jl3bwswu80PjjokCgh0o2w5c2U4LhQAE57gj9cz1kzA"}],"unpackedSize":106695},"type":"module","engines":{"node":">=18"},"exports":"./dist/index.js","gitHead":"7de2ef78ceade22f304adac85231e2139f21fcb7","scripts":{"test":"NODE_ENV=test ava","build":"tsc","cover":"NODE_ENV=test c8 ava"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"10.7.0","description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and HTTP/2 support.","directories":{},"_nodeVersion":"18.20.3","dependencies":{"got":"^14.4.1","tslog":"^4.9.3","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"devDependencies":{"c8":"^10.1.2","ava":"^6.1.3","tsx":"^4.15.7","nock":"^13.5.4","sinon":"^18.0.0","eslint":"~9.4.0","globals":"^15.6.0","@eslint/js":"^9.5.0","@types/got":"^9.6.12","typescript":"^5.4.5","@types/node":"^20.14.8","tough-cookie":"^4.1.4","typescript-eslint":"8.0.0-alpha.30"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.2_1721104592764_0.23711184004814423","host":"s3://npm-registry-packages"}},"2.0.3-beta.1":{"name":"crawler","version":"2.0.3-beta.1","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"license":"MIT","_id":"crawler@2.0.3-beta.1","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"c8":{"clean":true,"reporter":["lcov","html","text"]},"ava":{"files":["test/*.js","!test/*test.js"],"timeout":"20s","verbose":true,"extensions":{"js":true}},"dist":{"shasum":"dc33bc8bf42677cbd29ac1b080f5c1acf07501a6","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.3-beta.1.tgz","fileCount":5,"integrity":"sha512-F7HDsOSJBrWbDYuqDFdZ1StpoyucsPBrrJBbrMuzdOEqhCQERBJdyY9xJnsk+eR/1HNkCRjqb273U9LaWmKwoQ==","signatures":[{"sig":"MEQCIC4IlCVEDRNJZVLYQRaoP0PoEwIMSI21qatC8YalGlelAiB5TX5E7bZHG0EUiDvXtXEH4M6Tgv6bqh5P75IZbbXRwA==","keyid":"SHA256:DhQ8wR5APBvFHLF/+Tc+AYvPOdTpcIDqOhxsBHRwC7U"}],"unpackedSize":76659},"type":"module","engines":{"node":">=18"},"exports":{".":{"import":"./dist/index.mjs","require":"./dist/index.js"}},"gitHead":"cc640b677e50b3819dcd2e57783e05b76f78c30f","scripts":{"test":"NODE_ENV=test ava","build":"tsup src/index.ts --format cjs,esm --clean","cover":"NODE_ENV=test c8 ava","prepublishOnly":"npm run build"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"8.19.4","description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and HTTP/2 support.","directories":{},"_nodeVersion":"16.20.2","dependencies":{"got":"^14.4.2","tsup":"^8.4.0","tslog":"^4.9.3","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"c8":"^10.1.2","ava":"^6.1.3","tsx":"^4.16.3","nock":"^13.5.4","sinon":"^18.0.0","eslint":"^9.8.0","globals":"^15.8.0","@eslint/js":"^9.8.0","@types/got":"^9.6.12","typescript":"^5.5.4","@types/node":"^20.14.13","tough-cookie":"^4.1.4","typescript-eslint":"^8.0.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.3-beta.1_1745740287033_0.2093185771933972","host":"s3://npm-registry-packages-npm-production"}},"2.0.3-beta.2":{"name":"crawler","version":"2.0.3-beta.2","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"license":"MIT","_id":"crawler@2.0.3-beta.2","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"homepage":"https://github.com/bda-research/node-crawler#readme","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"c8":{"clean":true,"reporter":["lcov","html","text"]},"ava":{"files":["test/*.js","!test/*test.js"],"timeout":"20s","verbose":true,"extensions":{"js":true}},"dist":{"shasum":"00d4acb739574c9dda55ebcccadeb0e0df10a393","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.3-beta.2.tgz","fileCount":5,"integrity":"sha512-ofzbEMRZ2jB5TMxzZ2kg1d+05kgl3DbTeTmqPkBgGIujoSjBvw/wFgMF50d85k5B9pWIr4brvD3saljc/mY2tQ==","signatures":[{"sig":"MEQCIEzspK4E0JA3HDf6Nlz2K0FKCAg80JMkSCldT8qka175AiAjl98RBnimTpWlRPBpPbhazVVWq/rWt9ALxDs088EpzA==","keyid":"SHA256:DhQ8wR5APBvFHLF/+Tc+AYvPOdTpcIDqOhxsBHRwC7U"}],"unpackedSize":76659},"type":"module","engines":{"node":">=18"},"exports":{".":{"import":"./dist/index.js","require":"./dist/index.cjs"}},"gitHead":"cc640b677e50b3819dcd2e57783e05b76f78c30f","scripts":{"test":"NODE_ENV=test ava","build":"tsup src/index.ts --format cjs,esm --clean","cover":"NODE_ENV=test c8 ava","prepublishOnly":"npm run build"},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"_npmVersion":"8.19.4","description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and HTTP/2 support.","directories":{},"_nodeVersion":"16.20.2","dependencies":{"got":"^14.4.2","tsup":"^8.4.0","tslog":"^4.9.3","cheerio":"1.0.0-rc.12","hpagent":"^1.2.0","seenreq":"^3.0.0","iconv-lite":"^0.6.3","http2-wrapper":"^2.2.1"},"_hasShrinkwrap":false,"readmeFilename":"README.md","devDependencies":{"c8":"^10.1.2","ava":"^6.1.3","tsx":"^4.16.3","nock":"^13.5.4","sinon":"^18.0.0","eslint":"^9.8.0","globals":"^15.8.0","@eslint/js":"^9.8.0","@types/got":"^9.6.12","typescript":"^5.5.4","@types/node":"^20.14.13","tough-cookie":"^4.1.4","typescript-eslint":"^8.0.0"},"_npmOperationalInternal":{"tmp":"tmp/crawler_2.0.3-beta.2_1745741822818_0.9034570524468515","host":"s3://npm-registry-packages-npm-production"}},"2.0.3-beta.3":{"name":"crawler","version":"2.0.3-beta.3","description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and HTTP/2 support.","repository":{"type":"git","url":"git+https://github.com/bda-research/node-crawler.git"},"exports":{".":{"require":"./dist/index.cjs","import":"./dist/index.js"}},"scripts":{"build":"tsup src/index.ts --format cjs,esm --clean","prepublishOnly":"npm run build","test":"NODE_ENV=test ava","cover":"NODE_ENV=test c8 ava"},"engines":{"node":">=18"},"type":"module","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"license":"MIT","dependencies":{"cheerio":"1.0.0-rc.12","got":"^14.4.2","hpagent":"^1.2.0","http2-wrapper":"^2.2.1","iconv-lite":"^0.6.3","seenreq":"^3.0.0","tslog":"^4.9.3","tsup":"^8.4.0"},"devDependencies":{"@eslint/js":"^9.8.0","@types/got":"^9.6.12","@types/node":"^20.14.13","ava":"^6.1.3","c8":"^10.1.2","eslint":"^9.8.0","globals":"^15.8.0","nock":"^13.5.4","sinon":"^18.0.0","tough-cookie":"^4.1.4","tsx":"^4.16.3","typescript":"^5.5.4","typescript-eslint":"^8.0.0"},"ava":{"files":["test/*.js","!test/*test.js"],"timeout":"20s","extensions":{"js":true},"verbose":true},"c8":{"reporter":["lcov","html","text"],"clean":true},"readmeFilename":"README.md","gitHead":"5f6219c02925299953918de9d39954895e17b187","bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"homepage":"https://github.com/bda-research/node-crawler#readme","_id":"crawler@2.0.3-beta.3","_nodeVersion":"16.20.2","_npmVersion":"8.19.4","dist":{"integrity":"sha512-EoUwxmzHgxD+5CTHjnkwwYKYRCmRzrKyL6NlFVs1BYVJi6uGX3g8d5swP0dk3xBsH+2GHg2HGBJGNGkrjdbi+A==","shasum":"1ac6015964a9876fb757534fda1b23469412e5ab","tarball":"https://registry.npmjs.org/crawler/-/crawler-2.0.3-beta.3.tgz","fileCount":5,"unpackedSize":78527,"signatures":[{"keyid":"SHA256:DhQ8wR5APBvFHLF/+Tc+AYvPOdTpcIDqOhxsBHRwC7U","sig":"MEYCIQD7ubBnjYT5vJYghlGma2Q+zgPPTEPhFubZ6UqtMb0jTwIhAKCMi4Yk6X1WnemGh/cY/NVxPxjnLsoTjbOxwxEIL66j"}]},"_npmUser":{"name":"mike442144","email":"mike442144@hotmail.com"},"directories":{},"maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"_npmOperationalInternal":{"host":"s3://npm-registry-packages-npm-production","tmp":"tmp/crawler_2.0.3-beta.3_1748424960856_0.11232684138716498"},"_hasShrinkwrap":false}},"time":{"created":"2012-09-10T08:41:01.230Z","modified":"2025-05-28T09:36:01.187Z","0.0.1":"2012-09-10T08:41:01.230Z","0.0.2":"2012-09-10T08:41:01.230Z","0.2.0":"2012-12-06T16:07:37.238Z","0.2.1":"2012-12-06T21:31:17.775Z","0.2.2":"2012-12-09T19:18:13.728Z","0.2.3":"2013-02-22T08:55:39.932Z","0.2.4":"2013-02-26T14:57:36.485Z","0.2.5":"2013-08-01T16:14:29.802Z","0.2.6":"2014-07-04T08:23:19.072Z","0.2.7":"2014-09-04T08:08:06.445Z","0.3.0":"2014-10-22T00:43:17.165Z","0.3.1":"2014-10-23T06:00:02.473Z","0.4.0":"2014-11-08T03:19:51.793Z","0.4.1":"2014-11-14T05:53:37.073Z","0.4.2":"2015-02-20T00:56:09.988Z","0.4.3":"2015-03-17T00:54:46.102Z","1.0.0":"2016-12-01T10:50:43.686Z","1.0.1":"2016-12-09T14:14:35.561Z","1.0.2":"2017-01-03T03:50:57.747Z","1.0.3":"2017-02-21T05:33:21.175Z","1.0.4":"2017-04-10T05:35:35.920Z","1.0.5":"2017-05-10T11:51:47.863Z","1.1.1":"2017-10-20T02:35:33.965Z","1.1.2":"2017-12-11T11:38:38.849Z","1.1.3":"2018-05-23T13:19:13.653Z","1.1.4":"2018-05-31T10:45:20.983Z","1.2.0":"2018-08-14T13:32:19.843Z","1.2.1":"2019-04-16T09:10:47.345Z","1.2.2":"2020-06-09T14:02:21.101Z","1.3.0":"2020-11-23T13:28:00.922Z","1.4.0":"2022-12-30T02:39:43.847Z","1.5.0":"2024-01-04T08:16:33.372Z","2.0.0-beta.0":"2024-05-10T02:35:29.072Z","2.0.0-beta.1":"2024-05-11T09:32:39.266Z","2.0.0-beta.3":"2024-05-14T03:38:24.434Z","2.0.0-beta.4":"2024-05-15T10:15:03.616Z","2.0.0-beta.5":"2024-05-27T04:53:55.044Z","2.0.0":"2024-06-06T04:55:07.952Z","2.0.1":"2024-06-25T02:33:21.548Z","2.0.2":"2024-07-16T04:36:32.938Z","2.0.3-beta.1":"2025-04-27T07:51:27.202Z","2.0.3-beta.2":"2025-04-27T08:17:03.014Z","2.0.3-beta.3":"2025-05-28T09:36:01.035Z"},"bugs":{"url":"https://github.com/bda-research/node-crawler/issues"},"license":"MIT","homepage":"https://github.com/bda-research/node-crawler#readme","keywords":["javascript","crawler","spider","scraper","scraping","jquery","nodejs","http","https","http2","got","request","url","network","gzip"],"repository":{"url":"git+https://github.com/bda-research/node-crawler.git","type":"git"},"description":"Crawler is a ready-to-use web spider that works with proxies, asynchrony, rate limit, configurable request pools, jQuery, and HTTP/2 support.","maintainers":[{"name":"sylvinus","email":"sylvain@sylvainzimmer.com"},{"name":"paulvalla","email":"bonjour@pol.ninja"},{"name":"mike442144","email":"mike442144@hotmail.com"},{"name":"darrenqc","email":"darrenqc823@gmail.com"}],"readme":"","readmeFilename":"","users":{"yi":true,"etiv":true,"meeh":true,"nuer":true,"oldj":true,"pana":true,"astro":true,"ealen":true,"fissl":true,"hsian":true,"ineva":true,"mecal":true,"nelix":true,"ntl88":true,"scl33":true,"skrdv":true,"stany":true,"ukuli":true,"456wyc":true,"d-band":true,"dainov":true,"fpmweb":true,"gkodes":true,"googya":true,"iuykza":true,"kbakba":true,"mimmo1":true,"nuwaio":true,"womjoy":true,"8legged":true,"asaupup":true,"davidrc":true,"eludwig":true,"eterna2":true,"hugohua":true,"jculvey":true,"jerrywu":true,"kparkov":true,"mares29":true,"parties":true,"pdedkov":true,"rkielty":true,"ahvonenj":true,"asfrom30":true,"edum1978":true,"joeyblue":true,"leonning":true,"lucazweb":true,"mayorova":true,"me6iaton":true,"moimikey":true,"yxz_blue":true,"anmol1771":true,"arnoldask":true,"benjamins":true,"devonning":true,"dragonken":true,"kulakowka":true,"laserblue":true,"maalthous":true,"mastayoda":true,"sternelee":true,"timmytown":true,"villadora":true,"ahmadazimi":true,"jussipekka":true,"mike442144":true,"tomjamescn":true,"acarlos1000":true,"diegoperini":true,"ecilteodoro":true,"esraa-ammar":true,"wangnan0610":true,"yanqirenshi":true,"zafaransari":true,"timothykrell":true,"developerworks":true,"yoshimaa-tricot":true}}