Hi every one, How could I miss crawler 2.0 and posted 3.0 before this. Here I am posting 2.0 crawler with multiprocess facility.. šŸ˜‰ Actually 3.0 thread based crawler was easy to develop, and now it is the time for release of final 2.0.

Why I am making crawler,Ā  actually me and my friends Abhijeet and Zainab were thinking of making basic search engine. But we know there are already better than our’s. Then we thought we can do some better with this crawler thing, and now one more guy joined us, Mr Nirav quite high skilled person and work on highly critical projects.

Now I am more sure of finishing all this in time and make an automatic system that will post all new thing on bestindianwear.com. We can say it will be a basic AI (Artificial Intelligent) project. Abhijeet is working quite hard on it

Thanks Guys – I do not feel alone, and your efforts make our way enjoyable. Cheers to everyone we will be finishing this soon… šŸ™‚

https://github.com/vishvendrasingh/crawler/blob/master/crawler_2.0_stable.py

0

Hi Guys,

Yes as you read above, one of my colleague asked me what would be the speed of parallel, thread based crawler, Now I am posting this to so that you all can check out the speed. Now how to use this, its very simple, it is written in the file itself. check it out….. šŸ˜‰ enjoy and let me sleep now!

https://github.com/vishvendrasingh/crawler/blob/master/crawler_3.0_stable.py

0

Crazy day, I indexed 30GB file having 53 million lines of json data to elastic. Then I tried kibana with it it was really enjoyable after doing it with my drink. Link to kibana is shivalink.com:5601.

Link to exastic is shivalink.com:9200

the most tough was to unzip 5GB file using all cores, it was bz2 file. I used pbzip2 but it didn’t worked in my case. Then I found lbzip2 -d myfile.json. It was really fast and used my all cores efficiently. It turned out to be 30GB then. After that how could we insert it to elastic, as I am very new to this I found esbulk and started with this. I inserted 45 million entries then It became too slow. Now I had no option other and stopping it right there.

Than I came up with new idea of tail -n No of rest of the entries and inserted them back. I successfully did it. Now I can say I kind of know big big data….. šŸ™‚ feeling happy

0

crawler

Completed coding of recursive crawler, it was fun and a lot of hard work, some meditation, and lots of google. I finally did it. My friend Abhijeet asked to make recursive crawler and I was thinking how can I do that. So came up with this idea wo making two lists

1. processed list (All crawled urls are stored here)

2. unprocessed list (All new url are stored here)

Now if a new url exists in any of these lists then skip it and move furthur. Happy crawling guys…..:)

This program do the following thing

  1. store data in mongodb
  2. parse html in page title, meta data, meta keywords
  3. In case if page request fails error handling save it from breaking
  4. it does not follow any other domain except the given one

Here is the link https://github.com/vishvendrasingh/crawler.git

0

BRTS –Ā FastĀ busĀ serviceĀ inĀ ahmedabad ,Ā isĀ havingĀ theĀ bestĀ busĀ drivers. They drive very fast & save ass of @$/k guys onĀ road.Ā PeopleĀ hereĀ driveĀ inĀ BRTSĀ VIPĀ corridorĀ also,Ā whereĀ theyĀ shouldĀ notĀ neitherĀ permitted. TodayĀ BRTSĀ busĀ driverĀ saved aĀ kidĀ byĀ applyingĀ breaksĀ onĀ time &Ā waitedĀ forĀ himĀ untilĀ heĀ crossedĀ theĀ road.

0