Checking if your kit is complete... Looks good Writing Makefile for Scraper cp lib/WWW/Scraper/Opcode/TR.pm blib/lib/WWW/Scraper/Opcode/TR.pm cp lib/WWW/Scraper/Response/ScraperDiscovery/SELECT.pm blib/lib/WWW/Scraper/Response/ScraperDiscovery/SELECT.pm cp lib/WWW/Scraper/Opcode/GRUB.pm blib/lib/WWW/Scraper/Opcode/GRUB.pm cp lib/WWW/Scraper/ScraperDiscovery.pm blib/lib/WWW/Scraper/ScraperDiscovery.pm cp lib/WWW/Scraper/Response.pm blib/lib/WWW/Scraper/Response.pm cp lib/WWW/Scraper/Opcode/MACRO.pm blib/lib/WWW/Scraper/Opcode/MACRO.pm cp lib/WWW/Scraper/Response/ScraperDiscovery/OPTION.pm blib/lib/WWW/Scraper/Response/ScraperDiscovery/OPTION.pm cp lib/WWW/Scraper/Opcode.pm blib/lib/WWW/Scraper/Opcode.pm cp lib/WWW/Scraper/Opcode/FORM.pm blib/lib/WWW/Scraper/Opcode/FORM.pm cp lib/WWW/Scraper/Opcode/DISCOVERNEXT.pm blib/lib/WWW/Scraper/Opcode/DISCOVERNEXT.pm cp lib/WWW/Scraper/Dogpile.pm blib/lib/WWW/Scraper/Dogpile.pm cp lib/WWW/Scraper/TidyXML.pm blib/lib/WWW/Scraper/TidyXML.pm cp lib/WWW/Scraper/WSDL.pm blib/lib/WWW/Scraper/WSDL.pm cp lib/WWW/Scraper/CraigsList.pm blib/lib/WWW/Scraper/CraigsList.pm cp lib/WWW/Scraper/ZIPplus4.pm blib/lib/WWW/Scraper/ZIPplus4.pm cp lib/WWW/Scraper/Request.pm blib/lib/WWW/Scraper/Request.pm cp lib/WWW/Scraper/Opcode/REGEX.pm blib/lib/WWW/Scraper/Opcode/REGEX.pm cp lib/WWW/Scraper/Response/ScraperDiscovery/INPUT.pm blib/lib/WWW/Scraper/Response/ScraperDiscovery/INPUT.pm cp lib/WWW/Scraper/Opcode/MACROX.pm blib/lib/WWW/Scraper/Opcode/MACROX.pm cp lib/WWW/Scraper/FieldTranslation.pm blib/lib/WWW/Scraper/FieldTranslation.pm cp lib/WWW/Scraper/Opcode/BR.pm blib/lib/WWW/Scraper/Opcode/BR.pm cp lib/WWW/Scraper/Opcode/A.pm blib/lib/WWW/Scraper/Opcode/A.pm cp lib/WWW/Scraper/Opcode/INPUT.pm blib/lib/WWW/Scraper/Opcode/INPUT.pm cp lib/WWW/Scraper/Response/Sherlock.pm blib/lib/WWW/Scraper/Response/Sherlock.pm cp lib/WWW/Scraper.pm blib/lib/WWW/Scraper.pm cp lib/WWW/Scraper/Response/generic.pm blib/lib/WWW/Scraper/Response/generic.pm cp lib/WWW/Scraper/Sherlock.pm blib/lib/WWW/Scraper/Sherlock.pm cp lib/WWW/Scraper/CraigsList.Job.locations blib/lib/WWW/Scraper/CraigsList.Job.locations cp lib/WWW/Scraper/Lycos.pm blib/lib/WWW/Scraper/Lycos.pm cp lib/WWW/Scraper/Beaucoup.pm blib/lib/WWW/Scraper/Beaucoup.pm cp lib/WWW/Scraper/Google.pm blib/lib/WWW/Scraper/Google.pm cp lib/WWW/Scraper/Request/ZIPplus4.pm blib/lib/WWW/Scraper/Request/ZIPplus4.pm cp lib/WWW/Scraper/NorthernLight.pm blib/lib/WWW/Scraper/NorthernLight.pm cp lib/WWW/Scraper/Opcode/NEXT.pm blib/lib/WWW/Scraper/Opcode/NEXT.pm cp lib/WWW/Scraper/Grub.pm blib/lib/WWW/Scraper/Grub.pm cp lib/WWW/Scraper/Response/ScraperDiscovery.pm blib/lib/WWW/Scraper/Response/ScraperDiscovery.pm cp lib/WWW/Scraper/Opcode/BODY.pm blib/lib/WWW/Scraper/Opcode/BODY.pm cp prereqinst.pl blib/lib/prereqinst.pl cp lib/WWW/Scraper/Opcode/FONT.pm blib/lib/WWW/Scraper/Opcode/FONT.pm cp lib/WWW/Scraper/Response/ScraperDiscovery/FORM.pm blib/lib/WWW/Scraper/Response/ScraperDiscovery/FORM.pm cp lib/WWW/Scraper/Opcode/TD.pm blib/lib/WWW/Scraper/Opcode/TD.pm cp lib/WWW/ScraperPOD.pm blib/lib/WWW/ScraperPOD.pm cp lib/WWW/Scraper/Opcode/OPTION.pm blib/lib/WWW/Scraper/Opcode/OPTION.pm cp lib/WWW/Scraper/Opcode/REWRITE.pm blib/lib/WWW/Scraper/Opcode/REWRITE.pm cp lib/WWW/Scraper/Opcode/SELECT.pm blib/lib/WWW/Scraper/Opcode/SELECT.pm cp lib/WWW/Scraper/Opcode/TABLE.pm blib/lib/WWW/Scraper/Opcode/TABLE.pm cp lib/WWW/Scraper/Request/WSDL.pm blib/lib/WWW/Scraper/Request/WSDL.pm Manifying blib/man3/WWW::Scraper::Google.3 Manifying blib/man3/WWW::Scraper::Response::ScraperDiscovery::SELECT.3 Manifying blib/man3/WWW::Scraper::Request::ZIPplus4.3 Manifying blib/man3/WWW::Scraper::ScraperDiscovery.3 Manifying blib/man3/WWW::Scraper::Response::ScraperDiscovery::OPTION.3 Manifying blib/man3/WWW::Scraper::Response.3 Manifying blib/man3/WWW::Scraper::Opcode.3 Manifying blib/man3/WWW::Scraper::NorthernLight.3 Manifying blib/man3/WWW::Scraper::Dogpile.3 Manifying blib/man3/WWW::Scraper::TidyXML.3 Manifying blib/man3/WWW::Scraper::WSDL.3 Manifying blib/man3/WWW::Scraper::ZIPplus4.3 Manifying blib/man3/WWW::Scraper::CraigsList.3 Manifying blib/man3/WWW::Scraper::Response::ScraperDiscovery::INPUT.3 Manifying blib/man3/WWW::Scraper::Request.3 Manifying blib/man3/WWW::Scraper::Response::ScraperDiscovery.3 Manifying blib/man3/WWW::Scraper::Grub.3 Manifying blib/man3/WWW::Scraper::FieldTranslation.3 Manifying blib/man3/WWW::Scraper::Response::ScraperDiscovery::FORM.3 Manifying blib/man3/WWW::Scraper::Response::Sherlock.3 Manifying blib/man3/WWW::Scraper.3 Manifying blib/man3/WWW::Scraper::Response::generic.3 Manifying blib/man3/WWW::ScraperPOD.3 Manifying blib/man3/WWW::Scraper::Sherlock.3 Manifying blib/man3/WWW::Scraper::Lycos.3 Manifying blib/man3/WWW::Scraper::Beaucoup.3 Manifying blib/man3/WWW::Scraper::Request::WSDL.3 PERL_DL_NONLAZY=1 /home/cpanrun/ia64/build/5.8.2/bin/perl "-MExtUtils::Command::MM" "-e" "test_harness(1, 'blib/lib', 'blib/arch')" t/*.t t/t....1..1 ok 1 ok All tests successful. Files=1, Tests=1, 0 wallclock secs ( 0.01 cusr + 0.01 csys = 0.02 CPU) PERL_DL_NONLAZY=1 /home/cpanrun/ia64/build/5.8.2/bin/perl "-Iblib/lib" "-Iblib/arch" test.pl # VERSIONS OF MODULES ON WHICH SCRAPER DEPENDS # using HTML::Form(1.038); # using HTML::TreeBuilder(3.13); # using HTTP::Cookies(1.36); # using HTTP::Request(1.34); # using HTTP::Response(1.41); # using HTTP::Status(1.28); # using LWP(5.75); # using LWP::RobotUA(1.23); # using LWP::UserAgent(2.023); # using Storable(2.08); # using Text::ParseWords(3.21); # using Tie::Persistent(1.00); # using URI(1.27); # using URI::Escape(3.21); # using URI::URL(5.03); # using URI::http(); # using User(1.8); # using WWW::Search(2.550); # using XML::XPath(1.13); 1..11 ok 1 - 9 Scraper modules listed in MANIFEST (4,0,5) ok 2 - WWW::Scraper loaded ok 3 # skip This Scraper engine requires 'HTML Tidy' to scrub HTML before parsing. Get this program from 'http://tidy.sourceforge.net/docs/Overview.html#Download' Make sure it is in your execution search path. not ok 4 - CraigsList # Failed test (test.pl at line 129) ok 5 # skip This Scraper engine requires 'HTML Tidy' to scrub HTML before parsing. Get this program from 'http://tidy.sourceforge.net/docs/Overview.html#Download' Make sure it is in your execution search path. not ok 6 - Google # Failed test (test.pl at line 129) not ok 7 - Lycos # Failed test (test.pl at line 129) ok 8 # skip NorthernLight's search engine seems to be down these days!? ok 9 # skip Sherlock.pm is not working today; many changes out on the frontier that I haven't caught up with yet ok 10 # skip WSDL is still in a primitive experimental state. not ok 11 - ZIPplus4 # Failed test (test.pl at line 129) # Can't locate object method "zipcode" via package "WWW::Scraper::Request::ZIPplus4_" at blib/lib/WWW/Scraper/Request.pm line 207. # 4 tests had problems. See file 'test.trace' for details. ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## Operating system: hpux Perl version: 5.008002 VERSIONS OF MODULES ON WHICH SCRAPER DEPENDS using HTML::Form(1.038); using HTML::TreeBuilder(3.13); using HTTP::Cookies(1.36); using HTTP::Request(1.34); using HTTP::Response(1.41); using HTTP::Status(1.28); using LWP(5.75); using LWP::RobotUA(1.23); using LWP::UserAgent(2.023); using Storable(2.08); using Text::ParseWords(3.21); using Tie::Persistent(1.00); using URI(1.27); using URI::Escape(3.21); using URI::URL(5.03); using URI::http(); using User(1.8); using WWW::Search(2.550); using XML::XPath(1.13); ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## LIST SCRAPER SUB-CLASSES, FROM THE MANIFEST + Beaucoup(1.07) + CraigsList(1.16) + Dogpile(1.11) - FieldTranslation will not be tested: it is not a Scraper sub-class. + Google(1.23) - Grub will not be tested: it is not a Scraper sub-class. + Lycos(1.00) + NorthernLight(1.00) + Sherlock(1.00) - Response will not be tested: it is not a Scraper sub-class. - Request will not be tested: it is not a Scraper sub-class. - TidyXML will not be tested: it is not a Scraper sub-class. + WSDL(1.00) + ZIPplus4(1.09) - Opcode will not be tested: it is not a Scraper sub-class. - ScraperDiscovery will not be tested: it is not a Scraper sub-class. ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## Test #0: CraigsList Test #1: CraigsList 'bogus' search Test #2: CraigsList one-page search + got 0 results for 'Honda' --- got 0 results for CraigsList 'Honda', but expected 50 --- base URL: http://www.craigslist.org/cgi-bin/search? --- first URL: http://www.craigslist.org/cgi-bin/search?areaID=1&cat=all&catAbbreviation=car&group=J&max_ask=&min_ask=&new_cat=6&query=Honda&subAreaID=0&type_search= --- last URL: http://www.craigslist.org/cgi-bin/search?areaID=1&cat=all&catAbbreviation=car&group=J&max_ask=&min_ask=&new_cat=6&query=Honda&subAreaID=0&type_search= --- next URL: --- response message: 200 OK --- content size (bytes): 2150 --- ERRNO: --- Extended OS error: Scraper engine CraigsList failed once: Test #0: CraigsList Test #1: CraigsList 'bogus' search Test #2: CraigsList one-page search + got 0 results for 'Honda' --- got 0 results for CraigsList 'Honda', but expected 50 --- base URL: http://www.craigslist.org/cgi-bin/search? --- first URL: http://www.craigslist.org/cgi-bin/search?areaID=1&cat=all&catAbbreviation=car&group=J&max_ask=&min_ask=&new_cat=6&query=Honda&subAreaID=0&type_search= --- last URL: http://www.craigslist.org/cgi-bin/search?areaID=1&cat=all&catAbbreviation=car&group=J&max_ask=&min_ask=&new_cat=6&query=Honda&subAreaID=0&type_search= --- next URL: --- response message: 200 OK --- content size (bytes): 2150 --- ERRNO: --- Extended OS error: Scraper engine CraigsList failed twice: ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## Test #0: Google Test #1: Google 'bogus' search Test #2: Google one-page search + got 0 results for 'search scraper' --- got 0 results for Google 'search scraper', but expected 9 --- base URL: http://www.google.com/search? --- first URL: http://www.google.com/search?btnG=Google+Search&hl=en&ie=UTF%2D8&lr=&q=search+scraper&safe=active --- last URL: http://www.google.com/search?btnG=Google+Search&hl=en&ie=UTF%2D8&lr=&q=search+scraper&safe=active --- next URL: --- response message: 200 OK --- content size (bytes): 23136 --- ERRNO: --- Extended OS error: Scraper engine Google failed once: Test #0: Google Test #1: Google 'bogus' search Test #2: Google one-page search + got 0 results for 'search scraper' --- got 0 results for Google 'search scraper', but expected 9 --- base URL: http://www.google.com/search? --- first URL: http://www.google.com/search?btnG=Google+Search&hl=en&ie=UTF%2D8&lr=&q=search+scraper&safe=active --- last URL: http://www.google.com/search?btnG=Google+Search&hl=en&ie=UTF%2D8&lr=&q=search+scraper&safe=active --- next URL: --- response message: 200 OK --- content size (bytes): 22439 --- ERRNO: --- Extended OS error: Scraper engine Google failed twice: ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## Test #0: Lycos Test #1: Lycos 'bogus' search Test #2: Lycos one-page search + got 0 results for 'turntable' --- got 0 results for Lycos 'turntable', but expected 9 --- base URL: http://search.lycos.com/default.asp? --- first URL: http://search.lycos.com/default.asp?loc=searchhp&lpv=1&query=turntable&tab=web --- last URL: http://search.lycos.com/default.asp?loc=searchhp&lpv=1&query=turntable&tab=web --- next URL: --- response message: 200 OK --- content size (bytes): 53897 --- ERRNO: --- Extended OS error: Scraper engine Lycos failed once: Test #0: Lycos Test #1: Lycos 'bogus' search Test #2: Lycos one-page search + got 0 results for 'turntable' --- got 0 results for Lycos 'turntable', but expected 9 --- base URL: http://search.lycos.com/default.asp? --- first URL: http://search.lycos.com/default.asp?loc=searchhp&lpv=1&query=turntable&tab=web --- last URL: http://search.lycos.com/default.asp?loc=searchhp&lpv=1&query=turntable&tab=web --- next URL: --- response message: 200 OK --- content size (bytes): 53649 --- ERRNO: --- Extended OS error: Scraper engine Lycos failed twice: ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## ##_##_##_##_##_##_##_##_##_##_##_##_##_##_## Test #0: ZIPplus4 Test #1: ZIPplus4 'bogus' search Scraper engine ZIPplus4 failed once: Can't locate object method "zipcode" via package "WWW::Scraper::Request::ZIPplus4_" at blib/lib/WWW/Scraper/Request.pm line 207. Test #0: ZIPplus4 Test #1: ZIPplus4 'bogus' search Scraper engine ZIPplus4 failed twice: Can't locate object method "zipcode" via package "WWW::Scraper::Request::ZIPplus4_" at blib/lib/WWW/Scraper/Request.pm line 207. # Looks like you failed 4 tests of 11. make: *** [test_dynamic] Error 4