Added Zwarwald and AhoiPolloi

This commit is contained in:
Bastian Kleineidam 2013-03-07 23:51:55 +01:00
parent 8259a01d64
commit 1d7410c038
4 changed files with 30 additions and 4 deletions

View file

@ -1,9 +1,9 @@
Dosage 1.13 (released xx.xx.2013) Dosage 1.13 (released xx.xx.2013)
Features: Features:
- comics: Added comic strips AxeCop, Bearmageddon, DeadWinter, - comics: Added comic strips AhoiPolloi, AxeCop, Bearmageddon, DeadWinter,
HarkAVagrant, IAmArg, LoadingArtist, Nnewts, PHDComics, PokeyThePenguin, HarkAVagrant, IAmArg, LoadingArtist, Nnewts, PHDComics, PokeyThePenguin,
SnowFlame and WorldOfMrToast. SnowFlame and WorldOfMrToast, Zwarwald.
Fixes: Fixes:
- cmdline: Catch error when piping output to another - cmdline: Catch error when piping output to another

View file

@ -49,6 +49,20 @@ class AcademyVale(_BasicScraper):
help = 'Index format: nnn' help = 'Index format: nnn'
class AhoiPolloi(_BasicScraper):
url = 'http://ahoipolloi.blogger.de/'
stripUrl = url + '?day=%s'
firstStripUrl = stripUrl % '20060305'
multipleImagesPerStrip = True
imageSearch = compile(tagre('img', 'src', r'(/static/antville/ahoipolloi/images/[^"]+)'))
prevSearch = compile(tagre('a', 'href', r'(http://ahoipolloi\.blogger\.de/\?day=\d+)'))
help = 'Index format: yyyymmdd'
@classmethod
def namer(cls, imageUrl, pageUrl):
return imageUrl.rsplit('/', 1)[1]
class ALessonIsLearned(_BasicScraper): class ALessonIsLearned(_BasicScraper):
url = 'http://www.alessonislearned.com/' url = 'http://www.alessonislearned.com/'
prevSearch = compile(tagre("a", "href", r"(index\.php\?comic=\d+)", quote="'")+r"[^>]+previous") prevSearch = compile(tagre("a", "href", r"(index\.php\?comic=\d+)", quote="'")+r"[^>]+previous")
@ -279,7 +293,7 @@ class AxeCop(_BasicScraper):
starter = indirectStarter(url, compile(tagre("a", "href", r'(http://axecop\.com/index\.php/acepisodes/read/episode_\d+/)'))) starter = indirectStarter(url, compile(tagre("a", "href", r'(http://axecop\.com/index\.php/acepisodes/read/episode_\d+/)')))
stripUrl = url + 'index.php/acepisodes/read/episode_%s/' stripUrl = url + 'index.php/acepisodes/read/episode_%s/'
firstStripUrl = stripUrl % '0' firstStripUrl = stripUrl % '0'
imageSearch = compile(tagre("img", "src", r'(http://axecop\.com/images/uploads/axecop[^"]+)')) imageSearch = compile(tagre("img", "src", r'(http://axecop\.com/images/uploads/(?:axecop|acmarried)[^"]+)'))
prevSearch = compile(tagre("a", "href", r'(http://axecop\.com/index\.php/acepisodes/read/episode_\d+/)') + prevSearch = compile(tagre("a", "href", r'(http://axecop\.com/index\.php/acepisodes/read/episode_\d+/)') +
tagre("img", "src", r'http://axecop\.com/acimages/buttons/page_left\.png')) tagre("img", "src", r'http://axecop\.com/acimages/buttons/page_left\.png'))
help = 'Index format: number' help = 'Index format: number'

View file

@ -45,3 +45,15 @@ class ZombieHunters(_BasicScraper):
imageSearch = compile(tagre("img", "src", r'(/istrip_files/strips/[^"]+)')) imageSearch = compile(tagre("img", "src", r'(/istrip_files/strips/[^"]+)'))
prevSearch = compile(tagre("a", "href", r'(\?strip_id=\d+)') + tagre("img", "id", "prevcomic")) prevSearch = compile(tagre("a", "href", r'(\?strip_id=\d+)') + tagre("img", "id", "prevcomic"))
help = 'Index format: n(unpadded)' help = 'Index format: n(unpadded)'
class Zwarwald(_BasicScraper):
url = "http://www.zwarwald.de/"
stripUrl = url + 'index.php/page/%s/'
imageSearch = compile(tagre("img", "src", r'(http://www\.zwarwald\.de/images/\d+/\d+/[^"]+)'))
prevSearch = compile(tagre("a", "href", r'(http://www\.zwarwald\.de/index\.php/page/\d+/)') +
tagre("img", "src", r'http://zwarwald\.de/images/prev\.jpg', quote="'"))
help = 'Index format: number'
def shouldSkipUrl(self, url):
return url in (stripUrl % "112",)

View file

@ -122,7 +122,7 @@ def generate_comic_testers():
g = globals() g = globals()
if "TRAVIS" in os.environ: if "TRAVIS" in os.environ:
# Get limited number of scraper tests on Travis builds. # Get limited number of scraper tests on Travis builds.
max_scrapers = 200 max_scrapers = 300
scraperclasses = islice(scraper.get_scraperclasses(), 0, max_scrapers) scraperclasses = islice(scraper.get_scraperclasses(), 0, max_scrapers)
else: else:
scraperclasses = scraper.get_scraperclasses() scraperclasses = scraper.get_scraperclasses()