Convert all tests to py.test & cleanups.
This commit is contained in:
parent
b6a6a34a44
commit
2ca74d6e6b
9 changed files with 171 additions and 191 deletions
|
@ -3,6 +3,7 @@ python:
|
|||
- "2.7"
|
||||
- "3.3"
|
||||
- "3.4"
|
||||
- "3.5"
|
||||
# install libjpeg-dev for Pillow to handle JPEGs
|
||||
sudo: false
|
||||
addons:
|
||||
|
@ -14,9 +15,9 @@ addons:
|
|||
# command to install dependencies
|
||||
install:
|
||||
- pip install -r requirements.txt
|
||||
- pip install pytest-xdist
|
||||
- pip install pytest-xdist pytest-cov
|
||||
# command to run tests
|
||||
script: make test PYTESTOPTS="--tb=short -n10"
|
||||
script: make test PYTESTOPTS="--cov=dosage --cov=dosagelib --tb=short -n10"
|
||||
notifications:
|
||||
irc:
|
||||
channels:
|
||||
|
|
|
@ -1,22 +1,12 @@
|
|||
# -*- coding: iso-8859-1 -*-
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2013-2014 Bastian Kleineidam
|
||||
#
|
||||
# This program is free software: you can redistribute it and/or modify
|
||||
# it under the terms of the GNU General Public License as published by
|
||||
# the Free Software Foundation, either version 3 of the License, or
|
||||
# (at your option) any later version.
|
||||
#
|
||||
# This program is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
# GNU General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU General Public License
|
||||
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
# Copyright (C) 2016 Tobias Gruetzmacher
|
||||
|
||||
import os
|
||||
import subprocess
|
||||
import sys
|
||||
import pytest
|
||||
import shutil
|
||||
import subprocess
|
||||
import tempfile
|
||||
|
||||
basedir = os.path.dirname(__file__)
|
||||
dosage_cmd = os.path.join(os.path.dirname(basedir), "dosage")
|
||||
|
@ -40,24 +30,8 @@ def run_checked (cmd, ret_ok=(0,), **kwargs):
|
|||
return retcode
|
||||
|
||||
|
||||
# Python 3.x renamed the function name attribute
|
||||
if sys.version_info[0] > 2:
|
||||
fnameattr = '__name__'
|
||||
else:
|
||||
fnameattr = 'func_name'
|
||||
|
||||
def _need_func(testfunc, name, description):
|
||||
"""Decorator skipping test if given testfunc returns False."""
|
||||
def check_func(func):
|
||||
def newfunc(*args, **kwargs):
|
||||
if not testfunc(name):
|
||||
raise pytest.skip("%s %r is not available" % (description, name))
|
||||
return func(*args, **kwargs)
|
||||
setattr(newfunc, fnameattr, getattr(func, fnameattr))
|
||||
return newfunc
|
||||
return check_func
|
||||
|
||||
|
||||
def needs_os(name):
|
||||
"""Decorator skipping test if given operating system is not available."""
|
||||
return _need_func(lambda x: os.name == x, name, 'operating system')
|
||||
@pytest.yield_fixture
|
||||
def tmpdir():
|
||||
tmpdir = tempfile.mkdtemp()
|
||||
yield tmpdir
|
||||
shutil.rmtree(tmpdir)
|
||||
|
|
|
@ -1,21 +1,18 @@
|
|||
# -*- coding: iso-8859-1 -*-
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2012-2014 Bastian Kleineidam
|
||||
from unittest import TestCase
|
||||
# Copyright (C) 2016 Tobias Gruetzmacher
|
||||
|
||||
from dosagelib import scraper, util
|
||||
try:
|
||||
text_type = unicode
|
||||
except NameError:
|
||||
text_type = str
|
||||
|
||||
|
||||
class TestComicNames(TestCase):
|
||||
class TestComicNames(object):
|
||||
|
||||
def test_names(self):
|
||||
for scraperclass in scraper.get_scraperclasses():
|
||||
name = scraperclass.getName()
|
||||
self.assertTrue(name.count('/') <= 1, name)
|
||||
assert name.count('/') <= 1
|
||||
if '/' in name:
|
||||
comicname = name.split('/')[1]
|
||||
else:
|
||||
comicname = name
|
||||
self.assertEqual(util.asciify(comicname), comicname)
|
||||
assert util.asciify(comicname) == comicname
|
||||
|
|
|
@ -1,18 +1,17 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2004-2005 Tristan Seligmann and Jonathan Jacobs
|
||||
# Copyright (C) 2012-2014 Bastian Kleineidam
|
||||
# Copyright (C) 2015 Tobias Gruetzmacher
|
||||
import tempfile
|
||||
import shutil
|
||||
# Copyright (C) 2015-2016 Tobias Gruetzmacher
|
||||
|
||||
import re
|
||||
import os
|
||||
import multiprocessing
|
||||
import pytest
|
||||
try:
|
||||
from urllib.parse import urlsplit
|
||||
except ImportError:
|
||||
from urlparse import urlsplit
|
||||
from dosagelib import scraper
|
||||
from . import tmpdir # noqa
|
||||
|
||||
|
||||
def get_host(url):
|
||||
|
@ -24,6 +23,9 @@ def get_host(url):
|
|||
_locks = {}
|
||||
# Allowed number of connections per host
|
||||
MaxConnections = 4
|
||||
# Maximum number of strips to get to test a comic
|
||||
MaxStrips = 5
|
||||
|
||||
|
||||
def get_lock(host):
|
||||
"""Get bounded semphore for given host."""
|
||||
|
@ -31,24 +33,18 @@ def get_lock(host):
|
|||
_locks[host] = multiprocessing.BoundedSemaphore(MaxConnections)
|
||||
return _locks[host]
|
||||
|
||||
@pytest.yield_fixture
|
||||
def tmpdir():
|
||||
tmpdir = tempfile.mkdtemp()
|
||||
yield tmpdir
|
||||
shutil.rmtree(tmpdir)
|
||||
|
||||
def get_saved_images(tmpdir, scraper, filtertxt=False):
|
||||
def _get_saved_images(outdir, scraper):
|
||||
"""Get saved images."""
|
||||
dirs = tuple(scraper.getName().split('/'))
|
||||
files = os.listdir(os.path.join(tmpdir, *dirs))
|
||||
if filtertxt:
|
||||
files = os.listdir(os.path.join(outdir, *dirs))
|
||||
files = [x for x in files if not x.endswith(".txt")]
|
||||
return files
|
||||
|
||||
def test_comicmodule(tmpdir, scraperclass):
|
||||
# Test a scraper. It must be able to traverse backward for
|
||||
# at least 5 strips from the start, and find strip images
|
||||
# on at least 4 pages.
|
||||
|
||||
def test_comicmodule(tmpdir, scraperclass): # noqa
|
||||
'''Test a scraper. It must be able to traverse backward for at least 5
|
||||
strips from the start, and find strip images on at least 4 pages.'''
|
||||
scraperobj = scraperclass()
|
||||
# Limit number of connections to one host.
|
||||
host = get_host(scraperobj.url)
|
||||
|
@ -59,45 +55,59 @@ def test_comicmodule(tmpdir, scraperclass):
|
|||
# interprocess lock not supported
|
||||
_test_comic(tmpdir, scraperobj)
|
||||
|
||||
def _test_comic(tmpdir, scraperobj):
|
||||
|
||||
def _test_comic(outdir, scraperobj):
|
||||
num_strips = 0
|
||||
max_strips = 5
|
||||
strip = None
|
||||
for strip in scraperobj.getStrips(max_strips):
|
||||
images = []
|
||||
for image in strip.getImages():
|
||||
images.append(image.url)
|
||||
image.save(tmpdir)
|
||||
assert images, 'failed to find images at %s' % strip.stripUrl
|
||||
if not scraperobj.multipleImagesPerStrip:
|
||||
assert len(images) == 1, 'found more than 1 image at %s: %s' % (strip.stripUrl, images)
|
||||
for strip in scraperobj.getStrips(MaxStrips):
|
||||
_check_strip(outdir, strip, scraperobj.multipleImagesPerStrip)
|
||||
|
||||
if num_strips > 0 and scraperobj.prevUrlMatchesStripUrl:
|
||||
check_stripurl(strip, scraperobj)
|
||||
_check_stripurl(strip, scraperobj)
|
||||
num_strips += 1
|
||||
|
||||
if scraperobj.prevSearch and not scraperobj.hitFirstStripUrl:
|
||||
# check strips
|
||||
num_strips_expected = max_strips - len(scraperobj.skippedUrls)
|
||||
msg = 'Traversed %d strips instead of %d.' % (num_strips, num_strips_expected)
|
||||
# subtract the number of skipped URLs with no image from the expected
|
||||
# image number
|
||||
num_strips_expected = MaxStrips - len(scraperobj.skippedUrls)
|
||||
msg = 'Traversed %d strips instead of %d.' % (num_strips,
|
||||
num_strips_expected)
|
||||
if strip:
|
||||
msg += " Check the prevSearch pattern at %s" % strip.stripUrl
|
||||
assert num_strips == num_strips_expected, msg
|
||||
# check images
|
||||
if strip:
|
||||
check_scraperesult(tmpdir, num_strips_expected, strip, scraperobj)
|
||||
_check_scraperesult(outdir, num_strips_expected, strip, scraperobj)
|
||||
|
||||
def check_scraperesult(tmpdir, num_images_expected, strip, scraperobj):
|
||||
# Check that exactly or for multiple pages at least num_strips images are saved.
|
||||
# This checks saved files, ie. it detects duplicate filenames.
|
||||
saved_images = get_saved_images(tmpdir, scraperobj, filtertxt=bool(scraperobj.textSearch))
|
||||
|
||||
def _check_strip(outdir, strip, multipleImagesPerStrip):
|
||||
'''Check that a specific page yields images and the comic module correctly
|
||||
declares if there are multiple images per page.'''
|
||||
images = []
|
||||
for image in strip.getImages():
|
||||
images.append(image.url)
|
||||
image.save(outdir)
|
||||
assert images, 'failed to find images at %s' % strip.stripUrl
|
||||
if not multipleImagesPerStrip:
|
||||
assert len(images) == 1, 'found more than 1 image at %s: %s' % (
|
||||
strip.stripUrl, images)
|
||||
|
||||
|
||||
def _check_scraperesult(outdir, num_images_expected, strip, scraperobj):
|
||||
'''Check that exactly or for multiple pages at least num_strips images are
|
||||
saved. This checks saved files, ie. it detects duplicate filenames.'''
|
||||
saved_images = _get_saved_images(outdir, scraperobj)
|
||||
num_images = len(saved_images)
|
||||
# subtract the number of skipped URLs with no image from the expected image number
|
||||
attrs = (num_images, saved_images, num_images_expected, tmpdir)
|
||||
if scraperobj.multipleImagesPerStrip:
|
||||
assert num_images >= num_images_expected, 'saved %d %s instead of at least %d images in %s' % attrs
|
||||
else:
|
||||
assert num_images == num_images_expected, 'saved %d %s instead of %d images in %s' % attrs
|
||||
|
||||
def check_stripurl(strip, scraperobj):
|
||||
attrs = (num_images, saved_images, num_images_expected, outdir)
|
||||
if scraperobj.multipleImagesPerStrip:
|
||||
err = 'saved %d %s instead of at least %d images in %s' % attrs
|
||||
assert num_images >= num_images_expected, err
|
||||
else:
|
||||
err = 'saved %d %s instead of %d images in %s' % attrs
|
||||
assert num_images == num_images_expected, err
|
||||
|
||||
|
||||
def _check_stripurl(strip, scraperobj):
|
||||
if not scraperobj.stripUrl:
|
||||
# no indexing support
|
||||
return
|
||||
|
@ -107,7 +117,10 @@ def check_stripurl(strip, scraperobj):
|
|||
urlmatch = "^%s$" % urlmatch
|
||||
ro = re.compile(urlmatch)
|
||||
mo = ro.search(strip.stripUrl)
|
||||
assert mo is not None, 'strip URL %r does not match stripUrl pattern %s' % (strip.stripUrl, urlmatch)
|
||||
err = 'strip URL %r does not match stripUrl pattern %s' % (
|
||||
strip.stripUrl, urlmatch)
|
||||
assert mo is not None, err
|
||||
|
||||
|
||||
def get_test_scraperclasses():
|
||||
"""Return scrapers that should be tested."""
|
||||
|
@ -117,14 +130,18 @@ def get_test_scraperclasses():
|
|||
else:
|
||||
# Get limited number of scraper tests on Travis builds to make
|
||||
# it faster
|
||||
testscrapernames = ['AbstruseGoose', 'GoComics/CalvinandHobbes', 'xkcd']
|
||||
testscrapernames = [
|
||||
'AbstruseGoose',
|
||||
'GoComics/CalvinandHobbes',
|
||||
'xkcd'
|
||||
]
|
||||
scraperclasses = [
|
||||
scraperclass for scraperclass in scraper.get_scraperclasses()
|
||||
if scraperclass.getName() in testscrapernames
|
||||
]
|
||||
return scraperclasses
|
||||
|
||||
|
||||
def pytest_generate_tests(metafunc):
|
||||
if 'scraperclass' in metafunc.fixturenames:
|
||||
metafunc.parametrize('scraperclass', get_test_scraperclasses())
|
||||
|
||||
|
|
|
@ -1,68 +1,54 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2013-2014 Bastian Kleineidam
|
||||
#
|
||||
# This program is free software: you can redistribute it and/or modify
|
||||
# it under the terms of the GNU General Public License as published by
|
||||
# the Free Software Foundation, either version 3 of the License, or
|
||||
# (at your option) any later version.
|
||||
#
|
||||
# This program is distributed in the hope that it will be useful,
|
||||
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||
# GNU General Public License for more details.
|
||||
#
|
||||
# You should have received a copy of the GNU General Public License
|
||||
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||
import unittest
|
||||
# Copyright (C) 2016 Tobias Gruetzmacher
|
||||
|
||||
import pytest
|
||||
import sys
|
||||
import shutil
|
||||
import tempfile
|
||||
from . import dosage_cmd, run_checked
|
||||
from . import dosage_cmd, run_checked, tmpdir # noqa
|
||||
|
||||
|
||||
def run_with_options(options, cmd=dosage_cmd):
|
||||
"""Run dosage with given options."""
|
||||
run_checked([sys.executable, cmd] + options)
|
||||
run_checked([sys.executable, cmd, '--allow-multiple'] + options)
|
||||
|
||||
|
||||
class TestDosage (unittest.TestCase):
|
||||
class TestDosage(object):
|
||||
"""Test the dosage commandline client."""
|
||||
|
||||
def test_dosage(self):
|
||||
# list comics
|
||||
def test_list_comics(self):
|
||||
for option in ("-l", "--list", "--singlelist"):
|
||||
run_with_options([option])
|
||||
# display version
|
||||
|
||||
def test_display_version(self):
|
||||
run_with_options(["--version"])
|
||||
# display help
|
||||
|
||||
def test_display_help(self):
|
||||
for option in ("-h", "--help"):
|
||||
run_with_options([option])
|
||||
# module help
|
||||
|
||||
def test_module_help(self):
|
||||
run_with_options(["-m", "xkcd"])
|
||||
# no comics specified
|
||||
self.assertRaises(OSError, run_with_options, [])
|
||||
# unknown option
|
||||
self.assertRaises(OSError, run_with_options, ['--imadoofus'])
|
||||
# multiple comics match
|
||||
self.assertRaises(OSError, run_with_options, ['Garfield'])
|
||||
# create a temporary directory for images
|
||||
tmpdir = tempfile.mkdtemp()
|
||||
try:
|
||||
# fetch html and rss
|
||||
run_with_options(["-n", "2", "-v", "-b", tmpdir, "-o", "html", "-o", "rss", "xkcd"])
|
||||
finally:
|
||||
shutil.rmtree(tmpdir)
|
||||
# create a temporary directory for images
|
||||
tmpdir = tempfile.mkdtemp()
|
||||
try:
|
||||
# fetch html and rss 2
|
||||
run_with_options(["--numstrips", "2", "--baseurl", "bla", "--basepath", tmpdir, "--output", "rss", "--output", "html", "--adult", "oglaf"])
|
||||
finally:
|
||||
shutil.rmtree(tmpdir)
|
||||
# create a temporary directory for images
|
||||
tmpdir = tempfile.mkdtemp()
|
||||
try:
|
||||
# fetch indexed
|
||||
|
||||
def test_no_comics_specified(self):
|
||||
with pytest.raises(OSError):
|
||||
run_with_options([])
|
||||
|
||||
def test_unknown_option(self):
|
||||
with pytest.raises(OSError):
|
||||
run_with_options(['--imadoofus'])
|
||||
|
||||
def test_multiple_comics_match(self):
|
||||
with pytest.raises(OSError):
|
||||
run_with_options(['Garfield'])
|
||||
|
||||
def test_fetch_html_and_rss(self, tmpdir): # noqa
|
||||
run_with_options(["-n", "2", "-v", "-b", tmpdir, "-o", "html", "-o",
|
||||
"rss", "xkcd"])
|
||||
|
||||
def test_fetch_html_and_rss_2(self, tmpdir): # noqa
|
||||
run_with_options(["--numstrips", "2", "--baseurl", "bla",
|
||||
"--basepath", tmpdir, "--output", "rss", "--output",
|
||||
"html", "--adult", "oglaf"])
|
||||
|
||||
def test_fetch_indexed(self, tmpdir): # noqa
|
||||
run_with_options(["-n", "2", "-v", "-b", tmpdir, "xkcd:303"])
|
||||
finally:
|
||||
shutil.rmtree(tmpdir)
|
||||
|
|
|
@ -1,27 +1,28 @@
|
|||
# -*- coding: iso-8859-1 -*-
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2013-2014 Bastian Kleineidam
|
||||
# Copyright (C) 2015 Tobias Gruetzmacher
|
||||
from unittest import TestCase
|
||||
# Copyright (C) 2015-2016 Tobias Gruetzmacher
|
||||
|
||||
import pytest
|
||||
from dosagelib import scraper
|
||||
|
||||
|
||||
class ScraperTester(TestCase):
|
||||
class TestScraper(object):
|
||||
"""Test scraper module functions."""
|
||||
|
||||
def test_get_scraperclasses(self):
|
||||
for scraperclass in scraper.get_scraperclasses():
|
||||
scraperobj = scraperclass()
|
||||
scraperobj = scraperclass(indexes=["bla"])
|
||||
self.assertTrue(scraperobj.url,
|
||||
"missing url in %s" % scraperobj.getName())
|
||||
assert scraperobj.url, "missing url in %s" % scraperobj.getName()
|
||||
|
||||
def test_find_scraperclasses_single(self):
|
||||
result = scraper.find_scraperclasses("xkcd")
|
||||
self.assertEqual(len(result), 1)
|
||||
assert len(result) == 1
|
||||
|
||||
def test_find_scraperclasses_multi(self):
|
||||
result = scraper.find_scraperclasses("a", multiple_allowed=True)
|
||||
self.assertTrue(len(result) > 1)
|
||||
assert len(result) > 1
|
||||
|
||||
def test_find_scraperclasses_error(self):
|
||||
self.assertRaises(ValueError, scraper.find_scraperclasses, "")
|
||||
with pytest.raises(ValueError):
|
||||
scraper.find_scraperclasses("")
|
||||
|
|
|
@ -1,9 +1,9 @@
|
|||
# -*- coding: iso-8859-1 -*-
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copied from: https://github.com/pycontribs/tendo
|
||||
# License: PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2
|
||||
# Author: Sorin Sbarnea
|
||||
# Changes: changed logging and formatting
|
||||
from unittest import TestCase
|
||||
|
||||
from dosagelib import singleton
|
||||
from multiprocessing import Process
|
||||
|
||||
|
@ -12,13 +12,12 @@ def f(flavor_id):
|
|||
return singleton.SingleInstance(flavor_id=flavor_id, exit_code=1)
|
||||
|
||||
|
||||
class TestSingleton(TestCase):
|
||||
|
||||
class TestSingleton(object):
|
||||
def test_1(self):
|
||||
# test in current process
|
||||
me = singleton.SingleInstance(flavor_id="test-1")
|
||||
del me # now the lock should be removed
|
||||
self.assertTrue(True)
|
||||
assert True
|
||||
|
||||
def test_2(self):
|
||||
# test in current subprocess
|
||||
|
@ -26,7 +25,7 @@ class TestSingleton(TestCase):
|
|||
p.start()
|
||||
p.join()
|
||||
# the called function should succeed
|
||||
self.assertEqual(p.exitcode, 0)
|
||||
assert p.exitcode == 0
|
||||
|
||||
def test_3(self):
|
||||
# test in current process and subprocess with failure
|
||||
|
@ -36,10 +35,10 @@ class TestSingleton(TestCase):
|
|||
p = Process(target=f, args=("test-3",))
|
||||
p.start()
|
||||
p.join()
|
||||
self.assertEqual(p.exitcode, 1)
|
||||
assert p.exitcode == 1
|
||||
# third instance
|
||||
p = Process(target=f, args=("test-3",))
|
||||
p.start()
|
||||
p.join()
|
||||
self.assertEqual(p.exitcode, 1)
|
||||
assert p.exitcode == 1
|
||||
del me # now the lock should be removed
|
||||
|
|
|
@ -1,32 +1,35 @@
|
|||
# -*- coding: iso-8859-1 -*-
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2004-2005 Tristan Seligmann and Jonathan Jacobs
|
||||
# Copyright (C) 2012-2014 Bastian Kleineidam
|
||||
import re
|
||||
from unittest import TestCase
|
||||
# Copyright (C) 2016 Tobias Gruetzmacher
|
||||
|
||||
import pytest
|
||||
import re
|
||||
from dosagelib.util import normaliseURL, unescape, tagre, get_system_uid
|
||||
|
||||
|
||||
class URLTest(TestCase):
|
||||
class TestURL(object):
|
||||
"""
|
||||
Tests for URL utility functions.
|
||||
"""
|
||||
|
||||
def test_unescape(self):
|
||||
# Test HTML replacement.
|
||||
self.assertEqual(unescape(u'foo&bar'), u'foo&bar')
|
||||
self.assertEqual(unescape(u'foo bar'), u'foo\xa0bar')
|
||||
self.assertEqual(unescape(u'"foo"'), u'"foo"')
|
||||
assert unescape(u'foo&bar') == u'foo&bar'
|
||||
assert unescape(u'foo bar') == u'foo\xa0bar'
|
||||
assert unescape(u'"foo"') == u'"foo"'
|
||||
|
||||
def test_normalisation(self):
|
||||
# Test URL normalisation.
|
||||
self.assertEqual(normaliseURL('http://example.com//bar/baz&baz'),
|
||||
u'http://example.com/bar/baz&baz')
|
||||
assert normaliseURL('http://example.com//bar/baz&baz') == \
|
||||
u'http://example.com/bar/baz&baz'
|
||||
|
||||
|
||||
class RegexTest(TestCase):
|
||||
class TestRegex(object):
|
||||
|
||||
ValuePrefix = '/bla/'
|
||||
TagTests = (
|
||||
|
||||
@pytest.mark.parametrize("tag,value,domatch", [
|
||||
('<img src="%s">', ValuePrefix+'foo', True),
|
||||
('< img src = "%s" >', ValuePrefix, True),
|
||||
('<img class="prev" src="%s">', ValuePrefix+'...', True),
|
||||
|
@ -35,27 +38,27 @@ class RegexTest(TestCase):
|
|||
('<img SrC="%s">', ValuePrefix, True),
|
||||
('<img src="%s">', ValuePrefix[:-1], False),
|
||||
('<img class="prev" src="%s" a="b">', ValuePrefix, True),
|
||||
)
|
||||
|
||||
def test_regex(self):
|
||||
matcher = re.compile(tagre("img", "src", '(%s[^"]*)' % self.ValuePrefix))
|
||||
for tag, value, domatch in self.TagTests:
|
||||
])
|
||||
def test_regex(self, tag, value, domatch):
|
||||
matcher = re.compile(tagre("img", "src", '(%s[^"]*)' %
|
||||
self.ValuePrefix))
|
||||
self.match_tag(matcher, tag, value, domatch)
|
||||
|
||||
def match_tag(self, matcher, tag, value, domatch=True):
|
||||
text = tag % value
|
||||
match = matcher.search(text)
|
||||
if domatch:
|
||||
self.assertTrue(match, "%s should match %s" % (matcher.pattern, text))
|
||||
self.assertEqual(match.group(1), value)
|
||||
assert match, "%s should match %s" % (matcher.pattern, text)
|
||||
assert match.group(1) == value
|
||||
else:
|
||||
self.assertFalse(match, "%s should not match %s" % (matcher.pattern, text))
|
||||
assert not match, "%s should not match %s" % (matcher.pattern,
|
||||
text)
|
||||
|
||||
|
||||
class UidTest(TestCase):
|
||||
class TestUid(object):
|
||||
"""
|
||||
Tests for unique system IDs.
|
||||
"""
|
||||
|
||||
def test_system_uid(self):
|
||||
self.assertTrue(get_system_uid())
|
||||
assert get_system_uid()
|
||||
|
|
|
@ -1,14 +1,16 @@
|
|||
# -*- coding: iso-8859-1 -*-
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2013-2014 Bastian Kleineidam
|
||||
from unittest import TestCase
|
||||
# Copyright (C) 2016 Tobias Gruetzmacher
|
||||
|
||||
from dosagelib import scraper
|
||||
|
||||
|
||||
class ATestScraper(scraper._BasicScraper):
|
||||
name = 'Test_Test'
|
||||
|
||||
class TestVote(TestCase):
|
||||
|
||||
class TestVote(object):
|
||||
|
||||
def test_vote(self):
|
||||
answer = ATestScraper.vote()
|
||||
self.assertTrue(answer in ('counted', 'no'), 'invalid answer %r' % answer)
|
||||
assert answer in ('counted', 'no'), 'invalid answer %r' % answer
|
||||
|
|
Loading…
Reference in a new issue