unishark 0.3.2

Creator: bradpython12

Last updated:

0 purchases

TODO
Add to Cart

Description:

unishark 0.3.2

INTRODUCTION
unishark extends unittest (to be accurate, unittest2) in the
following ways:

Customizing test suites with dictionary config (or yaml/json like config).
Running the tests concurrently at different levels.
Generating polished test reports in HTML/XUnit formats.
Offering data-driven decorator to accelerate tests writing.

For existing unittests, the first three features could be obtained immediately with a single config, without changing any test code.

The Test Config
Here is an example config in YAML format (you could also write it
directly in a dict()):
suites:
my_suite_name_1:
package: my.package.name
groups:
my_group_1:
granularity: module
modules: [test_module1, test_module2]
except_classes: [test_module2.MyTestClass3]
except_methods: [test_module1.MyTestClass1.test_1]
my_group_2:
granularity: class
disable: False
classes: [test_module3.MyTestClass5]
except_methods: [test_module3.MyTestClass5.test_11]
concurrency:
level: module
max_workers: 2
my_suite_name_2:
package: my.package.name
groups:
my_group_1:
granularity: method
methods: [test_module3.MyTestClass6.test_13, test_module3.MyTestClass7.test_15]
concurrency:
level: class
max_workers: 2
my_suite_name_3:
package: another.package.name
groups:
group_1:
granularity: package
pattern: '(\w+\.){2}test\w*'
except_modules: [module1, module2]
except_classes: [module3.Class1, module3.Class3]
except_methods: [module3.Class2.test_1, module4.Class2.test_5]
concurrency:
level: method
max_workers: 20
reporters:
html:
class: unishark.HtmlReporter
kwargs:
dest: logs
overview_title: 'Example Report'
overview_description: 'This is an example report'
xunit:
class: unishark.XUnitReporter
kwargs:
summary_title: 'Example Report'

test:
suites: [my_suite_name_1, my_suite_name_2, my_suite_name_3]
concurrency:
type: processes
max_workers: 3
reporters: [html, xunit]
name_pattern: '^test\w*'
It configures 3 test suites with some of the test cases excluded, and running the defined set of tests concurrently, and generating both HTML and XUnit (default JUnit) format reports at the end of tests.
NOTE: In 0.2.x versions, ‘max_workers’ was set directly under ‘test’, and ‘max_workers’ and ‘concurrency_level’ were set directly under ‘{suite name}’.
To run it, simply add the following code:
import unishark
import yaml

if __name__ == '__main__':
with open('your_yaml_config_file', 'r') as f:
dict_conf = yaml.load(f.read()) # use a 3rd party yaml parser, e.g., PyYAML
program = unishark.DefaultTestProgram(dict_conf)
unishark.main(program)
A HTML report example can be found Here.


Data Driven
Here are some effects of using @unishark.data_driven.
‘Json’ style data-driven:
@unishark.data_driven(*[{'userid': 1, 'passwd': 'abc'}, {'userid': 2, 'passwd': 'def'}])
def test_data_driven(self, **param):
print('userid: %d, passwd: %s' % (param['userid'], param['passwd']))
Results:
userid: 1, passwd: abc
userid: 2, passwd: def
‘Args’ style data-driven:
@unishark.data_driven(userid=[1, 2, 3, 4], passwd=['a', 'b', 'c', 'd'])
def test_data_driven(self, **param):
print('userid: %d, passwd: %s' % (param['userid'], param['passwd']))
Results:
userid: 1, passwd: a
userid: 2, passwd: b
userid: 3, passwd: c
userid: 4, passwd: d
Cross-multiply data-driven:
@unishark.data_driven(left=list(range(10)))
@unishark.data_driven(right=list(range(10)))
def test_data_driven(self, **param):
l = param['left']
r = param['right']
print('%d x %d = %d' % (l, r, l * r))
Results:
0 x 1 = 0
0 x 2 = 0
...
1 x 0 = 0
1 x 1 = 1
1 x 2 = 2
...
...
9 x 8 = 72
9 x 9 = 81
You can get the permutations (with repetition) of the parameters values
by doing:
@unishark.data_driven(...)
@unishark.data_driven(...)
@unishark.data_driven(...)
...
Multi-threads data-driven in ‘json style’:
@unishark.multi_threading_data_driven(2, *[{'userid': 1, 'passwd': 'abc'}, {'userid': 2, 'passwd': 'def'}])
def test_data_driven(self, **param):
print('userid: %d, passwd: %s' % (param['userid'], param['passwd']))
Results: same results as using unishark.data_driven, but up to 2 threads are spawned, each running the test with a set of inputs (userid, passwd).
Multi-threads data-driven in ‘args style’:
@unishark.multi_threading_data_driven(5, time=[1, 1, 1, 1, 1, 1, 1, 1, 1, 1])
def test_data_driven(self, **param):
sleep(param['time'])
Results: 5 threads are spawned to run the test with 10 sets of inputs concurrently (only sleep 1 sec in each thread).
It takes about 2 sec in total (10 sec if using unishark.data_driven) to run.
For more information please visit the Project_Home and read README.md.



CHANGELOG

0.3.2 (2015-11-24)


added multiprocessing suites (which can bypass CPython’s GIL and utilize multi-cores).
modified result, runner and reporter classes to be picklable for multiprocessing.
supported running with Jython.




0.3.1 (2015-11-12)


fixed the issue of still running test methods even when setUpClass/setUpModule raises exception in concurrency mode.
fixed error descriptions of class or module level fixtures when they raise exceptions.




0.3.0 (2015-11-06)


rewrote concurrent execution model. Now test fixtures setUpModule/tearDownModule setUpClass/tearDownClass will be executed once and only once no matter what concurrency level(module/class/method) of a suite is. Fixed the problem that module fixtures were executed multiple times when concurrency level was ‘class’ or ‘method’, and class fixtures were executed multiple times when concurrency level was ‘method’.
changed the format of the concurrency-related settings in the dict config. Now ‘max_workers’ and ‘level’ are keys in the ‘concurrency’ sub-dict.
moved BufferedTestResult class from the runner module to the new result module which makes more sense.




0.2.3 (2015-10-01)


enabled ‘module’ and ‘method’ level concurrent execution in a suite.




0.2.2 (2015-08-12)


support loading tests from a package with pattern matching, and excluding modules/classes/methods from the loaded tests.
add load_tests_from_package and load_tests_from_modules api.
rename load_test_from_dict to load_tests_from_dict.
fix that verbose stdout mode does not print test method doc string.
fix that tests loaded with method granularity are not filtered by method name pattern.
less strict dependency versions.




0.2.1 (2015-05-11)


support data-driven with multi-threads.




0.2.0 (2015-04-04)


support running tests in parallel.
support configuring test suites, test reporters and concurrent tests in a single dict/yaml config.
improve HtmlReporter and XUnitReporter classes to be thread-safe.
allow user to generate reports with their own report templates.
allow user to filter loaded test cases by setting method name prefix in the test config.
bugs fix.
improve documentation.




0.1.2 (2015-03-25)


hotfix for setup.py (broken auto-downloading dependencies)
bugs fix.




0.1.1 (2015-03-24)


support loading customized test suites.
support a thread-safe string io buffer to buffer logging stream during the test running.
support writing logs, exceptions into generated HTML/XUnit reports.
offer a data-driven decorator.
initial setup (documentation, setup.py, travis CI, coveralls, etc.).

License

For personal and professional use. You cannot resell or redistribute these repositories in their original state.

Files:

Customer Reviews

There are no reviews.