UnitTest Bridge
Python’s built-in unittest framework has been a staple for test automation for years. Many teams have extensive test suites built with unittest, but as testing needs evolve, modern frameworks like Vedro offer powerful features that improve test execution, reporting, and stability.
Transitioning to a new testing framework can feel overwhelming, especially for teams with extensive unittest-based test suites. With vedro-unittest, you can gradually integrate Vedro into your testing process:
- Try Vedro without full commitment: Run your existing unittest test cases inside Vedro to explore its enhanced reporting, parallel execution, and built-in stability mechanisms without rewriting your tests.
- Migrate gradually: Keep unittest and Vedro tests separate while executing them together within a unified reporting system, allowing for a seamless transition.
- Enhance testing capabilities: Leverage Vedro’s ecosystem, including integrations with tools like Playwright, HTTPX, and Allure, to improve test automation while maintaining your unittest-based test logic.
Usage
- Quick
- Manual
For a quick installation, you can use a plugin manager like so:
$ vedro plugin install vedro-unittest
If you prefer a manual approach, follow these steps:
- Install the package using pip:
$ pip install vedro-unittest
- Then, enable the plugin in the
vedro.cfg.py
configuration file:
# ./vedro.cfg.py
import vedro
import vedro_unittest
class Config(vedro.Config):
class Plugins(vedro.Config.Plugins):
class VedroUnitTest(vedro_unittest.VedroUnitTest):
enabled = True
Once installed, you can run your existing unittest test cases with Vedro:
$ vedro run tests/
Vedro will automatically detect and execute unittest test cases as Vedro scenarios.
- vedro
- unittest
Scenarios
*
✗ [TestOrderProcessing] test order total (0.00s)
✗ do (0.00s)
╭────────────────────── Traceback (most recent call last) ───────────────────────╮
│ /order_service/tests/test_order.py:13 in test_order_total │
│ │
│ 10 def test_order_total(self): │
│ 11 order = {"price": 10, "quantity": 3} │
│ 12 result = process_order(order) │
│ ❱ 13 self.assertEqual(result["total"], 35) │
│ 14 │
│ ╭─────────────────── locals ────────────────────╮ │
│ │ order = {'price': 10, 'quantity': 3} │ │
│ │ result = {'total': 30, 'status': 'processed'} │ │
│ ╰───────────────────────────────────────────────╯ │
╰────────────────────────────────────────────────────────────────────────────────╯
AssertionError: 30 != 35
# 1 scenario, 0 passed, 1 failed, 0 skipped (0.00s)
======================================================================
FAIL: test_order_total (tests.test_order.TestOrderProcessing.test_order_total)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/order_service/tests/test_order.py", line 13, in test_order_total
self.assertEqual(result["total"], 35)
AssertionError: 30 != 35
----------------------------------------------------------------------
Ran 1 test in 0.000s
FAILED (failures=1)
Compatibility
vedro-unittest is designed to be fully compatible with unittest, maintaining the expected behavior while running tests within Vedro. However, there are a few nuances to be aware of. See the Limitations section for more details.
✅ Assertions
All assertion methods from unittest are fully supported, ensuring that existing test validations run without changes.
import unittest
class TestStringMethods(unittest.TestCase):
def test_upper(self):
self.assertEqual("banana".upper(), "BANANA")
✅ Fixtures (setUp & tearDown)
Lifecycle methods such as setUp()
and tearDown()
function as expected, ensuring that tests can properly prepare the environment before execution.
import unittest
class WidgetTestCase(unittest.TestCase):
def setUp(self):
self.widget = Widget('The widget')
def test_widget_resize(self):
self.widget.resize(100, 150)
self.assertEqual(self.widget.size(), (100, 150))
✅ Skipping Tests
Tests marked with @unittest.skip*
decorators will be skipped in Vedro just as they are in unittest, whether applied to individual test methods or entire test classes.
- Skipping a Test
- Skipping a Test Suite
import unittest
class MyTestCase(unittest.TestCase):
@unittest.skip("Feature not available")
def test_feature(self):
pass
@unittest.skipIf(mylib.__version__ < (1, 3), "Requires mylib version >= 1.3")
def test_new_feature(self):
pass
import unittest
@unittest.skip("Feature not available")
class MyTestCase(unittest.TestCase):
def test_feature(self):
pass
✅ Expected Failures
Vedro correctly handles @unittest.expectedFailure
, ensuring that tests expected to fail do not cause the scenario to fail.
import unittest
class ExpectedFailureTestCase(unittest.TestCase):
@unittest.expectedFailure
def test_fail(self):
self.assertEqual(1, 0, "This test is expected to fail")
Output:
Scenarios
*
✔ [ExpectedFailureTestCase] test fail (0.00s)
|> Expected Failure: Scenario passed because it failed as expected with AssertionError('1 != 0 : This test is expected to fail')
# 1 scenario, 1 passed, 0 failed, 0 skipped (0.00s)
If a test marked as an expected failure unexpectedly passes, vedro-unittest raises an UnexpectedSuccessError
.
Scenarios
*
✗ [ExpectedFailureTestCase] test fail (0.00s)
|> Unexpected Success: Scenario failed because it was expected to fail, but the scenario passed
✗ do (0.00s)
UnexpectedSuccessError: Scenario passed, but expected to fail
# 1 scenario, 0 passed, 1 failed, 0 skipped (0.00s)
✅ Parameterized Tests
Parameterized tests work as expected. The subTest()
feature and external libraries like parameterized remain fully supported.
- Using subTest
- Using parameterized
import unittest
class NumbersTest(unittest.TestCase):
def test_even(self):
for i in range(0, 6, 2):
with self.subTest(i=i):
self.assertEqual(i % 2, 0)
import unittest
from parameterized import parameterized
class NumbersTest(unittest.TestCase):
@parameterized.expand(range(0, 6, 2))
def test_even(self, value):
self.assertEqual(value % 2, 0)
✅ Async Support
vedro-unittest supports async test cases out of the box. Existing async def
test methods are executed correctly, running in a dedicated thread within Vedro’s async-aware environment.
import unittest
import asyncio
class AsyncTestCase(unittest.IsolatedAsyncioTestCase):
async def test_async_function(self):
result = await asyncio.sleep(0.1, result=42)
self.assertEqual(result, 42)
✅ Mocking
Since Vedro acts as a test runner rather than modifying test logic, the standard unittest.mock module remains fully functional. Existing @patch
decorators and context managers work as expected.
import unittest
from unittest.mock import Mock
class MockTestCase(unittest.TestCase):
def test_mock(self):
mock = Mock()
mock.method.return_value = "mocked"
self.assertEqual(mock.method(), "mocked")
Limitations
While vedro-unittest provides seamless integration of unittest test cases within Vedro, there are a few important limitations to be aware of. These limitations primarily stem from Vedro’s design philosophy, which prioritizes test independence, reliability, and maintainability.
🚫 Skipping Tests at Runtime
Vedro does not support dynamically skipping tests during execution using self.skipTest()
. This practice is considered an anti-pattern because it undermines the core principle of deterministic and independent test execution that Vedro enforces. All decisions about which tests to run should be made during test discovery (or configuration time), ensuring that the test suite behaves predictably and reliably, even in parallel execution environments.
For example, the following code, which attempts to skip a test based on a runtime condition, will fail under Vedro:
import unittest
class MyTestCase(unittest.TestCase):
def test_feature(self):
if not some_condition():
self.skipTest("Feature not available")
Instead of dynamically skipping tests within the test logic, use decorators such as @unittest.skipIf
or @unittest.skipUnless
to evaluate conditions at the time of test discovery. This method ensures that all test decisions are made upfront, aligning with Vedro’s design for predictable test execution.
⚠️ Custom Test Ordering
Vedro intentionally does not support defining a custom test execution order. This design decision reinforces the principle that each test should be independent and isolated. Enforcing a specific order can lead to hidden dependencies and stateful tests, often resulting in flaky behavior and reduced maintainability.
In some unittest workflows, tests are arranged to run in a predetermined sequence. For example:
import unittest
class MyTestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
# Set up shared state
...
def test_1(self):
# Relies on shared state
...
def test_2(self):
# Also relies on shared state
...
@classmethod
def tearDownClass(cls):
# Clean up shared state
...
In unittest, this code executes in a fixed sequence:
setUpClass
runs once before any tests.test_1
is executed.test_2
follows.tearDownClass
runs after all tests.
While this may work for certain scenarios, it introduces dependencies between tests. If additional tests are inserted or tests are run in parallel, these dependencies can lead to unpredictable outcomes.
To maintain compatibility with unittest, Vedro groups tests that share module-level (setUpModule
/tearDownModule
) or class-level (setUpClass
/tearDownClass
) fixtures into a single scenario. This grouping is a compatibility workaround, ensuring that shared setup and teardown routines are executed correctly, not an endorsement of custom test ordering.
This grouping means that:
- Each test still executes its own
setUp()
andtearDown()
methods. - If one test in a grouped scenario fails, the entire scenario is marked as failed (although all tests will still be executed, as in unittest).
- If multiple tests in a grouped scenario fail, the report will include a complete summary of all failures.
Gradual Migration to Vedro
Switching testing frameworks is rarely an overnight task, especially for large codebases. Before committing fully, it's important to experience the benefits firsthand and make an informed decision based on your own testing needs.
With vedro-unittest, you can transition smoothly, step by step, without disrupting your existing test suite.
Step 1: Start Running unittest Tests with Vedro
Instead of rewriting everything at once, start by running your existing unittest tests inside Vedro using the vedro-unittest plugin.
Benefit: No code changes required, your tests remain the same, but you immediately gain:
- Enhanced reporting for better visibility
- Improved test stability with built-in anti-flaky mechanisms
- Access to Vedro’s integrations (e.g., Playwright, HTTPX, Allure)
Step 2: Introduce New Tests in Vedro
As you develop new features, start writing new test scenarios in Vedro while still running legacy unittest tests. This lets your team gradually adopt Vedro’s scenario-based structure without disrupting development.
The easiest way to separate the test types:
- Run existing unittest tests via
vedro run tests/
(with vedro-unittest enabled) - Run native Vedro tests via
vedro run scenarios/
(with vedro-unittest disabled)
Benefit: Your team gets familiar with Vedro without modifying existing tests, reducing risk and friction.
For a quick reference on how unittest concepts translate to Vedro, see the unittest → Vedro Cheatsheet.
Step 3: Convert Legacy unittest Tests to Vedro Scenarios
Over time, start rewriting critical tests as native Vedro scenarios. This allows you to fully adopt Vedro’s structured, expressive, and maintainable testing approach.
Benefit:
- Clear and natural test structure that makes scenarios easier to read and understand.
- Independent and modular tests that simplify test management and reduce setup complexity.
- Built-in stability mechanisms that help mitigate flaky tests with retries and robust error handling.
Step 4: Remove vedro-unittest Once Fully Migrated
Once most tests have been rewritten as Vedro scenarios, you can phase out vedro-unittest and run only native Vedro tests.
Final Benefit: Your entire test suite is optimized for independence, reliability and maintainability, leveraging Vedro’s full potential.
Final Thoughts
Migrating to a new testing framework is a strategic decision, and Vedro makes it easier by allowing a gradual transition. With vedro-unittest, you can start small, gain the benefits of enhanced execution and reporting immediately, and adopt Vedro at your own pace.
By following this step-by-step approach, you minimize risk, improve test clarity and stability, and position your team for a more efficient and maintainable testing workflow.
If you're ready to take the next step, start running your unittest tests with Vedro today and experience the difference firsthand! 🚀