🎯

testing-python

🎯Skill

from eyh0602/skillshub

VibeIndex|
What it does

Writes and evaluates Python tests using pytest, covering test design, fixtures, parameterization, mocking, and async testing strategies.

πŸ“¦

Part of

eyh0602/skillshub(14 items)

testing-python

Installation

uv runRun with uv
uv run pytest -n auto # Run all tests in parallel
uv runRun with uv
uv run pytest -n auto -x # Stop on first failure
uv runRun with uv
uv run pytest path/to/test.py # Run specific file
uv runRun with uv
uv run pytest -k "test_name" # Run tests matching pattern
uv runRun with uv
uv run pytest -m "not integration" # Exclude integration tests
πŸ“– Extracted from docs: eyh0602/skillshub
1Installs
-
AddedFeb 4, 2026

Skill Details

SKILL.md

Write and evaluate effective Python tests using pytest. Use when writing tests, reviewing test code, debugging test failures, or improving test coverage. Covers test design, fixtures, parameterization, mocking, and async testing.

Overview

# Writing Effective Python Tests

Core Principles

Every test should be atomic, self-contained, and test single functionality. A test that tests multiple things is harder to debug and maintain.

Test Structure

Atomic unit tests

Each test should verify a single behavior. The test name should tell you what's broken when it fails. Multiple assertions are fine when they all verify the same behavior.

```python

# Good: Name tells you what's broken

def test_user_creation_sets_defaults():

user = User(name="Alice")

assert user.role == "member"

assert user.id is not None

assert user.created_at is not None

# Bad: If this fails, what behavior is broken?

def test_user():

user = User(name="Alice")

assert user.role == "member"

user.promote()

assert user.role == "admin"

assert user.can_delete_others()

```

Use parameterization for variations of the same concept

```python

import pytest

@pytest.mark.parametrize("input,expected", [

("hello", "HELLO"),

("World", "WORLD"),

("", ""),

("123", "123"),

])

def test_uppercase_conversion(input, expected):

assert input.upper() == expected

```

Use separate tests for different functionality

Don't parameterize unrelated behaviors. If the test logic differs, write separate tests.

Project-Specific Rules

No async markers needed

This project uses asyncio_mode = "auto" globally. Write async tests without decorators:

```python

# Correct

async def test_async_operation():

result = await some_async_function()

assert result == expected

# Wrong - don't add this

@pytest.mark.asyncio

async def test_async_operation():

...

```

Imports at module level

Put ALL imports at the top of the file:

```python

# Correct

import pytest

from fastmcp import FastMCP

from fastmcp.client import Client

async def test_something():

mcp = FastMCP("test")

...

# Wrong - no local imports

async def test_something():

from fastmcp import FastMCP # Don't do this

...

```

Use in-memory transport for testing

Pass FastMCP servers directly to clients:

```python

from fastmcp import FastMCP

from fastmcp.client import Client

mcp = FastMCP("TestServer")

@mcp.tool

def greet(name: str) -> str:

return f"Hello, {name}!"

async def test_greet_tool():

async with Client(mcp) as client:

result = await client.call_tool("greet", {"name": "World"})

assert result[0].text == "Hello, World!"

```

Only use HTTP transport when explicitly testing network features.

Inline snapshots for complex data

Use inline-snapshot for testing JSON schemas and complex structures:

```python

from inline_snapshot import snapshot

def test_schema_generation():

schema = generate_schema(MyModel)

assert schema == snapshot() # Will auto-populate on first run

```

Commands:

  • pytest --inline-snapshot=create - populate empty snapshots
  • pytest --inline-snapshot=fix - update after intentional changes

Fixtures

Prefer function-scoped fixtures

```python

@pytest.fixture

def client():

return Client()

async def test_with_client(client):

result = await client.ping()

assert result is not None

```

Use `tmp_path` for file operations

```python

def test_file_writing(tmp_path):

file = tmp_path / "test.txt"

file.write_text("content")

assert file.read_text() == "content"

```

Mocking

Mock at the boundary

```python

from unittest.mock import patch, AsyncMock

async def test_external_api_call():

with patch("mymodule.external_client.fetch", new_callable=AsyncMock) as mock:

mock.return_value = {"data": "test"}

result = await my_function()

assert result == {"data": "test"}

```

Don't mock what you own

Test your code with real implementations when possible. Mock external services, not internal classes.

Test Naming

Use descriptive names that explain the scenario:

```python

# Good

def test_login_fails_with_invalid_password():

def test_user_can_update_own_profile():

def test_admin_can_delete_any_user():

# Bad

def test_login():

def test_update():

def test_delete():

```

Error Testing

```python

import pytest

def test_raises_on_invalid_input():

with pytest.raises(ValueError, match="must be positive"):

calculate(-1)

async def test_async_raises():

with pytest.raises(ConnectionError):

await connect_to_invalid_host()

```

Running Tests

```bash

uv run pytest -n auto # Run all tests in parallel

uv run pytest -n auto -x # Stop on first failure

uv run pytest path/to/test.py # Run specific file

uv run pytest -k "test_name" # Run tests matching pattern

uv run pytest -m "not integration" # Exclude integration tests

```

Checklist

Before submitting tests:

  • [ ] Each test tests one thing
  • [ ] No @pytest.mark.asyncio decorators
  • [ ] Imports at module level
  • [ ] Descriptive test names
  • [ ] Using in-memory transport (not HTTP) unless testing networking
  • [ ] Parameterization for variations of same behavior
  • [ ] Separate tests for different behaviors

More from this repository10

🎯
paper-polish🎯Skill

paper-polish skill from eyh0602/skillshub

🎯
code-reviewer🎯Skill

Automates comprehensive code reviews across multiple languages, providing in-depth analysis, best practices, security checks, and detailed review reports.

🎯
write-unit-tests🎯Skill

Generates comprehensive unit and integration tests for tldraw SDK using Vitest, covering editor functionality, shapes, tools, and test file organization.

🎯
docstring🎯Skill

Generates comprehensive, Sphinx-formatted docstrings for PyTorch functions and methods, following official PyTorch documentation conventions.

🎯
docs-review🎯Skill

Reviews documentation changes against Metabase's writing style guide, providing structured feedback for markdown files in pull requests or local diffs.

🎯
read-repo-references🎯Skill

read-repo-references skill from eyh0602/skillshub

🎯
senior-data-scientist🎯Skill

Guides data science professionals through advanced machine learning techniques, model deployment, and complex data analysis workflows.

🎯
python-packaging🎯Skill

Guides developers through creating, structuring, and publishing Python packages using modern packaging tools and best practices.

🎯
uv-package-manager🎯Skill

Manages and installs Python packages from UV package manager with simplified command-line interactions and dependency resolution.

🎯
code-change-verification🎯Skill

Verifies code changes by running formatting, linting, type checking, and tests in the OpenAI Agents Python repository.