Skip to content

Test test_timeout_in_handler is flaky #1652

@vdusek

Description

@vdusek

https://github.com/apify/crawlee-python/actions/runs/20395848580/job/58611220603?pr=1635

=========================== short test summary info ============================
FAILED tests/unit/crawlers/_basic/test_basic_crawler.py::test_timeout_in_handler[async_sleep] - assert 0 == 1
 +  where 0 = StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2025, 12, 20, 14, 41, 27, 856866, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=946305), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0).requests_finished
 +    where StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2025, 12, 20, 14, 41, 21, 910561, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2025, 12, 20, 14, 41, 27, 856866, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=946305), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0) = <crawlee.statistics._statistics.Statistics object at 0x110733a10>.state
 +      where <crawlee.statistics._statistics.Statistics object at 0x110733a10> = <crawlee.crawlers._basic._basic_crawler.BasicCrawler object at 0x110732e40>.statistics
======================== 1 failed, 1 skipped in 16.87s =========================

https://github.com/apify/crawlee-python/actions/runs/20742769552/job/59552920657

FAILED tests/unit/crawlers/_basic/test_basic_crawler.py::test_timeout_in_handler[async_sleep] - assert 0 == 1
 +  where 0 = StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 37, 35, 531091, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=997703), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0).requests_finished
 +    where StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 37, 29, 533388, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 37, 35, 531091, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=997703), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0) = <crawlee.statistics._statistics.Statistics object at 0x130f02c90>.state
 +      where <crawlee.statistics._statistics.Statistics object at 0x130f02c90> = <crawlee.crawlers._basic._basic_crawler.BasicCrawler object at 0x130ced850>.statistics
======================== 1 failed, 1 skipped in 22.67s =========================

https://github.com/apify/crawlee-python/actions/runs/20742882768/job/59553465201

FAILED tests/unit/crawlers/_basic/test_basic_crawler.py::test_timeout_in_handler[async_sleep] - assert 0 == 1
 +  where 0 = StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 46, 56, 154946, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=959784), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0).requests_finished
 +    where StatisticsState(stats_id=0, requests_finished=0, requests_failed=0, requests_retries=0, requests_failed_per_minute=0, requests_finished_per_minute=0, request_min_duration=None, request_max_duration=None, request_total_failed_duration=datetime.timedelta(0), request_total_finished_duration=datetime.timedelta(0), crawler_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_last_started_at=datetime.datetime(2026, 1, 6, 8, 46, 50, 195162, tzinfo=datetime.timezone.utc), crawler_finished_at=datetime.datetime(2026, 1, 6, 8, 46, 56, 154946, tzinfo=datetime.timezone.utc), errors={}, retry_errors={}, requests_with_status_code={}, stats_persisted_at=None, request_retry_histogram={}, crawler_runtime_for_serialization=datetime.timedelta(seconds=5, microseconds=959784), request_total_duration=datetime.timedelta(0), request_avg_failed_duration=None, request_avg_finished_duration=None, requests_total=0) = <crawlee.statistics._statistics.Statistics object at 0x11ec4ac90>.state
 +      where <crawlee.statistics._statistics.Statistics object at 0x11ec4ac90> = <crawlee.crawlers._basic._basic_crawler.BasicCrawler object at 0x12956de10>.statistics
======================== 1 failed, 1 skipped in 21.65s =========================

Metadata

Metadata

Assignees

Labels

t-toolingIssues with this label are in the ownership of the tooling team.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions